Community Detection-based Analysis of the Human Interactome Network

The study of proteomics reached unexpected levels of interest, as a direct consequence of its discovered influence over some complex biological phenomena, such as problematic diseases like cancer. This paper presents a new technique that allows for an accurate analysis of the human interactome network. It is basically a two-step analysis process that involves, at first, the detection of each protein-s absolute importance through the betweenness centrality computation. Then, the second step determines the functionallyrelated communities of proteins. For this purpose, we use a community detection technique that is based on the edge betweenness calculation. The new technique was thoroughly tested on real biological data and the results prove some interesting properties of those proteins that are involved in the carcinogenesis process. Apart from its experimental usefulness, the novel technique is also computationally effective in terms of execution times. Based on the analysis- results, some topological features of cancer mutated proteins are presented and a possible optimization solution for cancer drugs design is suggested.

Automatic Lip Contour Tracking and Visual Character Recognition for Computerized Lip Reading

Computerized lip reading has been one of the most actively researched areas of computer vision in recent past because of its crime fighting potential and invariance to acoustic environment. However, several factors like fast speech, bad pronunciation, poor illumination, movement of face, moustaches and beards make lip reading difficult. In present work, we propose a solution for automatic lip contour tracking and recognizing letters of English language spoken by speakers using the information available from lip movements. Level set method is used for tracking lip contour using a contour velocity model and a feature vector of lip movements is then obtained. Character recognition is performed using modified k nearest neighbor algorithm which assigns more weight to nearer neighbors. The proposed system has been found to have accuracy of 73.3% for character recognition with speaker lip movements as the only input and without using any speech recognition system in parallel. The approach used in this work is found to significantly solve the purpose of lip reading when size of database is small.

Efficient DTW-Based Speech Recognition System for Isolated Words of Arabic Language

Despite the fact that Arabic language is currently one of the most common languages worldwide, there has been only a little research on Arabic speech recognition relative to other languages such as English and Japanese. Generally, digital speech processing and voice recognition algorithms are of special importance for designing efficient, accurate, as well as fast automatic speech recognition systems. However, the speech recognition process carried out in this paper is divided into three stages as follows: firstly, the signal is preprocessed to reduce noise effects. After that, the signal is digitized and hearingized. Consequently, the voice activity regions are segmented using voice activity detection (VAD) algorithm. Secondly, features are extracted from the speech signal using Mel-frequency cepstral coefficients (MFCC) algorithm. Moreover, delta and acceleration (delta-delta) coefficients have been added for the reason of improving the recognition accuracy. Finally, each test word-s features are compared to the training database using dynamic time warping (DTW) algorithm. Utilizing the best set up made for all affected parameters to the aforementioned techniques, the proposed system achieved a recognition rate of about 98.5% which outperformed other HMM and ANN-based approaches available in the literature.

Citizens- Expectations from Rural Telecentres: A Case Study of Implementation of Common Service Centres in Mushedpur Village, Haryana, India

Setting up of rural telecentres, popularly referred to as Common Service Centres (CSCs), are considered one of the initial forerunners of rural e-Governance initiatives under the Government of India-s National e-Governance Plan (NeGP). CSCs are implemented on public-private partnership (PPP) – where State governments play a major role in facilitating the establishment of CSCs and investments are made by private companies referred to as Service Centre Agencies (SCAs). CSC implementation is expected to help in improving public service delivery in a transparent and efficient manner. However, there is very little research undertaken to study the actual impact of CSC implementation at the grassroots level. This paper addresses the gap by identifying the circumstances, concerns and expectations from the point-of-view of citizens and examining the finer aspects of social processes in the context of rural e-Governance.

Exploring Dimensionality, Systematic Mutations and Number of Contacts in Simple HP ab-initio Protein Folding Using a Blackboard-based Agent Platform

A computational platform is presented in this contribution. It has been designed as a virtual laboratory to be used for exploring optimization algorithms in biological problems. This platform is built on a blackboard-based agent architecture. As a test case, the version of the platform presented here is devoted to the study of protein folding, initially with a bead-like description of the chain and with the widely used model of hydrophobic and polar residues (HP model). Some details of the platform design are presented along with its capabilities and also are revised some explorations of the protein folding problems with different types of discrete space. It is also shown the capability of the platform to incorporate specific tools for the structural analysis of the runs in order to understand and improve the optimization process. Accordingly, the results obtained demonstrate that the ensemble of computational tools into a single platform is worthwhile by itself, since experiments developed on it can be designed to fulfill different levels of information in a self-consistent fashion. By now, it is being explored how an experiment design can be useful to create a computational agent to be included within the platform. These inclusions of designed agents –or software pieces– are useful for the better accomplishment of the tasks to be developed by the platform. Clearly, while the number of agents increases the new version of the virtual laboratory thus enhances in robustness and functionality.

A Comparative Study of Rigid and Modified Simplex Methods for Optimal Parameter Settings of ACO for Noisy Non-Linear Surfaces

There are two common types of operational research techniques, optimisation and metaheuristic methods. The latter may be defined as a sequential process that intelligently performs the exploration and exploitation adopted by natural intelligence and strong inspiration to form several iterative searches. An aim is to effectively determine near optimal solutions in a solution space. In this work, a type of metaheuristics called Ant Colonies Optimisation, ACO, inspired by a foraging behaviour of ants was adapted to find optimal solutions of eight non-linear continuous mathematical models. Under a consideration of a solution space in a specified region on each model, sub-solutions may contain global or multiple local optimum. Moreover, the algorithm has several common parameters; number of ants, moves, and iterations, which act as the algorithm-s driver. A series of computational experiments for initialising parameters were conducted through methods of Rigid Simplex, RS, and Modified Simplex, MSM. Experimental results were analysed in terms of the best so far solutions, mean and standard deviation. Finally, they stated a recommendation of proper level settings of ACO parameters for all eight functions. These parameter settings can be applied as a guideline for future uses of ACO. This is to promote an ease of use of ACO in real industrial processes. It was found that the results obtained from MSM were pretty similar to those gained from RS. However, if these results with noise standard deviations of 1 and 3 are compared, MSM will reach optimal solutions more efficiently than RS, in terms of speed of convergence.

Optimization of Growth of Rhodobacter Sphaeroides Using Mixed Volatile Fatty Acidsby Response Surface Methodology

A combination of photosynthetic bacteria along with anaerobic acidogenic bacteria is an ideal option for efficient hydrogen production. In the present study, the optimum concentration of substrates for the growth of Rhodobacter sphaeroides was found by response surface methodology. The optimum combination of three individual fatty acids was determined by Box Behnken design. Increase of volatile fatty acid concentration decreased the growth. Combination of sodium acetate and sodium propionate was most significant for the growth of the organism. The results showed that a maximum biomass concentration of 0.916 g/l was obtained when the concentrations of acetate, propionate and butyrate were 0.73g/l,0.99g/l and 0.799g/l, respectively. The growth was studied under an optimum concentration of volatile fatty acids and at a light intensity of 3000 lux, initial pH of 7 and a temperature of 35°C.The maximum biomass concentration of 0.92g/l was obtained which verified the practicability of this optimization.

Photovoltaic Small-Scale Wastewater Treatment Project for Rural and New-Cultivated Areas in Egypt

The problem of wastewater treatment in Egypt is a two-fold problem; the first part concerning the existing rural areas, the second one dealing with new industrial/domestic areas. In Egypt several agricultural projects have been initiated by the government and the private sector as well, in order to change its infrastructure. As a reliable energy source, photovoltaic pumping systems have contributed to supply water for local rural communities worldwide; they can also be implemented to solve the problem “wastewater environment pollution". The solution of this problem can be categorised as recycle process. In addition, because of regional conditions past technologies are being reexamined to select a smallscale treatment system requiring low construction and maintenance costs. This paper gives the design guidelines of a Photovoltaic Small- Scale Wastewater Treatment Plant (PVSSWTP) based on technologies that can be transferred.

Optimization for Reducing Handoff Latency and Utilization of Bandwidth in ATM Networks

To support mobility in ATM networks, a number of technical challenges need to be resolved. The impact of handoff schemes in terms of service disruption, handoff latency, cost implications and excess resources required during handoffs needs to be addressed. In this paper, a one phase handoff and route optimization solution using reserved PVCs between adjacent ATM switches to reroute connections during inter-switch handoff is studied. In the second phase, a distributed optimization process is initiated to optimally reroute handoff connections. The main objective is to find the optimal operating point at which to perform optimization subject to cost constraint with the purpose of reducing blocking probability of inter-switch handoff calls for delay tolerant traffic. We examine the relation between the required bandwidth resources and optimization rate. Also we calculate and study the handoff blocking probability due to lack of bandwidth for resources reserved to facilitate the rapid rerouting.

Restarted Generalized Second-Order Krylov Subspace Methods for Solving Quadratic Eigenvalue Problems

This article is devoted to the numerical solution of large-scale quadratic eigenvalue problems. Such problems arise in a wide variety of applications, such as the dynamic analysis of structural mechanical systems, acoustic systems, fluid mechanics, and signal processing. We first introduce a generalized second-order Krylov subspace based on a pair of square matrices and two initial vectors and present a generalized second-order Arnoldi process for constructing an orthonormal basis of the generalized second-order Krylov subspace. Then, by using the projection technique and the refined projection technique, we propose a restarted generalized second-order Arnoldi method and a restarted refined generalized second-order Arnoldi method for computing some eigenpairs of largescale quadratic eigenvalue problems. Some theoretical results are also presented. Some numerical examples are presented to illustrate the effectiveness of the proposed methods.

Mathematical Modeling of an Avalanche Release and Estimation of Flow Parameters by Numerical Method

Avalanche release of snow has been modeled in the present studies. Snow is assumed to be represented by semi-solid and the governing equations have been studied from the concept of continuum approach. The dynamical equations have been solved for two different zones [starting zone and track zone] by using appropriate initial and boundary conditions. Effect of density (ρ), Eddy viscosity (η), Slope angle (θ), Slab depth (R) on the flow parameters have been observed in the present studies. Numerical methods have been employed for computing the non linear differential equations. One of the most interesting and fundamental innovation in the present studies is getting initial condition for the computation of velocity by numerical approach. This information of the velocity has obtained through the concept of fracture mechanics applicable to snow. The results on the flow parameters have found to be in qualitative agreement with the published results.

Non-destructive Watermelon Ripeness Determination Using Image Processing and Artificial Neural Network (ANN)

Agriculture products are being more demanding in market today. To increase its productivity, automation to produce these products will be very helpful. The purpose of this work is to measure and determine the ripeness and quality of watermelon. The textures on watermelon skin will be captured using digital camera. These images will be filtered using image processing technique. All these information gathered will be trained using ANN to determine the watermelon ripeness accuracy. Initial results showed that the best model has produced percentage accuracy of 86.51%, when measured at 32 hidden units with a balanced percentage rate of training dataset.

Spatial Query Localization Method in Limited Reference Point Environment

Task of object localization is one of the major challenges in creating intelligent transportation. Unfortunately, in densely built-up urban areas, localization based on GPS only produces a large error, or simply becomes impossible. New opportunities arise for the localization due to the rapidly emerging concept of a wireless ad-hoc network. Such network, allows estimating potential distance between these objects measuring received signal level and construct a graph of distances in which nodes are the localization objects, and edges - estimates of the distances between pairs of nodes. Due to the known coordinates of individual nodes (anchors), it is possible to determine the location of all (or part) of the remaining nodes of the graph. Moreover, road map, available in digital format can provide localization routines with valuable additional information to narrow node location search. However, despite abundance of well-known algorithms for solving the problem of localization and significant research efforts, there are still many issues that currently are addressed only partially. In this paper, we propose localization approach based on the graph mapped distances on the digital road map data basis. In fact, problem is reduced to distance graph embedding into the graph representing area geo location data. It makes possible to localize objects, in some cases even if only one reference point is available. We propose simple embedding algorithm and sample implementation as spatial queries over sensor network data stored in spatial database, allowing employing effectively spatial indexing, optimized spatial search routines and geometry functions.

Economical Analysis of Thermal Energy Storage by Partially Operation

Building Sector is the major electricity consumer and it is costly to building owners. Therefore the application of thermal energy storage (TES) has gained attractive to reduce energy cost. Many attractive tariff packages are being offered by the electricity provider to promote TES. The tariff packages offered higher cost of electricity during peak period and lower cost of electricity during off peak period. This paper presented the return of initial investment by implementing a centralized air-conditioning plant integrated with thermal energy storage with partially operation strategies. Building load profile will be calculated hourly according to building specification and building usage trend. TES operation conditions will be designed according to building load demand profile, storage capacity, tariff packages and peak/off peak period. The Payback Period analysis method was used to evaluate economic analysis. The investment is considered a good investment where by the initial cost is recovered less than ten than seven years.

On the Sphere Method of Linear Programming Using Multiple Interior Points Approach

The Sphere Method is a flexible interior point algorithm for linear programming problems. This was developed mainly by Professor Katta G. Murty. It consists of two steps, the centering step and the descent step. The centering step is the most expensive part of the algorithm. In this centering step we proposed some improvements such as introducing two or more initial feasible solutions as we solve for the more favorable new solution by objective value while working with the rigorous updates of the feasible region along with some ideas integrated in the descent step. An illustration is given confirming the advantage of using the proposed procedure.

Assessing and Managing Intellectual Capital to Support Open Innovation Paradigm

The objective of this paper is to support the application of Open Innovation practices in firms and organizations by the assessment and management of Intellectual Capital. Intellectual Capital constituents are analyzed in order to verify their capability of acting as key drivers of Open Innovation processes and, therefore, of creating value. A methodology is defined to settle a procedure which helps to select the most relevant Intellectual Capital value drivers and to provide Communities of Innovation with strategic and managerial guidelines in sustaining Open Innovation paradigm. An application of the methodology is developed within a specifically addressed project and its results are hereafter examined.

The Household Behavior on Solid Waste and Wastewater Management in Municipal Area with Cleanliness Policy Determined by Community

The Bangnanglee Sub-district Administrative Office, Thailand had initiated a policy to environmental protection with encouraging household waste management in order to promote civil responsibility for domestic hygienic. This research studied the household behaviors on solid waste and wastewater management. A sample population of 306 families answered a questionnaire. The study showed that, on average, domestic activities had produced 1.93 kilograms of waste per household per day. It has been found that 79% of the households made several attempts to reduce their own amount of waste. 80% of the households stationed their own garbage bins. 71% managed their waste by selling recyclable products. As for the rest of the waste, 51% burned them, while 29% disposed their waste in the nearby public trashcans and other 13% have them buried. As for wastewater, 60% of the households disposed it into the sewage, whereas 30% disposed them right from their elevated house.

Learning User Keystroke Patterns for Authentication

Keystroke authentication is a new access control system to identify legitimate users via their typing behavior. In this paper, machine learning techniques are adapted for keystroke authentication. Seven learning methods are used to build models to differentiate user keystroke patterns. The selected classification methods are Decision Tree, Naive Bayesian, Instance Based Learning, Decision Table, One Rule, Random Tree and K-star. Among these methods, three of them are studied in more details. The results show that machine learning is a feasible alternative for keystroke authentication. Compared to the conventional Nearest Neighbour method in the recent research, learning methods especially Decision Tree can be more accurate. In addition, the experiment results reveal that 3-Grams is more accurate than 2-Grams and 4-Grams for feature extraction. Also, combination of attributes tend to result higher accuracy.

Morphology and Risk Factors for Blunt Aortic Trauma in Car Accidents - An Autopsy Study

Background: Blunt aortic trauma (BAT) includes various morphological changes that occur during deceleration, acceleration and/or body compression in traffic accidents. The various forms of BAT, from limited laceration of the intima to complete transection of the aorta, depends on the force acting on the vessel wall and the tolerance of the aorta to injury. The force depends on the change in velocity, the dynamics of the accident and of the seating position in the car. Tolerance to aortic injury depends on the anatomy, histological structure and pathomorphological alterations due to aging or disease of the aortic wall. An overview of the literature and medical documentation reveals that different terms are used to describe certain forms of BAT, which can lead to misinterpretation of findings or diagnoses. We therefore, propose a classification that would enable uniform systematic screening of all forms of BAT. We have classified BAT into three morphologycal types: TYPE I (intramural), TYPE II (transmural) and TYPE III (multiple) aortic ruptures with appropriate subtypes. Methods: All car accident casualties examined at the Institute of Forensic Medicine from 2001 to 2009 were included in this retrospective study. Autopsy reports were used to determine the occurrence of each morphological type of BAT in deceased drivers, front seat passengers and other passengers in cars and to define the morphology of BAT in relation to the accident dynamics and the age of the fatalities. Results: A total of 391 fatalities in car accidents were included in the study. TYPE I, TYPE II and TYPE III BAT were observed in 10,9%, 55,6% and 33,5%, respectively. The incidence of BAT in drivers, front seat and other passengers was 36,7%, 43,1% and 28,6%, respectively. In frontal collisions, the incidence of BAT was 32,7%, in lateral collisions 54,2%, and in other traffic accidents 29,3%. The average age of fatalities with BAT was 42,8 years and of those without BAT 39,1 years. Conclusion: Identification and early recognition of the risk factors of BAT following a traffic accident is crucial for successful treatment of patients with BAT. Front seat passengers over 50 years of age who have been injured in a lateral collision are the most at risk of BAT.

A Robust Method for Hand Tracking Using Mean-shift Algorithm and Kalman Filter in Stereo Color Image Sequences

Real-time hand tracking is a challenging task in many computer vision applications such as gesture recognition. This paper proposes a robust method for hand tracking in a complex environment using Mean-shift analysis and Kalman filter in conjunction with 3D depth map. The depth information solve the overlapping problem between hands and face, which is obtained by passive stereo measuring based on cross correlation and the known calibration data of the cameras. Mean-shift analysis uses the gradient of Bhattacharyya coefficient as a similarity function to derive the candidate of the hand that is most similar to a given hand target model. And then, Kalman filter is used to estimate the position of the hand target. The results of hand tracking, tested on various video sequences, are robust to changes in shape as well as partial occlusion.