Architecture Integrating Wireless Body Area Networks with Web Services for Ubiquitous Healthcare Service Provisioning

Recent advancements in sensor technologies and Wireless Body Area Networks (WBANs) have led to the development of cost-effective healthcare devices which can be used to monitor and analyse a person-s physiological parameters from remote locations. These advancements provides a unique opportunity to overcome current healthcare challenges of low quality service provisioning, lack of easy accessibility to service varieties, high costs of services and increasing population of the elderly experienced globally. This paper reports on a prototype implementation of an architecture that seamlessly integrates Wireless Body Area Network (WBAN) with Web services (WS) to proactively collect physiological data of remote patients to recommend diagnostic services. Technologies based upon WBAN and WS can provide ubiquitous accessibility to a variety of services by allowing distributed healthcare resources to be massively reused to provide cost-effective services without individuals physically moving to the locations of those resources. In addition, these technologies can reduce costs of healthcare services by allowing individuals to access services to support their healthcare. The prototype uses WBAN body sensors implemented on arduino fio platforms to be worn by the patient and an android smart phone as a personal server. The physiological data are collected and uploaded through GPRS/internet to the Medical Health Server (MHS) to be analysed. The prototype monitors the activities, location and physiological parameters such as SpO2 and Heart Rate of the elderly and patients in rehabilitation. Medical practitioners would have real time access to the uploaded information through a web application.

A Digital Media e-Learning Training Strategy for Healthcare Employees: Cost effective Distance Learning by Collaborative offline / online Engagement and Assessment

Within the healthcare system, training and continued professional development although essential, can be effected by cost and logistical restraints due to the nature of healthcare provision e.g employee shift patterns, access to expertise, cost factors in releasing staff to attend training etc. The use of multimedia technology for the development of e-learning applications is also a major cost consideration for healthcare management staff, and this type of media whether optical or on line requires careful planning in order to remain inclusive of all staff with potentially varied access to multimedia computing. This paper discusses a project in which the use of DVD authoring technology has been successfully implemented to meet the needs of distance learning and user considerations, and is based on film production techniques and reduced product turnaround deadlines.

A New Load Frequency Controller based on Parallel Fuzzy PI with Conventional PD (FPI-PD)

The artificial intelligent controller in power system plays as most important rule for many applications such as system operation and its control specially Load Frequency Controller (LFC). The main objective of LFC is to keep the frequency and tie-line power close to their decidable bounds in case of disturbance. In this paper, parallel fuzzy PI adaptive with conventional PD technique for Load Frequency Control system was proposed. PSO optimization method used to optimize both of scale fuzzy PI and tuning of PD. Two equal interconnected power system areas were used as a test system. Simulation results show the effectiveness of the proposed controller compared with different PID and classical fuzzy PI controllers in terms of speed response and damping frequency.

Mathematical Modeling Experimental Approach of the Friction on the Tool-Chip Interface of Multicoated Carbide Turning Inserts

The importance of machining process in today-s industry requires the establishment of more practical approaches to clearly represent the intimate and severe contact on the tool-chipworkpiece interfaces. Mathematical models are developed using the measured force signals to relate each of the tool-chip friction components on the rake face to the operating cutting parameters in rough turning operation using multilayers coated carbide inserts. Nonlinear modeling proved to have high capability to detect the nonlinear functional variability embedded in the experimental data. While feedrate is found to be the most influential parameter on the friction coefficient and its related force components, both cutting speed and depth of cut are found to have slight influence. Greater deformed chip thickness is found to lower the value of friction coefficient as the sliding length on the tool-chip interface is reduced.

Effect of Heat-Moisture Treatment on the Formation and Properties of Resistant Starches From Mung Bean (Phaseolus radiatus) Starches

Mung bean starches were subjected to heat-moisture treatment (HMT) by different moisture contents (15%, 20%, 25%, 30% and 35%) at 120Ôäâ for 12h. The impact on the yields of resistant starch (RS), microstructure, physicochemical and functional properties was investigated. Compared to native starch, the RS content of heat-moisture treated starches increased significantly. The RS level of HMT-20 was the highest of all the starches. Birefringence was displayed clear at the center of native starch. For HMT starches, pronounced birefringence was exhibited on the periphery of starch granules; however, birefringence disappeared at the centre of some starch granules. The shape of HMT starches hadn-t been changed and the integrity of starch granules was preserved for all the conditions. Concavity could be observed on HMT starches under scanning electronic microscopy. After HMT, apparent amylose contents were increased and starch macromolecule was degraded in comparison with those of native starch. There was a reduction in swelling power on HMT starches, but the solubility of HMT starches was higher than that of native starch. Both of native and HMT starches showed A-type X-ray diffraction pattern. Furthermore, there is a higher intensity at the peak of 15.0 and 22.9 Å than those of native starch.

In Silico Analysis of Pax6 Interacting Proteins Indicates Missing Molecular Links in Development of Brain and Associated Disease

The PAX6, a transcription factor, is essential for the morphogenesis of the eyes, brain, pituitary and pancreatic islets. In rodents, the loss of Pax6 function leads to central nervous system defects, anophthalmia, and nasal hypoplasia. The haplo-insufficiency of Pax6 causes microphthalmia, aggression and other behavioral abnormalities. It is also required in brain patterning and neuronal plasticity. In human, heterozygous mutation of Pax6 causes loss of iris [aniridia], mental retardation and glucose intolerance. The 3- deletion in Pax6 leads to autism and aniridia. The phenotypes are variable in peneterance and expressivity. However, mechanism of function and interaction of PAX6 with other proteins during development and associated disease are not clear. It is intended to explore interactors of PAX6 to elucidated biology of PAX6 function in the tissues where it is expressed and also in the central regulatory pathway. This report describes In-silico approaches to explore interacting proteins of PAX6. The models show several possible proteins interacting with PAX6 like MITF, SIX3, SOX2, SOX3, IPO13, TRIM, and OGT. Since the Pax6 is a critical transcriptional regulator and master control gene of eye and brain development it might be interacting with other protein involved in morphogenesis [TGIF, TGF, Ras etc]. It is also presumed that matricelluar proteins [SPARC, thrombospondin-1 and osteonectin etc] are likely to interact during transport and processing of PAX6 and are somewhere its cascade. The proteins involved in cell survival and cell proliferation can also not be ignored.

Adequacy of Object-Oriented Framework System-Based Testing Techniques

An application framework provides a reusable design and implementation for a family of software systems. If the framework contains defects, the defects will be passed on to the applications developed from the framework. Framework defects are hard to discover at the time the framework is instantiated. Therefore, it is important to remove all defects before instantiating the framework. In this paper, two measures for the adequacy of an object-oriented system-based testing technique are introduced. The measures assess the usefulness and uniqueness of the testing technique. The two measures are applied to experimentally compare the adequacy of two testing techniques introduced to test objectoriented frameworks at the system level. The two considered testing techniques are the New Framework Test Approach and Testing Frameworks Through Hooks (TFTH). The techniques are also compared analytically in terms of their coverage power of objectoriented aspects. The comparison study results show that the TFTH technique is better than the New Framework Test Approach in terms of usefulness degree, uniqueness degree, and coverage power.

Comparison of Different Gas Turbine Inlet Air Cooling Methods

Gas turbine air inlet cooling is a useful method for increasing output for regions where significant power demand and highest electricity prices occur during the warm months. Inlet air cooling increases the power output by taking advantage of the gas turbine-s feature of higher mass flow rate when the compressor inlet temperature decreases. Different methods are available for reducing gas turbine inlet temperature. There are two basic systems currently available for inlet cooling. The first and most cost-effective system is evaporative cooling. Evaporative coolers make use of the evaporation of water to reduce the gas turbine-s inlet air temperature. The second system employs various ways to chill the inlet air. In this method, the cooling medium flows through a heat exchanger located in the inlet duct to remove heat from the inlet air. However, the evaporative cooling is limited by wet-bulb temperature while the chilling can cool the inlet air to temperatures that are lower than the wet bulb temperature. In the present work, a thermodynamic model of a gas turbine is built to calculate heat rate, power output and thermal efficiency at different inlet air temperature conditions. Computational results are compared with ISO conditions herein called "base-case". Therefore, the two cooling methods are implemented and solved for different inlet conditions (inlet temperature and relative humidity). Evaporative cooler and absorption chiller systems results show that when the ambient temperature is extremely high with low relative humidity (requiring a large temperature reduction) the chiller is the more suitable cooling solution. The net increment in the power output as a function of the temperature decrease for each cooling method is also obtained.

Static and Dynamic Complexity Analysis of Software Metrics

Software complexity metrics are used to predict critical information about reliability and maintainability of software systems. Object oriented software development requires a different approach to software complexity metrics. Object Oriented Software Metrics can be broadly classified into static and dynamic metrics. Static Metrics give information at the code level whereas dynamic metrics provide information on the actual runtime. In this paper we will discuss the various complexity metrics, and the comparison between static and dynamic complexity.

Designing a Rescue System for Earthquake-Stricken Area with the Aim of Facilitation and Accelerating Accessibilities (Case Study: City of Tehran)

Natural disasters, including earthquake, kill many people around the world every year. Society rescue actions, which start after the earthquake and are called LAST in abbreviation, include locating, access, stabilization and transportation. In the present article, we have studied the process of local accessibility to the injured and transporting them to health care centers. With regard the heavy traffic load due to earthquake, the destruction of connecting roads and bridges and the heavy debris in alleys and street, which put the lives of the injured and the people buried under the debris in danger, accelerating the rescue actions and facilitating the accessibilities are of great importance, obviously. Tehran, the capital of Iran, is among the crowded cities in the world and is the center of extensive economic, political, cultural and social activities. Tehran has a population of about 9.5 millions and because of the immigration of people from the surrounding cities. Furthermore, considering the fact that Tehran is located on two important and large faults, a 6 Richter magnitude earthquake in this city could lead to the greatest catastrophe during the entire human history. The present study is a kind of review and a major part of the required information for it, has been obtained from libraries all of the rescue vehicles around the world, including rescue helicopters, ambulances, fire fighting vehicles and rescue boats, and their applied technology, and also the robots specifically designed for the rescue system and the advantages and disadvantages of them, have been investigated. The studies show that there is a significant relationship between the rescue team-s arrival time at the incident zone and the number of saved people; so that, if the duration of burial under debris 30 minutes, the probability of survival is %99.3, after a day is %81, after 2days is %19 and after 5days is %7.4. The exiting transport systems all have some defects. If these defects are removed, more people could be saved each hour and the preparedness against natural disasters is increased. In this study, transport system has been designed for the rescue team and the injured; which could carry the rescue team to the incident zone and the injured to the health care centers. In addition, this system is able to fly in the air and move on the earth as well; so that the destruction of roads and the heavy traffic load could not prevent the rescue team from arriving early at the incident zone. The system also has the equipment required firebird for debris removing, optimum transport of the injured and first aid.

Polarization Modulation by free-Standing Asymmetric Hole Arrays

We theoretically demonstrate modulation of light polarization by a crossed rectangular hole array with asymmetric arm lengths. There are two waveguide modes that can modulate the x- and y- polarized incident waves independently. A specific structure is proposed to convert a left-hand incident wave to a right-hand outgoing wave by transmission.

Combined Simulated Annealing and Genetic Algorithm to Solve Optimization Problems

Combinatorial optimization problems arise in many scientific and practical applications. Therefore many researchers try to find or improve different methods to solve these problems with high quality results and in less time. Genetic Algorithm (GA) and Simulated Annealing (SA) have been used to solve optimization problems. Both GA and SA search a solution space throughout a sequence of iterative states. However, there are also significant differences between them. The GA mechanism is parallel on a set of solutions and exchanges information using the crossover operation. SA works on a single solution at a time. In this work SA and GA are combined using new technique in order to overcome the disadvantages' of both algorithms.

A New Iterative Method for Solving Nonlinear Equations

In this study, a new root-finding method for solving nonlinear equations is proposed. This method requires two starting values that do not necessarily bracketing a root. However, when the starting values are selected to be close to a root, the proposed method converges to the root quicker than the secant method. Another advantage over all iterative methods is that; the proposed method usually converges to two distinct roots when the given function has more than one root, that is, the odd iterations of this new technique converge to a root and the even iterations converge to another root. Some numerical examples, including a sine-polynomial equation, are solved by using the proposed method and compared with results obtained by the secant method; perfect agreements are found.

Recognition of Obstacles and Providing Different Guidelines and Promotion of Electronic Government in Iran

Electronic Government is one of the special concepts which has been performed successfully within recent decades. Electronic government is a digital, wall-free government with a virtual organization for presenting of online governmental services and further cooperation in different political/social activities. In order to have a successful implementation of electronic government strategy and benefiting from its complete potential and benefits and generally for establishment and applying of electronic government, it is necessary to have different infrastructures as the basics of electronic government with lack of which it is impossible to benefit from mentioned services. For this purpose, in this paper we have managed to recognize relevant obstacles for establishment of electronic government in Iran. All required data for recognition of obstacles were collected from statistical society of involved specialists of Ministry of Communications & Information Technology of Iran and Information Technology Organization of Tehran Municipality through questionnaire. Then by considering of five-point Likert scope and μ =3 as the index of relevant factors of proposed model, we could specify current obstacles against electronic government in Iran along with some guidelines and proposal in this regard. According to the results, mentioned obstacles for applying of electronic government in Iran are as follows: Technical & technological problems, Legal, judicial & safety problems, Economic problems and Humanistic Problems.

A Mixture Model of Two Different Distributions Approach to the Analysis of Heterogeneous Survival Data

In this paper we propose a mixture of two different distributions such as Exponential-Gamma, Exponential-Weibull and Gamma-Weibull to model heterogeneous survival data. Various properties of the proposed mixture of two different distributions are discussed. Maximum likelihood estimations of the parameters are obtained by using the EM algorithm. Illustrative example based on real data are also given.

Hybrid Feature and Adaptive Particle Filter for Robust Object Tracking

A hybrid feature based adaptive particle filter algorithm is presented for object tracking in real scenarios with static camera. The hybrid feature is combined by two effective features: the Grayscale Arranging Pairs (GAP) feature and the color histogram feature. The GAP feature has high discriminative ability even under conditions of severe illumination variation and dynamic background elements, while the color histogram feature has high reliability to identify the detected objects. The combination of two features covers the shortage of single feature. Furthermore, we adopt an updating target model so that some external problems such as visual angles can be overcame well. An automatic initialization algorithm is introduced which provides precise initial positions of objects. The experimental results show the good performance of the proposed method.

A New Algorithm to Stereo Correspondence Using Rank Transform and Morphology Based On Genetic Algorithm

This paper presents a novel algorithm of stereo correspondence with rank transform. In this algorithm we used the genetic algorithm to achieve the accurate disparity map. Genetic algorithms are efficient search methods based on principles of population genetic, i.e. mating, chromosome crossover, gene mutation, and natural selection. Finally morphology is employed to remove the errors and discontinuities.

SAF: A Substitution and Alignment Free Similarity Measure for Protein Sequences

The literature reports a large number of approaches for measuring the similarity between protein sequences. Most of these approaches estimate this similarity using alignment-based techniques that do not necessarily yield biologically plausible results, for two reasons. First, for the case of non-alignable (i.e., not yet definitively aligned and biologically approved) sequences such as multi-domain, circular permutation and tandem repeat protein sequences, alignment-based approaches do not succeed in producing biologically plausible results. This is due to the nature of the alignment, which is based on the matching of subsequences in equivalent positions, while non-alignable proteins often have similar and conserved domains in non-equivalent positions. Second, the alignment-based approaches lead to similarity measures that depend heavily on the parameters set by the user for the alignment (e.g., gap penalties and substitution matrices). For easily alignable protein sequences, it's possible to supply a suitable combination of input parameters that allows such an approach to yield biologically plausible results. However, for difficult-to-align protein sequences, supplying different combinations of input parameters yields different results. Such variable results create ambiguities and complicate the similarity measurement task. To overcome these drawbacks, this paper describes a novel and effective approach for measuring the similarity between protein sequences, called SAF for Substitution and Alignment Free. Without resorting either to the alignment of protein sequences or to substitution relations between amino acids, SAF is able to efficiently detect the significant subsequences that best represent the intrinsic properties of protein sequences, those underlying the chronological dependencies of structural features and biochemical activities of protein sequences. Moreover, by using a new efficient subsequence matching scheme, SAF more efficiently handles protein sequences that contain similar structural features with significant meaning in chronologically non-equivalent positions. To show the effectiveness of SAF, extensive experiments were performed on protein datasets from different databases, and the results were compared with those obtained by several mainstream algorithms.

Effect of Scene Changing on Image Sequences Compression Using Zero Tree Coding

We study in this paper the effect of the scene changing on image sequences coding system using Embedded Zerotree Wavelet (EZW). The scene changing considered here is the full motion which may occurs. A special image sequence is generated where the scene changing occurs randomly. Two scenarios are considered: In the first scenario, the system must provide the reconstruction quality as best as possible by the management of the bit rate (BR) while the scene changing occurs. In the second scenario, the system must keep the bit rate as constant as possible by the management of the reconstruction quality. The first scenario may be motivated by the availability of a large band pass transmission channel where an increase of the bit rate may be possible to keep the reconstruction quality up to a given threshold. The second scenario may be concerned by the narrow band pass transmission channel where an increase of the bit rate is not possible. In this last case, applications for which the reconstruction quality is not a constraint may be considered. The simulations are performed with five scales wavelet decomposition using the 9/7-tap filter bank biorthogonal wavelet. The entropy coding is performed using a specific defined binary code book and EZW algorithm. Experimental results are presented and compared to LEAD H263 EVAL. It is shown that if the reconstruction quality is the constraint, the system increases the bit rate to obtain the required quality. In the case where the bit rate must be constant, the system is unable to provide the required quality if the scene change occurs; however, the system is able to improve the quality while the scene changing disappears.

Graphical Password Security Evaluation by Fuzzy AHP

In today's day and age, one of the important topics in information security is authentication. There are several alternatives to text-based authentication of which includes Graphical Password (GP) or Graphical User Authentication (GUA). These methods stems from the fact that humans recognized and remembers images better than alphanumerical text characters. This paper will focus on the security aspect of GP algorithms and what most researchers have been working on trying to define these security features and attributes. The goal of this study is to develop a fuzzy decision model that allows automatic selection of available GP algorithms by taking into considerations the subjective judgments of the decision makers who are more than 50 postgraduate students of computer science. The approach that is being proposed is based on the Fuzzy Analytic Hierarchy Process (FAHP) which determines the criteria weight as a linear formula.