Development for the Evaluation Index of an Anesthesia Depth using the Bispectrum Analysis

The linear SEF (Spectral Edge Frequency) parameter and spectrum analysis method can not reflect the non-linear of EEG. This method can not contribute to acquire real time analysis and obtain a high confidence in the clinic due to low discrimination. To solve the problems, the development of a new index is carried out using the bispectrum analyzing the EEG(electroencephalogram) including the non-linear characteristic. After analyzing the bispectrum of the 2 dimension, the most significant power spectrum density peaks appeared abundantly at the specific area in awakening and anesthesia state. These points are utilized to create the new index since many peaks appeared at the specific area in the frequency coordinate. The measured range of an index was 0-100. An index is 20-50 at an anesthesia, while the index is 90-60 at the awake. New index could afford to effectively discriminate the awake and anesthesia state.

Unscented Transformation for Estimating the Lyapunov Exponents of Chaotic Time Series Corrupted by Random Noise

Many systems in the natural world exhibit chaos or non-linear behavior, the complexity of which is so great that they appear to be random. Identification of chaos in experimental data is essential for characterizing the system and for analyzing the predictability of the data under analysis. The Lyapunov exponents provide a quantitative measure of the sensitivity to initial conditions and are the most useful dynamical diagnostic for chaotic systems. However, it is difficult to accurately estimate the Lyapunov exponents of chaotic signals which are corrupted by a random noise. In this work, a method for estimation of Lyapunov exponents from noisy time series using unscented transformation is proposed. The proposed methodology was validated using time series obtained from known chaotic maps. In this paper, the objective of the work, the proposed methodology and validation results are discussed in detail.

Application of Neural Network in User Authentication for Smart Home System

Security has been an important issue and concern in the smart home systems. Smart home networks consist of a wide range of wired or wireless devices, there is possibility that illegal access to some restricted data or devices may happen. Password-based authentication is widely used to identify authorize users, because this method is cheap, easy and quite accurate. In this paper, a neural network is trained to store the passwords instead of using verification table. This method is useful in solving security problems that happened in some authentication system. The conventional way to train the network using Backpropagation (BPN) requires a long training time. Hence, a faster training algorithm, Resilient Backpropagation (RPROP) is embedded to the MLPs Neural Network to accelerate the training process. For the Data Part, 200 sets of UserID and Passwords were created and encoded into binary as the input. The simulation had been carried out to evaluate the performance for different number of hidden neurons and combination of transfer functions. Mean Square Error (MSE), training time and number of epochs are used to determine the network performance. From the results obtained, using Tansig and Purelin in hidden and output layer and 250 hidden neurons gave the better performance. As a result, a password-based user authentication system for smart home by using neural network had been developed successfully.

Concept Abduction in Description Logics with Cardinality Restrictions

Recently the usefulness of Concept Abduction, a novel non-monotonic inference service for Description Logics (DLs), has been argued in the context of ontology-based applications such as semantic matchmaking and resource retrieval. Based on tableau calculus, a method has been proposed to realize this reasoning task in ALN, a description logic that supports simple cardinality restrictions as well as other basic constructors. However, in many ontology-based systems, the representation of ontology would require expressive formalisms for capturing domain-specific constraints, this language is not sufficient. In order to increase the applicability of the abductive reasoning method in such contexts, we would like to present in the scope of this paper an extension of the tableaux-based algorithm for dealing with concepts represented inALCQ, the description logic that extends ALN with full concept negation and quantified number restrictions.

Mean-Square Performance of Adaptive Filter Algorithms in Nonstationary Environments

Employing a recently introduced unified adaptive filter theory, we show how the performance of a large number of important adaptive filter algorithms can be predicted within a general framework in nonstationary environment. This approach is based on energy conservation arguments and does not need to assume a Gaussian or white distribution for the regressors. This general performance analysis can be used to evaluate the mean square performance of the Least Mean Square (LMS) algorithm, its normalized version (NLMS), the family of Affine Projection Algorithms (APA), the Recursive Least Squares (RLS), the Data-Reusing LMS (DR-LMS), its normalized version (NDR-LMS), the Block Least Mean Squares (BLMS), the Block Normalized LMS (BNLMS), the Transform Domain Adaptive Filters (TDAF) and the Subband Adaptive Filters (SAF) in nonstationary environment. Also, we establish the general expressions for the steady-state excess mean square in this environment for all these adaptive algorithms. Finally, we demonstrate through simulations that these results are useful in predicting the adaptive filter performance.

A New Approach for Counting Passersby Utilizing Space-Time Images

Understanding the number of people and the flow of the persons is useful for efficient promotion of the institution managements and company-s sales improvements. This paper introduces an automated method for counting passerby using virtualvertical measurement lines. The process of recognizing a passerby is carried out using an image sequence obtained from the USB camera. Space-time image is representing the human regions which are treated using the segmentation process. To handle the problem of mismatching, different color space are used to perform the template matching which chose automatically the best matching to determine passerby direction and speed. A relation between passerby speed and the human-pixel area is used to distinguish one or two passersby. In the experiment, the camera is fixed at the entrance door of the hall in a side viewing position. Finally, experimental results verify the effectiveness of the presented method by correctly detecting and successfully counting them in order to direction with accuracy of 97%.

Multirate Neural Control for AUV's Increased Situational Awareness during Diving Tasks Using Stochastic Model

This paper focuses on a critical component of the situational awareness (SA), the neural control of depth flight of an autonomous underwater vehicle (AUV). Constant depth flight is a challenging but important task for AUVs to achieve high level of autonomy under adverse conditions. With the SA strategy, we proposed a multirate neural control of an AUV trajectory for a nontrivial mid-small size AUV “r2D4" stochastic model. This control system has been demonstrated and evaluated by simulation of diving maneuvers using software package Simulink. From the simulation results it can be seen that the chosen AUV model is stable in the presence of noises, and also can be concluded that the proposed research technique will be useful for fast SA of similar AUV systems in real-time search-and-rescue operations.

Developing and Implementing Successful Key Performance Indicators

Measurement and the following evaluation of performance represent important part of management. The paper focuses on indicators as the basic elements of performance measurement system. It emphasizes a necessity of searching requirements for quality indicators so that they can become part of the useful system. It introduces standpoints for a systematic dividing of indicators so that they have as high as possible informative value of background sources for searching, analysis, designing and using of indicators. It draws attention to requirements for indicators' quality and at the same it deals with some dangers decreasing indicator's informative value. It submits a draft of questions that should be answered at the construction of indicator. It is obvious that particular indicators need to be defined exactly to stimulate the desired behavior in order to attain expected results. In the enclosure a concrete example of the defined indicator in the concrete conditions of a small firm is given. The authors of the paper pay attention to the fact that a quality indicator makes it possible to get to the basic causes of the problem and include the established facts into the company information system. At the same time they emphasize that developing of a quality indicator is a prerequisite for the utilization of the system of measurement in management.

Enhancing Seamless Communication Through a user Co-designed Wearable Device

This work aims to describe the process of developing services and applications of seamless communication within a Telecom Italia long-term research project, which takes as central aim the design of a wearable communication device. In particular, the objective was to design a wrist phone integrated into everyday life of people in full transparency. The methodology used to design the wristwatch was developed through several subsequent steps also involving the Personas Layering Framework. The data collected in this phases have been very useful for designing an improved version of the first two concepts of wrist phone going to change aspects related to the four critical points expressed by the users.

Frame Texture Classification Method (FTCM) Applied on Mammograms for Detection of Abnormalities

Texture classification is an important image processing task with a broad application range. Many different techniques for texture classification have been explored. Using sparse approximation as a feature extraction method for texture classification is a relatively new approach, and Skretting et al. recently presented the Frame Texture Classification Method (FTCM), showing very good results on classical texture images. As an extension of that work the FTCM is here tested on a real world application as detection of abnormalities in mammograms. Some extensions to the original FTCM that are useful in some applications are implemented; two different smoothing techniques and a vector augmentation technique. Both detection of microcalcifications (as a primary detection technique and as a last stage of a detection scheme), and soft tissue lesions in mammograms are explored. All the results are interesting, and especially the results using FTCM on regions of interest as the last stage in a detection scheme for microcalcifications are promising.

Design Optimization of Cutting Parameters when Turning Inconel 718 with Cermet Inserts

Inconel 718, a nickel based super-alloy is an extensively used alloy, accounting for about 50% by weight of materials used in an aerospace engine, mainly in the gas turbine compartment. This is owing to their outstanding strength and oxidation resistance at elevated temperatures in excess of 5500 C. Machining is a requisite operation in the aircraft industries for the manufacture of the components especially for gas turbines. This paper is concerned with optimization of the surface roughness when turning Inconel 718 with cermet inserts. Optimization of turning operation is very useful to reduce cost and time for machining. The approach is based on Response Surface Method (RSM). In this work, second-order quadratic models are developed for surface roughness, considering the cutting speed, feed rate and depth of cut as the cutting parameters, using central composite design. The developed models are used to determine the optimum machining parameters. These optimized machining parameters are validated experimentally, and it is observed that the response values are in reasonable agreement with the predicted values.

Handover Strategies Challenges in Wireless ATM Networks

To support user mobility for a wireless network new mechanisms are needed and are fundamental, such as paging, location updating, routing, and handover. Also an important key feature is mobile QoS offered by the WATM. Several ATM network protocols should be updated to implement mobility management and to maintain the already ATM QoS over wireless ATM networks. A survey of the various schemes and types of handover is provided. Handover procedure allows guarantee the terminal connection reestablishment when it moves between areas covered by different base stations. It is useful to satisfy user radio link transfer without interrupting a connection. However, failure to offer efficient solutions will result in handover important packet loss, severe delays and degradation of QoS offered to the applications. This paper reviews the requirements, characteristics and open issues of wireless ATM, particularly with regard to handover. It introduces key aspects of WATM and mobility extensions, which are added in the fixed ATM network. We propose a flexible approach for handover management that will minimize the QoS deterioration. Functional entities of this flexible approach are discussed in order to achieve minimum impact on the connection quality when a MT crosses the BS.

Geostatistical Analysis and Mapping of Groundlevel Ozone in a Medium Sized Urban Area

Ground-level tropospheric ozone is one of the air pollutants of most concern. It is mainly produced by photochemical processes involving nitrogen oxides and volatile organic compounds in the lower parts of the atmosphere. Ozone levels become particularly high in regions close to high ozone precursor emissions and during summer, when stagnant meteorological conditions with high insolation and high temperatures are common. In this work, some results of a study about urban ozone distribution patterns in the city of Badajoz, which is the largest and most industrialized city in Extremadura region (southwest Spain) are shown. Fourteen sampling campaigns, at least one per month, were carried out to measure ambient air ozone concentrations, during periods that were selected according to favourable conditions to ozone production, using an automatic portable analyzer. Later, to evaluate the ozone distribution at the city, the measured ozone data were analyzed using geostatistical techniques. Thus, first, during the exploratory analysis of data, it was revealed that they were distributed normally, which is a desirable property for the subsequent stages of the geostatistical study. Secondly, during the structural analysis of data, theoretical spherical models provided the best fit for all monthly experimental variograms. The parameters of these variograms (sill, range and nugget) revealed that the maximum distance of spatial dependence is between 302-790 m and the variable, air ozone concentration, is not evenly distributed in reduced distances. Finally, predictive ozone maps were derived for all points of the experimental study area, by use of geostatistical algorithms (kriging). High prediction accuracy was obtained in all cases as cross-validation showed. Useful information for hazard assessment was also provided when probability maps, based on kriging interpolation and kriging standard deviation, were produced.

A Green Chemical Technique for the Synthesis of Magnetic Nanoparticles by Magnetotactic Bacteria

Bacterial magnetic nanoparticles have great useful potential in biotechnological and biomedical applications. In this study, a liquid growth medium was modified for cultivation a fastidious magnetotactic bacterium that has been isolated from Anzali lagoon, Iran in our previous research. These modifications include change in vitamin, mineral, carbon sources and etcetera. In our experience, the serum bottles and designed air-tight laboratory bottles were used to create microaerobic conditions in order to development of a method for scale-up experiment. This information may serve as a guide to green chemistry based biological protocols for the synthesis of magnetic nanoparticles with control over the chemical composition, morphology and size.

The Relationship between Personality Characteristics and Driving Behavior

The present study investigated the relationship between personality characteristics of drivers and the number and amount of fines they have in a year .This study was carried out on 120 male taxi drivers that worked at least seven hours in a day in Lamerd - a city in the south of IRAN. Subjects were chosen voluntarily among those available. Predictive variables were the NEO –five great personality factors (1. conscientiousness 2. Openness to Experience 3.Neuroticism4 .Extraversion 5.Agreeableness ) thecriterion variables were the number and amount of fines the drivers have had the last three years. the result of regression analysis showed that conscientiousness factor was able to negatively predict the number and amount of financial fines the drivers had during the last three years. The openness factor positively predicted the number of fines they had in last 3 years and the amount of financial fines during the last year. The extraversion factor both meaningfully and positively could predict only the amount of financial fines they had during the last year. Increasing age was associated with decreasing driving offenses as well as financial loss.The findings can be useful in recognizing the high-risk drivers and leading them to counseling centers .They can also be used to inform the drivers about their personality and it’s relation with their accident rate. Such criteria would be of great importance in employing drivers in different places such as companies, offices etc…

Pseudo Last Useful Instant Queuing Strategy for Handovers in Low Earth Orbit Mobile Satellite Networks

This paper presents an alternative strategy of queuing handover called Pseudo Last Useful Instant PLUI scheme for Low Earth Orbit Mobile Satellite Systems LEO MSSs. The PLUI scheme uses the same approach as the Last Useful Instant LUI scheme previously proposed in literature, with less complex implementation. Simulation tests were carried out using Dynamic Channel Allocation DCA in order to evaluate the performance of this scheme and also an analytical approach has been presented to allow the performance evaluation of Fixed Channel Allocation FCA, with different handover queuing disciplines. The results show that performances achieved by the proposed strategy are close to those achieved using the LUI scheme.

Widening Students Perspective: Empowering Them with Systems Methodologies

Benefits to the organisation are just as important as technical ability when it comes to software success. The challenge is to provide industry with professionals who understand this. In other words: How to teach computer engineering students to look beyond technology, and at the benefits of software to organizations? This paper reports on the conceptual design of a section of the computer networks module aimed to sensitize the students to the organisational context. Checkland focuses on different worldviews represented by various role players in the organisation. He developed the Soft Systems Methodology that guides purposeful action in organisations, while incorporating different worldviews in the modeling process. If we can sensitize students to these methods, they are likely to appreciate the wider context of application of system software. This paper will provide literature on these concepts as well as detail on how the students will be guided to adopt these concepts.

Error Rate Probability for Coded MQAM with MRC Diversity in the Presence of Cochannel Interferers over Nakagami-Fading Channels

Exact expressions for bit-error probability (BEP) for coherent square detection of uncoded and coded M-ary quadrature amplitude modulation (MQAM) using an array of antennas with maximal ratio combining (MRC) in a flat fading channel interference limited system in a Nakagami-m fading environment is derived. The analysis assumes an arbitrary number of independent and identically distributed Nakagami interferers. The results for coded MQAM are computed numerically for the case of (24,12) extended Golay code and compared with uncoded MQAM by plotting error probabilities versus average signal-to-interference ratio (SIR) for various values of order of diversity N, number of distinct symbols M, in order to examine the effect of cochannel interferers on the performance of the digital communication system. The diversity gains and net gains are also presented in tabular form in order to examine the performance of digital communication system in the presence of interferers, as the order of diversity increases. The analytical results presented in this paper are expected to provide useful information needed for design and analysis of digital communication systems with space diversity in wireless fading channels.

Correlation between Capacitance and Dissipation Factor used for Assessment of Stator Insulation

Measurements of capacitance C and dissipation factor tand of the stator insulation system provide useful information about internal defects within the insulation. The index k is defined as the proportionality constant between the changes at high voltage of capacitance DC and of the dissipation factor Dtand . DC and Dtand values were highly correlated when small flat defects were within the insulation and that correlation was lost in the presence of large narrow defects like electrical treeing. The discrimination between small and large defects is made resorting to partial discharge PD phase angle analysis. For the validation of the results, C and tand measurements were carried out in a 15MVA 4160V steam turbine turbogenerator placed in a sugar mill. In addition, laboratory test results obtained by other authors were analyzed jointly. In such laboratory tests, model coil bars subjected to thermal cycling resulted highly degraded and DC and Dtand values were not correlated. Thus, the index k could not be calculated.

Coefficient of Parentage for Crop Hybridization

Hybridization refers to the crossing breeding of two plants. Coefficient of Parentage (COP) is used by the plant breeders to determine the genetic diversity across various varieties so as to incorporate the useful characters of the two varieties to develop a new crop variety with particular useful characters. Genetic Diversity is the prerequisite for any cultivar development program. Genetic Diversity depends upon the pedigree information of the varieties based on particular levels. Pedigree refers to the parents of a particular variety at various levels. This paper discusses the searching and analyses of different possible pairs of varieties selected on the basis of morphological characters, Climatic conditions and Nutrients so as to obtain the most optimal pair that can produce the required crossbreed variety. An algorithm was developed to determine the coefficient of parentage (COP) between the selected wheat varieties. Dummy values were used wherever actual data was not available.