Peer-to-Peer Epidemic Algorithms for Reliable Multicasting in Ad Hoc Networks

Characteristics of ad hoc networks and even their existence depend on the nodes forming them. Thus, services and applications designed for ad hoc networks should adapt to this dynamic and distributed environment. In particular, multicast algorithms having reliability and scalability requirements should abstain from centralized approaches. We aspire to define a reliable and scalable multicast protocol for ad hoc networks. Our target is to utilize epidemic techniques for this purpose. In this paper, we present a brief survey of epidemic algorithms for reliable multicasting in ad hoc networks, and describe formulations and analytical results for simple epidemics. Then, P2P anti-entropy algorithm for content distribution and our prototype simulation model are described together with our initial results demonstrating the behavior of the algorithm.

Analyzing Periurban Fringe with Rough Set

The distinction among urban, periurban and rural areas represents a classical example of uncertainty in land classification. Satellite images, geostatistical analysis and all kinds of spatial data are very useful in urban sprawl studies, but it is important to define precise rules in combining great amounts of data to build complex knowledge about territory. Rough Set theory may be a useful method to employ in this field. It represents a different mathematical approach to uncertainty by capturing the indiscernibility. Two different phenomena can be indiscernible in some contexts and classified in the same way when combining available information about them. This approach has been applied in a case of study, comparing the results achieved with both Map Algebra technique and Spatial Rough Set. The study case area, Potenza Province, is particularly suitable for the application of this theory, because it includes 100 municipalities with different number of inhabitants and morphologic features.

Predicting the Impact of the Defect on the Overall Environment in Function Based Systems

There is lot of work done in prediction of the fault proneness of the software systems. But, it is the severity of the faults that is more important than number of faults existing in the developed system as the major faults matters most for a developer and those major faults needs immediate attention. In this paper, we tried to predict the level of impact of the existing faults in software systems. Neuro-Fuzzy based predictor models is applied NASA-s public domain defect dataset coded in C programming language. As Correlation-based Feature Selection (CFS) evaluates the worth of a subset of attributes by considering the individual predictive ability of each feature along with the degree of redundancy between them. So, CFS is used for the selecting the best metrics that have highly correlated with level of severity of faults. The results are compared with the prediction results of Logistic Models (LMT) that was earlier quoted as the best technique in [17]. The results are recorded in terms of Accuracy, Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). The results show that Neuro-fuzzy based model provide a relatively better prediction accuracy as compared to other models and hence, can be used for the modeling of the level of impact of faults in function based systems.

Improved Approximation to the Derivative of a Digital Signal Using Wavelet Transforms for Crosstalk Analysis

The information revealed by derivatives can help to better characterize digital near-end crosstalk signatures with the ultimate goal of identifying the specific aggressor signal. Unfortunately, derivatives tend to be very sensitive to even low levels of noise. In this work we approximated the derivatives of both quiet and noisy digital signals using a wavelet-based technique. The results are presented for Gaussian digital edges, IBIS Model digital edges, and digital edges in oscilloscope data captured from an actual printed circuit board. Tradeoffs between accuracy and noise immunity are presented. The results show that the wavelet technique can produce first derivative approximations that are accurate to within 5% or better, even under noisy conditions. The wavelet technique can be used to calculate the derivative of a digital signal edge when conventional methods fail.

An Ant-based Clustering System for Knowledge Discovery in DNA Chip Analysis Data

Biological data has several characteristics that strongly differentiate it from typical business data. It is much more complex, usually large in size, and continuously changes. Until recently business data has been the main target for discovering trends, patterns or future expectations. However, with the recent rise in biotechnology, the powerful technology that was used for analyzing business data is now being applied to biological data. With the advanced technology at hand, the main trend in biological research is rapidly changing from structural DNA analysis to understanding cellular functions of the DNA sequences. DNA chips are now being used to perform experiments and DNA analysis processes are being used by researchers. Clustering is one of the important processes used for grouping together similar entities. There are many clustering algorithms such as hierarchical clustering, self-organizing maps, K-means clustering and so on. In this paper, we propose a clustering algorithm that imitates the ecosystem taking into account the features of biological data. We implemented the system using an Ant-Colony clustering algorithm. The system decides the number of clusters automatically. The system processes the input biological data, runs the Ant-Colony algorithm, draws the Topic Map, assigns clusters to the genes and displays the output. We tested the algorithm with a test data of 100 to1000 genes and 24 samples and show promising results for applying this algorithm to clustering DNA chip data.

Fingerprint Compression Using Contourlet Transform and Multistage Vector Quantization

This paper presents a new fingerprint coding technique based on contourlet transform and multistage vector quantization. Wavelets have shown their ability in representing natural images that contain smooth areas separated with edges. However, wavelets cannot efficiently take advantage of the fact that the edges usually found in fingerprints are smooth curves. This issue is addressed by directional transforms, known as contourlets, which have the property of preserving edges. The contourlet transform is a new extension to the wavelet transform in two dimensions using nonseparable and directional filter banks. The computation and storage requirements are the major difficulty in implementing a vector quantizer. In the full-search algorithm, the computation and storage complexity is an exponential function of the number of bits used in quantizing each frame of spectral information. The storage requirement in multistage vector quantization is less when compared to full search vector quantization. The coefficients of contourlet transform are quantized by multistage vector quantization. The quantized coefficients are encoded by Huffman coding. The results obtained are tabulated and compared with the existing wavelet based ones.

Evaluation and Analysis of Lean-Based Manufacturing Equipment and Technology System for Jordanian Industries

International markets driven forces are changing continuously, therefore companies need to gain a competitive edge in such markets. Improving the company's products, processes and practices is no longer auxiliary. Lean production is a production management philosophy that consolidates work tasks with minimum waste resulting in improved productivity. Lean production practices can be mapped into many production areas. One of these is Manufacturing Equipment and Technology (MET). Many lean production practices can be implemented in MET, namely, specific equipment configurations, total preventive maintenance, visual control, new equipment/ technologies, production process reengineering and shared vision of perfection.The purpose of this paper is to investigate the implementation level of these six practices in Jordanian industries. To achieve that a questionnaire survey has been designed according to five-point Likert scale. The questionnaire is validated through pilot study and through experts review. A sample of 350 Jordanian companies were surveyed, the response rate was 83%. The respondents were asked to rate the extent of implementation for each of practices. A relationship conceptual model is developed, hypotheses are proposed, and consequently the essential statistical analyses are then performed. An assessment tool that enables management to monitor the progress and the effectiveness of lean practices implementation is designed and presented. Consequently, the results show that the average implementation level of lean practices in MET is 77%, Jordanian companies are implementing successfully the considered lean production practices, and the presented model has Cronbach-s alpha value of 0.87 which is good evidence on model consistency and results validation.

Generating Concept Trees from Dynamic Self-organizing Map

Self-organizing map (SOM) provides both clustering and visualization capabilities in mining data. Dynamic self-organizing maps such as Growing Self-organizing Map (GSOM) has been developed to overcome the problem of fixed structure in SOM to enable better representation of the discovered patterns. However, in mining large datasets or historical data the hierarchical structure of the data is also useful to view the cluster formation at different levels of abstraction. In this paper, we present a technique to generate concept trees from the GSOM. The formation of tree from different spread factor values of GSOM is also investigated and the quality of the trees analyzed. The results show that concept trees can be generated from GSOM, thus, eliminating the need for re-clustering of the data from scratch to obtain a hierarchical view of the data under study.

Comparison of Vermicompost and Vermiwash Bio-Fertilizers from Vermicomposting Waste Corn Pulp

Vermicomposting is the conversion of organic waste into bio-fertilizers through the action of earthworm. This technology is widely used for organic solid waste management. Waste corn pulp blended with cow dung manure was vermicomposted over 30 days using Eisenia fetida earthworms species. pH, temperature, moisture content, and electrical conductivity were daily monitored. The feedstock, vermicompost and vermiwash were analyzed for nutrient composition. The average temperature and moisture content in the vermi-reactor was 22.5°C and 42.5% respectively. The vermicompost and vermiwash had an almost neutral pH whilst the electrical conductivity was 21% higher in the vermicompost. The nitrogen and potassium content was 57% and 79.6% richer in the vermicompost respectively compared to the vermiwash. However, the vermiwash was 84% richer in phosphorous as compared to vermicompost. Furthermore, the vermiwash was 89.1% and 97.6% richer in Ca and Mg respectively and was 97.8% richer in Na salts compared to the vermicompost. The vermiwash also indicated a significantly higher amount of micronutrients. Both bio-fertilizers were rich in nutrients specification for fertilizers.

Face Recognition: A Literature Review

The task of face recognition has been actively researched in recent years. This paper provides an up-to-date review of major human face recognition research. We first present an overview of face recognition and its applications. Then, a literature review of the most recent face recognition techniques is presented. Description and limitations of face databases which are used to test the performance of these face recognition algorithms are given. A brief summary of the face recognition vendor test (FRVT) 2002, a large scale evaluation of automatic face recognition technology, and its conclusions are also given. Finally, we give a summary of the research results.

Hospital Based Electrocardiogram Sensor Grid

The technological concepts such as wireless hospital and portable cardiac telemetry system require the development of physiological signal acquisition devices to be easily integrated into the hospital database. In this paper we present the low cost, portable wireless ECG acquisition hardware that transmits ECG signals to a dedicated computer.The front end of the system obtains and processes incoming signals, which are then transmitted via a microcontroller and wireless Bluetooth module. A monitoring purpose Bluetooth based end user application integrated with patient database management module is developed for the computers. The system will act as a continuous event recorder, which can be used to follow up patients who have been resuscitatedfrom cardiac arrest, ventricular tachycardia but also for diagnostic purposes for patients with arrhythmia symptoms. In addition, cardiac information can be saved into the patient-s database of the hospital.

Extraction of Knowledge Complexity in 3G Killer Application Construction for Telecommunications National Strategy

We review a knowledge extractor model in constructing 3G Killer Applications. The success of 3G is essential for Government as it became part of Telecommunications National Strategy. The 3G wireless technologies may reach larger area and increase country-s ICT penetration. In order to understand future customers needs, the operators require proper information (knowledge) lying inside. Our work approached future customers as complex system where the complex knowledge may expose regular behavior. The hidden information from 3G future customers is revealed by using fractal-based questionnaires. Afterward, further statistical analysis is used to match the results with operator-s strategic plan. The developments of 3G applications also consider its saturation time and further improvement of the application.

Cross Signal Identification for PSG Applications

The standard investigational method for obstructive sleep apnea syndrome (OSAS) diagnosis is polysomnography (PSG), which consists of a simultaneous, usually overnight recording of multiple electro-physiological signals related to sleep and wakefulness. This is an expensive, encumbering and not a readily repeated protocol, and therefore there is need for simpler and easily implemented screening and detection techniques. Identification of apnea/hypopnea events in the screening recordings is the key factor for the diagnosis of OSAS. The analysis of a solely single-lead electrocardiographic (ECG) signal for OSAS diagnosis, which may be done with portable devices, at patient-s home, is the challenge of the last years. A novel artificial neural network (ANN) based approach for feature extraction and automatic identification of respiratory events in ECG signals is presented in this paper. A nonlinear principal component analysis (NLPCA) method was considered for feature extraction and support vector machine for classification/recognition. An alternative representation of the respiratory events by means of Kohonen type neural network is discussed. Our prospective study was based on OSAS patients of the Clinical Hospital of Pneumology from Iaşi, Romania, males and females, as well as on non-OSAS investigated human subjects. Our computed analysis includes a learning phase based on cross signal PSG annotation.

Adhesion Strength Evaluation Methods in Thermally Sprayed Coatings

The techniques for estimating the adhesive and cohesive strength in high velocity oxy fuel (HVOF) thermal spray coatings have been discussed and compared. The development trend and the last investigation have been studied. We will focus on benefits and limitations of these methods in different process and materials.

Are PEG Molecules a Universal Protein Repellent?

Poly (ethylene glycol) (PEG) molecules attached to surfaces have shown high potential as a protein repellent due to their flexibility and highly water solubility. A quartz crystal microbalance recording frequency and dissipation changes (QCM-D) has been used to study the adsorption from aqueous solutions, of lysozyme and α-lactalbumin proteins (the last with and without calcium) onto modified stainless steel surfaces. Surfaces were coated with poly(ethylene imine) (PEI) and silicate before grafting on PEG molecules. Protein adsorption was also performed on the bare stainless steel surface as a control. All adsorptions were conducted at 23°C and pH 7.2. The results showed that the presence of PEG molecules significantly reduced the adsorption of lysozyme and α- lactalbumin (with calcium) onto the stainless steel surface. By contrast, and unexpected, PEG molecules enhanced the adsorption of α-lactalbumin (without calcium). It is suggested that the PEG -α- lactalbumin hydrophobic interaction plays a dominant role which leads to protein aggregation at the surface for this latter observation. The findings also lead to the general conclusion that PEG molecules are not a universal protein repellent. PEG-on-PEI surfaces were better at inhibiting the adsorption of lysozyme and α-lactalbumin (with calcium) than with PEG-on-silicate surfaces.

An Ontology Abstract Machine

As more people from non-technical backgrounds are becoming directly involved with large-scale ontology development, the focal point of ontology research has shifted from the more theoretical ontology issues to problems associated with the actual use of ontologies in real-world, large-scale collaborative applications. Recently the National Science Foundation funded a large collaborative ontology development project for which a new formal ontology model, the Ontology Abstract Machine (OAM), was developed to satisfy some unique functional and data representation requirements. This paper introduces the OAM model and the related algorithms that enable maintenance of an ontology that supports node-based user access. The successful software implementation of the OAM model and its subsequent acceptance by a large research community proves its validity and its real-world application value.

A Novel Approach for Protein Classification Using Fourier Transform

Discovering new biological knowledge from the highthroughput biological data is a major challenge to bioinformatics today. To address this challenge, we developed a new approach for protein classification. Proteins that are evolutionarily- and thereby functionally- related are said to belong to the same classification. Identifying protein classification is of fundamental importance to document the diversity of the known protein universe. It also provides a means to determine the functional roles of newly discovered protein sequences. Our goal is to predict the functional classification of novel protein sequences based on a set of features extracted from each protein sequence. The proposed technique used datasets extracted from the Structural Classification of Proteins (SCOP) database. A set of spectral domain features based on Fast Fourier Transform (FFT) is used. The proposed classifier uses multilayer back propagation (MLBP) neural network for protein classification. The maximum classification accuracy is about 91% when applying the classifier to the full four levels of the SCOP database. However, it reaches a maximum of 96% when limiting the classification to the family level. The classification results reveal that spectral domain contains information that can be used for classification with high accuracy. In addition, the results emphasize that sequence similarity measures are of great importance especially at the family level.

Managerial Styles of Asian Executives: The Case of Thailand

This research project is developed in order to study managerial styles of modern Thai executives. The thorough understanding will lead to continuous improvement and efficient performance of Thai business organizations. Regarding managerial skills, Thai executives focus heavily upon human skills. Also, the negotiator roles are most emphasis in their management. In addition, Thai executives pay most attention to the fundamental management principles including Harmony and Unity of Direction of the organizations. Moreover, the management techniques, consisting of Team work and Career Planning are of their main concern. Finally, Thai executives wish to enhance their firms- image and employees- morale through conducting the ethical and socially responsible activities. The major tactic deployed to stimulate employees- ethical behaviors and mindset is Code of Ethics development.

An Intelligent Optimization Model for Multi-objective Order Allocation Planning

This paper presents a multi-objective order allocation planning problem with the consideration of various real-world production features. A novel hybrid intelligent optimization model, integrating a multi-objective memetic optimization process, a Monte Carlo simulation technique and a heuristic pruning technique, is proposed to handle this problem. Experiments based on industrial data are conducted to validate the proposed model. Results show that (1) the proposed model can effectively solve the investigated problem by providing effective production decision-making solutions, which outperformsan NSGA-II-based optimization process and an industrial method.

Noise Removal from Surface Respiratory EMG Signal

The aim of this study was to remove the two principal noises which disturb the surface electromyography signal (Diaphragm). These signals are the electrocardiogram ECG artefact and the power line interference artefact. The algorithm proposed focuses on a new Lean Mean Square (LMS) Widrow adaptive structure. These structures require a reference signal that is correlated with the noise contaminating the signal. The noise references are then extracted : first with a noise reference mathematically constructed using two different cosine functions; 50Hz (the fundamental) function and 150Hz (the first harmonic) function for the power line interference and second with a matching pursuit technique combined to an LMS structure for the ECG artefact estimation. The two removal procedures are attained without the use of supplementary electrodes. These techniques of filtering are validated on real records of surface diaphragm electromyography signal. The performance of the proposed methods was compared with already conducted research results.