Biplot Analysis for Evaluation of Tolerance in Some Bean (Phaseolus vulgaris L.) Genotypes to Bean Common Mosaic Virus (BCMV)

The common bean is the most important grain legume for direct human consumption in the world and BCMV is one of the world's most serious bean diseases that can reduce yield and quality of harvested product. To determine the best tolerance index to BCMV and recognize tolerant genotypes, 2 experiments were conducted in field conditions. Twenty five common bean genotypes were sown in 2 separate RCB design with 3 replications under contamination and non-contamination conditions. On the basis of the results of indices correlations GMP, MP and HARM were determined as the most suitable tolerance indices. The results of principle components analysis indicated 2 first components totally explained 98.52% of variations among data. The first and second components were named potential yield and stress susceptible respectively. Based on the results of BCMV tolerance indices assessment and biplot analysis WA8563-4, WA8563-2 and Cardinal were the genotypes that exhibited potential seed yield under contamination and noncontamination conditions.

A Dynamic Composition of an Adaptive Course

The number of framework conceived for e-learning constantly increase, unfortunately the creators of learning materials and educational institutions engaged in e-formation adopt a “proprietor" approach, where the developed products (courses, activities, exercises, etc.) can be exploited only in the framework where they were conceived, their uses in the other learning environments requires a greedy adaptation in terms of time and effort. Each one proposes courses whose organization, contents, modes of interaction and presentations are unique for all learners, unfortunately the latter are heterogeneous and are not interested by the same information, but only by services or documents adapted to their needs. Currently the new tendency for the framework conceived for e-learning, is the interoperability of learning materials, several standards exist (DCMI (Dublin Core Metadata Initiative)[2], LOM (Learning Objects Meta data)[1], SCORM (Shareable Content Object Reference Model)[6][7][8], ARIADNE (Alliance of Remote Instructional Authoring and Distribution Networks for Europe)[9], CANCORE (Canadian Core Learning Resource Metadata Application Profiles)[3]), they converge all to the idea of learning objects. They are also interested in the adaptation of the learning materials according to the learners- profile. This article proposes an approach for the composition of courses adapted to the various profiles (knowledge, preferences, objectives) of learners, based on two ontologies (domain to teach and educational) and the learning objects.

Surface Flattening Assisted with 3D Mannequin Based On Minimum Energy

The topic of surface flattening plays a vital role in the field of computer aided design and manufacture. Surface flattening enables the production of 2D patterns and it can be used in design and manufacturing for developing a 3D surface to a 2D platform, especially in fashion design. This study describes surface flattening based on minimum energy methods according to the property of different fabrics. Firstly, through the geometric feature of a 3D surface, the less transformed area can be flattened on a 2D platform by geodesic. Then, strain energy that has accumulated in mesh can be stably released by an approximate implicit method and revised error function. In some cases, cutting mesh to further release the energy is a common way to fix the situation and enhance the accuracy of the surface flattening, and this makes the obtained 2D pattern naturally generate significant cracks. When this methodology is applied to a 3D mannequin constructed with feature lines, it enhances the level of computer-aided fashion design. Besides, when different fabrics are applied to fashion design, it is necessary to revise the shape of a 2D pattern according to the properties of the fabric. With this model, the outline of 2D patterns can be revised by distributing the strain energy with different results according to different fabric properties. Finally, this research uses some common design cases to illustrate and verify the feasibility of this methodology.

Canonical PSO based Nanorobot Control for Blood Vessel Repair

As nanotechnology advances, the use of nanotechnology for medical purposes in the field of nanomedicine seems more promising; the rise of nanorobots for medical diagnostics and treatments could be arriving in the near future. This study proposes a swarm intelligence based control mechanism for swarm nanorobots that operate as artificial platelets to search for wounds. The canonical particle swarm optimization algorithm is employed in this study. A simulation in the circulatory system is constructed and used for demonstrating the movement of nanorobots with essential characteristics to examine the performance of proposed control mechanism. The effects of three nanorobot capabilities including their perception range, maximum velocity and respond time are investigated. The results show that canonical particle swarm optimization can be used to control the early version nanorobots with simple behaviors and actions.

Using Data Mining Methodology to Build the Predictive Model of Gold Passbook Price

Gold passbook is an investing tool that is especially suitable for investors to do small investment in the solid gold. The gold passbook has the lower risk than other ways investing in gold, but its price is still affected by gold price. However, there are many factors can cause influences on gold price. Therefore, building a model to predict the price of gold passbook can both reduce the risk of investment and increase the benefits. This study investigates the important factors that influence the gold passbook price, and utilize the Group Method of Data Handling (GMDH) to build the predictive model. This method can not only obtain the significant variables but also perform well in prediction. Finally, the significant variables of gold passbook price, which can be predicted by GMDH, are US dollar exchange rate, international petroleum price, unemployment rate, whole sale price index, rediscount rate, foreign exchange reserves, misery index, prosperity coincident index and industrial index.

A Proposal of an Automatic Formatting Method for Transforming XML Data

PPX(Pretty Printer for XML) is a query language that offers a concise description method of formatting the XML data into HTML. In this paper, we propose a simple specification of formatting method that is a combination description of automatic layout operators and variables in the layout expression of the GENERATE clause of PPX. This method can automatically format irregular XML data included in a part of XML with layout decision rule that is referred to DTD. In the experiment, a quick comparison shows that PPX requires far less description compared to XSLT or XQuery programs doing same tasks.

Modeling and Simulating Reaction-Diffusion Systems with State-Dependent Diffusion Coefficients

The present models and simulation algorithms of intracellular stochastic kinetics are usually based on the premise that diffusion is so fast that the concentrations of all the involved species are homogeneous in space. However, recents experimental measurements of intracellular diffusion constants indicate that the assumption of a homogeneous well-stirred cytosol is not necessarily valid even for small prokaryotic cells. In this work a mathematical treatment of diffusion that can be incorporated in a stochastic algorithm simulating the dynamics of a reaction-diffusion system is presented. The movement of a molecule A from a region i to a region j of the space is represented as a first order reaction Ai k- ! Aj , where the rate constant k depends on the diffusion coefficient. The diffusion coefficients are modeled as function of the local concentration of the solutes, their intrinsic viscosities, their frictional coefficients and the temperature of the system. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the intrinsic reaction kinetics and diffusion dynamics. To demonstrate the method the simulation results of the reaction-diffusion system of chaperoneassisted protein folding in cytoplasm are shown.

An Evaluation of Sputum Smear Conversion and Haematological Parameter Alteration in Early Detection Period of New Pulmonary Tuberculosis (PTB) Patients

Sputum smear conversion after one month of antituberculosis therapy in new smear positive pulmonary tuberculosis patients (PTB+) is a vital indicator towards treatment success. The objective of this study is to determine the rate of sputum smear conversion in new PTB+ patients after one month under treatment of National Institute of Diseases of the Chest and Hospital (NIDCH). Analysis of sputum smear conversion was done by re-clinical examination with sputum smear microscopic test after one month. Socio-demographic and hematological parameters were evaluated to perceive the correlation with the disease status. Among all enrolled patients only 33.33% were available for follow up diagnosis and of them only 42.86% patients turned to smear negative. Probably this consequence is due to non-coherence to the proper disease management. 66.67% and 78.78% patients reported low haemoglobin and packed cell volume level respectively whereas 80% and 93.33% patients accounted accelerated platelet count and erythrocyte sedimentation rate correspondingly.

A Systems Approach to Gene Ranking from DNA Microarray Data of Cervical Cancer

In this paper we present a method for gene ranking from DNA microarray data. More precisely, we calculate the correlation networks, which are unweighted and undirected graphs, from microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to progression of the tumor. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth and, hence, indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.

Acoustic Noise Reduction in Single Phase SRM Drives by Random Switching Technique

It is known that if harmonic spectra are decreased, then acoustic noise also decreased. Hence, this paper deals with a new random switching strategy using DSP TMS320F2812 to decrease the harmonics spectra of single phase switched reluctance motor. The proposed method which combines random turn-on, turn-off angle technique and random pulse width modulation technique is shown. A harmonic spread factor (HSF) is used to evaluate the random modulation scheme. In order to confirm the effectiveness of the new method, the experimental results show that the harmonic intensity of output voltage for the proposed method is better than that for conventional methods.

Impact of Environmental Factors on Profit Efficiency of Rice Production: A Study in Vietnam-s Red River Delta

Environmental factors affect agriculture production productivity and efficiency resulted in changing of profit efficiency. This paper attempts to estimate the impacts of environmental factors to profitability of rice farmers in the Red River Delta of Vietnam. The dataset was extracted from 349 rice farmers using personal interviews. Both OLS and MLE trans-log profit functions were used in this study. Five production inputs and four environmental factors were included in these functions. The estimation of the stochastic profit frontier with a two-stage approach was used to measure profitability. The results showed that the profit efficiency was about 75% on the average and environmental factors change profit efficiency significantly beside farm specific characteristics. Plant disease, soil fertility, irrigation apply and water pollution were the four environmental factors cause profit loss in rice production. The result indicated that farmers should reduce household size, farm plots, apply row seeding technique and improve environmental factors to obtain high profit efficiency with special consideration is given for irrigation water quality improvement.

Temporally Coherent 3D Animation Reconstruction from RGB-D Video Data

We present a new method to reconstruct a temporally coherent 3D animation from single or multi-view RGB-D video data using unbiased feature point sampling. Given RGB-D video data, in form of a 3D point cloud sequence, our method first extracts feature points using both color and depth information. In the subsequent steps, these feature points are used to match two 3D point clouds in consecutive frames independent of their resolution. Our new motion vectors based dynamic alignement method then fully reconstruct a spatio-temporally coherent 3D animation. We perform extensive quantitative validation using novel error functions to analyze the results. We show that despite the limiting factors of temporal and spatial noise associated to RGB-D data, it is possible to extract temporal coherence to faithfully reconstruct a temporally coherent 3D animation from RGB-D video data.

The New Approach to Sustainability in the Design of Urban and Architectural Interiors – Elements of Composition Revised

Today we tend to go back to the past to our root relation to nature. Therefore in search of friendly spaces there are elements of natural environment introduced as elements of spatial composition. Though reinvented through the use of the new substance such as greenery, water etc. made possible by state of the art technologies, still, in principal, they remain the same. As a result, sustainable design, based upon the recognized means of composition in addition to the relation of architecture and urbanism vs. nature introduces a new aesthetical values into architectural and urban space.

Security Analysis of Password Hardened Multimodal Biometric Fuzzy Vault

Biometric techniques are gaining importance for personal authentication and identification as compared to the traditional authentication methods. Biometric templates are vulnerable to variety of attacks due to their inherent nature. When a person-s biometric is compromised his identity is lost. In contrast to password, biometric is not revocable. Therefore, providing security to the stored biometric template is very crucial. Crypto biometric systems are authentication systems, which blends the idea of cryptography and biometrics. Fuzzy vault is a proven crypto biometric construct which is used to secure the biometric templates. However fuzzy vault suffer from certain limitations like nonrevocability, cross matching. Security of the fuzzy vault is affected by the non-uniform nature of the biometric data. Fuzzy vault when hardened with password overcomes these limitations. Password provides an additional layer of security and enhances user privacy. Retina has certain advantages over other biometric traits. Retinal scans are used in high-end security applications like access control to areas or rooms in military installations, power plants, and other high risk security areas. This work applies the idea of fuzzy vault for retinal biometric template. Multimodal biometric system performance is well compared to single modal biometric systems. The proposed multi modal biometric fuzzy vault includes combined feature points from retina and fingerprint. The combined vault is hardened with user password for achieving high level of security. The security of the combined vault is measured using min-entropy. The proposed password hardened multi biometric fuzzy vault is robust towards stored biometric template attacks.

Performance Evaluation of an ANC-based Hybrid Algorithm for Multi-target Wideband Active Sonar Echolocation System

This paper evaluates performances of an adaptive noise cancelling (ANC) based target detection algorithm on a set of real test data supported by the Defense Evaluation Research Agency (DERA UK) for multi-target wideband active sonar echolocation system. The hybrid algorithm proposed is a combination of an adaptive ANC neuro-fuzzy scheme in the first instance and followed by an iterative optimum target motion estimation (TME) scheme. The neuro-fuzzy scheme is based on the adaptive noise cancelling concept with the core processor of ANFIS (adaptive neuro-fuzzy inference system) to provide an effective fine tuned signal. The resultant output is then sent as an input to the optimum TME scheme composed of twogauge trimmed-mean (TM) levelization, discrete wavelet denoising (WDeN), and optimal continuous wavelet transform (CWT) for further denosing and targets identification. Its aim is to recover the contact signals in an effective and efficient manner and then determine the Doppler motion (radial range, velocity and acceleration) at very low signal-to-noise ratio (SNR). Quantitative results have shown that the hybrid algorithm have excellent performance in predicting targets- Doppler motion within various target strength with the maximum false detection of 1.5%.

Control Chart Pattern Recognition Using Wavelet Based Neural Networks

Control chart pattern recognition is one of the most important tools to identify the process state in statistical process control. The abnormal process state could be classified by the recognition of unnatural patterns that arise from assignable causes. In this study, a wavelet based neural network approach is proposed for the recognition of control chart patterns that have various characteristics. The procedure of proposed control chart pattern recognizer comprises three stages. First, multi-resolution wavelet analysis is used to generate time-shape and time-frequency coefficients that have detail information about the patterns. Second, distance based features are extracted by a bi-directional Kohonen network to make reduced and robust information. Third, a back-propagation network classifier is trained by these features. The accuracy of the proposed method is shown by the performance evaluation with numerical results.

W-CAS: A Central Users Authentication and Authorization System for Enterprise Wide Web Applications

Centrally controlled authentication and authorization services can provide enterprise with an increase in security, more flexible access control solutions and an increased users' trust. By using redirections, users of all Web-based applications within an organization are authenticated at a single well known and secure Web site and using secure communication protocol. Users are first authenticated at the central server using their domain wide credentials before being redirected to a particular Web-based application. The central authentication server will then provide others with pertinence authorization related particulars and credentials of the authenticated user to the specific application. The trust between the clients and the server hosts is established by secure session keys exchange. Case- studies are provided to demonstrate the usefulness and flexibility of the proposed solution.

Fuzzy Relatives of the CLARANS Algorithm With Application to Text Clustering

This paper introduces new algorithms (Fuzzy relative of the CLARANS algorithm FCLARANS and Fuzzy c Medoids based on randomized search FCMRANS) for fuzzy clustering of relational data. Unlike existing fuzzy c-medoids algorithm (FCMdd) in which the within cluster dissimilarity of each cluster is minimized in each iteration by recomputing new medoids given current memberships, FCLARANS minimizes the same objective function minimized by FCMdd by changing current medoids in such away that that the sum of the within cluster dissimilarities is minimized. Computing new medoids may be effected by noise because outliers may join the computation of medoids while the choice of medoids in FCLARANS is dictated by the location of a predominant fraction of points inside a cluster and, therefore, it is less sensitive to the presence of outliers. In FCMRANS the step of computing new medoids in FCMdd is modified to be based on randomized search. Furthermore, a new initialization procedure is developed that add randomness to the initialization procedure used with FCMdd. Both FCLARANS and FCMRANS are compared with the robust and linearized version of fuzzy c-medoids (RFCMdd). Experimental results with different samples of the Reuter-21578, Newsgroups (20NG) and generated datasets with noise show that FCLARANS is more robust than both RFCMdd and FCMRANS. Finally, both FCMRANS and FCLARANS are more efficient and their outputs are almost the same as that of RFCMdd in terms of classification rate.

Analysis of a Mathematical Model for Dengue Disease in Pregnant Cases

Dengue fever is an important human arboviral disease. Outbreaks are now reported quite often from many parts of the world. The number of cases involving pregnant women and infant cases are increasing every year. The illness is often severe and complications may occur. Deaths often occur because of the difficulties in early diagnosis and in the improper management of the diseases. Dengue antibodies from pregnant women are passed on to infants and this protects the infants from dengue infections. Antibodies from the mother are transferred to the fetus when it is still in the womb. In this study, we formulate a mathematical model to describe the transmission of this disease in pregnant women. The model is formulated by dividing the human population into pregnant women and non-pregnant human (men and non-pregnant women). Each class is subdivided into susceptible (S), infectious (I) and recovered (R) subclasses. We apply standard dynamical analysis to our model. Conditions for the local stability of the equilibrium points are given. The numerical simulations are shown. The bifurcation diagrams of our model are discussed. The control of this disease in pregnant women is discussed in terms of the threshold conditions.

New Wavelet Indices to Assess Muscle Fatigue during Dynamic Contractions

The purpose of this study was to evaluate and compare new indices based on the discrete wavelet transform with another spectral parameters proposed in the literature as mean average voltage, median frequency and ratios between spectral moments applied to estimate acute exercise-induced changes in power output, i.e., to assess peripheral muscle fatigue during a dynamic fatiguing protocol. 15 trained subjects performed 5 sets consisting of 10 leg press, with 2 minutes rest between sets. Surface electromyography was recorded from vastus medialis (VM) muscle. Several surface electromyographic parameters were compared to detect peripheral muscle fatigue. These were: mean average voltage (MAV), median spectral frequency (Fmed), Dimitrov spectral index of muscle fatigue (FInsm5), as well as other five parameters obtained from the discrete wavelet transform (DWT) as ratios between different scales. The new wavelet indices achieved the best results in Pearson correlation coefficients with power output changes during acute dynamic contractions. Their regressions were significantly different from MAV and Fmed. On the other hand, they showed the highest robustness in presence of additive white gaussian noise for different signal to noise ratios (SNRs). Therefore, peripheral impairments assessed by sEMG wavelet indices may be a relevant factor involved in the loss of power output after dynamic high-loading fatiguing task.