Biplot Analysis for Evaluation of Tolerance in Some Bean (Phaseolus vulgaris L.) Genotypes to Bean Common Mosaic Virus (BCMV)

The common bean is the most important grain legume for direct human consumption in the world and BCMV is one of the world's most serious bean diseases that can reduce yield and quality of harvested product. To determine the best tolerance index to BCMV and recognize tolerant genotypes, 2 experiments were conducted in field conditions. Twenty five common bean genotypes were sown in 2 separate RCB design with 3 replications under contamination and non-contamination conditions. On the basis of the results of indices correlations GMP, MP and HARM were determined as the most suitable tolerance indices. The results of principle components analysis indicated 2 first components totally explained 98.52% of variations among data. The first and second components were named potential yield and stress susceptible respectively. Based on the results of BCMV tolerance indices assessment and biplot analysis WA8563-4, WA8563-2 and Cardinal were the genotypes that exhibited potential seed yield under contamination and noncontamination conditions.

Transmit Sub-aperture Optimization in MSTA Ultrasound Imaging Method

The paper presents the optimization problem for the multi-element synthetic transmit aperture method (MSTA) in ultrasound imaging applications. The optimal choice of the transmit aperture size is performed as a trade-off between the lateral resolution, penetration depth and the frame rate. Results of the analysis obtained by a developed optimization algorithm are presented. Maximum penetration depth and the best lateral resolution at given depths are chosen as the optimization criteria. The optimization algorithm was tested using synthetic aperture data of point reflectors simulated by Filed II program for Matlab® for the case of 5MHz 128-element linear transducer array with 0.48 mm pitch are presented. The visualization of experimentally obtained synthetic aperture data of a tissue mimicking phantom and in vitro measurements of the beef liver are also shown. The data were obtained using the SonixTOUCH Research systemequipped with a linear 4MHz 128 element transducerwith 0.3 mm element pitch, 0.28 mm element width and 70% fractional bandwidth was excited by one sine cycle pulse burst of transducer's center frequency.

A Dynamic Composition of an Adaptive Course

The number of framework conceived for e-learning constantly increase, unfortunately the creators of learning materials and educational institutions engaged in e-formation adopt a “proprietor" approach, where the developed products (courses, activities, exercises, etc.) can be exploited only in the framework where they were conceived, their uses in the other learning environments requires a greedy adaptation in terms of time and effort. Each one proposes courses whose organization, contents, modes of interaction and presentations are unique for all learners, unfortunately the latter are heterogeneous and are not interested by the same information, but only by services or documents adapted to their needs. Currently the new tendency for the framework conceived for e-learning, is the interoperability of learning materials, several standards exist (DCMI (Dublin Core Metadata Initiative)[2], LOM (Learning Objects Meta data)[1], SCORM (Shareable Content Object Reference Model)[6][7][8], ARIADNE (Alliance of Remote Instructional Authoring and Distribution Networks for Europe)[9], CANCORE (Canadian Core Learning Resource Metadata Application Profiles)[3]), they converge all to the idea of learning objects. They are also interested in the adaptation of the learning materials according to the learners- profile. This article proposes an approach for the composition of courses adapted to the various profiles (knowledge, preferences, objectives) of learners, based on two ontologies (domain to teach and educational) and the learning objects.

Surface Flattening Assisted with 3D Mannequin Based On Minimum Energy

The topic of surface flattening plays a vital role in the field of computer aided design and manufacture. Surface flattening enables the production of 2D patterns and it can be used in design and manufacturing for developing a 3D surface to a 2D platform, especially in fashion design. This study describes surface flattening based on minimum energy methods according to the property of different fabrics. Firstly, through the geometric feature of a 3D surface, the less transformed area can be flattened on a 2D platform by geodesic. Then, strain energy that has accumulated in mesh can be stably released by an approximate implicit method and revised error function. In some cases, cutting mesh to further release the energy is a common way to fix the situation and enhance the accuracy of the surface flattening, and this makes the obtained 2D pattern naturally generate significant cracks. When this methodology is applied to a 3D mannequin constructed with feature lines, it enhances the level of computer-aided fashion design. Besides, when different fabrics are applied to fashion design, it is necessary to revise the shape of a 2D pattern according to the properties of the fabric. With this model, the outline of 2D patterns can be revised by distributing the strain energy with different results according to different fabric properties. Finally, this research uses some common design cases to illustrate and verify the feasibility of this methodology.

The Effect of Soil Surface Slope on Splash Distribution under Water Drop Impact

The effects of down slope steepness on soil splash distribution under a water drop impact have been investigated in this study. The equipment used are the burette to simulate a water drop, a splash cup filled with sandy soil which forms the source area and a splash board to collect the ejected particles. The results found in this study have shown that the apparent mass increased with increasing downslope angle following a linear regression equation with high coefficient of determination. In the same way, the radial soil splash distribution over the distance has been analyzed statistically, and an exponential function was the best fit of the relationship for the different slope angles. The curves and the regressions equations validate the well known FSDF and extend the theory of Van Dijk.

Preparation of Nanosized Iron Oxide and their Photocatalytic Properties for Congo Red

Nanostructured Iron Oxide with different morphologies of rod-like and granular have been suc-cessfully prepared via a solid-state reaction in the presence of NaCl, NaBr, NaI and NaN3, respectively. The added salts not only prevent a drastic increase in the size of the products but also provide suitable conditions for the oriented growth of primary nanoparticles. The formation mechanisms of these materials by solid-state reaction at ambient temperature are proposed. The photocatalytic experiments for congo red (CR) have demonstrated that the mixture of α-Fe2O3 and Fe3O4 nanostructures were more efficient than α-Fe2O3 nanostructures.

Using Data Mining Methodology to Build the Predictive Model of Gold Passbook Price

Gold passbook is an investing tool that is especially suitable for investors to do small investment in the solid gold. The gold passbook has the lower risk than other ways investing in gold, but its price is still affected by gold price. However, there are many factors can cause influences on gold price. Therefore, building a model to predict the price of gold passbook can both reduce the risk of investment and increase the benefits. This study investigates the important factors that influence the gold passbook price, and utilize the Group Method of Data Handling (GMDH) to build the predictive model. This method can not only obtain the significant variables but also perform well in prediction. Finally, the significant variables of gold passbook price, which can be predicted by GMDH, are US dollar exchange rate, international petroleum price, unemployment rate, whole sale price index, rediscount rate, foreign exchange reserves, misery index, prosperity coincident index and industrial index.

Calculation of Wave Function at the Origin (WFO) for Heavy Mesons by Numerical Solving of the Schrodinger Equation

Many recent high energy physics calculations involving charm and beauty invoke wave function at the origin (WFO) for the meson bound state. Uncertainties of charm and beauty quark masses and different models for potentials governing these bound states require a simple numerical algorithm for evaluation of the WFO's for these bound states. We present a simple algorithm for this propose which provides WFO's with high precision compared with similar ones already obtained in the literature.

Linear-Operator Formalism in the Analysis of Omega Planar Layered Waveguides

A complete spectral representation for the electromagnetic field of planar multilayered waveguides inhomogeneously filled with omega media is presented. The problem of guided electromagnetic propagation is reduced to an eigenvalue equation related to a 2 ´ 2 matrix differential operator. Using the concept of adjoint waveguide, general bi-orthogonality relations for the hybrid modes (either from the discrete or from the continuous spectrum) are derived. For the special case of homogeneous layers the linear operator formalism is reduced to a simple 2 ´ 2 coupling matrix eigenvalue problem. Finally, as an example of application, the surface and the radiation modes of a grounded omega slab waveguide are analyzed.

A Proposal of an Automatic Formatting Method for Transforming XML Data

PPX(Pretty Printer for XML) is a query language that offers a concise description method of formatting the XML data into HTML. In this paper, we propose a simple specification of formatting method that is a combination description of automatic layout operators and variables in the layout expression of the GENERATE clause of PPX. This method can automatically format irregular XML data included in a part of XML with layout decision rule that is referred to DTD. In the experiment, a quick comparison shows that PPX requires far less description compared to XSLT or XQuery programs doing same tasks.

Using Ontology Search in the Design of Class Diagram from Business Process Model

Business process model describes process flow of a business and can be seen as the requirement for developing a software application. This paper discusses a BPM2CD guideline which complements the Model Driven Architecture concept by suggesting how to create a platform-independent software model in the form of a UML class diagram from a business process model. An important step is the identification of UML classes from the business process model. A technique for object-oriented analysis called domain analysis is borrowed and key concepts in the business process model will be discovered and proposed as candidate classes for the class diagram. The paper enhances this step by using ontology search to help identify important classes for the business domain. As ontology is a source of knowledge for a particular domain which itself can link to ontologies of related domains, the search can give a refined set of candidate classes for the resulting class diagram.

A Fast Code Acquisition Scheme for O-CDMA Systems

This paper proposes a fast code acquisition scheme for optical code division multiple access (O-CDMA) systems. Unlike the conventional scheme, the proposed scheme employs multiple thresholds providing a shorter mean acquisition time (MAT) performance. The simulation results show that the MAT of the proposed scheme is shorter than that of the conventional scheme.

Cell Growth and Metabolites Produced by Fluorescent Pseudomonad R62 in Modified Chemically Defined Medium

Chemically defined Schlegel-s medium was modified to improve production of cell growth and other metabolites that are produced by fluorescent pseudomonad R62 strain. The modified medium does not require pH control as pH changes are kept within ± 0.2 units of the initial pH 7.1 during fermentation. The siderophore production was optimized for the fluorescent pseudomonad strain in the modified medium containing 1% glycerol as a major carbon source supplemented with 0.05% succinic acid and 0.5% Ltryptophan. Indole-3 acetic acid (IAA) production was higher when L-tryptophan was used at 0.5%. The 2,4- diacetylphloroglucinol (DAPG) was higher with amended three trace elements in medium. The optimized medium produced 2.28 g/l of dry cell mass and 900 mg/l of siderophore at the end of 36 h cultivation, while the production levels of IAA and DAPG were 65 mg/l and 81 mg/l respectively at the end of 48 h cultivation.

The Adoption of Halal Transportations Technologies for Halal Logistics Service Providers in Malaysia

The purpose of this study is i) to investigate the driving factors and barriers of the adoption of Information and Communication Technology (ICT) in Halal logistic and ii) to develop an ICT adoption framework for Halal logistic service provider. The Halal LSPs selected for the study currently used ICT service platforms, such as accounting and management system for Halal logistic business. The study categorizes the factors influencing the adoption decision and process by LSPs into four groups: technology related factors, organizational and environmental factors, Halal assurance related factors, and government related factors. The major contribution in this study is the discovery that technology related factors (ICT compatibility with Halal requirement) and Halal assurance related factors are the most crucial factors among the Halal LSPs applying ICT for Halal control in transportation-s operation. Among the government related factors, ICT requirement for monitoring Halal included in Halal Logistic Standard on Transportation (MS2400:2010) are the most influencing factors in the adoption of ICT with the support of the government. In addition, the government related factors are very important in the reducing the main barriers and the creation of the atmosphere of ICT adoption in Halal LSP sector.

Robust Camera Calibration using Discrete Optimization

Camera calibration is an indispensable step for augmented reality or image guided applications where quantitative information should be derived from the images. Usually, a camera calibration is obtained by taking images of a special calibration object and extracting the image coordinates of projected calibration marks enabling the calculation of the projection from the 3d world coordinates to the 2d image coordinates. Thus such a procedure exhibits typical steps, including feature point localization in the acquired images, camera model fitting, correction of distortion introduced by the optics and finally an optimization of the model-s parameters. In this paper we propose to extend this list by further step concerning the identification of the optimal subset of images yielding the smallest overall calibration error. For this, we present a Monte Carlo based algorithm along with a deterministic extension that automatically determines the images yielding an optimal calibration. Finally, we present results proving that the calibration can be significantly improved by automated image selection.

Meta Model Based EA for Complex Optimization

Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, many real life optimization problems often require finding optimal solution to complex high dimensional, multimodal problems involving computationally very expensive fitness function evaluations. Use of evolutionary algorithms in such problem domains is thus practically prohibitive. An attractive alternative is to build meta models or use an approximation of the actual fitness functions to be evaluated. These meta models are order of magnitude cheaper to evaluate compared to the actual function evaluation. Many regression and interpolation tools are available to build such meta models. This paper briefly discusses the architectures and use of such meta-modeling tools in an evolutionary optimization context. We further present two evolutionary algorithm frameworks which involve use of meta models for fitness function evaluation. The first framework, namely the Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model [14] reduces computation time by controlled use of meta-models (in this case approximate model generated by Support Vector Machine regression) to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the metamodel are generated from a single uniform model. This does not take into account uncertain scenarios involving noisy fitness functions. The second model, DAFHEA-II, an enhanced version of the original DAFHEA framework, incorporates a multiple-model based learning approach for the support vector machine approximator to handle noisy functions [15]. Empirical results obtained by evaluating the frameworks using several benchmark functions demonstrate their efficiency

FT-IR Study of Stabilized PAN Fibers for Fabrication of Carbon Fibers

In this investigation, types of commercial and special polyacrylonitrile (PAN) fibers contain sodium 2-methyl-2- acrylamidopropane sulfonate (SAMPS) and itaconic acid (IA) comonomers were studied by fourier transform infrared (FT-IR) spectroscopy. The study of FT-IR spectra of PAN fibers samples with different comonomers shows that during stabilization of PAN fibers, the peaks related to C≡N bonds and CH2 are reduced sharply. These reductions are related to cyclization of nitrile groups and stabilization procedure. This reduction in PAN fibers contain IA comonomer is very intense in comparison with PAN fibers contain SAMPS comonomer. This fact indicates the cycling and stabilization for sample contain IA comonomer have been conducted more completely. Therefore the carbon fibers produced from this material have higher tensile strength due to suitable stabilization.

An Improved Learning Algorithm based on the Conjugate Gradient Method for Back Propagation Neural Networks

The conjugate gradient optimization algorithm usually used for nonlinear least squares is presented and is combined with the modified back propagation algorithm yielding a new fast training multilayer perceptron (MLP) algorithm (CGFR/AG). The approaches presented in the paper consist of three steps: (1) Modification on standard back propagation algorithm by introducing gain variation term of the activation function, (2) Calculating the gradient descent on error with respect to the weights and gains values and (3) the determination of the new search direction by exploiting the information calculated by gradient descent in step (2) as well as the previous search direction. The proposed method improved the training efficiency of back propagation algorithm by adaptively modifying the initial search direction. Performance of the proposed method is demonstrated by comparing to the conjugate gradient algorithm from neural network toolbox for the chosen benchmark. The results show that the number of iterations required by the proposed method to converge is less than 20% of what is required by the standard conjugate gradient and neural network toolbox algorithm.

Thermal and Mechanical Buckling of Short and Long Functionally Graded Cylindrical Shells Using First Order Shear Deformation Theory

This paper presents the buckling analysis of short and long functionally graded cylindrical shells under thermal and mechanical loads. The shell properties are assumed to vary continuously from the inner surface to the outer surface of the shell. The equilibrium and stability equations are derived using the total potential energy equations, Euler equations and first order shear deformation theory assumptions. The resulting equations are solved for simply supported boundary conditions. The critical temperature and pressure loads are calculated for both short and long cylindrical shells. Comparison studies show the effects of functionally graded index, loading type and shell geometry on critical buckling loads of short and long functionally graded cylindrical shells.

Online Computing System for Cctuple-Precision Computation with Fortran

Computations with higher than the IEEE 754 standard double-precision (about 16 significant digits) are required recently. Although there are available software routines in Fortran and C for high-precision computation, users are required to implement such routines in their own computers with detailed knowledges about them. We have constructed an user-friendly online system for octupleprecision computation. In our Web system users with no knowledges about high-precision computation can easily perform octupleprecision computations, by choosing mathematical functions with argument(s) inputted, by writing simple mathematical expression(s) or by uploading C program(s). In this paper we enhance the Web system above by adding the facility of uploading Fortran programs, which have been widely used in scientific computing. To this end we construct converter routines in two stages.