Sorting Primitives and Genome Rearrangementin Bioinformatics: A Unified Perspective

Bioinformatics and computational biology involve the use of techniques including applied mathematics, informatics, statistics, computer science, artificial intelligence, chemistry, and biochemistry to solve biological problems usually on the molecular level. Research in computational biology often overlaps with systems biology. Major research efforts in the field include sequence alignment, gene finding, genome assembly, protein structure alignment, protein structure prediction, prediction of gene expression and proteinprotein interactions, and the modeling of evolution. Various global rearrangements of permutations, such as reversals and transpositions,have recently become of interest because of their applications in computational molecular biology. A reversal is an operation that reverses the order of a substring of a permutation. A transposition is an operation that swaps two adjacent substrings of a permutation. The problem of determining the smallest number of reversals required to transform a given permutation into the identity permutation is called sorting by reversals. Similar problems can be defined for transpositions and other global rearrangements. In this work we perform a study about some genome rearrangement primitives. We show how a genome is modelled by a permutation, introduce some of the existing primitives and the lower and upper bounds on them. We then provide a comparison of the introduced primitives.

Two States Mapping Based Neural Network Model for Decreasing of Prediction Residual Error

The objective of this paper is to design a model of human vital sign prediction for decreasing prediction error by using two states mapping based time series neural network BP (back-propagation) model. Normally, lot of industries has been applying the neural network model by training them in a supervised manner with the error back-propagation algorithm for time series prediction systems. However, it still has a residual error between real value and prediction output. Therefore, we designed two states of neural network model for compensation of residual error which is possible to use in the prevention of sudden death and metabolic syndrome disease such as hypertension disease and obesity. We found that most of simulations cases were satisfied by the two states mapping based time series prediction model compared to normal BP. In particular, small sample size of times series were more accurate than the standard MLP model. We expect that this algorithm can be available to sudden death prevention and monitoring AGENT system in a ubiquitous homecare environment.

Study of a BVAR(p) Process Applied to U.S. Commodity Market Data

The paper presents an applied study of a multivariate AR(p) process fitted to daily data from U.S. commodity futures markets with the use of Bayesian statistics. In the first part a detailed description of the methods used is given. In the second part two BVAR models are chosen one with assumption of lognormal, the second with normal distribution of prices conditioned on the parameters. For a comparison two simple benchmark models are chosen that are commonly used in todays Financial Mathematics. The article compares the quality of predictions of all the models, tries to find an adequate rate of forgetting of information and questions the validity of Efficient Market Hypothesis in the semi-strong form.

Model-free Prediction based on Tracking Theory and Newton Form of Polynomial

The majority of existing predictors for time series are model-dependent and therefore require some prior knowledge for the identification of complex systems, usually involving system identification, extensive training, or online adaptation in the case of time-varying systems. Additionally, since a time series is usually generated by complex processes such as the stock market or other chaotic systems, identification, modeling or the online updating of parameters can be problematic. In this paper a model-free predictor (MFP) for a time series produced by an unknown nonlinear system or process is derived using tracking theory. An identical derivation of the MFP using the property of the Newton form of the interpolating polynomial is also presented. The MFP is able to accurately predict future values of a time series, is stable, has few tuning parameters and is desirable for engineering applications due to its simplicity, fast prediction speed and extremely low computational load. The performance of the proposed MFP is demonstrated using the prediction of the Dow Jones Industrial Average stock index.

Application of Seismic Wave Method in Early Estimation of Wencheng Earthquake

This paper introduces the application of seismic wave method in earthquake prediction and early estimation. The advantages of the seismic wave method over the traditional earthquake prediction method are demonstrated. An example is presented in this study to show the accuracy and efficiency of using the seismic wave method in predicting a medium-sized earthquake swarm occurred in Wencheng, Zhejiang, China. By applying this method, correct predictions were made on the day after this earthquake swarm started and the day the maximum earthquake occurred, which provided scientific bases for governmental decision-making.

System Reliability by Prediction of Generator Output and Losses in a Competitive Energy Market

In a competitive energy market, system reliability should be maintained at all times. Power system operation being of online in nature, the energy balance requirements must be satisfied to ensure reliable operation the system. To achieve this, information regarding the expected status of the system, the scheduled transactions and the relevant inputs necessary to make either a transaction contract or a transmission contract operational, have to be made available in real time. The real time procedure proposed, facilitates this. This paper proposes a quadratic curve learning procedure, which enables a generator-s contribution to the retailer demand, power loss of transaction in a line at the retail end and its associated losses for an oncoming operating scenario to be predicted. Matlab program was used to test in on a 24-bus IEE Reliability Test System, and the results are found to be acceptable.

Phase Equilibrium of Volatile Organic Compounds in Polymeric Solvents Using Group Contribution Methods

Group contribution methods such as the UNIFAC are of major interest to researchers and engineers involved synthesis, feasibility studies, design and optimization of separation processes as well as other applications of industrial use. Reliable knowledge of the phase equilibrium behavior is crucial for the prediction of the fate of the chemical in the environment and other applications. The objective of this study was to predict the solubility of selected volatile organic compounds (VOCs) in glycol polymers and biodiesel. Measurements can be expensive and time consuming, hence the need for thermodynamic models. The results obtained in this study for the infinite dilution activity coefficients compare very well those published in literature obtained through measurements. It is suggested that in preliminary design or feasibility studies of absorption systems for the abatement of volatile organic compounds, prediction procedures should be implemented while accurate fluid phase equilibrium data should be obtained from experiment.

Adsorption of Inorganic Salt by Granular Activated Carbon and Related Prediction Models

In recent years, the underground water sources in southern Taiwan have become salinized because of saltwater intrusions. This study explores the adsorption characteristics of activated carbon on salinizing inorganic salts using isothermal adsorption experiments and provides a model analysis. The temperature range for the isothermal adsorption experiments ranged between 5 to 45 ℃, and the amount adsorbed varied between 28.21 to 33.87 mg/g. All experimental data of adsorption can be fitted to both the Langmuir and the Freundlich models. The thermodynamic parameters for per chlorate onto granular activated carbon were calculated as -0.99 to -1.11 kcal/mol for DG°, -0.6 kcal/mol for DH°, and 1.21 to 1.84 kcal/mol for DS°. This shows that the adsorption process of granular activated carbon is spontaneously exothermic. The observation of adsorption behaviors under low ionic strength, low pH values, and low temperatures is beneficial to the adsorption removal of perchlorate with granular activated carbon.

Effect of Turbulence Models on Simulated Iced Aircraft Airfoil

The present work describes a computational study of aerodynamic characteristics of GLC305 airfoil clean and with 16.7 min ice shape (rime 212) and 22.5 min ice shape (glaze 944).The performance of turbulence models SA, Kε, Kω Std, and Kω SST model are observed against experimental flow fields at different Mach numbers 0.12, 0.21, 0.28 in a range of Reynolds numbers 3x106, 6x106, and 10.5x106 on clean and iced aircraft airfoil GLC305. Numerical predictions include lift, drag and pitching moment coefficients at different Mach numbers and at different angle of attacks were done. Accuracy of solutions with respect to the effects of turbulence models, variation of Mach number, initial conditions, grid resolution and grid spacing near the wall made the study much sensitive. Navier Stokes equation based computational technique is used. Results are very close to the experimental results. It has seen that SA and SST models are more efficient than Kε and Kω standard in under study problem.

A Quantitative Analysis of GSM Air Interface Based on Radiating Columns and Prediction Model

This paper explains the cause of nonlinearity in floor attenuation hither to left unexplained. The performance degradation occurring in air interface for GSM signals is quantitatively analysed using the concept of Radiating Columns of buildings. The signal levels were measured using Wireless Network Optimising Drive Test Tool (E6474A of Agilent Technologies). The measurements were taken in reflected signal environment under usual fading conditions on actual GSM signals radiated from base stations. A mathematical model is derived from the measurements to predict the GSM signal levels in different floors. It was applied on three buildings and found that the predicted signal levels deviated from the measured levels with in +/- 2 dB for all floors. It is more accurate than the prediction models based on Floor Attenuation Factor. It can be used for planning proper indoor coverage in multi storey buildings.

Effect of Ply Orientation on Roughness for the Trimming Process of CFRP Laminates

The machining of Carbon Fiber Reinforced Plastics has come to constitute a significant challenge for many fields of industry. The resulting surface finish of machined parts is of primary concern for several reasons, including contact quality and impact on the assembly. Therefore, the characterization and prediction of roughness based on machining parameters are crucial for costeffective operations. In this study, a PCD tool comprised of two straight flutes was used to trim 32-ply carbon fiber laminates in a bid to analyze the effects of the feed rate and the cutting speed on the surface roughness. The results show that while the speed has but a slight impact on the surface finish, the feed rate for its part affects it strongly. A detailed study was also conducted on the effect of fiber orientation on surface roughness, for quasi-isotropic laminates used in aerospace. The resulting roughness profiles for the four-ply orientation lay-up were compared, and it was found that fiber angle is a critical parameter relating to surface roughness. One of the four orientations studied led to very poor surface finishes, and characteristic roughness profiles were identified and found to only relate to the ply orientations of multilayer carbon fiber laminates.

Prediction of Compressive Strength of SCC Containing Bottom Ash using Artificial Neural Networks

The paper presents a comparative performance of the models developed to predict 28 days compressive strengths using neural network techniques for data taken from literature (ANN-I) and data developed experimentally for SCC containing bottom ash as partial replacement of fine aggregates (ANN-II). The data used in the models are arranged in the format of six and eight input parameters that cover the contents of cement, sand, coarse aggregate, fly ash as partial replacement of cement, bottom ash as partial replacement of sand, water and water/powder ratio, superplasticizer dosage and an output parameter that is 28-days compressive strength and compressive strengths at 7 days, 28 days, 90 days and 365 days, respectively for ANN-I and ANN-II. The importance of different input parameters is also given for predicting the strengths at various ages using neural network. The model developed from literature data could be easily extended to the experimental data, with bottom ash as partial replacement of sand with some modifications.

Simulation of Enhanced Biomass Gasification for Hydrogen Production using iCON

Due to the environmental and price issues of current energy crisis, scientists and technologists around the globe are intensively searching for new environmentally less-impact form of clean energy that will reduce the high dependency on fossil fuel. Particularly hydrogen can be produced from biomass via thermochemical processes including pyrolysis and gasification due to the economic advantage and can be further enhanced through in-situ carbon dioxide removal using calcium oxide. This work focuses on the synthesis and development of the flowsheet for the enhanced biomass gasification process in PETRONAS-s iCON process simulation software. This hydrogen prediction model is conducted at operating temperature between 600 to 1000oC at atmospheric pressure. Effects of temperature, steam-to-biomass ratio and adsorbent-to-biomass ratio were studied and 0.85 mol fraction of hydrogen is predicted in the product gas. Comparisons of the results are also made with experimental data from literature. The preliminary economic potential of developed system is RM 12.57 x 106 which equivalent to USD 3.77 x 106 annually shows economic viability of this process.

Memory Estimation of Internet Server Using Queuing Theory: Comparative Study between M/G/1, G/M/1 and G/G/1 Queuing Model

How to effectively allocate system resource to process the Client request by Gateway servers is a challenging problem. In this paper, we propose an improved scheme for autonomous performance of Gateway servers under highly dynamic traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.

Meteorological Data Study and Forecasting Using Particle Swarm Optimization Algorithm

Weather systems use enormously complex combinations of numerical tools for study and forecasting. Unfortunately, due to phenomena in the world climate, such as the greenhouse effect, classical models may become insufficient mostly because they lack adaptation. Therefore, the weather forecast problem is matched for heuristic approaches, such as Evolutionary Algorithms. Experimentation with heuristic methods like Particle Swarm Optimization (PSO) algorithm can lead to the development of new insights or promising models that can be fine tuned with more focused techniques. This paper describes a PSO approach for analysis and prediction of data and provides experimental results of the aforementioned method on realworld meteorological time series.

An Artificial Intelligent Technique for Robust Digital Watermarking in Multiwavelet Domain

In this paper, an artificial intelligent technique for robust digital image watermarking in multiwavelet domain is proposed. The embedding technique is based on the quantization index modulation technique and the watermark extraction process does not require the original image. We have developed an optimization technique using the genetic algorithms to search for optimal quantization steps to improve the quality of watermarked image and robustness of the watermark. In addition, we construct a prediction model based on image moments and back propagation neural network to correct an attacked image geometrically before the watermark extraction process begins. The experimental results show that the proposed watermarking algorithm yields watermarked image with good imperceptibility and very robust watermark against various image processing attacks.

Modelling the Occurrence of Defects and Change Requests during User Acceptance Testing

Software developed for a specific customer under contract typically undergoes a period of testing by the customer before acceptance. This is known as user acceptance testing and the process can reveal both defects in the system and requests for changes to the product. This paper uses nonhomogeneous Poisson processes to model a real user acceptance data set from a recently developed system. In particular a split Poisson process is shown to provide an excellent fit to the data. The paper explains how this model can be used to aid the allocation of resources through the accurate prediction of occurrences both during the acceptance testing phase and before this activity begins.

Convergence Analysis of a Prediction based Adaptive Equalizer for IIR Channels

This paper presents the convergence analysis of a prediction based blind equalizer for IIR channels. Predictor parameters are estimated by using the recursive least squares algorithm. It is shown that the prediction error converges almost surely (a.s.) toward a scalar multiple of the unknown input symbol sequence. It is also proved that the convergence rate of the parameter estimation error is of the same order as that in the iterated logarithm law.

Prediction of the Performance of a Bar-Type Piezoelectric Vibration Actuator Depending on the Frequency Using an Equivalent Circuit Analysis

This paper has been investigated a technique that predicts the performance of a bar-type unimorph piezoelectric vibration actuator depending on the frequency. This paper has been proposed an equivalent circuit that can be easily analyzed for the bar-type unimorph piezoelectric vibration actuator. In the dynamic analysis, rigidity and resonance frequency, which are important mechanical elements, were derived using the basic beam theory. In the equivalent circuit analysis, the displacement and bandwidth of the piezoelectric vibration actuator depending on the frequency were predicted. Also, for the reliability of the derived equations, the predicted performance depending on the shape change was compared with the result of a finite element analysis program.

Health Monitoring of Power Transformers by Dissolved Gas Analysis using Regression Method and Study the Effect of Filtration on Oil

Economically transformers constitute one of the largest investments in a Power system. For this reason, transformer condition assessment and management is a high priority task. If a transformer fails, it would have a significant negative impact on revenue and service reliability. Monitoring the state of health of power transformers has traditionally been carried out using laboratory Dissolved Gas Analysis (DGA) tests performed at periodic intervals on the oil sample, collected from the transformers. DGA of transformer oil is the single best indicator of a transformer-s overall condition and is a universal practice today, which started somewhere in the 1960s. Failure can occur in a transformer due to different reasons. Some failures can be limited or prevented by maintenance. Oil filtration is one of the methods to remove the dissolve gases and prevent the deterioration of the oil. In this paper we analysis the DGA data by regression method and predict the gas concentration in the oil in the future. We bring about a comparative study of different traditional methods of regression and the errors generated out of their predictions. With the help of these data we can deduce the health of the transformer by finding the type of fault if it has occurred or will occur in future. Additional in this paper effect of filtration on the transformer health is highlight by calculating the probability of failure of a transformer with and without oil filtrating.