Abstract: In this study, a mathematical model was proposed and
the accuracy of this model was assessed to predict the growth of
Pseudomonas aeruginosa and rhamnolipid production under nitrogen
limiting (sodium nitrate) fed-batch fermentation. All of the
parameters used in this model were achieved individually without
using any data from the literature.
The overall growth kinetic of the strain was evaluated using a
dual-parallel substrate Monod equation which was described by
several batch experimental data. Fed-batch data under different
glycerol (as the sole carbon source, C/N=10) concentrations and feed
flow rates were used to describe the proposed fed-batch model and
other parameters. In order to verify the accuracy of the proposed
model several verification experiments were performed in a vast
range of initial glycerol concentrations. While the results showed an
acceptable prediction for rhamnolipid production (less than 10%
error), in case of biomass prediction the errors were less than 23%. It
was also found that the rhamnolipid production by P. aeruginosa was
more sensitive at low glycerol concentrations.
Based on the findings of this work, it was concluded that the
proposed model could effectively be employed for rhamnolipid
production by this strain under fed-batch fermentation on up to 80 g l-
1 glycerol.
Abstract: Dynamic bandwidth allocation in EPONs can be
generally separated into inter-ONU scheduling and intra-ONU scheduling. In our previous work, the active intra-ONU scheduling
(AS) utilizes multiple queue reports (QRs) in each report message to cooperate with the inter-ONU scheduling and makes the granted
bandwidth fully utilized without leaving unused slot remainder (USR).
This scheme successfully solves the USR problem originating from the
inseparability of Ethernet frame. However, without proper setting of
threshold value in AS, the number of QRs constrained by the IEEE
802.3ah standard is not enough, especially in the unbalanced traffic
environment. This limitation may be solved by enlarging the threshold
value. The large threshold implies the large gap between the adjacent QRs, thus resulting in the large difference between the best granted bandwidth and the real granted bandwidth. In this paper, we integrate
AS with a cooperative prediction mechanism and distribute multiple
QRs to reduce the penalty brought by the prediction error.
Furthermore, to improve the QoS and save the usage of queue reports,
the highest priority (EF) traffic which comes during the waiting time is
granted automatically by OLT and is not considered in the requested
bandwidth of ONU. The simulation results show that the proposed
scheme has better performance metrics in terms of bandwidth
utilization and average delay for different classes of packets.
Abstract: Compression algorithms reduce the redundancy in
data representation to decrease the storage required for that data.
Lossless compression researchers have developed highly
sophisticated approaches, such as Huffman encoding, arithmetic
encoding, the Lempel-Ziv (LZ) family, Dynamic Markov
Compression (DMC), Prediction by Partial Matching (PPM), and
Burrows-Wheeler Transform (BWT) based algorithms.
Decompression is also required to retrieve the original data by
lossless means. A compression scheme for text files coupled with
the principle of dynamic decompression, which decompresses only
the section of the compressed text file required by the user instead of
decompressing the entire text file. Dynamic decompressed files offer
better disk space utilization due to higher compression ratios
compared to most of the currently available text file formats.
Abstract: Feature selection has recently been the subject of intensive research in data mining, specially for datasets with a large number of attributes. Recent work has shown that feature selection can have a positive effect on the performance of machine learning algorithms. The success of many learning algorithms in their attempts to construct models of data, hinges on the reliable identification of a small set of highly predictive attributes. The inclusion of irrelevant, redundant and noisy attributes in the model building process phase can result in poor predictive performance and increased computation. In this paper, a novel feature search procedure that utilizes the Ant Colony Optimization (ACO) is presented. The ACO is a metaheuristic inspired by the behavior of real ants in their search for the shortest paths to food sources. It looks for optimal solutions by considering both local heuristics and previous knowledge. When applied to two different classification problems, the proposed algorithm achieved very promising results.
Abstract: Segmentation, filtering out of measurement errors and
identification of breakpoints are integral parts of any analysis of
microarray data for the detection of copy number variation (CNV).
Existing algorithms designed for these tasks have had some successes
in the past, but they tend to be O(N2) in either computation time or
memory requirement, or both, and the rapid advance of microarray
resolution has practically rendered such algorithms useless. Here we
propose an algorithm, SAD, that is much faster and much less thirsty
for memory – O(N) in both computation time and memory requirement
-- and offers higher accuracy. The two key ingredients of SAD are the
fundamental assumption in statistics that measurement errors are
normally distributed and the mathematical relation that the product of
two Gaussians is another Gaussian (function). We have produced a
computer program for analyzing CNV based on SAD. In addition to
being fast and small it offers two important features: quantitative
statistics for predictions and, with only two user-decided parameters,
ease of use. Its speed shows little dependence on genomic profile.
Running on an average modern computer, it completes CNV analyses
for a 262 thousand-probe array in ~1 second and a 1.8 million-probe
array in 9 seconds
Abstract: Software maintenance is extremely important activity in software development life cycle. It involves a lot of human efforts, cost and time. Software maintenance may be further subdivided into different activities such as fault prediction, fault detection, fault prevention, fault correction etc. This topic has gained substantial attention due to sophisticated and complex applications, commercial hardware, clustered architecture and artificial intelligence. In this paper we surveyed the work done in the field of software maintenance. Software fault prediction has been studied in context of fault prone modules, self healing systems, developer information, maintenance models etc. Still a lot of things like modeling and weightage of impact of different kind of faults in the various types of software systems need to be explored in the field of fault severity.
Abstract: Time series models have been used to make predictions of academic enrollments, weather, road accident, casualties and stock prices, etc. Based on the concepts of quartile regression models, we have developed a simple time variant quantile based fuzzy time series forecasting method. The proposed method bases the forecast using prediction of future trend of the data. In place of actual quantiles of the data at each point, we have converted the statistical concept into fuzzy concept by using fuzzy quantiles using fuzzy membership function ensemble. We have given a fuzzy metric to use the trend forecast and calculate the future value. The proposed model is applied for TAIFEX forecasting. It is shown that proposed method work best as compared to other models when compared with respect to model complexity and forecasting accuracy.
Abstract: In this paper, we propose a new hybrid learning model for stock market indices prediction by adding a passive congregation term to the standard hybrid model comprising Particle Swarm Optimization (PSO) with Genetic Algorithm (GA) operators in training Neural Networks (NN). This new passive congregation term is based on the cooperation between different particles in determining new positions rather than depending on the particles selfish thinking without considering other particles positions, thus it enables PSO to perform both the local and global search instead of only doing the local search. Experiment study carried out on the most famous European stock market indices in both long term and short term prediction shows significantly the influence of the passive congregation term in improving the prediction accuracy compared to standard hybrid model.
Abstract: Music Information Retrieval (MIR) and modern data mining techniques are applied to identify style markers in midi music for stylometric analysis and author attribution. Over 100 attributes are extracted from a library of 2830 songs then mined using supervised learning data mining techniques. Two attributes are identified that provide high informational gain. These attributes are then used as style markers to predict authorship. Using these style markers the authors are able to correctly distinguish songs written by the Beatles from those that were not with a precision and accuracy of over 98 per cent. The identification of these style markers as well as the architecture for this research provides a foundation for future research in musical stylometry.
Abstract: Fossil fuels are the major source to meet the world
energy requirements but its rapidly diminishing rate and adverse
effects on our ecological system are of major concern. Renewable
energy utilization is the need of time to meet the future challenges.
Ocean energy is the one of these promising energy resources. Threefourths
of the earth-s surface is covered by the oceans. This enormous
energy resource is contained in the oceans- waters, the air above the
oceans, and the land beneath them. The renewable energy source of
ocean mainly is contained in waves, ocean current and offshore solar
energy. Very fewer efforts have been made to harness this reliable
and predictable resource. Harnessing of ocean energy needs detail
knowledge of underlying mathematical governing equation and their
analysis. With the advent of extra ordinary computational resources
it is now possible to predict the wave climatology in lab simulation.
Several techniques have been developed mostly stem from numerical
analysis of Navier Stokes equations. This paper presents a brief over
view of such mathematical model and tools to understand and
analyze the wave climatology. Models of 1st, 2nd and 3rd generations
have been developed to estimate the wave characteristics to assess the
power potential. A brief overview of available wave energy
technologies is also given. A novel concept of on-shore wave energy
extraction method is also presented at the end. The concept is based
upon total energy conservation, where energy of wave is transferred
to the flexible converter to increase its kinetic energy. Squeezing
action by the external pressure on the converter body results in
increase velocities at discharge section. High velocity head then can
be used for energy storage or for direct utility of power generation.
This converter utilizes the both potential and kinetic energy of the
waves and designed for on-shore or near-shore application. Increased
wave height at the shore due to shoaling effects increases the
potential energy of the waves which is converted to renewable
energy. This approach will result in economic wave energy
converter due to near shore installation and more dense waves due to
shoaling. Method will be more efficient because of tapping both
potential and kinetic energy of the waves.
Abstract: Text Mining is around applying knowledge discovery
techniques to unstructured text is termed knowledge discovery in text
(KDT), or Text data mining or Text Mining. In decision tree
approach is most useful in classification problem. With this
technique, tree is constructed to model the classification process.
There are two basic steps in the technique: building the tree and
applying the tree to the database. This paper describes a proposed
C5.0 classifier that performs rulesets, cross validation and boosting
for original C5.0 in order to reduce the optimization of error ratio.
The feasibility and the benefits of the proposed approach are
demonstrated by means of medial data set like hypothyroid. It is
shown that, the performance of a classifier on the training cases from
which it was constructed gives a poor estimate by sampling or using a
separate test file, either way, the classifier is evaluated on cases that
were not used to build and evaluate the classifier are both are large. If
the cases in hypothyroid.data and hypothyroid.test were to be
shuffled and divided into a new 2772 case training set and a 1000
case test set, C5.0 might construct a different classifier with a lower
or higher error rate on the test cases. An important feature of see5 is
its ability to classifiers called rulesets. The ruleset has an error rate
0.5 % on the test cases. The standard errors of the means provide an
estimate of the variability of results. One way to get a more reliable
estimate of predictive is by f-fold –cross- validation. The error rate of
a classifier produced from all the cases is estimated as the ratio of the
total number of errors on the hold-out cases to the total number of
cases. The Boost option with x trials instructs See5 to construct up to
x classifiers in this manner. Trials over numerous datasets, large and
small, show that on average 10-classifier boosting reduces the error
rate for test cases by about 25%.
Abstract: The purpose of this research was to analyze and compare the instability of a contact surface between Copper and Nickel an alloy cathode in vacuum, the different ratio of Copper and Copper were conducted at 1%, 2% and 4% by using the cathode spot model. The transient recovery voltage is predicted. The cathode spot region is recognized as the collisionless space charge sheath connected with singly ionized collisional plasma. It was found that the transient voltage is decreased with increasing the percentage of an amount of Nickel in cathode materials.
Abstract: Minimum Quantity Lubrication (MQL) technique
obtained a significant attention in machining processes to reduce
environmental loads caused by usage of conventional cutting fluids.
Recently nanofluids are finding an extensive application in the field
of mechanical engineering because of their superior lubrication and
heat dissipation characteristics. This paper investigates the use of a
nanofluid under MQL mode to improve grinding characteristics of
Ti-6Al-4V alloy. Taguchi-s experimental design technique has been
used in the present investigation and a second order model has been
established to predict grinding forces and surface roughness.
Different concentrations of water based Al2O3 nanofluids were
applied in the grinding operation through MQL setup developed in
house and the results have been compared with those of conventional
coolant and pure water. Experimental results showed that grinding
forces reduced significantly when nano cutting fluid was used even at
low concentration of the nano particles and surface finish has been
found to improve with higher concentration of the nano particles.
Abstract: The paper presents a one-dimensional transient
mathematical model of compressible non-isothermal multicomponent
fluid mixture flow in a pipe. The set of the mass,
momentum and enthalpy conservation equations for gas phase is
solved in the model. Thermo-physical properties of multi-component
gas mixture are calculated by solving the Equation of State (EOS)
model. The Soave-Redlich-Kwong (SRK-EOS) model is chosen. Gas
mixture viscosity is calculated on the basis of the Lee-Gonzales-
Eakin (LGE) correlation. Numerical analysis of rapid gas
decompression process in rich and base natural gases is made on the
basis of the proposed mathematical model. The model is successfully
validated on the experimental data [1]. The proposed mathematical
model shows a very good agreement with the experimental data [1] in
a wide range of pressure values and predicts the decompression in
rich and base gas mixtures much better than analytical and
mathematical models, which are available from the open source
literature.
Abstract: Intercropping is one of the sustainable agricultural
factors. The SPAD meter can be used to predict nitrogen index
reliably, it may also be a useful tool for assessing the relative impact
of weeds on crops. In order to study the effect of weeds on SPAD in
corn (Zea mays L.), sweet basil (Ocimum basilicum L.) and borage
(Borago officinalis L.) in intercropping system, a factorial experiment
was conducted in three replications in 2011. Experimental factors
were included intercropping of corn with sweet basil and borage in
different ratios (100:0, 75:25, 50:50, 25:75 and 0:100 corn: borage or
sweet basil) and weed infestation (weed control and weed
interference). The results showed that intercropping of corn with
sweet basil and borage increased the SPAD value of corn compare to
monoculture in weed interference condition. Sweet basil SPAD value
in weed control treatments (43.66) was more than weed interference
treatments (40.17). Corn could increase the borage SPAD value
compare to monoculture in weed interference treatments.
Abstract: The perfect operation of common Active Filters is depended on accuracy of identification system distortion. Also, using a suitable method in current injection and reactive power compensation, leads to increased filter performance. Due to this fact, this paper presents a method based on predictive current control theory in shunt active filter applications. The harmonics of the load current is identified by using o–d–q reference frame on load current and eliminating the DC part of d–q components. Then, the rest of these components deliver to predictive current controller as a Threephase reference current by using Park inverse transformation. System is modeled in discreet time domain. The proposed method has been tested using MATLAB model for a nonlinear load (with Total Harmonic Distortion=20%). The simulation results indicate that the proposed filter leads to flowing a sinusoidal current (THD=0.15%) through the source. In addition, the results show that the filter tracks the reference current accurately.
Abstract: In this paper, based on the past project cost and time
performance, a model for forecasting project cost performance is
developed. This study presents a probabilistic project control concept
to assure an acceptable forecast of project cost performance. In this
concept project activities are classified into sub-groups entitled
control accounts. Then obtain the Stochastic S-Curve (SS-Curve), for
each sub-group and the project SS-Curve is obtained by summing
sub-groups- SS-Curves. In this model, project cost uncertainties are
considered through Beta distribution functions of the project
activities costs required to complete the project at every selected time
sections through project accomplishment, which are extracted from a
variety of sources. Based on this model, after a percentage of the
project progress, the project performance is measured via Earned
Value Management to adjust the primary cost probability distribution
functions. Then, accordingly the future project cost performance is
predicted by using the Monte-Carlo simulation method.
Abstract: Heart disease (HD) is a major cause of morbidity and mortality in the modern society. Medical diagnosis is an important but complicated task that should be performed accurately and efficiently and its automation would be very useful. All doctors are unfortunately not equally skilled in every sub specialty and they are in many places a scarce resource. A system for automated medical diagnosis would enhance medical care and reduce costs. In this paper, a new approach based on coactive neuro-fuzzy inference system (CANFIS) was presented for prediction of heart disease. The proposed CANFIS model combined the neural network adaptive capabilities and the fuzzy logic qualitative approach which is then integrated with genetic algorithm to diagnose the presence of the disease. The performances of the CANFIS model were evaluated in terms of training performances and classification accuracies and the results showed that the proposed CANFIS model has great potential in predicting the heart disease.
Abstract: Classes on creativity, innovation, and entrepreneurship
are becoming quite popular at universities throughout the world.
However, it is not easy for business students to get involved to
innovative activities, especially patent application. The present study
investigated how to enhance business students- intention to participate
in innovative activities and which incentives universities should
consider. A 22-item research scale was used, and confirmatory factor
analysis was conducted to verify its reliability and validity. Multiple
regression and discriminant analyses were also conducted. The results
demonstrate the effect of growth-need strength on innovative behavior
and indicate that the theory of planned behavior can explain and
predict business students- intention to participate in innovative
activities. Additionally, the results suggest that applying our proposed
model in practice would effectively strengthen business students-
intentions to engage in innovative activities.
Abstract: Wireless Sensor Network (WSN) comprises of sensor
nodes which are designed to sense the environment, transmit sensed
data back to the base station via multi-hop routing to reconstruct
physical phenomena. Since physical phenomena exists significant
overlaps between temporal redundancy and spatial redundancy, it is
necessary to use Redundancy Suppression Algorithms (RSA) for sensor
node to lower energy consumption by reducing the transmission
of redundancy. A conventional algorithm of RSAs is threshold-based
RSA, which sets threshold to suppress redundant data. Although
many temporal and spatial RSAs are proposed, temporal-spatial RSA
are seldom to be proposed because it is difficult to determine when
to utilize temporal or spatial RSAs. In this paper, we proposed a
novel temporal-spatial redundancy suppression algorithm, Codebookbase
Redundancy Suppression Mechanism (CRSM). CRSM adopts
vector quantization to generate a codebook, which is easily used to
implement temporal-spatial RSA. CRSM not only achieves power
saving and reliability for WSN, but also provides the predictability
of network lifetime. Simulation result shows that the network lifetime
of CRSM outperforms at least 23% of that of other RSAs.