Abstract: The aim of this paper is to rank the impact of Object
Oriented(OO) metrics in fault prediction modeling using Artificial
Neural Networks(ANNs). Past studies on empirical validation of
object oriented metrics as fault predictors using ANNs have focused
on the predictive quality of neural networks versus standard
statistical techniques. In this empirical study we turn our attention to
the capability of ANNs in ranking the impact of these explanatory
metrics on fault proneness. In ANNs data analysis approach, there is
no clear method of ranking the impact of individual metrics. Five
ANN based techniques are studied which rank object oriented
metrics in predicting fault proneness of classes. These techniques are
i) overall connection weights method ii) Garson-s method iii) The
partial derivatives methods iv) The Input Perturb method v) the
classical stepwise methods. We develop and evaluate different
prediction models based on the ranking of the metrics by the
individual techniques. The models based on overall connection
weights and partial derivatives methods have been found to be most
accurate.
Abstract: Antiseismic property of telecommunication equipment
is very important for the grasp of the damage and the restoration after
earthquake. Telecommunication business operators are regulating
seismic standard for their equipments. These standards are organized
to simulate the real seismic situations and usually define the minimum
value of first natural frequency of the equipments or the allowable
maximum displacement of top of the equipments relative to bottom.
Using the finite element analysis, natural frequency can be obtained
with high accuracy but the relative displacement of top of the
equipments is difficult to predict accurately using the analysis.
Furthermore, in the case of simulating the equipments with access
floor, predicting the relative displacement of top of the equipments
become more difficult.
In this study, using enormous experimental datum, an empirical
formula is suggested to forecast the relative displacement of top of the
equipments. Also it can be known that which physical quantities are
related with the relative displacement.
Abstract: Employing a recently introduced unified adaptive filter
theory, we show how the performance of a large number of important
adaptive filter algorithms can be predicted within a general framework
in nonstationary environment. This approach is based on energy conservation
arguments and does not need to assume a Gaussian or white
distribution for the regressors. This general performance analysis can
be used to evaluate the mean square performance of the Least Mean
Square (LMS) algorithm, its normalized version (NLMS), the family
of Affine Projection Algorithms (APA), the Recursive Least Squares
(RLS), the Data-Reusing LMS (DR-LMS), its normalized version
(NDR-LMS), the Block Least Mean Squares (BLMS), the Block
Normalized LMS (BNLMS), the Transform Domain Adaptive Filters
(TDAF) and the Subband Adaptive Filters (SAF) in nonstationary
environment. Also, we establish the general expressions for the
steady-state excess mean square in this environment for all these
adaptive algorithms. Finally, we demonstrate through simulations that
these results are useful in predicting the adaptive filter performance.
Abstract: A new analytical model is developed which provides
close-formed solutions for both transient indoor and envelope
temperature changes in buildings. Time-dependent boundary
temperature is presented as Fourier series which can approximate real
weather conditions. The final close-formed solutions are simple,
concise, and comprehensive. The model was compared with
numerical results and good accuracy was obtained. The model can
be used as design and control guidelines in engineering applications
for analysing mechanical heat transfer properties for buildings.
Abstract: Simulation of occlusal function during laboratory
material-s testing becomes essential in predicting long-term
performance before clinical usage. The aim of the study was to assess
the influence of chamfer preparation depth on failure risk of heat
pressed ceramic crowns with and without zirconia framework by
means of finite element analysis. 3D models of maxillary central
incisor, prepared for full ceramic crowns with different depths of the
chamfer margin (between 0.8 and 1.2 mm) and 6-degree tapered
walls together with the overlying crowns were generated using
literature data (Fig. 1, 2). The crowns were designed with and
without a zirconia framework with a thickness of 0.4 mm. For all
preparations and crowns, stresses in the pressed ceramic crown,
zirconia framework, pressed ceramic veneer, and dentin were
evaluated separately. The highest stresses were registered in the
dentin. The depth of the preparations had no significant influence on
the stress values of the teeth and pressed ceramics for the studied
cases, only for the zirconia framework. The zirconia framework
decreases the stress values in the veneer.
Abstract: Cutting tools are widely used in manufacturing processes and drilling is the most commonly used machining process. Although drill-bits used in drilling may not be expensive, their breakage can cause damage to expensive work piece being drilled and at the same time has major impact on productivity. Predicting drill-bit breakage, therefore, is important in reducing cost and improving productivity. This study uses twenty features extracted from two degradation signals viz., thrust force and torque. The methodology used involves developing and comparing decision tree, random forest, and multinomial logistic regression models for classifying and predicting drill-bit breakage using degradation signals.
Abstract: The purpose of this study was to examine to what
extend classroom management efficacy, marital status, gender, and
teaching experience predict burnout among primary school teachers.
Participants of this study were 523 (345 female, 178 male) teachers
who completed inventories. The results of multiple regression
analysis indicated that three dimensions of teacher burnout
(Emotional Exhaustion, Depersonalization, Personal
Accomplishment) were affected differently from four predictor
variables. Findings indicated that for the emotional exhaustion,
classroom management efficacy, marital status and teaching
experience; for depersonalization dimension, classroom management
efficacy and marital status and finally for the personal
accomplishment dimension, classroom management efficacy, gender,
and teaching experience were significant predictors.
Abstract: High Strength Concrete (HSC) is defined as concrete
that meets special combination of performance and uniformity
requirements that cannot be achieved routinely using conventional
constituents and normal mixing, placing, and curing procedures. It is
a highly complex material, which makes modeling its behavior a very
difficult task. This paper aimed to show possible applicability of
Neural Networks (NN) to predict the slump in High Strength
Concrete (HSC). Neural Network models is constructed, trained and
tested using the available test data of 349 different concrete mix
designs of High Strength Concrete (HSC) gathered from a particular
Ready Mix Concrete (RMC) batching plant. The most versatile
Neural Network model is selected to predict the slump in concrete.
The data used in the Neural Network models are arranged in a format
of eight input parameters that cover the Cement, Fly Ash, Sand,
Coarse Aggregate (10 mm), Coarse Aggregate (20 mm), Water,
Super-Plasticizer and Water/Binder ratio. Furthermore, to test the
accuracy for predicting slump in concrete, the final selected model is
further used to test the data of 40 different concrete mix designs of
High Strength Concrete (HSC) taken from the other batching plant.
The results are compared on the basis of error function (or
performance function).
Abstract: Protecting is the sources of drinking water is the first
barrier of contamination of drinking water. The Feitsui Reservoir
watershed of Taiwan supplies domestic water for around 5 million
people in the Taipei metropolitan area. Understanding the spatial
patterns of water quality trends in this watershed is an important
agenda for management authorities. This study examined 7 sites in the
watershed for water quality parameters regulated in the standard for
drinking water source. The non-parametric seasonal Mann-Kendall-s
test was used to determine significant trends for each parameter.
Significant trends of increasing pH occurred at the sampling station in
the uppermost stream watershed, and in total phosphorus at 4 sampling
stations in the middle and downstream watershed. Additionally, the
multi-scale land cover assessment and average land slope were used to
explore the influence on the water quality in the watershed. Regression
models for predicting water quality were also developed.
Abstract: In this study, numerical simulations on laminar flow in
sinusoidal wavy shaped tubes were conducted for mean Reynolds
number of 250, which is in the range of physiological flow-rate and
investigated flow structures, pressure distribution and particle
trajectories both in steady and periodic inflow conditions. For
extensive comparisons, various wave lengths and amplitudes of sine
function for geometry of tube models were employed. The results
showed that small amplitude secondary curvature has significant
influence on the nature of flow patterns and particle mixing
mechanism. This implies that characterizing accurate geometry is
essential in accurate predicting of in vivo hemodynamics and may
motivate further study on any possibility of reflection of secondary
flow on vascular remodeling and pathophysiology.
Abstract: Using Dynamic Bayesian Networks (DBN) to model genetic regulatory networks from gene expression data is one of the major paradigms for inferring the interactions among genes. Averaging a collection of models for predicting network is desired, rather than relying on a single high scoring model. In this paper, two kinds of model searching approaches are compared, which are Greedy hill-climbing Search with Restarts (GSR) and Markov Chain Monte Carlo (MCMC) methods. The GSR is preferred in many papers, but there is no such comparison study about which one is better for DBN models. Different types of experiments have been carried out to try to give a benchmark test to these approaches. Our experimental results demonstrated that on average the MCMC methods outperform the GSR in accuracy of predicted network, and having the comparable performance in time efficiency. By proposing the different variations of MCMC and employing simulated annealing strategy, the MCMC methods become more efficient and stable. Apart from comparisons between these approaches, another objective of this study is to investigate the feasibility of using DBN modeling approaches for inferring gene networks from few snapshots of high dimensional gene profiles. Through synthetic data experiments as well as systematic data experiments, the experimental results revealed how the performances of these approaches can be influenced as the target gene network varies in the network size, data size, as well as system complexity.
Abstract: Dengue fever is prevalent in Malaysia with numerous
cases including mortality recorded over the years. Public education
on the prevention of the desease through various means has been
carried out besides the enforcement of legal means to eradicate
Aedes mosquitoes, the dengue vector breeding ground. Hence, other
means need to be explored, such as predicting the seasonal peak
period of the dengue outbreak and identifying related climate factors
contributing to the increase in the number of mosquitoes. Simulation
model can be employed for this purpose. In this study, we created a
simulation of system dynamic to predict the spread of dengue
outbreak in Hulu Langat, Selangor Malaysia. The prototype was
developed using STELLA 9.1.2 software. The main data input are
rainfall, temperature and denggue cases. Data analysis from the graph
showed that denggue cases can be predicted accurately using these
two main variables- rainfall and temperature. However, the model
will be further tested over a longer time period to ensure its
accuracy, reliability and efficiency as a prediction tool for dengue
outbreak.
Abstract: As the majority of faults are found in a few of its
modules so there is a need to investigate the modules that are
affected severely as compared to other modules and proper
maintenance need to be done in time especially for the critical
applications. As, Neural networks, which have been already applied
in software engineering applications to build reliability growth
models predict the gross change or reusability metrics. Neural
networks are non-linear sophisticated modeling techniques that are
able to model complex functions. Neural network techniques are
used when exact nature of input and outputs is not known. A key
feature is that they learn the relationship between input and output
through training. In this present work, various Neural Network Based
techniques are explored and comparative analysis is performed for
the prediction of level of need of maintenance by predicting level
severity of faults present in NASA-s public domain defect dataset.
The comparison of different algorithms is made on the basis of Mean
Absolute Error, Root Mean Square Error and Accuracy Values. It is
concluded that Generalized Regression Networks is the best
algorithm for classification of the software components into different
level of severity of impact of the faults. The algorithm can be used to
develop model that can be used for identifying modules that are
heavily affected by the faults.
Abstract: The energy consumption and delay in read/write
operation of conventional SRAM is investigated analytically as well
as by simulation. Explicit analytical expressions for the energy
consumption and delay in read and write operation as a function of
device parameters and supply voltage are derived. The expressions are
useful in predicting the effect of parameter changes on the energy
consumption and speed as well as in optimizing the design of
conventional SRAM. HSPICE simulation in standard 0.25μm CMOS
technology confirms precision of analytical expressions derived from
this paper.
Abstract: Energy dissipation in drops has been investigated by
physical models. After determination of effective parameters on the
phenomenon, three drops with different heights have been
constructed from Plexiglas. They have been installed in two existing
flumes in the hydraulic laboratory. Several runs of physical models
have been undertaken to measured required parameters for
determination of the energy dissipation. Results showed that the
energy dissipation in drops depend on the drop height and discharge.
Predicted relative energy dissipations varied from 10.0% to 94.3%.
This work has also indicated that the energy loss at drop is mainly
due to the mixing of the jet with the pool behind the jet that causes
air bubble entrainment in the flow. Statistical model has been
developed to predict the energy dissipation in vertical drops denotes
nonlinear correlation between effective parameters. Further an
artificial neural networks (ANNs) approach was used in this paper to
develop an explicit procedure for calculating energy loss at drops
using NeuroSolutions. Trained network was able to predict the
response with R2 and RMSE 0.977 and 0.0085 respectively. The
performance of ANN was found effective when compared to
regression equations in predicting the energy loss.
Abstract: Determination of wellbore problems during a
production/injection process might be evaluated thorough
temperature log analysis. Other applications of this kind of log
analysis may also include evaluation of fluid distribution analysis
along the wellbore and identification of anomalies encountered
during production/injection process. While the accuracy of such
prediction is paramount, the common method of determination of a
wellbore temperature log includes use of steady-state energy balance
equations, which hardly describe the real conditions as observed in
typical oil and gas flowing wells during production operation; and
thus increase level of uncertainties. In this study, a practical method
has been proposed through development of a simplified semianalytical
model to apply for predicting temperature profile along the
wellbore. The developed model includes an overall heat transfer
coefficient accounting all modes of heat transferring mechanism,
which has been focused on the prediction of a temperature profile as
a function of depth for the injection/production wells. The model has
been validated with the results obtained from numerical simulation.
Abstract: In this study, cometabolic biodegradation of
chloroform was experimented with mixed cultures in the presence of
various organic solvents like methanol, ethanol, isopropanol, acetone,
acetonitrile and toluene as these are predominant discharges in
pharmaceutical industries. Toluene and acetone showed higher
specific chloroform degradation rate when compared to other
compounds. Cometabolic degradation of chloroform was further
confirmed by observation of free chloride ions in the medium. An
extended Haldane model, incorporating the inhibition due to
chloroform and the competitive inhibition between primary
substrates, was developed to predict the biodegradation of primary
substrates, cometabolic degradation of chloroform and the biomass
growth. The proposed model is based on the use of biokinetic
parameters obtained from single substrate degradation studies. The
model was able to satisfactorily predict the experimental results of
ternary and quaternary mixtures. The proposed model can be used for
predicting the performance of bioreactors treating discharges from
pharmaceutical industries.
Abstract: A multiple-option analytical model for the evaluation of the energy performance and distribution of aerodynamic forces acting on a vertical-axis Darrieus wind turbine depending on both rotor architecture and operating conditions is presented. For this purpose, a numerical algorithm, capable of generating the desired rotor conformation depending on design geometric parameters, is coupled to a Single/Double-Disk Multiple-Streamtube Blade Element – Momentum code. Both single and double-disk configurations are analyzed and model predictions are compared to literature experimental data in order to test the capability of the code for predicting rotor performance. Effective airfoil characteristics based on local blade Reynolds number are obtained through interpolation of literature low-Reynolds airfoil databases. Some corrections are introduced inside the original model with the aim of simulating also the effects of blade dynamic stall, rotor streamtube expansion and blade finite aspect ratio, for which a new empirical relationship to better fit the experimental data is proposed. In order to predict also open field rotor operation, a freestream wind shear profile is implemented, reproducing the effect of atmospheric boundary layer.
Abstract: A multi-agent system is developed here to predict
monthly details of the upcoming peak of the 24th solar magnetic
cycle. While studies typically predict the timing and magnitude of
cycle peaks using annual data, this one utilizes the unsmoothed
monthly sunspot number instead. Monthly numbers display more
pronounced fluctuations during periods of strong solar magnetic
activity than the annual sunspot numbers. Because strong magnetic
activities may cause significant economic damages, predicting
monthly variations should provide different and perhaps helpful
information for decision-making purposes. The multi-agent system
developed here operates in two stages. In the first, it produces twelve
predictions of the monthly numbers. In the second, it uses those
predictions to deliver a final forecast. Acting as expert agents, genetic
programming and neural networks produce the twelve fits and
forecasts as well as the final forecast. According to the results
obtained, the next peak is predicted to be 156 and is expected to
occur in October 2011- with an average of 136 for that year.
Abstract: There is lot of work done in prediction of the fault proneness of the software systems. But, it is the severity of the faults that is more important than number of faults existing in the developed system as the major faults matters most for a developer and those major faults needs immediate attention. In this paper, we tried to predict the level of impact of the existing faults in software systems. Neuro-Fuzzy based predictor models is applied NASA-s public domain defect dataset coded in C programming language. As Correlation-based Feature Selection (CFS) evaluates the worth of a subset of attributes by considering the individual predictive ability of each feature along with the degree of redundancy between them. So, CFS is used for the selecting the best metrics that have highly correlated with level of severity of faults. The results are compared with the prediction results of Logistic Models (LMT) that was earlier quoted as the best technique in [17]. The results are recorded in terms of Accuracy, Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). The results show that Neuro-fuzzy based model provide a relatively better prediction accuracy as compared to other models and hence, can be used for the modeling of the level of impact of faults in function based systems.