Abstract: As the majority of faults are found in a few of its
modules so there is a need to investigate the modules that are
affected severely as compared to other modules and proper
maintenance need to be done in time especially for the critical
applications. As, Neural networks, which have been already applied
in software engineering applications to build reliability growth
models predict the gross change or reusability metrics. Neural
networks are non-linear sophisticated modeling techniques that are
able to model complex functions. Neural network techniques are
used when exact nature of input and outputs is not known. A key
feature is that they learn the relationship between input and output
through training. In this present work, various Neural Network Based
techniques are explored and comparative analysis is performed for
the prediction of level of need of maintenance by predicting level
severity of faults present in NASA-s public domain defect dataset.
The comparison of different algorithms is made on the basis of Mean
Absolute Error, Root Mean Square Error and Accuracy Values. It is
concluded that Generalized Regression Networks is the best
algorithm for classification of the software components into different
level of severity of impact of the faults. The algorithm can be used to
develop model that can be used for identifying modules that are
heavily affected by the faults.
Abstract: Due to the high percentage of induction motors in industrial market, there exist a large opportunity for energy savings. Replacement of working induction motors with more efficient ones can be an important resource for energy savings. A calculation of energy savings and payback periods, as a result of such a replacement, based on nameplate motor efficiency or manufacture-s data can lead to large errors [1]. Efficiency of induction motors (IMs) can be extracted using some procedures that use the no-load test results. In the cases that we must estimate the efficiency on-line, some of these procedures can-t be efficient. In some cases the efficiency estimates using the rating values of the motor, but these procedures can have errors due to the different working condition of the motor. In this paper the efficiency of an IM estimated by using the genetic algorithm. The results are compared with the measured values of the torque and power. The results show smaller errors for this procedure compared with the conventional classical procedures, hence the cost of the equipments is reduced and on-line estimation of the efficiency can be made.
Abstract: In this paper, the application of neural networks to study the design of short-term temperature forecasting (STTF) Systems for Kermanshah city, west of Iran was explored. One important architecture of neural networks named Multi-Layer Perceptron (MLP) to model STTF systems is used. Our study based on MLP was trained and tested using ten years (1996-2006) meteorological data. The results show that MLP network has the minimum forecasting error and can be considered as a good method to model the STTF systems.
Abstract: In this study, aerobic digestion of tannery industry
wastewater was carried out using mixed culture obtained from
common effluent treatment plant treating tannery wastewater. The
effect of pH, temperature, inoculum concentration, agitation speed
and initial substrate concentration on the reduction of organic matters
were found. The optimum conditions for COD reduction was found
to be pH - 7 (60%), temperature - 30ÔùªC (61%), inoculum
concentration - 2% (61%), agitation speed - 150rpm (65%) and initial
substrate concentration - 1560 mg COD/L (74%). Kinetics studies
were carried by using Monod model, First order, Diffusional model
and Singh model. From the results it was found that the Monod
model suits well for the degradation of tannery wastewater using
mixed microbial consortium.
Abstract: The UML modeling of complex distributed systems often is a great challenge due to the large amount of parallel real-time operating components. In this paper the problems of verification of such systems are discussed. ECPN, an Extended Colored Petri Net is defined to formally describe state transitions of components and interactions among components. The relationship between sequence diagrams and Free Choice Petri Nets is investigated. Free Choice Petri Net theory helps verifying the liveness of sequence diagrams. By converting sequence diagrams to ECPNs and then comparing behaviors of sequence diagram ECPNs and statecharts, the consistency among models is analyzed. Finally, a verification process for an example model is demonstrated.
Abstract: Large scale systems such as computational Grid is
a distributed computing infrastructure that can provide globally
available network resources. The evolution of information processing
systems in Data Grid is characterized by a strong decentralization of
data in several fields whose objective is to ensure the availability and
the reliability of the data in the reason to provide a fault tolerance
and scalability, which cannot be possible only with the use of the
techniques of replication. Unfortunately the use of these techniques
has a height cost, because it is necessary to maintain consistency
between the distributed data. Nevertheless, to agree to live with
certain imperfections can improve the performance of the system by
improving competition. In this paper, we propose a multi-layer protocol
combining the pessimistic and optimistic approaches conceived
for the data consistency maintenance in large scale systems. Our
approach is based on a hierarchical representation model with tree
layers, whose objective is with double vocation, because it initially
makes it possible to reduce response times compared to completely
pessimistic approach and it the second time to improve the quality
of service compared to an optimistic approach.
Abstract: In the present study Schwertmannite (an iron oxide
hydroxide) is selected as an adsorbent for defluoridation of water.
The adsorbent was prepared by wet chemical process and was
characterized by SEM, XRD and BET. The fluoride adsorption
efficiency of the prepared adsorbent was determined with respect to
contact time, initial fluoride concentration, adsorbent dose and pH of
the solution. The batch adsorption data revealed that the fluoride
adsorption efficiency was highly influenced by the studied factors.
Equilibrium was attained within one hour of contact time indicating
fast kinetics and the adsorption data followed pseudo second order
kinetic model. Equilibrium isotherm data fitted to both Langmuir and
Freundlich isotherm models for a concentration range of 5-30 mg/L.
The adsorption system followed Langmuir isotherm model with
maximum adsorption capacity of 11.3 mg/g. The high adsorption
capacity of Schwertmannite points towards the potential of this
adsorbent for fluoride removal from aqueous medium.
Abstract: Pattern recognition and image recognition methods are commonly developed and tested using testbeds, which contain known responses to a query set. Until now, testbeds available for image analysis and content-based image retrieval (CBIR) have been scarce and small-scale. Here we present the one million images CEA-List Image Collection (CLIC) testbed that we have produced, and report on our use of this testbed to evaluate image analysis merging techniques. This testbed will soon be made publicly available through the EU MUSCLE Network of Excellence.
Abstract: Ensemble learning algorithms such as AdaBoost and
Bagging have been in active research and shown improvements in
classification results for several benchmarking data sets with mainly
decision trees as their base classifiers. In this paper we experiment to
apply these Meta learning techniques with classifiers such as random
forests, neural networks and support vector machines. The data sets
are from MAGIC, a Cherenkov telescope experiment. The task is to
classify gamma signals from overwhelmingly hadron and muon
signals representing a rare class classification problem. We compare
the individual classifiers with their ensemble counterparts and
discuss the results. WEKA a wonderful tool for machine learning has
been used for making the experiments.
Abstract: In developing a text-to-speech system, it is well
known that the accuracy of information extracted from a text is
crucial to produce high quality synthesized speech. In this paper, a
new scheme for converting text into its equivalent phonetic spelling
is introduced and developed. This method is applicable to many
applications in text to speech converting systems and has many
advantages over other methods. The proposed method can also
complement the other methods with a purpose of improving their
performance. The proposed method is a probabilistic model and is
based on Smooth Ergodic Hidden Markov Model. This model can be
considered as an extension to HMM. The proposed method is applied
to Persian language and its accuracy in converting text to speech
phonetics is evaluated using simulations.
Abstract: The wide increase and diffusion on telecommunication
technologies have caused a huge spread of electromagnetic sources
in most European Countries. Since the public is continuously being
exposed to electromagnetic radiation the possible health effects have
become the focus of population concerns. As a result, electromagnetic
field monitoring stations which control field strength in commercial
frequency bands are being placed on the flat roof of many buildings.
However there is no guidance on where to place them. This paper
presents an analysis of frequency, polarization and angles of incidence
of a plane wave which impinges on a flat roof security wall and its
dependence on electromagnetic field strength meters placement.
Abstract: Injection forging is a Nett-shape manufacturing
process in which one or two punches move axially causing a radial
flow into a die cavity in a form which is prescribed by the exitgeometry,
such as pulley, flanges, gears and splines on a shaft. This
paper presents an experimental and numerical study of the injection
forging of splines in terms of load requirement and material flow.
Three dimensional finite element analyses are used to investigate the
effect of some important parameters in this process. The experiment
has been carried out using solid commercial lead billets with two
different billet diameters and four different dies.
Abstract: One of the most important requirements for the
operation and planning activities of an electrical utility is the
prediction of load for the next hour to several days out, known as
short term load forecasting. This paper presents the development of
an artificial neural network based short-term load forecasting model.
The model can forecast daily load profiles with a load time of one
day for next 24 hours. In this method can divide days of year with
using average temperature. Groups make according linearity rate of
curve. Ultimate forecast for each group obtain with considering
weekday and weekend. This paper investigates effects of temperature
and humidity on consuming curve. For forecasting load curve of
holidays at first forecast pick and valley and then the neural network
forecast is re-shaped with the new data. The ANN-based load models
are trained using hourly historical. Load data and daily historical
max/min temperature and humidity data. The results of testing the
system on data from Yazd utility are reported.
Abstract: This paper presents the applicability of artificial
neural networks for 24 hour ahead solar power generation forecasting
of a 20 kW photovoltaic system, the developed forecasting is suitable
for a reliable Microgrid energy management. In total four neural
networks were proposed, namely: multi-layred perceptron, radial
basis function, recurrent and a neural network ensemble consisting in
ensemble of bagged networks. Forecasting reliability of the proposed
neural networks was carried out in terms forecasting error
performance basing on statistical and graphical methods. The
experimental results showed that all the proposed networks achieved
an acceptable forecasting accuracy. In term of comparison the neural
network ensemble gives the highest precision forecasting comparing
to the conventional networks. In fact, each network of the ensemble
over-fits to some extent and leads to a diversity which enhances the
noise tolerance and the forecasting generalization performance
comparing to the conventional networks.
Abstract: In over deployed sensor networks, one approach
to Conserve energy is to keep only a small subset of sensors
active at Any instant. For the coverage problems, the monitoring
area in a set of points that require sensing, called demand points, and
consider that the node coverage area is a circle of range R, where R
is the sensing range, If the Distance between a demand point and
a sensor node is less than R, the node is able to cover this point. We
consider a wireless sensor network consisting of a set of sensors
deployed randomly. A point in the monitored area is covered if it is
within the sensing range of a sensor. In some applications, when the
network is sufficiently dense, area coverage can be approximated by
guaranteeing point coverage. In this case, all the points of wireless
devices could be used to represent the whole area, and the working
sensors are supposed to cover all the sensors. We also introduce
Hybrid Algorithm and challenges related to coverage in sensor
networks.
Abstract: Online news websites are one of the main and wide areas of Mass Media. Since the nineties several Jordanian newspapers were introduced to the World Wide Web to reach various and large numbers of audiances. Examples of these newspapers that have online version are Al-Rai, Ad-Dustor and AlGhad. Other pure online news websites include Ammon and Rum. The main aim of this study is to evaluate online newspaper websites using two assessment measures; usability and web content. This aim is achieved by using a questionnaire based evaluation which is based on the definition of usability and web content in the ISO document as the standard number 9241-part 11. The results are obtained based on 204 audiences- responses. The results of the research showed that the usability factor is relatively good for all Jordanian online newspapers whereas the web content factor is moderate.
Abstract: Position based routing protocols are the kinds of
routing protocols, which they use of nodes location information,
instead of links information to routing. In position based routing
protocols, it supposed that the packet source node has position
information of itself and it's neighbors and packet destination node.
Greedy is a very important position based routing protocol. In one of
it's kinds, named MFR (Most Forward Within Radius), source node
or packet forwarder node, sends packet to one of it's neighbors with
most forward progress towards destination node (closest neighbor to
destination). Using distance deciding metric in Greedy to forward
packet to a neighbor node, is not suitable for all conditions. If closest
neighbor to destination node, has high speed, in comparison with
source node or intermediate packet forwarder node speed or has very
low remained battery power, then packet loss probability is
increased. Proposed strategy uses combination of metrics distancevelocity
similarity-power, to deciding about giving the packet to
which neighbor. Simulation results show that the proposed strategy
has lower lost packets average than Greedy, so it has more reliability.
Abstract: The new technology of fuzzy neural networks for identification of parameters for mathematical models of geofields is proposed and checked. The effectiveness of that soft computing technology is demonstrated, especially in the early stage of modeling, when the information is uncertain and limited.
Abstract: This research is a collaborative narrative research, which is mixed with issues of selected papers and researcher's experience as an anonymous user on social networking sites. The objective of this research is to understand the reasons of the regular users who reject to contact with anonymous users, and to study the communication traditions used in the selected studies. Anonymous users are rejected by regular users, because of the fear of cyber bully, the fear of unpleasant behaviors, and unwillingness of changing communication norm. The suggestion for future research design is to use longitudinal design or quantitative design; and the theory in rhetorical tradition should be able to help develop a strong trust message.
Abstract: The emergence of information technology has
resulted in an ever-increasing demand to use computers for the
efficient management and dissemination of information. Keeping in
view the strong need of farmers to collect important and updated
information for interactive, flexible and quick decision-making, a
model of Decision Support System for Farm Management is
developed. The paper discusses the use of Internet technology for the
farmers to take decisions. A model is developed for the farmers to
access online interactive and flexible information for their farm
management. The workflow of the model is presented highlighting
the information transfer between different modules.