Abstract: This work is to study a roll of the fluctuating density
gradient in the compressible flows for the computational fluid dynamics
(CFD). A new anisotropy tensor with the fluctuating density
gradient is introduced, and is used for an invariant modeling technique
to model the turbulent density gradient correlation equation derived
from the continuity equation. The modeling equation is decomposed
into three groups: group proportional to the mean velocity, and that
proportional to the mean strain rate, and that proportional to the mean
density. The characteristics of the correlation in a wake are extracted
from the results by the two dimensional direct simulation, and shows
the strong correlation with the vorticity in the wake near the body.
Thus, it can be concluded that the correlation of the density gradient
is a significant parameter to describe the quick generation of the
turbulent property in the compressible flows.
Abstract: The hydrologic time series data display periodic
structure and periodic autoregressive process receives considerable
attention in modeling of such series. In this communication long
term record of monthly waste flow of Lyari river is utilized to
quantify by using PAR modeling technique. The parameters of
model are estimated by using Frances & Paap methodology. This
study shows that periodic autoregressive model of order 2 is the most
parsimonious model for assessing periodicity in waste flow of the
river. A careful statistical analysis of residuals of PAR (2) model is
used for establishing goodness of fit. The forecast by using proposed
model confirms significance and effectiveness of the model.
Abstract: An alternative approach to the use of Discrete Fourier
Transform (DFT) for Magnetic Resonance Imaging (MRI) reconstruction
is the use of parametric modeling technique. This method
is suitable for problems in which the image can be modeled by
explicit known source functions with a few adjustable parameters.
Despite the success reported in the use of modeling technique as an
alternative MRI reconstruction technique, two important problems
constitutes challenges to the applicability of this method, these are
estimation of Model order and model coefficient determination. In
this paper, five of the suggested method of evaluating the model
order have been evaluated, these are: The Final Prediction Error
(FPE), Akaike Information Criterion (AIC), Residual Variance (RV),
Minimum Description Length (MDL) and Hannan and Quinn (HNQ)
criterion. These criteria were evaluated on MRI data sets based on the
method of Transient Error Reconstruction Algorithm (TERA). The
result for each criterion is compared to result obtained by the use of a
fixed order technique and three measures of similarity were evaluated.
Result obtained shows that the use of MDL gives the highest measure
of similarity to that use by a fixed order technique.
Abstract: Many digital signal processing, techniques have been used to automatically distinguish protein coding regions (exons) from non-coding regions (introns) in DNA sequences. In this work, we have characterized these sequences according to their nonlinear dynamical features such as moment invariants, correlation dimension, and largest Lyapunov exponent estimates. We have applied our model to a number of real sequences encoded into a time series using EIIP sequence indicators. In order to discriminate between coding and non coding DNA regions, the phase space trajectory was first reconstructed for coding and non-coding regions. Nonlinear dynamical features are extracted from those regions and used to investigate a difference between them. Our results indicate that the nonlinear dynamical characteristics have yielded significant differences between coding (CR) and non-coding regions (NCR) in DNA sequences. Finally, the classifier is tested on real genes where coding and non-coding regions are well known.
Abstract: Recent advances in both the testing and verification of software based on formal specifications of the system to be built have reached a point where the ideas can be applied in a powerful way in the design of agent-based systems. The software engineering research has highlighted a number of important issues: the importance of the type of modeling technique used; the careful design of the model to enable powerful testing techniques to be used; the automated verification of the behavioural properties of the system; the need to provide a mechanism for translating the formal models into executable software in a simple and transparent way. This paper introduces the use of the X-machine formalism as a tool for modeling biology inspired agents proposing the use of the techniques built around X-machine models for the construction of effective, and reliable agent-based software systems.
Abstract: Selecting the data modeling technique for an
information system is determined by the objective of the resultant
data model. Dimensional modeling is the preferred modeling
technique for data destined for data warehouses and data mining,
presenting data models that ease analysis and queries which are in
contrast with entity relationship modeling. The establishment of data
warehouses as components of information system landscapes in
many organizations has subsequently led to the development of
dimensional modeling. This has been significantly more developed
and reported for the commercial database management systems as
compared to the open sources thereby making it less affordable for
those in resource constrained settings. This paper presents
dimensional modeling of HIV patient information using open source
modeling tools. It aims to take advantage of the fact that the most
affected regions by the HIV virus are also heavily resource
constrained (sub-Saharan Africa) whereas having large quantities of
HIV data. Two HIV data source systems were studied to identify
appropriate dimensions and facts these were then modeled using two
open source dimensional modeling tools. Use of open source would
reduce the software costs for dimensional modeling and in turn make
data warehousing and data mining more feasible even for those in
resource constrained settings but with data available.
Abstract: A new method identifies coupled fluid-structure system with a reduced set of state variables is presented. Assuming that the structural model is known a priori either from an analysis or a test and using linear transformations between structural and aeroelastic states, it is possible to deduce aerodynamic information from sampled time histories of the aeroelastic system. More specifically given a finite set of structural modes the method extracts generalized aerodynamic force matrix corresponding to these mode shapes. Once the aerodynamic forces are known, an aeroelastic reduced-order model can be constructed in discrete-time, state-space format by coupling the structural model and the aerodynamic system. The resulting reduced-order model is suitable for constant Mach, varying density analysis.
Abstract: As the majority of faults are found in a few of its
modules so there is a need to investigate the modules that are
affected severely as compared to other modules and proper
maintenance need to be done in time especially for the critical
applications. As, Neural networks, which have been already applied
in software engineering applications to build reliability growth
models predict the gross change or reusability metrics. Neural
networks are non-linear sophisticated modeling techniques that are
able to model complex functions. Neural network techniques are
used when exact nature of input and outputs is not known. A key
feature is that they learn the relationship between input and output
through training. In this present work, various Neural Network Based
techniques are explored and comparative analysis is performed for
the prediction of level of need of maintenance by predicting level
severity of faults present in NASA-s public domain defect dataset.
The comparison of different algorithms is made on the basis of Mean
Absolute Error, Root Mean Square Error and Accuracy Values. It is
concluded that Generalized Regression Networks is the best
algorithm for classification of the software components into different
level of severity of impact of the faults. The algorithm can be used to
develop model that can be used for identifying modules that are
heavily affected by the faults.
Abstract: The paper investigates the potential of support vector
machines and Gaussian process based regression approaches to
model the oxygen–transfer capacity from experimental data of
multiple plunging jets oxygenation systems. The results suggest the
utility of both the modeling techniques in the prediction of the
overall volumetric oxygen transfer coefficient (KLa) from operational
parameters of multiple plunging jets oxygenation system. The
correlation coefficient root mean square error and coefficient of
determination values of 0.971, 0.002 and 0.945 respectively were
achieved by support vector machine in comparison to values of
0.960, 0.002 and 0.920 respectively achieved by Gaussian process
regression. Further, the performances of both these regression
approaches in predicting the overall volumetric oxygen transfer
coefficient was compared with the empirical relationship for multiple
plunging jets. A comparison of results suggests that support vector
machines approach works well in comparison to both empirical
relationship and Gaussian process approaches, and could successfully
be employed in modeling oxygen-transfer.
Abstract: The success of an e-learning system is highly
dependent on the quality of its educational content and how effective,
complete, and simple the design tool can be for teachers. Educational
modeling languages (EMLs) are proposed as design languages
intended to teachers for modeling diverse teaching-learning
experiences, independently of the pedagogical approach and in
different contexts. However, most existing EMLs are criticized for
being too abstract and too complex to be understood and manipulated
by teachers. In this paper, we present a visual EML that simplifies the
process of designing learning scenarios for teachers with no
programming background. Based on the conceptual framework of the
activity theory, our resulting visual EML focuses on using Domainspecific
modeling techniques to provide a pedagogical level of
abstraction in the design process.
Abstract: In determining the electromagnetic properties of
magnetic materials, hysteresis modeling is of high importance. Many
models are available to investigate those characteristics but they tend
to be complex and difficult to implement. In this paper a new
qualitative hysteresis model for ferromagnetic core is presented,
based on the function approximation capabilities of adaptive neuro
fuzzy inference system (ANFIS). The proposed ANFIS model
combined the neural network adaptive capabilities and the fuzzy
logic qualitative approach can restored the hysteresis curve with a
little RMS error. The model accuracy is good and can be easily
adapted to the requirements of the application by extending or
reducing the network training set and thus the required amount of
measurement data.
Abstract: Though nonlinear dynamic analysis using a specialized
hydro-code such as AUTODYN is accurate and useful tool for
progressive collapse assessment of a multi-story building subjected to
blast load, it takes too much time to be applied to a practical simulation
of progressive collapse of a tall building. In this paper, blast analysis of
a RC frame structure using a simplified model with Reinforcement
Contact technique provided in Ansys Workbench was introduced and
investigated on its accuracy. Even though the simplified model has a
fraction of elements of the detailed model, the simplified model with
this modeling technique shows similar structural behavior under the
blast load to the detailed model. The proposed modeling method can
be effectively applied to blast loading progressive collapse analysis of
a RC frame structure.
Abstract: The neural network's performance can be measured by efficiency and accuracy. The major disadvantages of neural network approach are that the generalization capability of neural networks is often significantly low, and it may take a very long time to tune the weights in the net to generate an accurate model for a highly complex and nonlinear systems. This paper presents a novel Neuro-fuzzy architecture based on Extended Kalman filter. To test the performance and applicability of the proposed neuro-fuzzy model, simulation study of nonlinear complex dynamic system is carried out. The proposed method can be applied to an on-line incremental adaptive learning for the prediction of financial time series. A benchmark case studie is used to demonstrate that the proposed model is a superior neuro-fuzzy modeling technique.
Abstract: Today's business environment requires that companies have access to highly relevant information in a matter of seconds.
Modern Business Intelligence tools rely on data structured mostly in traditional dimensional database schemas, typically represented by
star schemas. Dimensional modeling is already recognized as a
leading industry standard in the field of data warehousing although
several drawbacks and pitfalls were reported. This paper focuses on
the analysis of another data warehouse modeling technique - the
anchor modeling, and its characteristics in context with the standardized dimensional modeling technique from a query performance perspective. The results of the analysis show
information about performance of queries executed on database
schemas structured according to principles of each database modeling
technique.
Abstract: The aim of this paper is to determine the stress levels
at the end of a long slender shaft such as a drilling assembly used in
the oil or gas industry using a mathematical model in real-time. The
torsional deflection experienced by this type of drilling shaft (about 4
KM length and 20 cm diameter hollow shaft with a thickness of 1
cm) can only be determined using a distributed modeling technique.
The main objective of this project is to calculate angular velocity and
torque at the end of the shaft by TLM method and also analyzing of
the behavior of the system by transient response. The obtained result
is compared with lumped modeling technique the importance of these
results will be evident only after the mentioned comparison. Two
systems have different transient responses and in this project because
of the length of the shaft transient response is very important.
Abstract: The paper compares different channel models used for
modeling Broadband Power-Line Communication (BPLC) system.
The models compared are Zimmermann and Dostert, Philipps,
Anatory et al and Anatory et al generalized Transmission Line (TL)
model. The validity of each model was compared in time domain
with ATP-EMTP software which uses transmission line approach. It
is found that for a power-line network with minimum number of
branches all the models give similar signal/pulse time responses
compared with ATP-EMTP software; however, Zimmermann and
Dostert model indicates the same amplitude but different time delay.
It is observed that when the numbers of branches are increased only
generalized TL theory approach results are comparable with ATPEMTP
results. Also the Multi-Carrier Spread Spectrum (MC-SS)
system was applied to check the implication of such behavior on the
modulation schemes. It is observed that using Philipps on the
underground cable can predict the performance up to 25dB better
than other channel models which can misread the actual performance
of the system. Also modified Zimmermann and Dostert under
multipath can predict a better performance of about 5dB better than
the actual predicted by Generalized TL theory. It is therefore
suggested for a realistic BPLC system design and analyses the model
based on generalized TL theory be used.
Abstract: This work concerns the topological optimization
problem for determining the optimal petroleum refinery
configuration. We are interested in further investigating and
hopefully advancing the existing optimization approaches and
strategies employing logic propositions to conceptual process
synthesis problems. In particular, we seek to contribute to this
increasingly exciting area of chemical process modeling by
addressing the following potentially important issues: (a) how the
formulation of design specifications in a mixed-logical-and-integer
optimization model can be employed in a synthesis problem to enrich
the problem representation by incorporating past design experience,
engineering knowledge, and heuristics; and (b) how structural
specifications on the interconnectivity relationships by space (states)
and by function (tasks) in a superstructure should be properly
formulated within a mixed-integer linear programming (MILP)
model. The proposed modeling technique is illustrated on a case
study involving the alternative processing routes of naphtha, in which
significant improvement in the solution quality is obtained.
Abstract: This study was conducted to explore the effects of two
countries model comparison program in Taiwan and Singapore in
TIMSS database. The researchers used Multi-Group Hierarchical
Linear Modeling techniques to compare the effects of two different
country models and we tested our hypotheses on 4,046 Taiwan
students and 4,599 Singapore students in 2007 at two levels: the class
level and student (individual) level. Design quality is a class level
variable. Student level variables are achievement and self-confidence.
The results challenge the widely held view that retention has a positive
impact on self-confidence. Suggestions for future research are
discussed.
Abstract: In this work a surgical simulator is produced which
enables a training otologist to conduct a virtual, real-time prosthetic
insertion. The simulator provides the Ear, Nose and Throat surgeon
with real-time visual and haptic responses during virtual cochlear
implantation into a 3D model of the human Scala Tympani (ST). The
parametric model is derived from measured data as published in the
literature and accounts for human morphological variance, such as
differences in cochlear shape, enabling patient-specific pre- operative
assessment. Haptic modeling techniques use real physical data and
insertion force measurements, to develop a force model which
mimics the physical behavior of an implant as it collides with the ST
walls during an insertion. Output force profiles are acquired from the
insertion studies conducted in the work, to validate the haptic model.
The simulator provides the user with real-time, quantitative insertion
force information and associated electrode position as user inserts the
virtual implant into the ST model. The information provided by this
study may also be of use to implant manufacturers for design
enhancements as well as for training specialists in optimal force
administration, using the simulator. The paper reports on the methods
for anatomical modeling and haptic algorithm development, with
focus on simulator design, development, optimization and validation.
The techniques may be transferrable to other medical applications
that involve prosthetic device insertions where user vision is
obstructed.