Abstract: Prediction of viscosity of natural gas is an important parameter in the energy industries such as natural gas storage and transportation. In this study viscosity of different compositions of natural gas is modeled by using an artificial neural network (ANN) based on back-propagation method. A reliable database including more than 3841 experimental data of viscosity for testing and training of ANN is used. The designed neural network can predict the natural gas viscosity using pseudo-reduced pressure and pseudo-reduced temperature with AARD% of 0.221. The accuracy of designed ANN has been compared to other published empirical models. The comparison indicates that the proposed method can provide accurate results.
Abstract: The study was a case study analysis about Thai Asia
Pacific Brewery Company. The purpose was to analyze the
company’s marketing objective, marketing strategy at company level,
and marketing mix before liquor liberalization in 2000. Methods used
in this study were qualitative and descriptive research approach
which demonstrated the following results of the study demonstrated
as follows: (1) Marketing objective was to increase market share of
Heineken and Amtel, (2) the company’s marketing strategies were
brand building strategy and distribution strategy. Additionally, the
company also conducted marketing mix strategy as follows. Product
strategy: The company added more beer brands namely Amstel and
Tiger to provide additional choice to consumers, product and
marketing research, and product development. Price strategy: the
company had taken the following into consideration: cost,
competitor, market, economic situation and tax. Promotion strategy:
the company conducted sales promotion and advertising. Distribution
strategy: the company extended channels its channels of distribution
into food shops, pubs and various entertainment places. This strategy
benefited interested persons and people who were engaged in the beer
business.
Abstract: This paper presents a method to support dynamic
packing in cases when no collision-free path can be found. The
method, which is primarily based on path planning and shrinking of
geometries, suggests a minimal geometry design change that results
in a collision-free assembly path. A supplementing approach to
optimize geometry design change with respect to redesign cost is
described. Supporting this dynamic packing method, a new method
to shrink geometry based on vertex translation, interweaved with
retriangulation, is suggested. The shrinking method requires neither
tetrahedralization nor calculation of medial axis and it preserves the
topology of the geometry, i.e. holes are neither lost nor introduced.
The proposed methods are successfully applied on industrial
geometries.
Abstract: This research presented in this paper is an on-going
project of an application of neural network and fuzzy models to
evaluate the sociological factors which affect the educational
performance of the students in Sri Lanka. One of its major goals is to
prepare the grounds to device a counseling tool which helps these
students for a better performance at their examinations, especially at
their G.C.E O/L (General Certificate of Education-Ordinary Level)
examination. Closely related sociological factors are collected as raw
data and the noise of these data are filtered through the fuzzy
interface and the supervised neural network is being utilized to
recognize the performance patterns against the chosen social factors.
Abstract: This paper presents Genetic Algorithm (GA) based
approach for the allocation of FACTS (Flexible AC Transmission
System) devices for the improvement of Power transfer capacity in an
interconnected Power System. The GA based approach is applied on
IEEE 30 BUS System. The system is reactively loaded starting from
base to 200% of base load. FACTS devices are installed in the
different locations of the power system and system performance is
noticed with and without FACTS devices. First, the locations, where
the FACTS devices to be placed is determined by calculating active
and reactive power flows in the lines. Genetic Algorithm is then
applied to find the amount of magnitudes of the FACTS devices. This
approach of GA based placement of FACTS devices is tremendous
beneficial both in terms of performance and economy is clearly
observed from the result obtained.
Abstract: This paper proposes an effective algorithm approach to hybrid control systems combining fuzzy logic and conventional control techniques of controlling the speed of induction motor assumed to operate in high-performance drives environment. The introducing of fuzzy logic in the control systems helps to achieve good dynamical response, disturbance rejection and low sensibility to parameter variations and external influences. Some fundamentals of the fuzzy logic control are preliminary illustrated. The developed control algorithm is robust, efficient and simple. It also assures precise trajectory tracking with the prescribed dynamics. Experimental results have shown excellent tracking performance of the proposed control system, and have convincingly demonstrated the validity and the usefulness of the hybrid fuzzy controller in high-performance drives with parameter and load uncertainties. Satisfactory performance was observed for most reference tracks.
Abstract: Despite so many years- development, the mainstream of workflow solutions from IT industries has not made ad-hoc workflow-support easy or inexpensive in MIS. Moreover, most of academic approaches tend to make their resulted BPM (Business Process Management) more complex and clumsy since they used to necessitate modeling workflow. To cope well with various ad-hoc or casual requirements on workflows while still keeping things simple and inexpensive, the author puts forth first the TSM design pattern that can provide a flexible workflow control while minimizing demand of predefinitions and modeling workflow, which introduces a generic approach for building BPM in workflow-aware MISs (Management Information Systems) with low development and running expenses.
Abstract: The Aggregate Production Plan (APP) is a schedule of
the organization-s overall operations over a planning horizon to
satisfy demand while minimizing costs. It is the baseline for any
further planning and formulating the master production scheduling,
resources, capacity and raw material planning. This paper presents a
methodology to model the Aggregate Production Planning problem,
which is combinatorial in nature, when optimized with Genetic
Algorithms. This is done considering a multitude of constraints of
contradictory nature and the optimization criterion – overall cost,
made up of costs with production, work force, inventory, and
subcontracting. A case study of substantial size, used to develop the
model, is presented, along with the genetic operators.
Abstract: The paper investigates the potential of support vector
machines and Gaussian process based regression approaches to
model the oxygen–transfer capacity from experimental data of
multiple plunging jets oxygenation systems. The results suggest the
utility of both the modeling techniques in the prediction of the
overall volumetric oxygen transfer coefficient (KLa) from operational
parameters of multiple plunging jets oxygenation system. The
correlation coefficient root mean square error and coefficient of
determination values of 0.971, 0.002 and 0.945 respectively were
achieved by support vector machine in comparison to values of
0.960, 0.002 and 0.920 respectively achieved by Gaussian process
regression. Further, the performances of both these regression
approaches in predicting the overall volumetric oxygen transfer
coefficient was compared with the empirical relationship for multiple
plunging jets. A comparison of results suggests that support vector
machines approach works well in comparison to both empirical
relationship and Gaussian process approaches, and could successfully
be employed in modeling oxygen-transfer.
Abstract: A framework to estimate the state of dynamically
varying environment where data are generated from heterogeneous
sources possessing partial knowledge about the environment is presented.
This is entirely derived within Dempster-Shafer and Evidence
Filtering frameworks. The belief about the current state is expressed
as belief and plausibility functions. An addition to Single Input
Single Output Evidence Filter, Multiple Input Single Output Evidence
Filtering approach is introduced. Variety of applications such as
situational estimation of an emergency environment can be developed
within the framework successfully. Fire propagation scenario is used
to justify the proposed framework, simulation results are presented.
Abstract: Process capability index Cpk is the most widely
used index in making managerial decisions since it provides bounds
on the process yield for normally distributed processes. However,
existent methods for assessing process performance which
constructed by statistical inference may unfortunately lead to fine
results, because uncertainties exist in most real-world applications.
Thus, this study adopts fuzzy inference to deal with testing of Cpk .
A brief score is obtained for assessing a supplier’s process instead of
a severe evaluation.
Abstract: Addition of milli or micro sized particles to the heat
transfer fluid is one of the many techniques employed for improving
heat transfer rate. Though this looks simple, this method has
practical problems such as high pressure loss, clogging and erosion
of the material of construction. These problems can be overcome by
using nanofluids, which is a dispersion of nanosized particles in a
base fluid. Nanoparticles increase the thermal conductivity of the
base fluid manifold which in turn increases the heat transfer rate.
Nanoparticles also increase the viscosity of the basefluid resulting in
higher pressure drop for the nanofluid compared to the base fluid. So
it is imperative that the Reynolds number (Re) and the volume
fraction have to be optimum for better thermal hydraulic
effectiveness. In this work, the heat transfer enhancement using
aluminium oxide nanofluid using low and high volume fraction
nanofluids in turbulent pipe flow with constant wall temperature has
been studied by computational fluid dynamic modeling of the
nanofluid flow adopting the single phase approach. Nanofluid, up till
a volume fraction of 1% is found to be an effective heat transfer
enhancement technique. The Nusselt number (Nu) and friction factor
predictions for the low volume fractions (i.e. 0.02%, 0.1 and 0.5%)
agree very well with the experimental values of Sundar and Sharma
(2010). While, predictions for the high volume fraction nanofluids
(i.e. 1%, 4% and 6%) are found to have reasonable agreement with
both experimental and numerical results available in the literature.
So the computationally inexpensive single phase approach can be
used for heat transfer and pressure drop prediction of new nanofluids.
Abstract: The success of an e-learning system is highly
dependent on the quality of its educational content and how effective,
complete, and simple the design tool can be for teachers. Educational
modeling languages (EMLs) are proposed as design languages
intended to teachers for modeling diverse teaching-learning
experiences, independently of the pedagogical approach and in
different contexts. However, most existing EMLs are criticized for
being too abstract and too complex to be understood and manipulated
by teachers. In this paper, we present a visual EML that simplifies the
process of designing learning scenarios for teachers with no
programming background. Based on the conceptual framework of the
activity theory, our resulting visual EML focuses on using Domainspecific
modeling techniques to provide a pedagogical level of
abstraction in the design process.
Abstract: This paper presents a new optimization technique based on quantum computing principles to solve a security constrained power system economic dispatch problem (SCED). The proposed technique is a population-based algorithm, which uses some quantum computing elements in coding and evolving groups of potential solutions to reach the optimum following a partially directed random approach. The SCED problem is formulated as a constrained optimization problem in a way that insures a secure-economic system operation. Real Coded Quantum-Inspired Evolution Algorithm (RQIEA) is then applied to solve the constrained optimization formulation. Simulation results of the proposed approach are compared with those reported in literature. The outcome is very encouraging and proves that RQIEA is very applicable for solving security constrained power system economic dispatch problem (SCED).
Abstract: This paper introduces a framework based on the collaboration of multi agent and hyper-heuristics to find a solution of the real single machine production problem. There are many techniques used to solve this problem. Each of it has its own advantages and disadvantages. By the collaboration of multi agent system and hyper-heuristics, we can get more optimal solution. The hyper-heuristics approach operates on a search space of heuristics rather than directly on a search space of solutions. The proposed framework consists of some agents, i.e. problem agent, trainer agent, algorithm agent (GPHH, GAHH, and SAHH), optimizer agent, and solver agent. Some low level heuristics used in this paper are MRT, SPT, LPT, EDD, LDD, and MON
Abstract: Fractional delay FIR filters design method based on
the differential evolution algorithm is presented. Differential evolution
is an evolutionary algorithm for solving a global optimization problems in the continuous search space. In the proposed approach,
an evolutionary algorithm is used to determine the coefficients of
a fractional delay FIR filter based on the Farrow structure. Basic
differential evolution is enhanced with a restricted mating technique,
which improves the algorithm performance in terms of convergence
speed and obtained solution. Evolutionary optimization is carried out by minimizing an objective function which is based on the amplitude
response and phase delay errors. Experimental results show that the proposed algorithm leads to a reduction in the amplitude response and phase delay errors relative to those achieved with the Least-Squares
method.
Abstract: Quality Function Deployment (QFD) is an expounded, multi-step planning method for delivering commodity, services, and processes to customers, both external and internal to an organization. It is a way to convert between the diverse customer languages expressing demands (Voice of the Customer), and the organization-s languages expressing results that sate those demands. The policy is to establish one or more matrices that inter-relate producer and consumer reciprocal expectations. Due to its visual presence is called the “House of Quality" (HOQ). In this paper, we assumed HOQ in multi attribute decision making (MADM) pattern and through a proposed MADM method, rank technical specifications. Thereafter compute satisfaction degree of customer requirements and for it, we apply vagueness and uncertainty conditions in decision making by fuzzy set theory. This approach would propound supervised neural network (perceptron) for MADM problem solving.
Abstract: The purpose of this paper is to present two different
approaches of financial distress pre-warning models appropriate for
risk supervisors, investors and policy makers. We examine a sample
of the financial institutions and electronic companies of Taiwan
Security Exchange (TSE) market from 2002 through 2008. We
present a binary logistic regression with paned data analysis. With
the pooled binary logistic regression we build a model including
more variables in the regression than with random effects, while the
in-sample and out-sample forecasting performance is higher in
random effects estimation than in pooled regression. On the other
hand we estimate an Adaptive Neuro-Fuzzy Inference System
(ANFIS) with Gaussian and Generalized Bell (Gbell) functions and
we find that ANFIS outperforms significant Logit regressions in both
in-sample and out-of-sample periods, indicating that ANFIS is a
more appropriate tool for financial risk managers and for the
economic policy makers in central banks and national statistical
services.
Abstract: Magnesium is used implant material potentially for
non-toxicity to the human body. Due to the excellent
bio-compatibility, Mg alloys is applied to implants avoiding removal
second surgery. However, it is found commercial magnesium alloys
including aluminum has low corrosion resistance, resulting
subcutaneous gas bubbles and consequently the approach as
permanent bio-materials. Generally, Aluminum is known to pollution
substance, and it raises toxicity to nervous system. Therefore
especially Mg-35Zn-3Ca alloy is prepared for new biodegradable
materials in this study. And the pulsed power is used in
constant-current mode of DC power kinds of anodization. Based on
the aforementioned study, it examines corrosion resistance and
biocompatibility by effect of current and frequency variation. The
surface properties and thickness were compared using scanning
electronic microscopy. Corrosion resistance was assessed via
potentiodynamic polarization and the effect of oxide layer on the body
was assessed cell viability. Anodized Mg-35Zn-3Ca alloy has good
biocompatibility in vitro by current and frequency variation.
Abstract: One important problem in today organizations is the
existence of non-integrated information systems, inconsistency and
lack of suitable correlations between legacy and modern systems.
One main solution is to transfer the local databases into a global one.
In this regards we need to extract the data structures from the legacy
systems and integrate them with the new technology systems. In
legacy systems, huge amounts of a data are stored in legacy
databases. They require particular attention since they need more
efforts to be normalized, reformatted and moved to the modern
database environments. Designing the new integrated (global)
database architecture and applying the reverse engineering requires
data normalization. This paper proposes the use of database reverse
engineering in order to integrate legacy and modern databases in
organizations. The suggested approach consists of methods and
techniques for generating data transformation rules needed for the
data structure normalization.