Abstract: In this paper we use data mining techniques to investigate factors that contribute significantly to enhancing the risk of acute coronary syndrome. We assume that the dependent variable is diagnosis – with dichotomous values showing presence or absence of disease. We have applied binary regression to the factors affecting the dependent variable. The data set has been taken from two different cardiac hospitals of Karachi, Pakistan. We have total sixteen variables out of which one is assumed dependent and other 15 are independent variables. For better performance of the regression model in predicting acute coronary syndrome, data reduction techniques like principle component analysis is applied. Based on results of data reduction, we have considered only 14 out of sixteen factors.
Abstract: Surface roughness (Ra) is one of the most important requirements in machining process. In order to obtain better surface roughness, the proper setting of cutting parameters is crucial before the process take place. This research presents the development of mathematical model for surface roughness prediction before milling process in order to evaluate the fitness of machining parameters; spindle speed, feed rate and depth of cut. 84 samples were run in this study by using FANUC CNC Milling α-Τ14ιE. Those samples were randomly divided into two data sets- the training sets (m=60) and testing sets(m=24). ANOVA analysis showed that at least one of the population regression coefficients was not zero. Multiple Regression Method was used to determine the correlation between a criterion variable and a combination of predictor variables. It was established that the surface roughness is most influenced by the feed rate. By using Multiple Regression Method equation, the average percentage deviation of the testing set was 9.8% and 9.7% for training data set. This showed that the statistical model could predict the surface roughness with about 90.2% accuracy of the testing data set and 90.3% accuracy of the training data set.
Abstract: Static analysis of source code is used for auditing web
applications to detect the vulnerabilities. In this paper, we propose a
new algorithm to analyze the PHP source code for detecting LFI and
RFI potential vulnerabilities. In our approach, we first define some
patterns for finding some functions which have potential to be abused
because of unhandled user inputs. More precisely, we use regular
expression as a fast and simple method to define some patterns for
detection of vulnerabilities. As inclusion functions could be also used
in a safe way, there could occur many false positives (FP). The first
cause of these FP-s could be that the function does not use a usersupplied
variable as an argument. So, we extract a list of usersupplied
variables to be used for detecting vulnerable lines of code.
On the other side, as vulnerability could spread among the variables
like by multi-level assignment, we also try to extract the hidden usersupplied
variables. We use the resulted list to decrease the false
positives of our method. Finally, as there exist some ways to prevent
the vulnerability of inclusion functions, we define also some patterns
to detect them and decrease our false positives.
Abstract: Anaerobic Digestion has become a promising
technology for biological transformation of organic fraction of the
municipal solid wastes (MSW). In order to represent the kinetic
behavior of such biological process and thereby to design a reactor
system, development of a mathematical model is essential.
Addressing this issue, a simplistic mathematical model has been
developed for anaerobic digestion of MSW in a continuous flow
reactor unit under homogeneous steady state condition. Upon
simulated hydrolysis, the kinetics of biomass growth and substrate
utilization rate are assumed to follow first order reaction kinetics.
Simulation of this model has been conducted by studying sensitivity
of various process variables. The model was simulated using typical
kinetic data of anaerobic digestion MSW and typical MSW
characteristics of Kolkata. The hydraulic retention time (HRT) and
solid retention time (SRT) time were mainly estimated by varying
different model parameters like efficiency of reactor, influent
substrate concentration and biomass concentration. Consequently,
design table and charts have also been prepared for ready use in the
actual plant operation.
Abstract: The Random Coefficient Dynamic Regression (RCDR)
model is to developed from Random Coefficient Autoregressive
(RCA) model and Autoregressive (AR) model. The RCDR model
is considered by adding exogenous variables to RCA model. In this
paper, the concept of the Maximum Likelihood (ML) method is used
to estimate the parameter of RCDR(1,1) model. Simulation results
have shown the AIC and BIC criterion to compare the performance of
the the RCDR(1,1) model. The variables as the stationary and weakly
stationary data are good estimates where the exogenous variables
are weakly stationary. However, the model selection indicated that
variables are nonstationarity data based on the stationary data of the
exogenous variables.
Abstract: This research was conducted for the first time at the
southeastern coasts of the Caspian Sea in order to evaluate the
performance of osteichthyes cooperatives through production (catch)
function. Using one of the indirect valuation methods in this research,
contributory factors in catch were identified and were inserted into
the function as independent variables. In order to carry out this
research, the performance of 25 Osteichthyes catching cooperatives
in the utilization year of 2009 which were involved in fishing in
Miankale wildlife refuge region. The contributory factors in catch
were divided into groups of economic, ecological and biological
factors. In the mentioned function, catch rate of the cooperative were
inserted into as the dependant variable and fourteen partial variables
in terms of nine general variables as independent variables. Finally,
after function estimation, seven variables were rendered significant at
99 percent reliably level. The results of the function estimation
indicated that human resource (fisherman quantity) had the greatest
positive effect on catch rate with an influence coefficient of 1.7 while
weather conditions had the greatest negative effect on the catch rate
of cooperatives with an influence coefficient of -2.07. Moreover,
factors like member's share, experience and fisherman training and
fishing effort played the main roles in the catch rate of cooperative
with influence coefficients of 0.81, 0.5 and 0.21, respectively.
Abstract: In this paper, a tooth shape optimization method for
cogging torque reduction in Permanent Magnet (PM) motors is
developed by using the Reduced Basis Technique (RBT) coupled by
Finite Element Analysis (FEA) and Design of Experiments (DOE)
methods. The primary objective of the method is to reduce the
enormous number of design variables required to define the tooth
shape. RBT is a weighted combination of several basis shapes. The
aim of the method is to find the best combination using the weights
for each tooth shape as the design variables. A multi-level design
process is developed to find suitable basis shapes or trial shapes at
each level that can be used in the reduced basis technique. Each level
is treated as a separated optimization problem until the required
objective – minimum cogging torque – is achieved. The process is
started with geometrically simple basis shapes that are defined by
their shape co-ordinates. The experimental design of Taguchi method
is used to build the approximation model and to perform
optimization. This method is demonstrated on the tooth shape
optimization of a 8-poles/12-slots PM motor.
Abstract: In this paper, some practical solid transportation models are formulated considering per trip capacity of each type of conveyances with crisp and rough unit transportation costs. This is applicable for the system in which full vehicles, e.g. trucks, rail coaches are to be booked for transportation of products so that transportation cost is determined on the full of the conveyances. The models with unit transportation costs as rough variables are transformed into deterministic forms using rough chance constrained programming with the help of trust measure. Numerical examples are provided to illustrate the proposed models in crisp environment as well as with unit transportation costs as rough variables.
Abstract: The Prediction of aerodynamic characteristics and
shape optimization of airfoil under the ground effect have been carried
out by integration of computational fluid dynamics and the multiobjective
Pareto-based genetic algorithm. The main flow
characteristics around an airfoil of WIG craft are lift force, lift-to-drag
ratio and static height stability (H.S). However, they show a strong
trade-off phenomenon so that it is not easy to satisfy the design
requirements simultaneously. This difficulty can be resolved by the
optimal design. The above mentioned three characteristics are chosen
as the objective functions and NACA0015 airfoil is considered as a
baseline model in the present study. The profile of airfoil is
constructed by Bezier curves with fourteen control points and these
control points are adopted as the design variables. For multi-objective
optimization problems, the optimal solutions are not unique but a set
of non-dominated optima and they are called Pareto frontiers or Pareto
sets. As the results of optimization, forty numbers of non- dominated
Pareto optima can be obtained at thirty evolutions.
Abstract: Cooktop burners are widely used nowadays. In
cooktop burner design, nozzle efficiency and greenhouse
gas(GHG) emissions mainly depend on heat transfer from the
premixed flame to the impinging surface. This is a complicated
issue depending on the individual and combined effects of various
input combustion variables. Optimal operating conditions for
sustainable burner design were rarely addressed, especially in the
case of multiple slot-jet burners. Through evaluating the optimal
combination of combustion conditions for a premixed slot-jet
array, this paper develops a practical approach for the sustainable
design of gas cooktop burners. Efficiency, CO and NOx emissions
in respect of an array of slot jets using premixed flames were
analysed. Response surface experimental design were applied to
three controllable factors of the combustion process, viz.
Reynolds number, equivalence ratio and jet-to-vessel distance.
Desirability Function Approach(DFA) is the analytic technique
used for the simultaneous optimization of the efficiency and
emission responses.
Abstract: This paper describes the shape optimization of impeller
blades for a anti-heeling bidirectional axial flow pump used in ships.
In general, a bidirectional axial pump has an efficiency much lower
than the classical unidirectional pump because of the symmetry of the
blade type. In this paper, by focusing on a pump impeller, the shape of
blades is redesigned to reach a higher efficiency in a bidirectional axial
pump. The commercial code employed in this simulation is CFX v.13.
CFD result of pump torque, head, and hydraulic efficiency was
compared. The orthogonal array (OA) and analysis of variance
(ANOVA) techniques and surrogate model based optimization using
orthogonal polynomial, are employed to determine the main effects
and their optimal design variables. According to the optimal design,
we confirm an effective design variable in impeller blades and explain
the optimal solution, the usefulness for satisfying the constraints of
pump torque and head.
Abstract: Fuzzy linear programming is an application of fuzzy set theory in linear decision making problems and most of these problems are related to linear programming with fuzzy variables. A convenient method for solving these problems is based on using of auxiliary problem. In this paper a new method for solving fuzzy variable linear programming problems directly using linear ranking functions is proposed. This method uses simplex tableau which is used for solving linear programming problems in crisp environment before.
Abstract: Hypernetworks are a generalized graph structure
representing higher-order interactions between variables. We present a
method for self-organizing hypernetworks to learn an associative
memory of sentences and to recall the sentences from this memory.
This learning method is inspired by the “mental chemistry" model of
cognition and the “molecular self-assembly" technology in
biochemistry. Simulation experiments are performed on a corpus of
natural-language dialogues of approximately 300K sentences
collected from TV drama captions. We report on the sentence
completion performance as a function of the order of word-interaction
and the size of the learning corpus, and discuss the plausibility of this
architecture as a cognitive model of language learning and memory.
Abstract: In this paper, all variables are supposed to be integer
and positive. In this modern method, objective function is assumed to
be maximized or minimized but constraints are always explained like
less or equal to. In this method, choosing a dual combination of ideal
nonequivalent and omitting one of variables. With continuing this
act, finally, having one nonequivalent with (n-m+1) unknown
quantities in which final nonequivalent, m is counter for constraints,
n is counter for variables of decision.
Abstract: Data Mining aims at discovering knowledge out of
data and presenting it in a form that is easily comprehensible to
humans. One of the useful applications in Egypt is the Cancer
management, especially the management of Acute Lymphoblastic
Leukemia or ALL, which is the most common type of cancer in
children.
This paper discusses the process of designing a prototype that can
help in the management of childhood ALL, which has a great
significance in the health care field. Besides, it has a social impact
on decreasing the rate of infection in children in Egypt. It also
provides valubale information about the distribution and
segmentation of ALL in Egypt, which may be linked to the possible
risk factors.
Undirected Knowledge Discovery is used since, in the case of this
research project, there is no target field as the data provided is
mainly subjective. This is done in order to quantify the subjective
variables. Therefore, the computer will be asked to identify
significant patterns in the provided medical data about ALL. This
may be achieved through collecting the data necessary for the
system, determimng the data mining technique to be used for the
system, and choosing the most suitable implementation tool for the
domain.
The research makes use of a data mining tool, Clementine, so as to
apply Decision Trees technique. We feed it with data extracted from
real-life cases taken from specialized Cancer Institutes. Relevant
medical cases details such as patient medical history and diagnosis
are analyzed, classified, and clustered in order to improve the disease
management.
Abstract: We study the spatial design of experiment and we want to select a most informative subset, having prespecified size, from a set of correlated random variables. The problem arises in many applied domains, such as meteorology, environmental statistics, and statistical geology. In these applications, observations can be collected at different locations and possibly at different times. In spatial design, when the design region and the set of interest are discrete then the covariance matrix completely describe any objective function and our goal is to choose a feasible design that minimizes the resulting uncertainty. The problem is recast as that of maximizing the determinant of the covariance matrix of the chosen subset. This problem is NP-hard. For using these designs in computer experiments, in many cases, the design space is very large and it's not possible to calculate the exact optimal solution. Heuristic optimization methods can discover efficient experiment designs in situations where traditional designs cannot be applied, exchange methods are ineffective and exact solution not possible. We developed a GA algorithm to take advantage of the exploratory power of this algorithm. The successful application of this method is demonstrated in large design space. We consider a real case of design of experiment. In our problem, design space is very large and for solving the problem, we used proposed GA algorithm.
Abstract: An enhanced particle swarm optimization algorithm
(PSO) is presented in this work to solve the non-convex OPF
problem that has both discrete and continuous optimization variables.
The objective functions considered are the conventional quadratic
function and the augmented quadratic function. The latter model
presents non-differentiable and non-convex regions that challenge
most gradient-based optimization algorithms. The optimization
variables to be optimized are the generator real power outputs and
voltage magnitudes, discrete transformer tap settings, and discrete
reactive power injections due to capacitor banks. The set of equality
constraints taken into account are the power flow equations while the
inequality ones are the limits of the real and reactive power of the
generators, voltage magnitude at each bus, transformer tap settings,
and capacitor banks reactive power injections. The proposed
algorithm combines PSO with Newton-Raphson algorithm to
minimize the fuel cost function. The IEEE 30-bus system with six
generating units is used to test the proposed algorithm. Several cases
were investigated to test and validate the consistency of detecting
optimal or near optimal solution for each objective. Results are
compared to solutions obtained using sequential quadratic
programming and Genetic Algorithms.
Abstract: The optimal control problem for the viscoelastic melt
spinning process has not been reported yet in the literature. In this
study, an optimal control problem for a mathematical model of a
viscoelastic melt spinning process is considered. Maxwell-Oldroyd
model is used to describe the rheology of the polymeric material, the
fiber is made of. The extrusion velocity of the polymer at the spinneret
as well as the velocity and the temperature of the quench air and the
fiber length serve as control variables. A constrained optimization
problem is derived and the first–order optimality system is set up
to obtain the adjoint equations. Numerical solutions are carried out
using a steepest descent algorithm. A computer program in MATLAB
is developed for simulations.
Abstract: In this paper bi-annual time series data on unemployment rates (from the Labour Force Survey) are expanded to quarterly rates and linked to quarterly unemployment rates (from the Quarterly Labour Force Survey). The resultant linked series and the consumer price index (CPI) series are examined using Johansen’s cointegration approach and vector error correction modeling. The study finds that both the series are integrated of order one and are cointegrated. A statistically significant co-integrating relationship is found to exist between the time series of unemployment rates and the CPI. Given this significant relationship, the study models this relationship using Vector Error Correction Models (VECM), one with a restriction on the deterministic term and the other with no restriction.
A formal statistical confirmation of the existence of a unique linear and lagged relationship between inflation and unemployment for the period between September 2000 and June 2011 is presented. For the given period, the CPI was found to be an unbiased predictor of the unemployment rate. This relationship can be explored further for the development of appropriate forecasting models incorporating other study variables.
Abstract: The aim of this paper is to provide an empirical
evidence about the effects that the management of continuous
training have on employability (or employment stability) in the
Spanish labour market. With this purpose a binary logit model with
interaction effect is been used. The dependent variable includes two
situations of the active workers: continuous and discontinuous
employability. To distinguish between them an Employability Index
Stability (ESI) was calculated taking into account two factors: time
worked and job security. Various aspects of the continuous training
and personal workers data are used as independent variables. The
data obtained from a survey of a sample of 918 employed have
revealed a relationship between the likelihood of continuous
employability and continuous training received. The empirical results
support the positive and significant relationship between various
aspects of the training provided by firms and employability
likelihood of the workers, postulate alike from a theoretical point of
view.