Abstract: In general, issues related to design and maintenance are considered in an independent manner. However, the decisions made in these two sets influence each other. The design for maintenance is considered an opportunity to optimize the life cycle cost of a product, particularly in the nuclear or aeronautical field, where maintenance expenses represent more than 60% of life cycle costs. The design of large-scale systems starts with product architecture, a choice of components in terms of cost, reliability, weight and other attributes, corresponding to the specifications. On the other hand, the design must take into account maintenance by improving, in particular, real-time monitoring of equipment through the integration of new technologies such as connected sensors and intelligent actuators. We noticed that different approaches used in the Design For Maintenance (DFM) methods are limited to the simultaneous characterization of the reliability and maintainability of a multi-component system. This article proposes a method of DFM that assists designers to propose dynamic maintenance for multi-component industrial systems. The term "dynamic" refers to the ability to integrate available monitoring data to adapt the maintenance decision in real time. The goal is to maximize the availability of the system at a given life cycle cost. This paper presents an approach for simultaneous optimization of the design and maintenance of multi-component systems. Here the design is characterized by four decision variables for each component (reliability level, maintainability level, redundancy level, and level of monitoring data). The maintenance is characterized by two decision variables (the dates of the maintenance stops and the maintenance operations to be performed on the system during these stops). The DFM model helps the designers choose technical solutions for the large-scale industrial products. Large-scale refers to the complex multi-component industrial systems and long life-cycle, such as trains, aircraft, etc. The method is based on a two-level hybrid algorithm for simultaneous optimization of design and maintenance, using genetic algorithms. The first level is to select a design solution for a given system that considers the life cycle cost and the reliability. The second level consists of determining a dynamic and optimal maintenance plan to be deployed for a design solution. This level is based on the Maintenance Free Operating Period (MFOP) concept, which takes into account the decision criteria such as, total reliability, maintenance cost and maintenance time. Depending on the life cycle duration, the desired availability, and the desired business model (sales or rental), this tool provides visibility of overall costs and optimal product architecture.
Abstract: In this paper, the concept of a non-dominated sorting multi-objective particle swarm optimization with local search (NSPSO-LS) is presented for the optimal design of multimachine power system stabilizers (PSSs). The controller design is formulated as an optimization problem in order to shift the system electromechanical modes in a pre-specified region in the s-plan. A composite set of objective functions comprising the damping factor and the damping ratio of the undamped and lightly damped electromechanical modes is considered. The performance of the proposed optimization algorithm is verified for the 3-machine 9-bus system. Simulation results based on eigenvalue analysis and nonlinear time-domain simulation show the potential and superiority of the NSPSO-LS algorithm in tuning PSSs over a wide range of loading conditions and large disturbance compared to the classic PSO technique and genetic algorithms.
Abstract: This paper presents the effect of actuation schemes on the performance of parallel manipulators and also how the singularity loci have been changed in the reachable workspace of the manipulator with the choice of actuation scheme to drive the manipulator. The performance of the eight possible actuation schemes of 3RRR planar parallel manipulator is compared with each other. The optimal design problem is formulated to find the manipulator geometry that maximizes the singularity free conditioned workspace for all the eight actuation cases, the optimization problem is solved by using genetic algorithms.
Abstract: Power systems are operating under stressed condition
due to continuous increase in demand of load. This can lead to
voltage instability problem when face additional load increase or
contingency. In order to avoid voltage instability suitable size of
reactive power compensation at optimal location in the system is
required which improves the load margin. This work aims at
obtaining optimal size as well as location of compensation in the 39-
bus New England system with the help of Bacteria Foraging and
Genetic algorithms. To reduce the computational time the work
identifies weak candidate buses in the system, and then picks only
two of them to take part in the optimization. The objective function is
based on a recently proposed voltage stability index which takes into
account the weighted average sensitivity index is a simpler and faster
approach than the conventional CPF algorithm. BFOA has been
found to give better results compared to GA.
Abstract: The Choquet integral is a tool for the information fusion that is very effective in the case where fuzzy measures associated with it are well chosen. In this paper, we propose a new approach for calculating fuzzy measures associated with the Choquet integral in a context of data fusion in multimodal biometrics. The proposed approach is based on genetic algorithms. It has been validated in two databases: the first base is relative to synthetic scores and the second one is biometrically relating to the face, fingerprint and palmprint. The results achieved attest the robustness of the proposed approach.
Abstract: In this paper, we present a comparative study of the
genetic algorithms and Hessian-s methods for optimal research of the
active powers in an electric network of power. The objective function
which is the performance index of production of electrical energy is
minimized by satisfying the constraints of the equality type and
inequality type initially by the Hessian-s methods and in the second
time by the genetic Algorithms. The results found by the application
of AG for the minimization of the electric production costs of power
are very encouraging. The algorithms seem to be an effective
technique to solve a great number of problems and which are in
constant evolution. Nevertheless it should be specified that the
traditional binary representation used for the genetic algorithms
creates problems of optimization of management of the large-sized
networks with high numerical precision.
Abstract: Fine-grained data replication over the Internet allows duplication of frequently accessed data objects, as opposed to entire sites, to certain locations so as to improve the performance of largescale content distribution systems. In a distributed system, agents representing their sites try to maximize their own benefit since they are driven by different goals such as to minimize their communication costs, latency, etc. In this paper, we will use game theoretical techniques and in particular auctions to identify a bidding mechanism that encapsulates the selfishness of the agents, while having a controlling hand over them. In essence, the proposed game theory based mechanism is the study of what happens when independent agents act selfishly and how to control them to maximize the overall performance. A bidding mechanism asks how one can design systems so that agents- selfish behavior results in the desired system-wide goals. Experimental results reveal that this mechanism provides excellent solution quality, while maintaining fast execution time. The comparisons are recorded against some well known techniques such as greedy, branch and bound, game theoretical auctions and genetic algorithms.
Abstract: Fuzzy C-means Clustering algorithm (FCM) is a
method that is frequently used in pattern recognition. It has the
advantage of giving good modeling results in many cases, although,
it is not capable of specifying the number of clusters by itself. In
FCM algorithm most researchers fix weighting exponent (m) to a
conventional value of 2 which might not be the appropriate for all
applications. Consequently, the main objective of this paper is to use
the subtractive clustering algorithm to provide the optimal number of
clusters needed by FCM algorithm by optimizing the parameters of
the subtractive clustering algorithm by an iterative search approach
and then to find an optimal weighting exponent (m) for the FCM
algorithm. In order to get an optimal number of clusters, the iterative
search approach is used to find the optimal single-output Sugenotype
Fuzzy Inference System (FIS) model by optimizing the
parameters of the subtractive clustering algorithm that give minimum
least square error between the actual data and the Sugeno fuzzy
model. Once the number of clusters is optimized, then two
approaches are proposed to optimize the weighting exponent (m) in
the FCM algorithm, namely, the iterative search approach and the
genetic algorithms. The above mentioned approach is tested on the
generated data from the original function and optimal fuzzy models
are obtained with minimum error between the real data and the
obtained fuzzy models.
Abstract: In this paper, we introduce an effective strategy for
subgoal division and ordering based upon recursive subgoals and
combine this strategy with a genetic-based planning approach. This
strategy can be applied to domains with conjunctive goals. The main
idea is to recursively decompose a goal into a set of serializable
subgoals and to specify a strict ordering among the subgoals.
Empirical results show that the recursive subgoal strategy reduces the
size of the search space and improves the quality of solutions to
planning problems.
Abstract: An accurate optimal design of laminated composite
structures may present considerable difficulties due to the complexity
and multi-modality of the functional design space. The Big Bang
– Big Crunch (BB-BC) optimization method is a relatively new
technique and has already proved to be a valuable tool for structural
optimization. In the present study the exceptional efficiency of the
method is demonstrated by an example of the lay-up optimization
of multilayered anisotropic cylinders based on a three-dimensional
elasticity solution. It is shown that, due to its simplicity and speed,
the BB-BC is much more efficient for this class of problems when
compared to the genetic algorithms.
Abstract: This paper presents an approach for the design of
fuzzy logic power system stabilizers using genetic algorithms. In the
proposed fuzzy expert system, speed deviation and its derivative
have been selected as fuzzy inputs. In this approach the parameters of
the fuzzy logic controllers have been tuned using genetic algorithm.
Incorporation of GA in the design of fuzzy logic power system
stabilizer will add an intelligent dimension to the stabilizer and
significantly reduces computational time in the design process. It is
shown in this paper that the system dynamic performance can be
improved significantly by incorporating a genetic-based searching
mechanism. To demonstrate the robustness of the genetic based
fuzzy logic power system stabilizer (GFLPSS), simulation studies on
multimachine system subjected to small perturbation and three-phase
fault have been carried out. Simulation results show the superiority
and robustness of GA based power system stabilizer as compare to
conventionally tuned controller to enhance system dynamic
performance over a wide range of operating conditions.
Abstract: In this paper a genetic algorithms approach for solving the linear and quadratic fuzzy equations Ãx̃=B̃ and Ãx̃2 + B̃x̃=C̃ , where Ã, B̃, C̃ and x̃ are fuzzy numbers is proposed by genetic algorithms. Our genetic based method initially starts with a set of random fuzzy solutions. Then in each generation of genetic algorithms, the solution candidates converge more to better fuzzy solution x̃b . In this proposed method the final reached x̃b is not only restricted to fuzzy triangular and it can be fuzzy number.
Abstract: This work presents a recursive identification algorithm. This algorithm relates to the identification of closed loop system with Variable Structure Controller. The approach suggested includes two stages. In the first stage a genetic algorithm is used to obtain the parameters of switching function which gives a control signal rich in commutations (i.e. a control signal whose spectral characteristics are closest possible to those of a white noise signal). The second stage consists in the identification of the system parameters by the instrumental variable method and using the optimal switching function parameters obtained with the genetic algorithm. In order to test the validity of this algorithm a simulation example is presented.
Abstract: In this paper, we propose a selective mutation method
for improving the performances of genetic algorithms. In selective
mutation, individuals are first ranked and then additionally mutated
one bit in a part of their strings which is selected corresponding to
their ranks. This selective mutation helps genetic algorithms to fast
approach the global optimum and to quickly escape local optima.
This results in increasing the performances of genetic algorithms.
We measured the effects of selective mutation with four function
optimization problems. It was found from extensive experiments that
the selective mutation can significantly enhance the performances of
genetic algorithms.
Abstract: In this contribution an innovative platform is being
presented that integrates intelligent agents and evolutionary
computation techniques in legacy e-learning environments. It
introduces the design and development of a scalable and
interoperable integration platform supporting:
I) various assessment agents for e-learning environments,
II) a specific resource retrieval agent for the provision of
additional information from Internet sources matching the
needs and profile of the specific user and
III) a genetic algorithm designed to extract efficient information
(classifying rules) based on the students- answering input
data.
The agents are implemented in order to provide intelligent
assessment services based on computational intelligence techniques
such as Bayesian Networks and Genetic Algorithms.
The proposed Genetic Algorithm (GA) is used in order to extract
efficient information (classifying rules) based on the students-
answering input data. The idea of using a GA in order to fulfil this
difficult task came from the fact that GAs have been widely used in
applications including classification of unknown data.
The utilization of new and emerging technologies like web
services allows integrating the provided services to any web based
legacy e-learning environment.
Abstract: Fine-grained data replication over the Internet allows duplication of frequently accessed data objects, as opposed to entire sites, to certain locations so as to improve the performance of largescale content distribution systems. In a distributed system, agents representing their sites try to maximize their own benefit since they are driven by different goals such as to minimize their communication costs, latency, etc. In this paper, we will use game theoretical techniques and in particular auctions to identify a bidding mechanism that encapsulates the selfishness of the agents, while having a controlling hand over them. In essence, the proposed game theory based mechanism is the study of what happens when independent agents act selfishly and how to control them to maximize the overall performance. A bidding mechanism asks how one can design systems so that agents- selfish behavior results in the desired system-wide goals. Experimental results reveal that this mechanism provides excellent solution quality, while maintaining fast execution time. The comparisons are recorded against some well known techniques such as greedy, branch and bound, game theoretical auctions and genetic algorithms.
Abstract: The main goal of data mining is to extract accurate, comprehensible and interesting knowledge from databases that may be considered as large search spaces. In this paper, a new, efficient type of Genetic Algorithm (GA) called uniform two-level GA is proposed as a search strategy to discover truly interesting, high-level prediction rules, a difficult problem and relatively little researched, rather than discovering classification knowledge as usual in the literatures. The proposed method uses the advantage of uniform population method and addresses the task of generalized rule induction that can be regarded as a generalization of the task of classification. Although the task of generalized rule induction requires a lot of computations, which is usually not satisfied with the normal algorithms, it was demonstrated that this method increased the performance of GAs and rapidly found interesting rules.
Abstract: This paper proposes a novel game theoretical
technique to address the problem of data object replication in largescale
distributed computing systems. The proposed technique draws
inspiration from computational economic theory and employs the
extended Vickrey auction. Specifically, players in a non-cooperative
environment compete for server-side scarce memory space to
replicate data objects so as to minimize the total network object
transfer cost, while maintaining object concurrency. Optimization of
such a cost in turn leads to load balancing, fault-tolerance and
reduced user access time. The method is experimentally evaluated
against four well-known techniques from the literature: branch and
bound, greedy, bin-packing and genetic algorithms. The experimental
results reveal that the proposed approach outperforms the four
techniques in both the execution time and solution quality.
Abstract: We board the problem of creating a seismic alert
system, based upon artificial neural networks, trained by using the
well-known back-propagation and genetic algorithms, in order to emit
the alarm for the population located into a specific city, about an
eminent earthquake greater than 4.5 Richter degrees, and avoiding
disasters and human loses. In lieu of using the propagation wave, we
employed the magnitude of the earthquake, to establish a correlation
between the recorded magnitudes from a controlled area and the city,
where we want to emit the alarm. To measure the accuracy of the
posed method, we use a database provided by CIRES, which contains
the records of 2500 quakes incoming from the State of Guerrero
and Mexico City. Particularly, we performed the proposed method to
generate an issue warning in Mexico City, employing the magnitudes
recorded in the State of Guerrero.
Abstract: Investment in a constructed facility represents a cost in
the short term that returns benefits only over the long term use of the
facility. Thus, the costs occur earlier than the benefits, and the owners
of facilities must obtain the capital resources to finance the costs of
construction. A project cannot proceed without an adequate
financing, and the cost of providing an adequate financing can be
quite large. For these reasons, the attention to the project finance is an
important aspect of project management. Finance is also a concern to
the other organizations involved in a project such as the general
contractor and material suppliers. Unless an owner immediately and
completely covers the costs incurred by each participant, these
organizations face financing problems of their own. At a more
general level, the project finance is the only one aspect of the general
problem of corporate finance. If numerous projects are considered
and financed together, then the net cash flow requirements constitute
the corporate financing problem for capital investment. Whether
project finance is performed at the project or at the corporate level
does not alter the basic financing problem .In this paper, we will first
consider facility financing from the owner's perspective, with due
consideration for its interaction with other organizations involved in a
project. Later, we discuss the problems of construction financing
which are crucial to the profitability and solvency of construction
contractors. The objective of this paper is to present the steps utilized
to determine the best combination of minimum project financing.
The proposed model considers financing; schedule and maximum net
area .The proposed model is called Project Financing and Schedule
Integration using Genetic Algorithms "PFSIGA". This model
intended to determine more steps (maximum net area) for any project
with a subproject. An illustrative example will demonstrate the
feature of this technique. The model verification and testing are put
into consideration.