Abstract: We provide a maximum norm analysis of a finite
element Schwarz alternating method for a nonlinear elliptic boundary
value problem of the form -Δu = f(u), on two overlapping sub
domains with non matching grids. We consider a domain which is
the union of two overlapping sub domains where each sub domain
has its own independently generated grid. The two meshes being
mutually independent on the overlap region, a triangle belonging to
one triangulation does not necessarily belong to the other one. Under
a Lipschitz assumption on the nonlinearity, we establish, on each sub
domain, an optimal L∞ error estimate between the discrete Schwarz
sequence and the exact solution of the boundary value problem.
Abstract: Every commercial bank optimises its asset portfolio
depending on the profitability of assets and chosen or imposed
constraints. This paper proposes and applies a stylized model for
optimising banks' asset and liability structure, reflecting profitability
of different asset categories and their risks as well as costs associated
with different liability categories and reserve requirements. The level
of detail for asset and liability categories is chosen to create a
suitably parsimonious model and to include the most important
categories in the model. It is shown that the most appropriate
optimisation criterion for the model is the maximisation of the ratio
of net interest income to assets. The maximisation of this ratio is
subject to several constraints. Some are accounting identities or
dictated by legislative requirements; others vary depending on the
market objectives for a particular bank. The model predicts variable
amount of assets allocated to loan provision.
Abstract: The demand of the energy management systems (EMS) set forth by modern power systems requires fast energy management systems. Contingency analysis is among the functions in EMS which is time consuming. In order to handle this limitation, this paper introduces agent based technology in the contingency analysis. The main function of agents is to speed up the performance. Negotiations process in decision making is explained and the issue set forth is the minimization of the operating costs. The IEEE 14 bus system and its line outage have been used in the research and simulation results are presented.
Abstract: This paper presents a novel approach for tuning unified power flow controller (UPFC) based damping controller in order to enhance the damping of power system low frequency oscillations. The design problem of damping controller is formulated as an optimization problem according to the eigenvalue-based objective function which is solved using iteration particle swarm optimization (IPSO). The effectiveness of the proposed controller is demonstrated through eigenvalue analysis and nonlinear time-domain simulation studies under a wide range of loading conditions. The simulation study shows that the designed controller by IPSO performs better than CPSO in finding the solution. Moreover, the system performance analysis under different operating conditions show that the δE based controller is superior to the mB based controller.
Abstract: This work presents a multiple objective linear programming (MOLP) model based on the desirability function approach for solving the aggregate production planning (APP) decision problem upon Masud and Hwang-s model. The proposed model minimises total production costs, carrying or backordering costs and rates of change in labor levels. An industrial case demonstrates the feasibility of applying the proposed model to the APP problems with three scenarios of inventory levels. The proposed model yields an efficient compromise solution and the overall levels of DM satisfaction with the multiple combined response levels. There has been a trend to solve complex planning problems using various metaheuristics. Therefore, in this paper, the multi-objective APP problem is solved by hybrid metaheuristics of the hunting search (HuSIHSA) and firefly (FAIHSA) mechanisms on the improved harmony search algorithm. Results obtained from the solution of are then compared. It is observed that the FAIHSA can be used as a successful alternative solution mechanism for solving APP problems over three scenarios. Furthermore, the FAIHSA provides a systematic framework for facilitating the decision-making process, enabling a decision maker interactively to modify the desirability function approach and related model parameters until a good optimal solution is obtained with proper selection of control parameters when compared.
Abstract: Efficient preprocessing is very essential for automatic
recognition of handwritten documents. In this paper, techniques on
segmenting words in handwritten Arabic text are presented. Firstly,
connected components (ccs) are extracted, and distances among
different components are analyzed. The statistical distribution of this
distance is then obtained to determine an optimal threshold for words
segmentation. Meanwhile, an improved projection based method is
also employed for baseline detection. The proposed method has been
successfully tested on IFN/ENIT database consisting of 26459
Arabic words handwritten by 411 different writers, and the results
were promising and very encouraging in more accurate detection of
the baseline and segmentation of words for further recognition.
Abstract: In modern telecommunications industry, demand &
supply chain management (DSCM) needs reliable design and
versatile tools to control the material flow. The objective for efficient
DSCM is reducing inventory, lead times and related costs in order to
assure reliable and on-time deliveries from manufacturing units
towards customers. In this paper the multi-rate expert system based
methodology for developing simulation tools that would enable
optimal DSCM for multi region, high volume and high complexity
manufacturing environment was proposed.
Abstract: The permanent magnet synchronous motor (PMSM) is
very useful in many applications. Vector control of PMSM is popular
kind of its control. In this paper, at first an optimal vector control for
PMSM is designed and then results are compared with conventional
vector control. Then, it is assumed that the measurements are noisy
and linear quadratic Gaussian (LQG) methodology is used to filter
the noises. The results of noisy optimal vector control and filtered
optimal vector control are compared to each other. Nonlinearity of
PMSM and existence of inverter in its control circuit caused that the
system is nonlinear and time-variant. With deriving average model,
the system is changed to nonlinear time-invariant and then the
nonlinear system is converted to linear system by linearization of
model around average values. This model is used to optimize vector
control then two optimal vector controls are compared to each other.
Simulation results show that the performance and robustness to noise
of the control system has been highly improved.
Abstract: Cross sections of As radionuclides in the interaction of natGe with 14-30 MeV protons have been deduced by off-line y-ray spectroscopy to find optimal reaction channels leading to radiotracers for positron emission tomography. The experimental results were compared with the previous results and those estimated by the compound nucleus reaction model.
Abstract: A multicriteria linear programming problem with integer variables and parameterized optimality principle "from lexicographic to Slater" is considered. A situation in which initial coefficients of penalty cost functions are not fixed but may be potentially a subject to variations is studied. For any efficient solution, appropriate measures of the quality are introduced which incorporate information about variations of penalty cost function coefficients. These measures correspond to the so-called stability and accuracy functions defined earlier for efficient solutions of a generic multicriteria combinatorial optimization problem with Pareto and lexicographic optimality principles. Various properties of such functions are studied and maximum norms of perturbations for which an efficient solution preserves the property of being efficient are calculated.
Abstract: The continuity in the electric supply of the electric installations is becoming one of the main requirements of the electric supply network (generation, transmission, and distribution of the electric energy). The achievement of this requirement depends from one side on the structure of the electric network and on the other side on the avaibility of the reserve source provided to maintain the supply in case of failure of the principal one. The avaibility of supply does not only depends on the reliability parameters of the both sources (principal and reserve) but it also depends on the reliability of the circuit breaker which plays the role of interlocking the reserve source in case of failure of the principal one. In addition, the principal source being under operation, its control can be ideal and sure, however, for the reserve source being in stop, a preventive maintenances which proceed on time intervals (periodicity) and for well defined lengths of time are envisaged, so that this source will always available in case of the principal source failure. The choice of the periodicity of preventive maintenance of the source of reserve influences directly the reliability of the electric feeder system In this work and on the basis of the semi- markovian's processes, the influence of the time of interlocking the reserve source upon the reliability of an industrial electric network is studied and is given the optimal time of interlocking the reserve source in case of failure the principal one, also the influence of the periodicity of the preventive maintenance of the source of reserve is studied and is given the optimal periodicity.
Abstract: This paper presents Genetic Algorithm (GA) based
approach for the allocation of FACTS (Flexible AC Transmission
System) devices for the improvement of Power transfer capacity in an
interconnected Power System. The GA based approach is applied on
IEEE 30 BUS System. The system is reactively loaded starting from
base to 200% of base load. FACTS devices are installed in the
different locations of the power system and system performance is
noticed with and without FACTS devices. First, the locations, where
the FACTS devices to be placed is determined by calculating active
and reactive power flows in the lines. Genetic Algorithm is then
applied to find the amount of magnitudes of the FACTS devices. This
approach of GA based placement of FACTS devices is tremendous
beneficial both in terms of performance and economy is clearly
observed from the result obtained.
Abstract: In comparison to the original SVM, which involves a
quadratic programming task; LS–SVM simplifies the required
computation, but unfortunately the sparseness of standard SVM is
lost. Another problem is that LS-SVM is only optimal if the training
samples are corrupted by Gaussian noise. In Least Squares SVM
(LS–SVM), the nonlinear solution is obtained, by first mapping the
input vector to a high dimensional kernel space in a nonlinear
fashion, where the solution is calculated from a linear equation set. In
this paper a geometric view of the kernel space is introduced, which
enables us to develop a new formulation to achieve a sparse and
robust estimate.
Abstract: In this paper, an accurate theoretical analysis for the achievable average channel capacity (in the Shannon sense) per user of a hybrid cellular direct-sequence/fast frequency hopping code-division multiple-access (DS/FFH-CDMA) system operating in a Rayleigh fading environment is presented. The analysis covers the downlink operation and leads to the derivation of an exact mathematical expression between the normalized average channel capacity available to each system-s user, under simultaneous optimal power and rate adaptation and the system-s parameters, as the number of hops per bit, the processing gain applied, the number of users per cell and the received signal-tonoise power ratio over the signal bandwidth. Finally, numerical results are presented to illustrate the proposed mathematical analysis.
Abstract: In this paper, we present an experimental testing for
a new algorithm that determines an optimal controller-s coefficients
for output variance reduction related to Linear Time Invariant (LTI)
Systems. The algorithm features simplicity in calculation, generalization
to minimal and non-minimal phase systems, and could be
configured to achieve reference tracking as well as variance reduction
after compromising with the output variance. An experiment of DCmotor
velocity control demonstrates the application of this new
algorithm in designing the controller. The results show that the
controller achieves minimum variance and reference tracking for a
preset velocity reference relying on an identified model of the motor.
Abstract: In rapidly changing market environment, firms are investing a lot of time and resources into new product development (NPD) projects to make profit and to obtain competitive advantage. However, failure rate of NPD projects is becoming high due to various internal and external risks which hinder successful NPD projects. To reduce the failure rate, it is critical that risks have to be managed effectively and efficiently through good strategy, and treated by optimal responses to minimize risk cost. Four strategies are adopted to handle the risks in this study. The optimal responses are characterized by high reduction of risk costs with high efficiency. This study suggests a framework to decide the optimal responses considering the core risks, risk costs, response efficiency and response costs for successful NPD projects. Both binary particles warm optimization (BPSO) and multi-objective particle swarm optimization (MOPSO) methods are mainly used in the framework. Although several limitations exist in use for real industries, the frame work shows good strength for handling the risks with highly scientific ways through an example.
Abstract: This paper introduces a framework based on the collaboration of multi agent and hyper-heuristics to find a solution of the real single machine production problem. There are many techniques used to solve this problem. Each of it has its own advantages and disadvantages. By the collaboration of multi agent system and hyper-heuristics, we can get more optimal solution. The hyper-heuristics approach operates on a search space of heuristics rather than directly on a search space of solutions. The proposed framework consists of some agents, i.e. problem agent, trainer agent, algorithm agent (GPHH, GAHH, and SAHH), optimizer agent, and solver agent. Some low level heuristics used in this paper are MRT, SPT, LPT, EDD, LDD, and MON
Abstract: The objective of this research is to calculate the
optimal inventory lot-sizing for each supplier and minimize the total
inventory cost which includes joint purchase cost of the products,
transaction cost for the suppliers, and holding cost for remaining
inventory. Genetic algorithms (GAs) are applied to the multi-product
and multi-period inventory lot-sizing problems with supplier
selection under storage space. Also a maximum storage space for the
decision maker in each period is considered. The decision maker
needs to determine what products to order in what quantities with
which suppliers in which periods. It is assumed that demand of
multiple products is known over a planning horizon. The problem is
formulated as a mixed integer programming and is solved with the
GAs. The detailed computation results are presented.
Abstract: Roundabout work on the principle of circulation and
entry flows, where the maximum entry flow rates depend largely on
circulating flow bearing in mind that entry flows must give away to
circulating flows. Where an existing roundabout has a road hump
installed at the entry arm, it can be hypothesized that the kinematics
of vehicles may prevent the entry arm from achieving optimum
performance. Road humps are traffic calming devices placed across
road width solely as speed reduction mechanism. They are the
preferred traffic calming option in Malaysia and often used on single
and dual carriageway local routes. The speed limit on local routes is
30mph (50 km/hr). Road humps in their various forms achieved the
biggest mean speed reduction (based on a mean speed before traffic
calming of 30mph) of up to 10mph or 16 km/hr according to the UK
Department of Transport. The underlying aim of reduced speed
should be to achieve a 'safe' distribution of speeds which reflects the
function of the road and the impacts on the local community.
Constraining safe distribution of speeds may lead to poor drivers
timing and delayed reflex reaction that can probably cause accident.
Previous studies on road hump impact have focused mainly on speed
reduction, traffic volume, noise and vibrations, discomfort and delay
from the use of road humps. The paper is aimed at optimal entry and
circulating flow induced by road humps. Results show that
roundabout entry and circulating flow perform better in
circumstances where there is no road hump at entrance.
Abstract: The main focus of the work was concerned with hydrodynamic and thermal analysis of the plate heat exchanger channel with corrugation patterns suggested to be triangular, sinusoidal, and square corrugation. This study was to numerically model and validate the triangular corrugated channel with dimensions/parameters taken from open literature, and then model/analyze both sinusoidal, and square corrugated channel referred to the triangular model. Initially, 2D modeling with local extensive analysis for triangular corrugated channel was carried out. By that, all local pressure drop, wall shear stress, friction factor, static temperature, heat flux, Nusselt number, and surface heat coefficient, were analyzed to interpret the hydrodynamic and thermal phenomena occurred in the flow. Furthermore, in order to facilitate confidence in this model, a comparison between the values predicted, and experimental results taken from literature for almost the same case, was done. Moreover, a holistic numerical study for sinusoidal and square channels together with global comparisons with triangular corrugation under the same condition, were handled. Later, a comparison between electric, and fluid cooling through varying the boundary condition was achieved. The constant wall temperature and constant wall heat flux boundary conditions were employed, and the different resulted Nusselt numbers as a consequence were justified. The results obtained can be used to come up with an optimal design, a 'compromise' between heat transfer and pressure drop.