Abstract: This paper aims to select the optimal location and
setting parameters of TCSC (Thyristor Controlled Series
Compensator) controller using Particle Swarm Optimization (PSO)
and Genetic Algorithm (GA) to mitigate small signal oscillations in a
multimachine power system. Though Power System Stabilizers
(PSSs) are prime choice in this issue, installation of FACTS device
has been suggested here in order to achieve appreciable damping of
system oscillations. However, performance of any FACTS devices
highly depends upon its parameters and suitable location in the
power network. In this paper PSO as well as GA based techniques are
used separately and compared their performances to investigate this
problem. The results of small signal stability analysis have been
represented employing eigenvalue as well as time domain response in
face of two common power system disturbances e.g., varying load
and transmission line outage. It has been revealed that the PSO based
TCSC controller is more effective than GA based controller even
during critical loading condition.
Abstract: Heavy metal pollution is an environmental concern.
Phytoremediation is a low-cost, environmental-friendly approach to
solve this problem. Mustard has the potential in reducing heavy metal
contents in soils. Among mustard (Brassica juncea (L.) Czern &
Coss) genotypes in Sri Lanka, accessions 7788, 8831 and 5088 give
significantly a high yield. Therefore, present study was conducted to
quantify the phytoextractive potential among these local mustard
accessions and to assess the interaction of heavy metals, Pb, Co, Mn
on phytoextraction. A pot experiment was designed with acid washed
sand (quartz) and a series of heavy metal solutions of 0, 25, 50, 75
and 100 μg/g. Experiment was carried out with factorial
experimental design. Mustard accessions were tolerant to heavy
metals and could be successfully used in removal of Pb, Co and Mn
and they are capable of accumulating significant quantities of heavy
metals in vegetative and reproductive organs. The order of the
accumulative potential of Pb, Co and Mn in mustard accessions is,
root > shoot >seed.
Abstract: In this paper, optimal generation expansion planning (GEP) is investigated considering purchase prices, profits of independent power producers (IPPs) and reliability criteria using a new method based on hybrid coded Genetic Algorithm (GA) and Particle Swarm Optimization (PSO). In this approach, optimal purchase price of each IPP is obtained by HCGA and reliability criteria are calculated by PSO technique. It should be noted that reliability criteria and the rate of carbon dioxide (CO2) emission have been considered as constraints of the GEP problem. Finally, the proposed method has been tested on the case study system. The results evaluation show that the proposed method can simply obtain optimal purchase prices of IPPs and is a fast method for calculation of reliability criteria in expansion planning. Also, considering the optimal purchase prices and profits of IPPs in generation expansion planning are caused that the expansion costs are decreased and the problem is solved more exactly.
Abstract: The problem of bin-packing in two dimensions (2BP) consists in placing a given set of rectangular items in a minimum number of rectangular and identical containers, called bins. This article treats the case of objects with a free orientation of 90Ôùª. We propose an approach of resolution combining optimization by colony of ants (ACO) and the heuristic method IMA to resolve this NP-Hard problem.
Abstract: Software testability is proposed to address the problem of increasing cost of test and the quality of software. Testability measure provides a quantified way to denote the testability of software. Since 1990s, many testability measure models are proposed to address the problem. By discussing the contradiction between domain testability and domain range ratio (DRR), a new testability measure, semantic fault distance, is proposed. Its validity is discussed.
Abstract: Asthma is a condition that causing chronic health problems in children. In addition to basic therapy against disease, we must try to reduce the impact of chronic health problems and also optimize their medical aspect of growth and development. A boy with mild asthma attack frequent episode did not showed any improvement with medical treatment and his asthma control test was 11. From radiologic examination he got hyperaerated lung and billateral sinusitis maxillaris; skin test results were house dust, food and pet allergy; an overweight body; bad school grades; psychological and environmental problem. We followed and evaluated this boy in 6 months, treated holistically. Even we could not do much on environmental but no more psychological and school problems, his on a good bodyweight and his asthma control test was 22. A case of a child with mild asthma attack frequent episode was reported. Asthma clinical course show no significant improvement when other predisposing factor is not well-controlled and a child’s growth and development may be affected. Improving condition of the patient can be created with the help of loving and caring way of nurturing from the parents and supportive peer group. Therefore, continuous and consistent monitoring is required because prognosis of asthma is generally good when regularly and properly controlled.
Abstract: Predictions of flow and heat transfer characteristics and shape optimization in internally finned circular tubes have been performed on three-dimensional periodically fully developed turbulent flow and thermal fields. For a trapezoidal fin profile, the effects of fin height h, upper fin widths d1, lower fin widths d2, and helix angle of fin ? on transport phenomena are investigated for the condition of fin number of N = 30. The CFD and mathematical optimization technique are coupled in order to optimize the shape of internally finned tube. The optimal solutions of the design variables (i.e., upper and lower fin widths, fin height and helix angle) are numerically obtained by minimizing the pressure loss and maximizing the heat transfer rate, simultaneously, for the limiting conditions of d1 = 0.5~1.5 mm, d2 = 0.5~1.5 mm, h= 0.5~1.5mm, ? = 10~30 degrees. The fully developed flow and thermal fields are predicted using the finite volume method and the optimization is carried out by means of the multi-objective genetic algorithm that is widely used in the constrained nonlinear optimization problem.
Abstract: In this study, the density dependent nonlinear reactiondiffusion
equation, which arises in the insect dispersal models, is
solved using the combined application of differential quadrature
method(DQM) and implicit Euler method. The polynomial based
DQM is used to discretize the spatial derivatives of the problem. The
resulting time-dependent nonlinear system of ordinary differential
equations(ODE-s) is solved by using implicit Euler method. The
computations are carried out for a Cauchy problem defined by a onedimensional
density dependent nonlinear reaction-diffusion equation
which has an exact solution. The DQM solution is found to be in a
very good agreement with the exact solution in terms of maximum
absolute error. The DQM solution exhibits superior accuracy at large
time levels tending to steady-state. Furthermore, using an implicit
method in the solution procedure leads to stable solutions and larger
time steps could be used.
Abstract: Given a simple connected unweighted undirected graph G = (V (G), E(G)) with |V (G)| = n and |E(G)| = m, we present a new algorithm for the all-pairs shortest-path (APSP) problem. The running time of our algorithm is in O(n2 log n). This bound is an improvement over previous best known O(n2.376) time bound of Raimund Seidel (1995) for general graphs. The algorithm presented does not rely on fast matrix multiplication. Our algorithm with slight modifications, enables us to compute the APSP problem for unweighted directed graph in time O(n2 log n), improving a previous best known O(n2.575) time bound of Uri Zwick (2002).
Abstract: In this paper two models using a functional network
were employed to solving classification problem. Functional networks
are generalized neural networks, which permit the specification of
their initial topology using knowledge about the problem at hand. In
this case, and after analyzing the available data and their relations, we
systematically discuss a numerical analysis method used for
functional network, and apply two functional network models to
solving XOR problem. The XOR problem that cannot be solved with
two-layered neural network can be solved by two-layered functional
network, which reveals a potent computational power of functional
networks, and the performance of the proposed model was validated
using classification problems.
Abstract: Over the past decades, automatic face recognition has become a highly active research area, mainly due to the countless application possibilities in both the private as well as the public sector. Numerous algorithms have been proposed in the literature to cope with the problem of face recognition, nevertheless, a group of methods commonly referred to as appearance based have emerged as the dominant solution to the face recognition problem. Many comparative studies concerned with the performance of appearance based methods have already been presented in the literature, not rarely with inconclusive and often with contradictory results. No consent has been reached within the scientific community regarding the relative ranking of the efficiency of appearance based methods for the face recognition task, let alone regarding their susceptibility to appearance changes induced by various environmental factors. To tackle these open issues, this paper assess the performance of the three dominant appearance based methods: principal component analysis, linear discriminant analysis and independent component analysis, and compares them on equal footing (i.e., with the same preprocessing procedure, with optimized parameters for the best possible performance, etc.) in face verification experiments on the publicly available XM2VTS database. In addition to the comparative analysis on the XM2VTS database, ten degraded versions of the database are also employed in the experiments to evaluate the susceptibility of the appearance based methods on various image degradations which can occur in "real-life" operating conditions. Our experimental results suggest that linear discriminant analysis ensures the most consistent verification rates across the tested databases.
Abstract: The overlay approach has been widely used by many service providers for Traffic Engineering (TE) in large Internet backbones. In the overlay approach, logical connections are set up between edge nodes to form a full mesh virtual network on top of the physical topology. IP routing is then run over the virtual network. Traffic engineering objectives are achieved through carefully routing logical connections over the physical links. Although the overlay approach has been implemented in many operational networks, it has a number of well-known scaling issues. This paper proposes a new approach to achieve traffic engineering without full-mesh overlaying with the help of integrated approach and equal subset split method. Traffic engineering needs to determine the optimal routing of traffic over the existing network infrastructure by efficiently allocating resource in order to optimize traffic performance on an IP network. Even though constraint-based routing [1] of Multi-Protocol Label Switching (MPLS) is developed to address this need, since it is not widely tested or debugged, Internet Service Providers (ISPs) resort to TE methods under Open Shortest Path First (OSPF), which is the most commonly used intra-domain routing protocol. Determining OSPF link weights for optimal network performance is an NP-hard problem. As it is not possible to solve this problem, we present a subset split method to improve the efficiency and performance by minimizing the maximum link utilization in the network via a small number of link weight modifications. The results of this method are compared against results of MPLS architecture [9] and other heuristic methods.
Abstract: Dynamics of a vapour bubble generated due to a high local energy input near a circular thin bronze plate in the absence of the buoyancy forces is numerically investigated in this paper. The bubble is generated near a thin bronze plate and during the growth and collapse of the bubble, it deforms the nearby plate. The Boundary Integral Equation Method is employed for numerical simulation of the problem. The fluid is assumed to be incompressible, irrotational and inviscid and the surface tension on the bubble boundary is neglected. Therefore the fluid flow around the vapour bubble can be assumed as a potential flow. Furthermore, the thin bronze plate is assumed to have perfectly plastic behaviour. Results show that the displacement of the circular thin bronze plate has considerable effect on the dynamics of its nearby vapour bubble. It is found that by decreasing the thickness of the thin bronze plate, the growth and collapse rate of the bubble becomes higher and consequently the lifetime of the bubble becomes shorter.
Abstract: Wireless channels are characterized by more serious
bursty and location-dependent errors. Many packet scheduling
algorithms have been proposed for wireless networks to guarantee
fairness and delay bounds. However, most existing schemes do not
consider the difference of traffic natures among packet flows. This
will cause the delay-weight coupling problem. In particular, serious
queuing delays may be incurred for real-time flows. In this paper, it
is proposed a scheduling algorithm that takes traffic types of flows
into consideration when scheduling packets and also it is provided
scheduling flexibility by trading off video quality to meet the
playback deadline.
Abstract: Sharing consistent and correct master data among
disparate applications in a reverse-logistics chain has long been
recognized as an intricate problem. Although a master data
management (MDM) system can surely assume that responsibility,
applications that need to co-operate with it must comply with
proprietary query interfaces provided by the specific MDM system. In
this paper, we present a RFID-ready MDM system which makes
master data readily available for any participating applications in a
reverse-logistics chain. We propose a RFID-wrapper as a part of our
MDM. It acts as a gateway between any data retrieval request and
query interfaces that process it. With the RFID-wrapper, any
participating applications in a reverse-logistics chain can easily
retrieve master data in a way that is analogous to retrieval of any other
RFID-based logistics transactional data.
Abstract: Residue Number System (RNS) is a modular representation and is proved to be an instrumental tool in many digital signal processing (DSP) applications which require high-speed computations. RNS is an integer and non weighted number system; it can support parallel, carry-free, high-speed and low power arithmetic. A very interesting correspondence exists between the concepts of Multiple Valued Logic (MVL) and Residue Number Arithmetic. If the number of levels used to represent MVL signals is chosen to be consistent with the moduli which create the finite rings in the RNS, MVL becomes a very natural representation for the RNS. There are two concerns related to the application of this Number System: reaching the most possible speed and the largest dynamic range. There is a conflict when one wants to resolve both these problem. That is augmenting the dynamic range results in reducing the speed in the same time. For achieving the most performance a method is considere named “One-Hot Residue Number System" in this implementation the propagation is only equal to one transistor delay. The problem with this method is the huge increase in the number of transistors they are increased in order m2 . In real application this is practically impossible. In this paper combining the Multiple Valued Logic and One-Hot Residue Number System we represent a new method to resolve both of these two problems. In this paper we represent a novel design of an OHRNS-based adder circuit. This circuit is useable for Multiple Valued Logic moduli, in comparison to other RNS design; this circuit has considerably improved the number of transistors and power consumption.
Abstract: Ensemble learning algorithms such as AdaBoost and
Bagging have been in active research and shown improvements in
classification results for several benchmarking data sets with mainly
decision trees as their base classifiers. In this paper we experiment to
apply these Meta learning techniques with classifiers such as random
forests, neural networks and support vector machines. The data sets
are from MAGIC, a Cherenkov telescope experiment. The task is to
classify gamma signals from overwhelmingly hadron and muon
signals representing a rare class classification problem. We compare
the individual classifiers with their ensemble counterparts and
discuss the results. WEKA a wonderful tool for machine learning has
been used for making the experiments.
Abstract: This paper presents an optimal design of linear phase
digital high pass finite impulse response (FIR) filter using Improved
Particle Swarm Optimization (IPSO). In the design process, the filter
length, pass band and stop band frequencies, feasible pass band and
stop band ripple sizes are specified. FIR filter design is a multi-modal
optimization problem. An iterative method is introduced to find the
optimal solution of FIR filter design problem. Evolutionary
algorithms like real code genetic algorithm (RGA), particle swarm
optimization (PSO), improved particle swarm optimization (IPSO)
have been used in this work for the design of linear phase high pass
FIR filter. IPSO is an improved PSO that proposes a new definition
for the velocity vector and swarm updating and hence the solution
quality is improved. A comparison of simulation results reveals the
optimization efficacy of the algorithm over the prevailing
optimization techniques for the solution of the multimodal, nondifferentiable,
highly non-linear, and constrained FIR filter design
problems.
Abstract: In this work, I present a review on Sparse Distributed
Memory for Small Cues (SDMSCue), a variant of Sparse Distributed
Memory (SDM) that is capable of handling small cues. I then conduct
and show some cognitive experiments on SDMSCue to test its
cognitive soundness compared to SDM. Small cues refer to input
cues that are presented to memory for reading associations; but have
many missing parts or fields from them. The original SDM failed to
handle such a problem. SDMSCue handles and overcomes this
pitfall. The main idea in SDMSCue; is the repeated projection of the
semantic space on smaller subspaces; that are selected based on the
input cue length and pattern. This process allows for Read/Write
operations using an input cue that is missing a large portion.
SDMSCue is augmented with the use of genetic algorithms for
memory allocation and initialization. I claim that SDM functionality
is a subset of SDMSCue functionality.
Abstract: This paper presents a modified version of the
maximum urgency first scheduling algorithm. The maximum
urgency algorithm combines the advantages of fixed and dynamic
scheduling to provide the dynamically changing systems with
flexible scheduling. This algorithm, however, has a major
shortcoming due to its scheduling mechanism which may cause a
critical task to fail. The modified maximum urgency first scheduling
algorithm resolves the mentioned problem. In this paper, we propose
two possible implementations for this algorithm by using either
earliest deadline first or modified least laxity first algorithms for
calculating the dynamic priorities. These two approaches are
compared together by simulating the two algorithms. The earliest
deadline first algorithm as the preferred implementation is then
recommended. Afterwards, we make a comparison between our
proposed algorithm and maximum urgency first algorithm using
simulation and results are presented. It is shown that modified
maximum urgency first is superior to maximum urgency first, since it
usually has less task preemption and hence, less related overhead. It
also leads to less failed non-critical tasks in overloaded situations.