Abstract: A reduced-bit multiplication algorithm based on the ancient Vedic multiplication formulae is proposed in this paper. Both the Vedic multiplication formulae, Urdhva tiryakbhyam and Nikhilam, are first discussed in detail. Urdhva tiryakbhyam, being a general multiplication formula, is equally applicable to all cases of multiplication. It is applied to the digital arithmetic and is shown to yield a multiplier architecture which is very similar to the popular array multiplier. Due to its structure, it leads to a high carry propagation delay in case of multiplication of large numbers. Nikhilam Sutra, on the other hand, is more efficient in the multiplication of large numbers as it reduces the multiplication of two large numbers to that of two smaller numbers. The framework of the proposed algorithm is taken from this Sutra and is further optimized by use of some general arithmetic operations such as expansion and bit-shifting to take advantage of bit-reduction in multiplication. We illustrate the proposed algorithm by reducing a general 4x4-bit multiplication to a single 2 x 2-bit multiplication operation.
Abstract: Since the actuator capacity is limited, in the real
application of active control systems under sever earthquakes it is
conceivable that the actuators saturate, hence the actuator saturation
should be considered as a constraint in design of optimal controllers.
In this paper optimal design of active controllers for nonlinear
structures by considering actuator saturation, has been studied. The
proposed method for designing optimal controllers is based on
defining an optimization problem which the objective has been to
minimize the maximum displacement of structure when a limited
capacity for actuator has been used. To this end a single degree of
freedom (SDF) structure with a bilinear hysteretic behavior has been
simulated under a white noise ground acceleration of different
amplitudes. Active tendon control mechanism, comprised of prestressed
tendons and an actuator, and extended nonlinear Newmark
method based instantaneous optimal control algorithm have been
used. To achieve the best results, the weights corresponding to
displacement, velocity, acceleration and control force in the
performance index have been optimized by the Distributed Genetic
Algorithm (DGA). Results show the effectiveness of the proposed
method in considering actuator saturation. Also based on the
numerical simulations it can be concluded that the actuator capacity
and the average value of required control force are two important
factors in designing nonlinear controllers which consider the actuator
saturation.
Abstract: Most routing protocols (DSR, AODV etc.) that have
been designed for wireless adhoc networks incorporate the broadcasting
operation in their route discovery scheme. Probabilistic broadcasting
techniques have been developed to optimize the broadcast operation
which is otherwise very expensive in terms of the redundancy
and the traffic it generates. In this paper we have explored percolation
theory to gain a different perspective on probabilistic broadcasting
schemes which have been actively researched in the recent years.
This theory has helped us estimate the value of broadcast probability
in a wireless adhoc network as a function of the size of the network.
We also show that, operating at those optimal values of broadcast
probability there is at least 25-30% reduction in packet regeneration
during successful broadcasting.
Abstract: The response surface methodology (RSM) is a
collection of mathematical and statistical techniques useful in the
modeling and analysis of problems in which the dependent variable
receives the influence of several independent variables, in order to
determine which are the conditions under which should operate these
variables to optimize a production process. The RSM estimated a
regression model of first order, and sets the search direction using the
method of maximum / minimum slope up / down MMS U/D.
However, this method selects the step size intuitively, which can
affect the efficiency of the RSM. This paper assesses how the step
size affects the efficiency of this methodology. The numerical
examples are carried out through Monte Carlo experiments,
evaluating three response variables: efficiency gain function, the
optimum distance and the number of iterations. The results in the
simulation experiments showed that in response variables efficiency
and gain function at the optimum distance were not affected by the
step size, while the number of iterations is found that the efficiency if
it is affected by the size of the step and function type of test used.
Abstract: In the recent years multimedia traffic and in particular
VoIP services are growing dramatically. We present a new algorithm
to control the resource utilization and to optimize the voice codec
selection during SIP call setup on behalf of the traffic condition
estimated on the network path.
The most suitable methodologies and the tools that perform realtime
evaluation of the available bandwidth on a network path have
been integrated with our proposed algorithm: this selects the best
codec for a VoIP call in function of the instantaneous available
bandwidth on the path. The algorithm does not require any explicit
feedback from the network, and this makes it easily deployable over
the Internet. We have also performed intensive tests on real network
scenarios with a software prototype, verifying the algorithm
efficiency with different network topologies and traffic patterns
between two SIP PBXs.
The promising results obtained during the experimental validation
of the algorithm are now the basis for the extension towards a larger
set of multimedia services and the integration of our methodology
with existing PBX appliances.
Abstract: Most high-performance ac drives utilize a current
controller. The controller switches a voltage source inverter (VSI)
such that the motor current follows a set of reference current
waveforms. Fixed-band hysteresis (FBH) current control has been
widely used for the PWM inverter. We want to apply the same
controller for the PWM AC chopper. The aims of the controller is to
optimize the harmonic content at both input and output sides, while
maintaining acceptable losses in the ac chopper and to control in
wide range the fundamental output voltage. Fixed band controller has
been simulated and analyzed for a single-phase AC chopper and are
easily extended to three-phase systems. Simulation confirmed the
advantages and the excellent performance of the modulation method
applied for the AC chopper.
Abstract: The remediation of water resources pollution in
developing countries requires the application of alternative
sustainable cheaper and efficient end-of-pipe wastewater treatment
technologies. The feasibility of use of South African cheap and
abundant pine tree (Pinus patula) sawdust for development of lowcost
AC of comparable quality to expensive commercial ACs in the
abatement of water pollution was investigated. AC was developed at
optimized two-stage N2-superheated steam activation conditions in a
fixed bed reactor, and characterized for proximate and ultimate
properties, N2-BET surface area, pore size distribution, SEM, pHPZC
and FTIR. The sawdust pyrolysis activation energy was evaluated by
TGA. Results indicated that the chars prepared at 800oC and 2hrs
were suitable for development of better quality AC at 800oC and 47%
burn-off having BET surface area (1086m2/g), micropore volume
(0.26cm3/g), and mesopore volume (0.43cm3/g) comparable to
expensive commercial ACs, and suitable for water contaminants
removal. The developed AC showed basic surface functionality at
pHPZC at 10.3, and a phenol adsorption capacity that was higher than
that of commercial Norit (RO 0.8) AC. Thus, it is feasible to develop
better quality low-cost AC from (Pinus patula) sawdust using twostage
N2-steam activation in fixed-bed reactor.
Abstract: A neurofuzzy approach for a given set of input-output training data is proposed in two phases. Firstly, the data set is partitioned automatically into a set of clusters. Then a fuzzy if-then rule is extracted from each cluster to form a fuzzy rule base. Secondly, a fuzzy neural network is constructed accordingly and parameters are tuned to increase the precision of the fuzzy rule base. This network is able to learn and optimize the rule base of a Sugeno like Fuzzy inference system using Hybrid learning algorithm, which combines gradient descent, and least mean square algorithm. This proposed neurofuzzy system has the advantage of determining the number of rules automatically and also reduce the number of rules, decrease computational time, learns faster and consumes less memory. The authors also investigate that how neurofuzzy techniques can be applied in the area of control theory to design a fuzzy controller for linear and nonlinear dynamic systems modelling from a set of input/output data. The simulation analysis on a wide range of processes, to identify nonlinear components on-linely in a control system and a benchmark problem involving the prediction of a chaotic time series is carried out. Furthermore, the well-known examples of linear and nonlinear systems are also simulated under the Matlab/Simulink environment. The above combination is also illustrated in modeling the relationship between automobile trips and demographic factors.
Abstract: The tree structured approach of non-uniform filterbank
(NUFB) is normally used in perfect reconstruction (PR). The PR is
not always feasible due to certain limitations, i.e, constraints in
selecting design parameters, design complexity and some times
output is severely affected by aliasing error if necessary and
sufficient conditions of PR is not satisfied perfectly. Therefore, there
has been generalized interest of researchers to go for near perfect
reconstruction (NPR). In this proposed work, an optimized tree
structure technique is used for the design of NPR non-uniform
filterbank. Window functions of Blackman family are used to design
the prototype FIR filter. A single variable linear optimization is used
to minimize the amplitude distortion. The main feature of the
proposed design is its simplicity with linear phase property.
Abstract: Open urban public spaces comprise an important
element for the development of social, cultural and economic
activities of the population in the modern cities. These spaces are also
considered regulators of the region-s climate conditions, providing
better thermal, visual and auditory conditions which can be optimized
by the application of appropriate strategies of bioclimatic design. The
paper focuses on the analysis and evaluation of the recent unification
of the open spaces in the centre of Xanthi, a medium – size city in
northern Greece, from a bioclimatic perspective, as well as in the
creation of suitable methodology. It is based both on qualitative
observation of the interventions by fieldwork research and
assessment and on quantitative analysis and modeling of the research
area.
Abstract: This paper presents a tested research concept that
implements a complex evolutionary algorithm, genetic algorithm
(GA), in a multi-microcontroller environment. Parallel Distributed
Genetic Algorithm (PDGA) is employed in adaptive beam forming
technique to reduce power usage of adaptive antenna at WCDMA
base station. Adaptive antenna has dynamic beam that requires more
advanced beam forming algorithm such as genetic algorithm which
requires heavy computation and memory space. Microcontrollers are
low resource platforms that are normally not associated with GAs,
which are typically resource intensive. The aim of this project was to
design a cooperative multiprocessor system by expanding the role of
small scale PIC microcontrollers to optimize WCDMA base station
transmitter power. Implementation results have shown that PDGA
multi-microcontroller system returned optimal transmitted power
compared to conventional GA.
Abstract: This paper reports on the results of experimental investigations on the performance of a jet pump operated under selected primary flows to optimize the related parameters. For this purpose a two-phase flow jet pump was used employing various profiles of nozzles as the primary device which was designed, fabricated and used along with the combination of mixing tube and diffuser. The profiles employed were circular, conical, and elliptical. The diameter of the nozzle used was 4 mm. The area ratio of the jet pump was 0.16. The test facility created for this purpose was an open loop continuous circulation system. Performance of the jet pump was obtained as iso-efficiency curves on characteristic curves drawn for various water flow rates. To perform the suction capability, evacuation test was conducted at best efficiency point for all the profiles.
Abstract: In recent years, copulas have become very popular in
financial research and actuarial science as they are more flexible in
modelling the co-movements and relationships of risk factors as compared
to the conventional linear correlation coefficient by Pearson.
However, a precise estimation of the copula parameters is vital in
order to correctly capture the (possibly nonlinear) dependence structure
and joint tail events. In this study, we employ two optimization
heuristics, namely Differential Evolution and Threshold Accepting to
tackle the parameter estimation of multivariate t distribution models
in the EML approach. Since the evolutionary optimizer does not rely
on gradient search, the EML approach can be applied to estimation of
more complicated copula models such as high-dimensional copulas.
Our experimental study shows that the proposed method provides
more robust and more accurate estimates as compared to the IFM
approach.
Abstract: We proposed a technique to identify road traffic
congestion levels from velocity of mobile sensors with high accuracy
and consistent with motorists- judgments. The data collection utilized
a GPS device, a webcam, and an opinion survey. Human perceptions
were used to rate the traffic congestion levels into three levels: light,
heavy, and jam. Then the ratings and velocity were fed into a
decision tree learning model (J48). We successfully extracted vehicle
movement patterns to feed into the learning model using a sliding
windows technique. The parameters capturing the vehicle moving
patterns and the windows size were heuristically optimized. The
model achieved accuracy as high as 99.68%. By implementing the
model on the existing traffic report systems, the reports will cover
comprehensive areas. The proposed method can be applied to any
parts of the world.
Abstract: This paper examined the influence of matching
students- learning preferences with the teaching methodology
adopted, on their academic performance in an accounting course in
two types of learning environment in one university in Lebanon:
classes with PowerPoint (PPT) vs. conventional classes. Learning
preferences were either for PPT or for Conventional methodology. A
statistically significant increase in academic achievement is found in
the conventionally instructed group as compared to the group taught
with PPT. This low effectiveness of PPT might be attributed to the
learning preferences of Lebanese students. In the PPT group, better
academic performance was found among students with
learning/teaching match as compared with students with
learning/teaching mismatch. Since the majority of students display a
preference for the conventional methodology, the result might
suggest that Lebanese students- performance is not optimized by PPT
in the accounting classrooms, not because of PPT itself, but because
it is not matching the Lebanese students- learning preferences in such
a quantitative course.
Abstract: In this paper an algorithm is used to detect the color defects of ceramic tiles. First the image of a normal tile is clustered using GCMA; Genetic C-means Clustering Algorithm; those results in best cluster centers. C-means is a common clustering algorithm which optimizes an objective function, based on a measure between data points and the cluster centers in the data space. Here the objective function describes the mean square error. After finding the best centers, each pixel of the image is assigned to the cluster with closest cluster center. Then, the maximum errors of clusters are computed. For each cluster, max error is the maximum distance between its center and all the pixels which belong to it. After computing errors all the pixels of defected tile image are clustered based on the centers obtained from normal tile image in previous stage. Pixels which their distance from their cluster center is more than the maximum error of that cluster are considered as defected pixels.
Abstract: In this paper, to optimize the “Characteristic Straight Line Method" which is used in the soil displacement analysis, a “best estimate" of the geodetic leveling observations has been achieved by taking in account the concept of 'Height systems'. This concept has been discussed in detail and consequently the concept of “height". In landslides dynamic analysis, the soil is considered as a mosaic of rigid blocks. The soil displacement has been monitored and analyzed by using the “Characteristic Straight Line Method". Its characteristic components have been defined constructed from a “best estimate" of the topometric observations. In the measurement of elevation differences, we have used the most modern leveling equipment available. Observational procedures have also been designed to provide the most effective method to acquire data. In addition systematic errors which cannot be sufficiently controlled by instrumentation or observational techniques are minimized by applying appropriate corrections to the observed data: the level collimation correction minimizes the error caused by nonhorizontality of the leveling instrument's line of sight for unequal sight lengths, the refraction correction is modeled to minimize the refraction error caused by temperature (density) variation of air strata, the rod temperature correction accounts for variation in the length of the leveling rod' s Invar/LO-VAR® strip which results from temperature changes, the rod scale correction ensures a uniform scale which conforms to the international length standard and the introduction of the concept of the 'Height systems' where all types of height (orthometric, dynamic, normal, gravity correction, and equipotential surface) have been investigated. The “Characteristic Straight Line Method" is slightly more convenient than the “Characteristic Circle Method". It permits to evaluate a displacement of very small magnitude even when the displacement is of an infinitesimal quantity. The inclination of the landslide is given by the inverse of the distance reference point O to the “Characteristic Straight Line". Its direction is given by the bearing of the normal directed from point O to the Characteristic Straight Line (Fig..6). A “best estimate" of the topometric observations was used to measure the elevation of points carefully selected, before and after the deformation. Gross errors have been eliminated by statistical analyses and by comparing the heights within local neighborhoods. The results of a test using an area where very interesting land surface deformation occurs are reported. Monitoring with different options and qualitative comparison of results based on a sufficient number of check points are presented.
Abstract: The paper discusses a computationally efficient
method for the design of prototype filters required for the
implementation of an M-band cosine modulated filter bank. The
prototype filter is formulated as an optimum interpolated FIR filter.
The optimum interpolation factor requiring minimum number of
multipliers is used. The model filter as well as the image suppressor
will be designed using the Kaiser window. The method will seek to
optimize a single parameter namely cutoff frequency to minimize the
distortion in the overlapping passband.
Abstract: In this paper, cloud resource broker using goalbased
request in medical application is proposed. To handle recent
huge production of digital images and data in medical informatics
application, the cloud resource broker could be used by medical
practitioner for proper process in discovering and selecting correct
information and application. This paper summarizes several
reviewed articles to relate medical informatics application with
current broker technology and presents a research work in applying
goal-based request in cloud resource broker to optimize the use of
resources in cloud environment. The objective of proposing a new
kind of resource broker is to enhance the current resource
scheduling, discovery, and selection procedures. We believed that
it could help to maximize resources allocation in medical
informatics application.
Abstract: This paper examines many mathematical methods for
molding the hourly price forward curve (HPFC); the model will be
constructed by numerous regression methods, like polynomial
regression, radial basic function neural networks & a furrier series.
Examination the models goodness of fit will be done by means of
statistical & graphical tools. The criteria for choosing the model will
depend on minimize the Root Mean Squared Error (RMSE), using the
correlation analysis approach for the regression analysis the optimal
model will be distinct, which are robust against model
misspecification. Learning & supervision technique employed to
determine the form of the optimal parameters corresponding to each
measure of overall loss. By using all the numerical methods that
mentioned previously; the explicit expressions for the optimal model
derived and the optimal designs will be implemented.