Abstract: Based on the classical algorithm LSQR for solving (unconstrained) LS problem, an iterative method is proposed for the least-squares like-minimum-norm symmetric solution of AXB+CYD=E. As the application of this algorithm, an iterative method for the least-squares like-minimum-norm biymmetric solution of AXB=E is also obtained. Numerical results are reported that show the efficiency of the proposed methods.
Abstract: The distinction among urban, periurban and rural areas represents a classical example of uncertainty in land classification. Satellite images, geostatistical analysis and all kinds of spatial data are very useful in urban sprawl studies, but it is important to define precise rules in combining great amounts of data to build complex knowledge about territory. Rough Set theory may be a useful method to employ in this field. It represents a different mathematical approach to uncertainty by capturing the indiscernibility. Two different phenomena can be indiscernible in some contexts and classified in the same way when combining available information about them. This approach has been applied in a case of study, comparing the results achieved with both Map Algebra technique and Spatial Rough Set. The study case area, Potenza Province, is particularly suitable for the application of this theory, because it includes 100 municipalities with different number of inhabitants and morphologic features.
Abstract: Resource-constrained project scheduling is an NPhard
optimisation problem. There are many different heuristic
strategies how to shift activities in time when resource requirements
exceed their available amounts. These strategies are frequently based
on priorities of activities. In this paper, we assume that a suitable
heuristic has been chosen to decide which activities should be
performed immediately and which should be postponed and
investigate the resource-constrained project scheduling problem
(RCPSP) from the implementation point of view. We propose an
efficient routine that, instead of shifting the activities, extends their
duration. It makes it possible to break down their duration into active
and sleeping subintervals. Then we can apply the classical Critical
Path Method that needs only polynomial running time. This
algorithm can simply be adapted for multiproject scheduling with
limited resources.
Abstract: In text categorization problem the most used method
for documents representation is based on words frequency vectors
called VSM (Vector Space Model). This representation is based only
on words from documents and in this case loses any “word context"
information found in the document. In this article we make a
comparison between the classical method of document representation
and a method called Suffix Tree Document Model (STDM) that is
based on representing documents in the Suffix Tree format. For the
STDM model we proposed a new approach for documents
representation and a new formula for computing the similarity
between two documents. Thus we propose to build the suffix tree
only for any two documents at a time. This approach is faster, it has
lower memory consumption and use entire document representation
without using methods for disposing nodes. Also for this method is
proposed a formula for computing the similarity between documents,
which improves substantially the clustering quality. This
representation method was validated using HAC - Hierarchical
Agglomerative Clustering. In this context we experiment also the
stemming influence in the document preprocessing step and highlight
the difference between similarity or dissimilarity measures to find
“closer" documents.
Abstract: We seek exact solutions of the coupled Klein-Gordon-Schrödinger equation with variable coefficients with the aid of Lie classical approach. By using the Lie classical method, we are able to derive symmetries that are used for reducing the coupled system of partial differential equations into ordinary differential equations. From reduced differential equations we have derived some new exact solutions of coupled Klein-Gordon-Schrödinger equations involving some special functions such as Airy wave functions, Bessel functions, Mathieu functions etc.
Abstract: Analytical investigation of the free vibration behavior
of circular functionally graded (FG) plates integrated with two
uniformly distributed actuator layers made of piezoelectric (PZT4)
material on the top and bottom surfaces of the circular FG plate
based on the classical plate theory (CPT) is presented in this paper.
The material properties of the functionally graded substrate plate are
assumed to be graded in the thickness direction according to the
power-law distribution in terms of the volume fractions of the
constituents and the distribution of electric potential field along the
thickness direction of piezoelectric layers is simulated by a quadratic
function. The differential equations of motion are solved analytically
for clamped edge boundary condition of the plate. The detailed
mathematical derivations are presented and Numerical investigations
are performed for FG plates with two surface-bonded piezoelectric
layers. Emphasis is placed on investigating the effect of varying the
gradient index of FG plate on the free vibration characteristics of the
structure. The results are verified by those obtained from threedimensional
finite element analyses.
Abstract: Neighborhood Rough Sets (NRS) has been proven to
be an efficient tool for heterogeneous attribute reduction. However,
most of researches are focused on dealing with complete and noiseless
data. Factually, most of the information systems are noisy, namely,
filled with incomplete data and inconsistent data. In this paper, we
introduce a generalized neighborhood rough sets model, called
VPTNRS, to deal with the problem of heterogeneous attribute
reduction in noisy system. We generalize classical NRS model with
tolerance neighborhood relation and the probabilistic theory.
Furthermore, we use the neighborhood dependency to evaluate the
significance of a subset of heterogeneous attributes and construct a
forward greedy algorithm for attribute reduction based on it.
Experimental results show that the model is efficient to deal with noisy
data.
Abstract: Rarefied gas flows are often occurred in micro electro
mechanical systems and classical CFD could not precisely anticipate
the flow and thermal behavior due to the high Knudsen number.
Therefore, the heat transfer and the fluid dynamics characteristics of
rarefied gas flows in both a two-dimensional simple microchannel
and geometry similar to single Knudsen compressor have been
investigated with a goal of increasing performance of a actual
Knudsen compressor by using a particle simulation method. Thermal
transpiration and thermal creep, which are rarefied gas dynamic
phenomena, that cause movement of the flow from less to higher
temperature is generated by using two different longitude temperature
gradients (Linear, Step) along the walls of the flow microchannel. In
this study the influence of amount of temperature gradient and
governing pressure in various Knudsen numbers and length-to-height
ratios have been examined.
Abstract: In this paper, we present a novel statistical approach to
corpus-based speech synthesis. Classically, phonetic information is
defined and considered as acoustic reference to be respected. In this
way, many studies were elaborated for acoustical unit classification.
This type of classification allows separating units according to their
symbolic characteristics. Indeed, target cost and concatenation cost
were classically defined for unit selection.
In Corpus-Based Speech Synthesis System, when using large text
corpora, cost functions were limited to a juxtaposition of symbolic
criteria and the acoustic information of units is not exploited in the
definition of the target cost.
In this manuscript, we token in our consideration the unit phonetic
information corresponding to acoustic information. This would be realized
by defining a probabilistic linguistic Bi-grams model basically
used for unit selection. The selected units would be extracted from
the English TIMIT corpora.
Abstract: Several models of vulnerability assessment have been proposed. The selection of one of these models depends on the objectives of the study. The classical methodologies for seismic vulnerability analysis, as a part of seismic risk analysis, have been formulated with statistical criteria based on a rapid observation. The information relating to the buildings performance is statistically elaborated. In this paper, we use the European Macroseismic Scale EMS-98 to define the relationship between damage and macroseismic intensity to assess the seismic vulnerability. Applying to Algiers area, the first step is to identify building typologies and to assign vulnerability classes. In the second step, damages are investigated according to EMS-98.
Abstract: Based on assumptions of neo-classical economics and
rational choice / public choice theory, this paper investigates the
regulation of industrial land use in Taiwan by homeowners
associations (HOAs) as opposed to traditional government
administration. The comparison, which applies the transaction cost
theory and a polynomial regression analysis, manifested that HOAs
are superior to conventional government administration in terms of
transaction costs and overall efficiency. A case study that compares
Taiwan-s commonhold industrial park, NangKang Software Park, to
traditional government counterparts using limited data on the costs
and returns was analyzed. This empirical study on the relative
efficiency of governmental and private institutions justified the
important theoretical proposition. Numerical results prove the
efficiency of the established model.
Abstract: this paper presents a novel neural network controller
with composite adaptation low to improve the trajectory tracking
problems of biped robots comparing with classical controller. The
biped model has 5_link and 6 degrees of freedom and actuated by
Plated Pneumatic Artificial Muscle, which have a very high power to
weight ratio and it has large stoke compared to similar actuators. The
proposed controller employ a stable neural network in to approximate
unknown nonlinear functions in the robot dynamics, thereby
overcoming some limitation of conventional controllers such as PD
or adaptive controllers and guarantee good performance. This NN
controller significantly improve the accuracy requirements by
retraining the basic PD/PID loop, but adding an inner adaptive loop
that allows the controller to learn unknown parameters such as
friction coefficient, therefore improving tracking accuracy.
Simulation results plus graphical simulation in virtual reality show
that NN controller tracking performance is considerably better than
PD controller tracking performance.
Abstract: This paper presents a new feature based dense stereo
matching algorithm to obtain the dense disparity map via dynamic
programming. After extraction of some proper features, we use some
matching constraints such as epipolar line, disparity limit, ordering
and limit of directional derivative of disparity as well. Also, a coarseto-
fine multiresolution strategy is used to decrease the search space
and therefore increase the accuracy and processing speed. The
proposed method links the detected feature points into the chains and
compares some of the feature points from different chains, to
increase the matching speed. We also employ color stereo matching
to increase the accuracy of the algorithm. Then after feature
matching, we use the dynamic programming to obtain the dense
disparity map. It differs from the classical DP methods in the stereo
vision, since it employs sparse disparity map obtained from the
feature based matching stage. The DP is also performed further on a
scan line, between any matched two feature points on that scan line.
Thus our algorithm is truly an optimization method. Our algorithm
offers a good trade off in terms of accuracy and computational
efficiency. Regarding the results of our experiments, the proposed
algorithm increases the accuracy from 20 to 70%, and reduces the
running time of the algorithm almost 70%.
Abstract: This paper presents a model for the evaluation of
energy performance and aerodynamic forces acting on a three-bladed
small vertical axis Darrieus wind turbine depending on blade chord
curvature with respect to rotor axis.
The adopted survey methodology is based on an analytical code
coupled to a solid modeling software, capable of generating the
desired blade geometry depending on the blade design geometric
parameters, which is linked to a finite volume CFD code for the
calculation of rotor performance.
After describing and validating the model with experimental data,
the results of numerical simulations are proposed on the bases of two
different blade profile architectures, which are respectively
characterized by a straight chord and by a curved one, having a chord
radius equal to rotor external circumference. A CFD campaign of
analysis is completed for three blade-candidate airfoil sections, that is
the recently-developed DU 06-W-200 cambered blade profile, a
classical symmetrical NACA 0021 and its derived cambered airfoil,
characterized by a curved chord, having a chord radius equal to rotor
external circumference.
The effects of blade chord curvature on angle of attack, blade
tangential and normal forces are first investigated and then the
overall rotor torque and power are analyzed as a function of blade
azimuthal position, achieving a numerical quantification of the
influence of blade camber on overall rotor performance.
Abstract: This work presents a neural network model for the
clustering analysis of data based on Self Organizing Maps (SOM).
The model evolves during the training stage towards a hierarchical
structure according to the input requirements. The hierarchical structure
symbolizes a specialization tool that provides refinements of the
classification process. The structure behaves like a single map with
different resolutions depending on the region to analyze. The benefits
and performance of the algorithm are discussed in application to the
Iris dataset, a classical example for pattern recognition.
Abstract: In this paper, many techniques for blind identification of moving average (MA) process are presented. These methods utilize third- and fourth-order cumulants of the noisy observations of the system output. The system is driven by an independent and identically distributed (i.i.d) non-Gaussian sequence that is not observed. Two nonlinear optimization algorithms, namely the Gradient Descent and the Gauss-Newton algorithms are exposed. An algorithm based on the joint-diagonalization of the fourth-order cumulant matrices (FOSI) is also considered, as well as an improved version of the classical C(q, 0, k) algorithm based on the choice of the Best 1-D Slice of fourth-order cumulants. To illustrate the effectiveness of our methods, various simulation examples are presented.
Abstract: This paper studies the dependability of componentbased
applications, especially embedded ones, from the diagnosis
point of view. The principle of the diagnosis technique is to
implement inter-component tests in order to detect and locate the
faulty components without redundancy. The proposed approach for
diagnosing faulty components consists of two main aspects. The first
one concerns the execution of the inter-component tests which
requires integrating test functionality within a component. This is the
subject of this paper. The second one is the diagnosis process itself
which consists of the analysis of inter-component test results to
determine the fault-state of the whole system. Advantage of this
diagnosis method when compared to classical redundancy faulttolerant
techniques are application autonomy, cost-effectiveness and
better usage of system resources. Such advantage is very important
for many systems and especially for embedded ones.
Abstract: The identification and elimination of bad
measurements is one of the basic functions of a robust state estimator
as bad data have the effect of corrupting the results of state
estimation according to the popular weighted least squares method.
However this is a difficult problem to handle especially when dealing
with multiple errors from the interactive conforming type. In this
paper, a self adaptive genetic based algorithm is proposed. The
algorithm utilizes the results of the classical linearized normal
residuals approach to tune the genetic operators thus instead of
making a randomized search throughout the whole search space it is
more likely to be a directed search thus the optimum solution is
obtained at very early stages(maximum of 5 generations). The
algorithm utilizes the accumulating databases of already computed
cases to reduce the computational burden to minimum. Tests are
conducted with reference to the standard IEEE test systems. Test
results are very promising.
Abstract: Most of the real queuing systems include special properties and constraints, which can not be analyzed directly by using the results of solved classical queuing models. Lack of Markov chains features, unexponential patterns and service constraints, are the mentioned conditions. This paper represents an applied general algorithm for analysis and optimizing the queuing systems. The algorithm stages are described through a real case study. It is consisted of an almost completed non-Markov system with limited number of customers and capacities as well as lots of common exception of real queuing networks. Simulation is used for optimizing this system. So introduced stages over the following article include primary modeling, determining queuing system kinds, index defining, statistical analysis and goodness of fit test, validation of model and optimizing methods of system with simulation.
Abstract: A road pricing game is a game where various stakeholders and/or regions with different (and usually conflicting) objectives compete for toll setting in a given transportation network to satisfy their individual objectives. We investigate some classical game theoretical solution concepts for the road pricing game. We establish results for the road pricing game so that stakeholders and/or regions playing such a game will beforehand know what is obtainable. This will save time and argument, and above all, get rid of the feelings of unfairness among the competing actors and road users. Among the classical solution concepts we investigate is Nash equilibrium. In particular, we show that no pure Nash equilibrium exists among the actors, and further illustrate that even “mixed Nash equilibrium" may not be achievable in the road pricing game. The paper also demonstrates the type of coalitions that are not only reachable, but also stable and profitable for the actors involved.