Abstract: In an open real-time system environment, the coexistence of different kinds of real-time and non real-time applications makes the system scheduling mechanism face new requirements and challenges. One two-level scheduling scheme of the open real-time systems is introduced, and points out that hard and soft real-time applications are scheduled non-distinctively as the same type real-time applications, the Quality of Service (QoS) cannot be guaranteed. It has two flaws: The first, it can not differentiate scheduling priorities of hard and soft real-time applications, that is to say, it neglects characteristic differences between hard real-time applications and soft ones, so it does not suit a more complex real-time environment. The second, the worst case execution time of soft real-time applications cannot be predicted exactly, so it is not worth while to cost much spending in order to assure all soft real-time applications not to miss their deadlines, and doing that may cause resource wasting. In order to solve this problem, a novel two-level real-time scheduling mechanism (including scheduling profile and scheduling algorithm) which adds the process of dealing with soft real-time applications is proposed. Finally, we verify real-time scheduling mechanism from two aspects of theory and experiment. The results indicate that our scheduling mechanism can achieve the following objectives. (1) It can reflect the difference of priority when scheduling hard and soft real-time applications. (2) It can ensure schedulability of hard real-time applications, that is, their rate of missing deadline is 0. (3) The overall rate of missing deadline of soft real-time applications can be less than 1. (4) The deadline of a non-real-time application is not set, whereas the scheduling algorithm that server 0 S uses can avoid the “starvation" of jobs and increase QOS. By doing that, our scheduling mechanism is more compatible with different types of applications and it will be applied more widely.
Abstract: The group mutual exclusion (GME) problem is an
interesting generalization of the mutual exclusion problem. Several
solutions of the GME problem have been proposed for message
passing distributed systems. However, none of these solutions is
suitable for real time distributed systems. In this paper, we propose a
token-based distributed algorithms for the GME problem in soft real
time distributed systems. The algorithm uses the concepts of priority
queue, dynamic request set and the process state. The algorithm uses
first come first serve approach in selecting the next session type
between the same priority levels and satisfies the concurrent
occupancy property. The algorithm allows all n processors to be
inside their CS provided they request for the same session. The
performance analysis and correctness proof of the algorithm has also
been included in the paper.
Abstract: The purpose of this study was to explore the correlation
between leisure participation and perceived wellness, with the students
of a nursing college in southern Taiwan as the subjects. One thousand
six hundred and ninety-six (1,696) surveys were sent, and 1,408
surveys were received for an 83.02% valid response rate. Using
canonical correlation analysis to analyze the data, the results showed
that the linear combination of the two sets of variable produces five
significant canonical variates. Out of the five canonical variates, only
the first has sufficient explanatory power. The canonical correlation
coefficient of first canonical variate is 0.660. This indicated that
leisure participation and perceived wellness are significantly
correlated.
Abstract: Block replacement algorithms to increase hit ratio
have been extensively used in cache memory management. Among
basic replacement schemes, LRU and FIFO have been shown to be
effective replacement algorithms in terms of hit rates. In this paper,
we introduce a flexible stack-based circuit which can be employed in
hardware implementation of both LRU and FIFO policies. We
propose a simple and efficient architecture such that stack-based
replacement algorithms can be implemented without the drawbacks
of the traditional architectures. The stack is modular and hence, a set
of stack rows can be cascaded depending on the number of blocks in
each cache set. Our circuit can be implemented in conjunction with
the cache controller and static/dynamic memories to form a cache
system. Experimental results exhibit that our proposed circuit
provides an average value of 26% improvement in storage bits and its
maximum operating frequency is increased by a factor of two
Abstract: In this researcha particle swarm optimization (PSO)
algorithm is proposedfor no-wait flowshopsequence dependent
setuptime scheduling problem with weighted earliness-tardiness
penalties as the criterion (|,
|Σ
"
).The
smallestposition value (SPV) rule is applied to convert the continuous
value of position vector of particles in PSO to job permutations.A
timing algorithm is generated to find the optimal schedule and
calculate the objective function value of a given sequence in PSO
algorithm. Twodifferent neighborhood structures are applied to
improve the solution quality of PSO algorithm.The first one is based
on variable neighborhood search (VNS) and the second one is a
simple one with invariable structure. In order to compare the
performance of two neighborhood structures, random test problems
are generated and solved by both neighborhood
approaches.Computational results show that the VNS algorithmhas
better performance than the other one especially for the large sized
problems.
Abstract: The purpose is to study the model and characteristic of
participation of the suitable community to lead to develop permanent
water marketing in Bang Noi Floating Market, Bangkonti District,
Samutsongkhram Province. A total of 342 survey questionnaire was
administered to potential respondents. The researchers interviewed
the leader of the community. Appreciation Influence Control (AIC)
was used to talk with 20 villagers on arena. The findings revealed
that overall, most people had the middle level of the participation in
developing the durable Bang Noi Floating Market, Bangkonti,
Samutsongkhram Province and in aspects of gaining benefits from
developing it with atmosphere and a beautiful view for tourism. For
example, the landscape is beautiful with public utilities. The
participation in preserving and developing Bang Noi Floating Market
remains in the former way of life. The basic factor of person affects
to the participation of people such as age, level of education, career,
and income per month. Most participants are the original hosts that
have houses and shops located in the marketing and neighbor. These
people involve with the benefits and have the power to make a water
marketing strategy, the major role to set the information database. It
also found that the leader and the villagers play the important role in
setting a five-physical database. Data include level of information
such as position of village, territory of village, road, river, and
premises. Information of culture consists of a two-level of
information, interesting point, and Itinerary. The information occurs
from presenting and practicing by the leader and villagers in the
community.All of phases are presented for listening and investigating
database together in both the leader and villagers in the process of
participation.
Abstract: Due to their high power-to-weight ratio and low cost,
pneumatic actuators are attractive for robotics and automation
applications; however, achieving fast and accurate control of their
position have been known as a complex control problem. A
methodology for obtaining high position accuracy with a linear
pneumatic actuator is presented. During experimentation with a
number of PID classical control approaches over many operations of
the pneumatic system, the need for frequent manual re-tuning of the
controller could not be eliminated. The reason for this problem is
thermal and energy losses inside the cylinder body due to the
complex friction forces developed by the piston displacements.
Although PD controllers performed very well over short periods, it
was necessary in our research project to introduce some form of
automatic gain-scheduling to achieve good long-term performance.
We chose a fuzzy logic system to do this, which proved to be an
easily designed and robust approach. Since the PD approach showed
very good behaviour in terms of position accuracy and settling time,
it was incorporated into a modified form of the 1st order Tagaki-
Sugeno fuzzy method to build an overall controller. This fuzzy gainscheduler
uses an input variable which automatically changes the PD
gain values of the controller according to the frequency of repeated
system operations. Performance of the new controller was
significantly improved and the need for manual re-tuning was
eliminated without a decrease in performance. The performance of
the controller operating with the above method is going to be tested
through a high-speed web network (GRID) for research purposes.
Abstract: The study of the generated defects on manufactured
parts shows the difficulty to maintain parts in their positions during
the machining process and to estimate them during the pre-process
plan. This work presents a contribution to the development of 3D
models for the optimization of the manufacturing tolerances. An
experimental study allows the measurement of the defects of part
positioning for the determination of ε and the choice of an optimal
setup of the part. An approach of 3D tolerance based on the small
displacements method permits the determination of the
manufacturing errors upstream. A developed tool, allows an
automatic generation of the tolerance intervals along the three axes.
Abstract: For gamma radiation detection, assemblies having
scintillation crystals and a photomultiplier tube, also there is a
preamplifier connected to the detector because the signals from
photomultiplier tube are of small amplitude. After pre-amplification
the signals are sent to the amplifier and then to the multichannel
analyser. The multichannel analyser sorts all incoming electrical
signals according to their amplitudes and sorts the detected photons
in channels covering small energy intervals. The energy range of
each channel depends on the gain settings of the multichannel
analyser and the high voltage across the photomultiplier tube. The
exit spectrum data of the two main isotopes studied ,putting data in
biomass program ,process it by Matlab program to get the solid
holdup image (solid spherical nuclear fuel)
Abstract: The objective of global optimization is to find the
globally best solution of a model. Nonlinear models are ubiquitous
in many applications and their solution often requires a global
search approach; i.e. for a function f from a set A ⊂ Rn to
the real numbers, an element x0 ∈ A is sought-after, such that
∀ x ∈ A : f(x0) ≤ f(x). Depending on the field of application,
the question whether a found solution x0 is not only a local minimum
but a global one is very important.
This article presents a probabilistic approach to determine the
probability of a solution being a global minimum. The approach is
independent of the used global search method and only requires a
limited, convex parameter domain A as well as a Lipschitz continuous
function f whose Lipschitz constant is not needed to be known.
Abstract: One of the most used assumptions in logic programming
and deductive databases is the so-called Closed World Assumption
(CWA), according to which the atoms that cannot be inferred
from the programs are considered to be false (i.e. a pessimistic
assumption). One of the most successful semantics of conventional
logic programs based on the CWA is the well-founded semantics.
However, the CWA is not applicable in all circumstances when
information is handled. That is, the well-founded semantics, if
conventionally defined, would behave inadequately in different cases.
The solution we adopt in this paper is to extend the well-founded
semantics in order for it to be based also on other assumptions. The
basis of (default) negative information in the well-founded semantics
is given by the so-called unfounded sets. We extend this concept
by considering optimistic, pessimistic, skeptical and paraconsistent
assumptions, used to complete missing information from a program.
Our semantics, called extended well-founded semantics, expresses
also imperfect information considered to be missing/incomplete,
uncertain and/or inconsistent, by using bilattices as multivalued
logics. We provide a method of computing the extended well-founded
semantics and show that Kripke-Kleene semantics is captured by
considering a skeptical assumption. We show also that the complexity
of the computation of our semantics is polynomial time.
Abstract: This paper presents an approach which is based on the
use of supervised feed forward neural network, namely multilayer
perceptron (MLP) neural network and finite element method (FEM)
to solve the inverse problem of parameters identification. The
approach is used to identify unknown parameters of ferromagnetic
materials. The methodology used in this study consists in the
simulation of a large number of parameters in a material under test,
using the finite element method (FEM). Both variations in relative
magnetic permeability and electrical conductivity of the material
under test are considered. Then, the obtained results are used to
generate a set of vectors for the training of MLP neural network.
Finally, the obtained neural network is used to evaluate a group of
new materials, simulated by the FEM, but not belonging to the
original dataset. Noisy data, added to the probe measurements is used
to enhance the robustness of the method. The reached results
demonstrate the efficiency of the proposed approach, and encourage
future works on this subject.
Abstract: In this paper, we suggest new product-type estimators for the population mean of the variable of interest exploiting the first or the third quartile of the auxiliary variable. We obtain mean square error equations and the bias for the estimators. We study the properties of these estimators using simple random sampling (SRS) and ranked set sampling (RSS) methods. It is found that, SRS and RSS produce approximately unbiased estimators of the population mean. However, the RSS estimators are more efficient than those obtained using SRS based on the same number of measured units for all values of the correlation coefficient.
Abstract: An epidemiological cross sectional study was
undertaken in Yaoundé in 2002 and updated in 2005. Focused on
health within the city, the objectives were to measure diarrheal
prevalence and to identify the risk factors associated with them.
Results of microbiological examinations have revealed an urban
average prevalence rate of 14.5%. Access to basic services in the
living environment appears to be an important risk factor for
diarrheas. Statistical and spatial analyses conducted have revealed
that prevalence of diarrheal diseases vary among the two main types
of settlement (informal and planned). More importantly, this study
shows that, diarrhea prevalence rates (notably bacterial and parasitic
diarrheas) vary according to the sub- category of settlements. The
study draws a number of theoretical and policy implications for
researchers and policy decision makers.
Abstract: The customary practice of identifying industrial sickness is a set traditional techniques which rely upon a range of manual monitoring and compilation of financial records. It makes the process tedious, time consuming and often are susceptible to manipulation. Therefore, certain readily available tools are required which can deal with such uncertain situations arising out of industrial sickness. It is more significant for a country like India where the fruits of development are rarely equally distributed. In this paper, we propose an approach based on Artificial Neural Network (ANN) to deal with industrial sickness with specific focus on a few such units taken from a less developed north-east (NE) Indian state like Assam. The proposed system provides decision regarding industrial sickness using eight different parameters which are directly related to the stages of sickness of such units. The mechanism primarily uses certain signals and symptoms of industrial health to decide upon the state of a unit. Specifically, we formulate an ANN based block with data obtained from a few selected units of Assam so that required decisions related to industrial health could be taken. The system thus formulated could become an important part of planning and development. It can also contribute towards computerization of decision support systems related to industrial health and help in better management.
Abstract: Because of importance of energy, optimization of
power generation systems is necessary. Gas turbine cycles are
suitable manner for fast power generation, but their efficiency is
partly low. In order to achieving higher efficiencies, some
propositions are preferred such as recovery of heat from exhaust
gases in a regenerator, utilization of intercooler in a multistage
compressor, steam injection to combustion chamber and etc.
However thermodynamic optimization of gas turbine cycle, even
with above components, is necessary. In this article multi-objective
genetic algorithms are employed for Pareto approach optimization of
Regenerative-Intercooling-Gas Turbine (RIGT) cycle. In the multiobjective
optimization a number of conflicting objective functions
are to be optimized simultaneously. The important objective
functions that have been considered for optimization are entropy
generation of RIGT cycle (Ns) derives using Exergy Analysis and
Gouy-Stodola theorem, thermal efficiency and the net output power
of RIGT Cycle. These objectives are usually conflicting with each
other. The design variables consist of thermodynamic parameters
such as compressor pressure ratio (Rp), excess air in combustion
(EA), turbine inlet temperature (TIT) and inlet air temperature (T0).
At the first stage single objective optimization has been investigated
and the method of Non-dominated Sorting Genetic Algorithm
(NSGA-II) has been used for multi-objective optimization.
Optimization procedures are performed for two and three objective
functions and the results are compared for RIGT Cycle. In order to
investigate the optimal thermodynamic behavior of two objectives,
different set, each including two objectives of output parameters, are
considered individually. For each set Pareto front are depicted. The
sets of selected decision variables based on this Pareto front, will
cause the best possible combination of corresponding objective
functions. There is no superiority for the points on the Pareto front
figure, but they are superior to any other point. In the case of three
objective optimization the results are given in tables.
Abstract: A two-dimensional moving mesh algorithm is developed to simulate the general motion of two rotating bodies with relative translational motion. The grid includes a background grid and two sets of grids around the moving bodies. With this grid arrangement rotational and translational motions of two bodies are handled separately, with no complications. Inter-grid boundaries are determined based on their distances from two bodies. In this method, the overset concept is applied to hybrid grid, and flow variables are interpolated using a simple stencil. To evaluate this moving mesh algorithm unsteady Euler flow is solved for different cases using dual-time method of Jameson. Numerical results show excellent agreement with experimental data and other numerical results. To demonstrate the capability of present algorithm for accurate solution of flow fields around moving bodies, some benchmark problems have been defined in this paper.
Abstract: Environmental awareness and the recent
environmental policies have forced many electric utilities to
restructure their operational practices to account for their emission
impacts. One way to accomplish this is by reformulating the
traditional economic dispatch problem such that emission effects are
included in the mathematical model. This paper presents a Particle
Swarm Optimization (PSO) algorithm to solve the Economic-
Emission Dispatch problem (EED) which gained recent attention due
to the deregulation of the power industry and strict environmental
regulations. The problem is formulated as a multi-objective one with
two competing functions, namely economic cost and emission
functions, subject to different constraints. The inequality constraints
considered are the generating unit capacity limits while the equality
constraint is generation-demand balance. A novel equality constraint
handling mechanism is proposed in this paper. PSO algorithm is
tested on a 30-bus standard test system. Results obtained show that
PSO algorithm has a great potential in handling multi-objective
optimization problems and is capable of capturing Pareto optimal
solution set under different loading conditions.
Abstract: With deep development of software reuse, componentrelated
technologies have been widely applied in the development of
large-scale complex applications. Component identification (CI) is
one of the primary research problems in software reuse, by analyzing
domain business models to get a set of business components with high
reuse value and good reuse performance to support effective reuse.
Based on the concept and classification of CI, its technical stack is
briefly discussed from four views, i.e., form of input business models,
identification goals, identification strategies, and identification
process. Then various CI methods presented in literatures are
classified into four types, i.e., domain analysis based methods,
cohesion-coupling based clustering methods, CRUD matrix based
methods, and other methods, with the comparisons between these
methods for their advantages and disadvantages. Additionally, some
insufficiencies of study on CI are discussed, and the causes are
explained subsequently. Finally, it is concluded with some
significantly promising tendency about research on this problem.
Abstract: Support vector machines (SVMs) are considered to be
the best machine learning algorithms for minimizing the predictive
probability of misclassification. However, their drawback is that for
large data sets the computation of the optimal decision boundary is a
time consuming function of the size of the training set. Hence several
methods have been proposed to speed up the SVM algorithm. Here
three methods used to speed up the computation of the SVM
classifiers are compared experimentally using a musical genre
classification problem. The simplest method pre-selects a random
sample of the data before the application of the SVM algorithm. Two
additional methods use proximity graphs to pre-select data that are
near the decision boundary. One uses k-Nearest Neighbor graphs and
the other Relative Neighborhood Graphs to accomplish the task.