Abstract: This paper presents a novel method for data hiding based on neighborhood pixels information to calculate the number of bits that can be used for substitution and modified Least Significant Bits technique for data embedding. The modified solution is independent of the nature of the data to be hidden and gives correct results along with un-noticeable image degradation. The technique, to find the number of bits that can be used for data hiding, uses the green component of the image as it is less sensitive to human eye and thus it is totally impossible for human eye to predict whether the image is encrypted or not. The application further encrypts the data using a custom designed algorithm before embedding bits into image for further security. The overall process consists of three main modules namely embedding, encryption and extraction cm.
Abstract: The client server systems using mobile
communications networks for data transmission became very
attractive for many economic agents, in the purpose of promoting and
offering electronic services to their clients. E-services are suitable for
business developing and financial benefits increasing. The products
or services can be efficiently delivered to a large number of clients,
using mobile Internet access technologies. The clients can have
access to e-services, anywhere and anytime, with the support of 3G,
GPRS, WLAN, etc., channels bandwidth, data services and protocols.
Based on the mobile communications networks evolution and
development, a convergence of technological and financial interests
of mobile operators, software developers, mobile terminals producers
and e-content providers is established. These will lead to a high level
of integration of IT&C resources and will facilitate the value added
services delivery through the mobile communications networks. In
this paper it is presented a client server system, for e-services access,
with Smartphones and PDA-s mobile software applications, installed
on Symbian and Windows Mobile operating systems.
Abstract: According to the density of the chips, designers are
trying to put so any facilities of computational and storage on single
chips. Along with the complexity of computational and storage
circuits, the designing, testing and debugging become more and more
complex and expensive. So, hardware design will be built by using
very high speed hardware description language, which is more
efficient and cost effective. This paper will focus on the
implementation of 32-bit ALU design based on Verilog hardware
description language. Adder and subtracter operate correctly on both
unsigned and positive numbers. In ALU, addition takes most of the
time if it uses the ripple-carry adder. The general strategy for
designing fast adders is to reduce the time required to form carry
signals. Adders that use this principle are called carry look- ahead
adder. The carry look-ahead adder is to be designed with combination
of 4-bit adders. The syntax of Verilog HDL is similar to the C
programming language. This paper proposes a unified approach to
ALU design in which both simulation and formal verification can
co-exist.
Abstract: This research investigates risk factors for defective products in autoparts factories. Under a Bayesian framework, a generalized linear mixed model (GLMM) in which the dependent variable, the number of defective products, has a Poisson distribution is adopted. Its performance is compared with the Poisson GLM under a Bayesian framework. The factors considered are production process, machines, and workers. The products coded RT50 are observed. The study found that the Poisson GLMM is more appropriate than the Poisson GLM. For the production Process factor, the highest risk of producing defective products is Process 1, for the Machine factor, the highest risk is Machine 5, and for the Worker factor, the highest risk is Worker 6.
Abstract: Global competitiveness has recently become the
biggest concern of both manufacturing and service companies.
Electronic commerce, as a key technology enables the firms to reach
all the potential consumers from all over the world. In this study, we
have presented commonly used electronic payment systems, and then
we have shown the evaluation of these systems in respect to different
criteria. The payment systems which are included in this research are
the credit card, the virtual credit card, the electronic money, the
mobile payment, the credit transfer and the debit instruments. We
have realized a systematic comparison of these systems in respect to
three main criteria: Technical, economical and social. We have
conducted a fuzzy multi-criteria decision making procedure to deal
with the multi-attribute nature of the problem. The subjectiveness
and imprecision of the evaluation process are modeled using
triangular fuzzy numbers.
Abstract: This paper presents an improved variable ordering method to obtain the minimum number of nodes in Reduced Ordered Binary Decision Diagrams (ROBDD). The proposed method uses the graph topology to find the best variable ordering. Therefore the input Boolean function is converted to a unidirectional graph. Three levels of graph parameters are used to increase the probability of having a good variable ordering. The initial level uses the total number of nodes (NN) in all the paths, the total number of paths (NP) and the maximum number of nodes among all paths (MNNAP). The second and third levels use two extra parameters: The shortest path among two variables (SP) and the sum of shortest path from one variable to all the other variables (SSP). A permutation of the graph parameters is performed at each level for each variable order and the number of nodes is recorded. Experimental results are promising; the proposed method is found to be more effective in finding the variable ordering for the majority of benchmark circuits.
Abstract: The Automatic Speech Recognition (ASR) applied to
Arabic language is a challenging task. This is mainly related to the
language specificities which make the researchers facing multiple
difficulties such as the insufficient linguistic resources and the very
limited number of available transcribed Arabic speech corpora. In
this paper, we are interested in the development of a HMM-based
ASR system for Standard Arabic (SA) language. Our fundamental
research goal is to select the most appropriate acoustic parameters
describing each audio frame, acoustic models and speech recognition
unit. To achieve this purpose, we analyze the effect of varying frame
windowing (size and period), acoustic parameter number resulting
from features extraction methods traditionally used in ASR, speech
recognition unit, Gaussian number per HMM state and number of
embedded re-estimations of the Baum-Welch Algorithm. To evaluate
the proposed ASR system, a multi-speaker SA connected-digits
corpus is collected, transcribed and used throughout all experiments.
A further evaluation is conducted on a speaker-independent continue
SA speech corpus. The phonemes recognition rate is 94.02% which is
relatively high when comparing it with another ASR system
evaluated on the same corpus.
Abstract: Block replacement algorithms to increase hit ratio
have been extensively used in cache memory management. Among
basic replacement schemes, LRU and FIFO have been shown to be
effective replacement algorithms in terms of hit rates. In this paper,
we introduce a flexible stack-based circuit which can be employed in
hardware implementation of both LRU and FIFO policies. We
propose a simple and efficient architecture such that stack-based
replacement algorithms can be implemented without the drawbacks
of the traditional architectures. The stack is modular and hence, a set
of stack rows can be cascaded depending on the number of blocks in
each cache set. Our circuit can be implemented in conjunction with
the cache controller and static/dynamic memories to form a cache
system. Experimental results exhibit that our proposed circuit
provides an average value of 26% improvement in storage bits and its
maximum operating frequency is increased by a factor of two
Abstract: Wireless sensor networks include small nodes which
have sensing ability; calculation and connection extend themselves
everywhere soon. Such networks have source limitation on
connection, calculation and energy consumption. So, since the nodes
have limited energy in sensor networks, the optimized energy
consumption in these networks is of more importance and has created
many challenges. The previous works have shown that by organizing
the network nodes in a number of clusters, the energy consumption
could be reduced considerably. So the lifetime of the network would
be increased. In this paper, we used the Queen-bee algorithm to
create energy efficient clusters in wireless sensor networks. The
Queen-bee (QB) is similar to nature in that the queen-bee plays a
major role in reproduction process. The QB is simulated with J-sim
simulator. The results of the simulation showed that the clustering by
the QB algorithm decreases the energy consumption with regard to
the other existing algorithms and increases the lifetime of the
network.
Abstract: Due to their high power-to-weight ratio and low cost,
pneumatic actuators are attractive for robotics and automation
applications; however, achieving fast and accurate control of their
position have been known as a complex control problem. A
methodology for obtaining high position accuracy with a linear
pneumatic actuator is presented. During experimentation with a
number of PID classical control approaches over many operations of
the pneumatic system, the need for frequent manual re-tuning of the
controller could not be eliminated. The reason for this problem is
thermal and energy losses inside the cylinder body due to the
complex friction forces developed by the piston displacements.
Although PD controllers performed very well over short periods, it
was necessary in our research project to introduce some form of
automatic gain-scheduling to achieve good long-term performance.
We chose a fuzzy logic system to do this, which proved to be an
easily designed and robust approach. Since the PD approach showed
very good behaviour in terms of position accuracy and settling time,
it was incorporated into a modified form of the 1st order Tagaki-
Sugeno fuzzy method to build an overall controller. This fuzzy gainscheduler
uses an input variable which automatically changes the PD
gain values of the controller according to the frequency of repeated
system operations. Performance of the new controller was
significantly improved and the need for manual re-tuning was
eliminated without a decrease in performance. The performance of
the controller operating with the above method is going to be tested
through a high-speed web network (GRID) for research purposes.
Abstract: Pentachlorophenol (PCP) is a polychlorinated
aromatic compound that is widespread in industrial effluents and is
considered to be a serious pollutant. Among the variety of industrial
effluents encountered, effluents from tanning industry are very
important and have a serious pollution potential. PCP is also formed
unintentionally in effluents of paper and pulp industries. It is highly
persistent in soils and is lethal to a wide variety of beneficial
microorganisms and insects, human beings and animals. The natural
processes that breakdown toxic chemicals in the environment have
become the focus of much attention to develop safe and environmentfriendly
deactivation technologies. Microbes and plants are among
the most important biological agents that remove and degrade waste
materials to enable their recycling in the environment. The present
investigation was carried out with the aim of developing a microbial
system for bioremediation of PCP polluted soils. A number of plant
species were evaluated for their ability to tolerate different
concentrations of pentachlorophenol (PCP) in the soil. The
experiment was conducted for 30 days under pot culture conditions.
The toxic effect of PCP on plants was studied by monitoring seed
germination, plant growth and biomass. As the concentration of PCP
was increased to 50 ppm, the inhibition of seed germination, plant
growth and biomass was also increased. Although PCP had a
negative effect on all plant species tested, maize and groundnut
showed the maximum tolerance to PCP. Other tolerating crops
included wheat, safflower, sunflower, and soybean. From the
rhizosphere soil of the tolerant seedlings, as many as twenty seven
PCP tolerant bacteria were isolated. From soybean, 8; sunflower, 3;
safflower 8; maize 2; groundnut and wheat, 3 each isolates were
made. They were screened for their PCP degradation potentials.
HPLC analyses of PCP degradation revealed that the isolate MAZ-2
degraded PCP completely. The isolate MAZ-1 was the next best
isolate with 90 per cent PCP degradation. These strains hold promise
to be used in the bioremediation of PCP polluted soils.
Abstract: The objective of global optimization is to find the
globally best solution of a model. Nonlinear models are ubiquitous
in many applications and their solution often requires a global
search approach; i.e. for a function f from a set A ⊂ Rn to
the real numbers, an element x0 ∈ A is sought-after, such that
∀ x ∈ A : f(x0) ≤ f(x). Depending on the field of application,
the question whether a found solution x0 is not only a local minimum
but a global one is very important.
This article presents a probabilistic approach to determine the
probability of a solution being a global minimum. The approach is
independent of the used global search method and only requires a
limited, convex parameter domain A as well as a Lipschitz continuous
function f whose Lipschitz constant is not needed to be known.
Abstract: The software system goes through a number of stages
during its life and a software process model gives a standard format
for planning, organizing and running a project. The article presents a
new software development process model named as “Divide and
Conquer Process Model", based on the idea first it divides the things
to make them simple and then gathered them to get the whole work
done. The article begins with the backgrounds of different software
process models and problems in these models. This is followed by a
new divide and conquer process model, explanation of its different
stages and at the end edge over other models is shown.
Abstract: The hydrodynamic and thermal lattice Boltzmann
methods are applied to investigate the turbulent convective heat
transfer in the wavy channel flows. In this study, the turbulent
phenomena are modeling by large-eddy simulations with the
Smagorinsky model. As a benchmark, the laminar and turbulent
backward-facing step flows are simulated first. The results give good
agreement with other numerical and experimental data. For wavy
channel flows, the distribution of Nusselt number and the skin-friction
coefficients are calculated to evaluate the heat transfer effect and the
drag force. It indicates that the vortices at the trough would affect the
magnitude of drag and weaken the heat convection effects on the wavy
surface. In turbulent cases, if the amplitude of the wavy boundary is
large enough, the secondary vortices would be generated at troughs
and contribute to the heat convection. Finally, the effects of different
Re on the turbulent transport phenomena are discussed.
Abstract: Due to the increasing penetration of wind energy, it is
necessary to possess design tools that are able to simulate the impact
of these installations in utility grids. In order to provide a net
contribution to this issue a detailed wind park model has been
developed and is briefly presented. However, the computational costs
associated with the performance of such a detailed model in
describing the behavior of a wind park composed by a considerable
number of units may render its practical application very difficult. To
overcome this problem integral manifolds theory has been applied to
reduce the order of the detailed wind park model, and therefore
create the conditions for the development of a dynamic equivalent
which is able to retain the relevant dynamics with respect to the
existing a.c. system. In this paper integral manifold method has been
introduced for order reduction. Simulation results of the proposed
method represents that integral manifold method results fit the
detailed model results with a higher precision than singular
perturbation method.
Abstract: This paper presents an approach which is based on the
use of supervised feed forward neural network, namely multilayer
perceptron (MLP) neural network and finite element method (FEM)
to solve the inverse problem of parameters identification. The
approach is used to identify unknown parameters of ferromagnetic
materials. The methodology used in this study consists in the
simulation of a large number of parameters in a material under test,
using the finite element method (FEM). Both variations in relative
magnetic permeability and electrical conductivity of the material
under test are considered. Then, the obtained results are used to
generate a set of vectors for the training of MLP neural network.
Finally, the obtained neural network is used to evaluate a group of
new materials, simulated by the FEM, but not belonging to the
original dataset. Noisy data, added to the probe measurements is used
to enhance the robustness of the method. The reached results
demonstrate the efficiency of the proposed approach, and encourage
future works on this subject.
Abstract: In this paper, we suggest new product-type estimators for the population mean of the variable of interest exploiting the first or the third quartile of the auxiliary variable. We obtain mean square error equations and the bias for the estimators. We study the properties of these estimators using simple random sampling (SRS) and ranked set sampling (RSS) methods. It is found that, SRS and RSS produce approximately unbiased estimators of the population mean. However, the RSS estimators are more efficient than those obtained using SRS based on the same number of measured units for all values of the correlation coefficient.
Abstract: Fast retrieval of data has been a need of user in any
database application. This paper introduces a buffer based query
optimization technique in which queries are assigned weights
according to their number of execution in a query bank. These
queries and their optimized executed plans are loaded into the buffer
at the start of the database application. For every query the system
searches for a match in the buffer and executes the plan without
creating new plans.
Abstract: Gas hydrates form when a number of factors co-exist:
free water, hydrocarbon gas, cold temperatures and high pressures are typical of the near mud-line conditions in a deepwater drilling
operation. Subsequently, when drilling with water based muds, particularly on exploration wells, the risk of hydrate formation
associated with a gas influx is high. The consequences of gas hydrate
formation while drilling are severe, and as such, every effort should be made to ensure the risk of hydrate formation is either eliminated
or significantly reduced. Thermodynamic inhibitors are used to reduce the free water content of a drilling mud, and thus suppress the
hydrate formation temperature. Very little experimental work has
been performed by oil and gas research companies on the evaluation
of gas hydrate formation in a water-based drilling mud. The main
objective of this paper is to investigate the experimental gas hydrate
formation for a mixture of methane, carbon dioxide & nitrogen in a
water-based drilling mud with or without presence of different
concentrations of thermodynamic inhibitors including pure salt and a
combination of salt with methanol or ethylene glycol at different
concentrations in a static loop apparatus. The experiments were
performed using a static loop apparatus consisting of a 2.4307 cm
inside diameter and 800 cm long pipe. All experiments were conducted at 2200 psia. The temperature in the loop was decreased at
a rate of 3.33 °F/h from initial temperature of 80 °F.
Abstract: An epidemiological cross sectional study was
undertaken in Yaoundé in 2002 and updated in 2005. Focused on
health within the city, the objectives were to measure diarrheal
prevalence and to identify the risk factors associated with them.
Results of microbiological examinations have revealed an urban
average prevalence rate of 14.5%. Access to basic services in the
living environment appears to be an important risk factor for
diarrheas. Statistical and spatial analyses conducted have revealed
that prevalence of diarrheal diseases vary among the two main types
of settlement (informal and planned). More importantly, this study
shows that, diarrhea prevalence rates (notably bacterial and parasitic
diarrheas) vary according to the sub- category of settlements. The
study draws a number of theoretical and policy implications for
researchers and policy decision makers.