Abstract: This paper presents the design and implements the prototype of an intelligent data processing framework in ubiquitous sensor networks. Much focus is put on how to handle the sensor data stream as well as the interoperability between the low-level sensor data and application clients. Our framework first addresses systematic middleware which mitigates the interaction between the application layer and low-level sensors, for the sake of analyzing a great volume of sensor data by filtering and integrating to create value-added context information. Then, an agent-based architecture is proposed for real-time data distribution to efficiently forward a specific event to the appropriate application registered in the directory service via the open interface. The prototype implementation demonstrates that our framework can host a sophisticated application on the ubiquitous sensor network and it can autonomously evolve to new middleware, taking advantages of promising technologies such as software agents, XML, cloud computing, and the like.
Abstract: In this paper, a reliable cooperative multipath routing
algorithm is proposed for data forwarding in wireless sensor networks
(WSNs). In this algorithm, data packets are forwarded towards the
base station (BS) through a number of paths, using a set of relay
nodes. In addition, the Rayleigh fading model is used to calculate
the evaluation metric of links. Here, the quality of reliability is
guaranteed by selecting optimal relay set with which the probability
of correct packet reception at the BS will exceed a predefined
threshold. Therefore, the proposed scheme ensures reliable packet
transmission to the BS. Furthermore, in the proposed algorithm,
energy efficiency is achieved by energy balancing (i.e. minimizing
the energy consumption of the bottleneck node of the routing path)
at the same time. This work also demonstrates that the proposed
algorithm outperforms existing algorithms in extending longevity of
the network, with respect to the quality of reliability. Given this, the
obtained results make possible reliable path selection with minimum
energy consumption in real time.
Abstract: Biomimicry has many potential benefits as many
technologies found in nature are superior to their man-made
counterparts. As technological device components approach the micro
and nanoscale, surface properties such as surface adhesion and friction
may need to be taken into account. Lowering surface adhesion by
manipulating chemistry alone might no longer be sufficient for such
components and thus physical manipulation may be required.
Adhesion reduction is only one of the many surface functions
displayed by micro/nano-structured cuticles of insects. Here, we
present a mini review of our understanding of insect cuticle structures
and the relationship between the structure dimensions and the
corresponding functional mechanisms. It may be possible to introduce
additional properties to material surfaces (indeed multi-functional
properties) based on the design of natural surfaces.
Abstract: Cognitive models allow predicting some aspects of utility
and usability of human machine interfaces (HMI), and simulating
the interaction with these interfaces. The action of predicting is based
on a task analysis, which investigates what a user is required to do
in terms of actions and cognitive processes to achieve a task. Task
analysis facilitates the understanding of the system-s functionalities.
Cognitive models are part of the analytical approaches, that do not
associate the users during the development process of the interface.
This article presents a study about the evaluation of a human
machine interaction with a contextual assistant-s interface using ACTR
and GOMS cognitive models. The present work shows how these
techniques may be applied in the evaluation of HMI, design and
research by emphasizing firstly the task analysis and secondly the
time execution of the task. In order to validate and support our
results, an experimental study of user performance is conducted at
the DOMUS laboratory, during the interaction with the contextual
assistant-s interface. The results of our models show that the GOMS
and ACT-R models give good and excellent predictions respectively
of users performance at the task level, as well as the object level.
Therefore, the simulated results are very close to the results obtained
in the experimental study.
Abstract: Vector quantization is a powerful tool for speech
coding applications. This paper deals with LPC Coding of speech
signals which uses a new technique called Multi Switched Split
Vector Quantization (MSSVQ), which is a hybrid of Multi, switched,
split vector quantization techniques. The spectral distortion
performance, computational complexity, and memory requirements
of MSSVQ are compared to split vector quantization (SVQ), multi
stage vector quantization(MSVQ) and switched split vector
quantization (SSVQ) techniques. It has been proved from results that
MSSVQ has better spectral distortion performance, lower
computational complexity and lower memory requirements when
compared to all the above mentioned product code vector
quantization techniques. Computational complexity is measured in
floating point operations (flops), and memory requirements is
measured in (floats).
Abstract: This study investigates the electrical performance of a
planar solid oxide fuel cell unit with cross-flow configuration when the fuel utilization gets higher and the fuel inlet flow are non-uniform.
A software package in this study solves two-dimensional,
simultaneous, partial differential equations of mass, energy, and
electro-chemistry, without considering stack direction variation. The
results show that the fuel utilization increases with a decrease in the molar flow rate, and the average current density decreases when the
molar flow rate drops. In addition, non-uniform Pattern A will induce more severe happening of non-reaction area in the corner of the fuel
exit and the air inlet. This non-reaction area deteriorates the average
current density and then deteriorates the electrical performance to –7%.
Abstract: This paper aims to develop a NOx emission model of
an acid gas incinerator using Nelder-Mead least squares support
vector regression (LS-SVR). Malaysia DOE is actively imposing the
Clean Air Regulation to mandate the installation of analytical
instrumentation known as Continuous Emission Monitoring System
(CEMS) to report emission level online to DOE . As a hardware
based analyzer, CEMS is expensive, maintenance intensive and often
unreliable. Therefore, software predictive technique is often
preferred and considered as a feasible alternative to replace the
CEMS for regulatory compliance. The LS-SVR model is built based
on the emissions from an acid gas incinerator that operates in a LNG
Complex. Simulated Annealing (SA) is first used to determine the
initial hyperparameters which are then further optimized based on the
performance of the model using Nelder-Mead simplex algorithm.
The LS-SVR model is shown to outperform a benchmark model
based on backpropagation neural networks (BPNN) in both training
and testing data.
Abstract: A challenging problem in radar signal processing is to
achieve reliable target detection in the presence of interferences. In
this paper, we propose a novel algorithm for automatic censoring of
radar interfering targets in log-normal clutter. The proposed
algorithm, termed the forward automatic censored cell averaging
detector (F-ACCAD), consists of two steps: removing the corrupted
reference cells (censoring) and the actual detection. Both steps are
performed dynamically by using a suitable set of ranked cells to
estimate the unknown background level and set the adaptive
thresholds accordingly. The F-ACCAD algorithm does not require
any prior information about the clutter parameters nor does it require
the number of interfering targets. The effectiveness of the F-ACCAD
algorithm is assessed by computing, using Monte Carlo simulations,
the probability of censoring and the probability of detection in
different background environments.
Abstract: The purpose of this study is to analyze the islands
tourist travel information sources, as well as for the satisfaction of the
tourist destination services. This study used questionnaires to the
island of Taiwan to the Penghu Islands to engage in tourism activities
tourist adopt the designated convenience sampling method, a total of
889 valid questionnaires were collected. After statistical analysis, this
study found that: 1. tourists to the Penghu Islands travel information
source for “friends and family came to Penghu". 2. Tourists feel the
service of the outlying islands of Penghu, the highest feelings of
“friendly local residents". 3. There are different demographic variables
affect the tourist travel information source and service satisfaction.
Based on the findings of this study not only for Penghu's tourism
industry with the unit in charge of the proposed operating and
suggestions for future research to other researchers.
Abstract: Security issue and the importance of the function of
police to provide practical and psychological contexts in the
community has been the main topics among researchers , police and
security circles and this subject require to review and analysis
mechanisms within the police and its interaction with other parts of
the system for providing community safety. This paper examine
national and social security in the Internet.
Abstract: Querying a data source and routing data towards sink
becomes a serious challenge in static wireless sensor networks if sink
and/or data source are mobile. Many a times the event to be observed
either moves or spreads across wide area making maintenance of
continuous path between source and sink a challenge. Also, sink can
move while query is being issued or data is on its way towards sink.
In this paper, we extend our already proposed Grid Based Data
Dissemination (GBDD) scheme which is a virtual grid based
topology management scheme restricting impact of movement of
sink(s) and event(s) to some specific cells of a grid. This obviates the
need for frequent path modifications and hence maintains continuous
flow of data while minimizing the network energy consumptions.
Simulation experiments show significant improvements in network
energy savings and average packet delay for a packet to reach at sink.
Abstract: In today-s new technology era, cluster has become a
necessity for the modern computing and data applications since many
applications take more time (even days or months) for computation.
Although after parallelization, computation speeds up, still time
required for much application can be more. Thus, reliability of the
cluster becomes very important issue and implementation of fault
tolerant mechanism becomes essential. The difficulty in designing a
fault tolerant cluster system increases with the difficulties of various
failures. The most imperative obsession is that the algorithm, which
avoids a simple failure in a system, must tolerate the more severe
failures. In this paper, we implemented the theory of watchdog timer
in a parallel environment, to take care of failures. Implementation of
simple algorithm in our project helps us to take care of different
types of failures; consequently, we found that the reliability of this
cluster improves.
Abstract: We board the problem of creating a seismic alert
system, based upon artificial neural networks, trained by using the
well-known back-propagation and genetic algorithms, in order to emit
the alarm for the population located into a specific city, about an
eminent earthquake greater than 4.5 Richter degrees, and avoiding
disasters and human loses. In lieu of using the propagation wave, we
employed the magnitude of the earthquake, to establish a correlation
between the recorded magnitudes from a controlled area and the city,
where we want to emit the alarm. To measure the accuracy of the
posed method, we use a database provided by CIRES, which contains
the records of 2500 quakes incoming from the State of Guerrero
and Mexico City. Particularly, we performed the proposed method to
generate an issue warning in Mexico City, employing the magnitudes
recorded in the State of Guerrero.
Abstract: Fast depth estimation from binocular vision is often
desired for autonomous vehicles, but, most algorithms could not easily
be put into practice because of the much time cost. We present an
image-processing technique that can fast estimate depth image from
binocular vision images. By finding out the lines which present the
best matched area in the disparity space image, the depth can be
estimated. When detecting these lines, an edge-emphasizing filter is
used. The final depth estimation will be presented after the smooth
filter. Our method is a compromise between local methods and global
optimization.
Abstract: Every commercial bank optimises its asset portfolio
depending on the profitability of assets and chosen or imposed
constraints. This paper proposes and applies a stylized model for
optimising banks' asset and liability structure, reflecting profitability
of different asset categories and their risks as well as costs associated
with different liability categories and reserve requirements. The level
of detail for asset and liability categories is chosen to create a
suitably parsimonious model and to include the most important
categories in the model. It is shown that the most appropriate
optimisation criterion for the model is the maximisation of the ratio
of net interest income to assets. The maximisation of this ratio is
subject to several constraints. Some are accounting identities or
dictated by legislative requirements; others vary depending on the
market objectives for a particular bank. The model predicts variable
amount of assets allocated to loan provision.
Abstract: Image synthesis is an important area in image processing.
To synthesize images various systems are proposed in
the literature. In this paper, we propose a bio-inspired system to
synthesize image and to study the generating power of the system, we
define the class of languages generated by our system. We call image
as array in this paper. We use a primitive called iso-array to synthesize
image/array. The operation is double splicing on iso-arrays. The
double splicing operation is used in DNA computing and we use
this to synthesize image. A comparison of the family of languages
generated by the proposed self restricted double splicing systems on
iso-arrays with the existing family of local iso-picture languages is
made. Certain closure properties such as union, concatenation and
rotation are studied for the family of languages generated by the
proposed model.
Abstract: The demand of the energy management systems (EMS) set forth by modern power systems requires fast energy management systems. Contingency analysis is among the functions in EMS which is time consuming. In order to handle this limitation, this paper introduces agent based technology in the contingency analysis. The main function of agents is to speed up the performance. Negotiations process in decision making is explained and the issue set forth is the minimization of the operating costs. The IEEE 14 bus system and its line outage have been used in the research and simulation results are presented.
Abstract: Object-oriented programming is a wonderful way to
make programming of huge real life tasks much easier than by using
procedural languages. In order to teach those ideas to students, it
is important to find a good task that shows the advantages of OOprogramming
very naturally. This paper gives an example, the game
Battleship, which seems to work excellent for teaching the OO ideas
(using Java, [1], [2], [3], [4]).
A three-step task is presented for how to teach OO-programming
using just one example suitable to convey many of the OO ideas.
Observations are given at the end and conclusions about how the
whole teaching course worked out.
Abstract: Heart-s electric field can be measured anywhere on
the surface of the body (ECG). When individuals touch, one person-s
ECG signal can be registered in other person-s EEG and elsewhere
on his body. Now, the aim of this study was to test the hypothesis
that physical contact (hand-holding) of two persons changes their
heart rate variability. Subjects were sixteen healthy female (age: 20-
26) which divided into eight sets. In each sets, we had two friends
that they passed intimacy test of J.sternberg. ECG of two subjects
(each set) acquired for 5 minutes before hand-holding (as control
group) and 5 minutes during they held their hands (as experimental
group). Then heart rate variability signals were extracted from
subjects' ECG and analyzed in linear feature space (time and
frequency domain) and nonlinear feature space. Considering the
results, we conclude that physical contact (hand-holding of two
friends) increases parasympathetic activity, as indicate by increase
SD1, SD1/SD2, HF and MF power (p
Abstract: Cell phone forensics to acquire and analyze data in the
cellular phone is nowadays being used in a national investigation
organization and a private company. In order to collect cellular phone
flash memory data, we have two methods. Firstly, it is a logical
method which acquires files and directories from the file system of the
cell phone flash memory. Secondly, we can get all data from bit-by-bit
copy of entire physical memory using a low level access method. In
this paper, we describe a forensic tool to acquire cell phone flash
memory data using a logical level approach. By our tool, we can get
EFS file system and peek memory data with an arbitrary region from
Korea CDMA cell phone.