Abstract: This paper proposes a method, combining color and layout features, for identifying documents captured from low-resolution handheld devices. On one hand, the document image color density surface is estimated and represented with an equivalent ellipse and on the other hand, the document shallow layout structure is computed and hierarchically represented. Our identification method first uses the color information in the documents in order to focus the search space on documents having a similar color distribution, and finally selects the document having the most similar layout structure in the remaining of the search space.
Abstract: In this paper we present semantic assistant agent
(SAA), an open source digital library agent which takes user query
for finding information in the digital library and takes resources-
metadata and stores it semantically. SAA uses Semantic Web to
improve browsing and searching for resources in digital library. All
metadata stored in the library are available in RDF format for
querying and processing by SemanSreach which is a part of SAA
architecture. The architecture includes a generic RDF-based model
that represents relationships among objects and their components.
Queries against these relationships are supported by an RDF triple
store.
Abstract: Appropriate description of business processes through
standard notations has become one of the most important assets for
organizations. Organizations must therefore deal with quality faults
in business process models such as the lack of understandability and
modifiability. These quality faults may be exacerbated if business
process models are mined by reverse engineering, e.g., from existing
information systems that support those business processes. Hence,
business process refactoring is often used, which change the internal
structure of business processes whilst its external behavior is
preserved. This paper aims to choose the most appropriate set of
refactoring operators through the quality assessment concerning
understandability and modifiability. These quality features are
assessed through well-proven measures proposed in the literature.
Additionally, a set of measure thresholds are heuristically established
for applying the most promising refactoring operators, i.e., those that
achieve the highest quality improvement according to the selected
measures in each case.
Abstract: Due to the stringent legislation for emission of diesel
engines and also increasing demand on fuel consumption, the
importance of detailed 3D simulation of fuel injection, mixing and
combustion have been increased in the recent years. In the present
work, FIRE code has been used to study the detailed modeling of
spray and mixture formation in a Caterpillar heavy-duty diesel
engine. The paper provides an overview of the submodels
implemented, which account for liquid spray atomization, droplet
secondary break-up, droplet collision, impingement, turbulent
dispersion and evaporation. The simulation was performed from
intake valve closing (IVC) to exhaust valve opening (EVO). The
predicted in-cylinder pressure is validated by comparing with
existing experimental data. A good agreement between the predicted
and experimental values ensures the accuracy of the numerical
predictions collected with the present work. Predictions of engine
emissions were also performed and a good quantitative agreement
between measured and predicted NOx and soot emission data were
obtained with the use of the present Zeldowich mechanism and
Hiroyasu model. In addition, the results reported in this paper
illustrate that the numerical simulation can be one of the most
powerful and beneficial tools for the internal combustion engine
design, optimization and performance analysis.
Abstract: Short Message Service (SMS) has grown in
popularity over the years and it has become a common way of
communication, it is a service provided through General System
for Mobile Communications (GSM) that allows users to send text
messages to others.
SMS is usually used to transport unclassified information, but
with the rise of mobile commerce it has become a popular tool for
transmitting sensitive information between the business and its
clients. By default SMS does not guarantee confidentiality and
integrity to the message content.
In the mobile communication systems, security (encryption)
offered by the network operator only applies on the wireless link.
Data delivered through the mobile core network may not be
protected. Existing end-to-end security mechanisms are provided
at application level and typically based on public key
cryptosystem.
The main concern in a public-key setting is the authenticity of
the public key; this issue can be resolved by identity-based (IDbased)
cryptography where the public key of a user can be derived
from public information that uniquely identifies the user.
This paper presents an encryption mechanism based on the IDbased
scheme using Elliptic curves to provide end-to-end security
for SMS. This mechanism has been implemented over the standard
SMS network architecture and the encryption overhead has been
estimated and compared with RSA scheme. This study indicates
that the ID-based mechanism has advantages over the RSA
mechanism in key distribution and scalability of increasing
security level for mobile service.
Abstract: Quantum computation using qubits made of two component Bose-Einstein condensates (BECs) is analyzed. We construct a general framework for quantum algorithms to be executed using the collective states of the BECs. The use of BECs allows for an increase of energy scales via bosonic enhancement, resulting in two qubit gate operations that can be performed at a time reduced by a factor of N, where N is the number of bosons per qubit. We illustrate the scheme by an application to Deutsch-s and Grover-s algorithms, and discuss possible experimental implementations. Decoherence effects are analyzed under both general conditions and for the experimental implementation proposed.
Abstract: Based on Rayleigh beam theory, the sub-impacts of a
free-free beam struck horizontally by a round-nosed rigid mass is
simulated by the finite difference method and the impact-separation
conditions. In order to obtain the sub-impact force, a uniaxial
compression elastic-plastic contact model is employed to analyze the
local deformation field on contact zone. It is found that the horizontal
impact is a complicated process including the elastic plastic
sub-impacts in sequence. There are two sub-zones of sub-impact. In
addition, it found that the elastic energy of the free-free beam is more
suitable for the Poisson collision hypothesis to explain compression
and recovery processes.
Abstract: This paper presents application artificial intelligent (AI) techniques, namely artificial neural network (ANN), adaptive neuro fuzzy interface system (ANFIS), to estimate the real power transfer between generators and loads. Since these AI techniques adopt supervised learning, it first uses modified nodal equation method (MNE) to determine real power contribution from each generator to loads. Then the results of MNE method and load flow information are utilized to estimate the power transfer using AI techniques. The 25-bus equivalent system of south Malaysia is utilized as a test system to illustrate the effectiveness of both AI methods compared to that of the MNE method. The mean squared error of the estimate of ANN and ANFIS power transfer allocation methods are 1.19E-05 and 2.97E-05, respectively. Furthermore, when compared to MNE method, ANN and ANFIS methods computes generator contribution to loads within 20.99 and 39.37msec respectively whereas the MNE method took 360msec for the calculation of same real power transfer allocation.
Abstract: Predicting short term wind speed is essential in order
to prevent systems in-action from the effects of strong winds. It also
helps in using wind energy as an alternative source of energy, mainly
for Electrical power generation. Wind speed prediction has
applications in Military and civilian fields for air traffic control,
rocket launch, ship navigation etc. The wind speed in near future
depends on the values of other meteorological variables, such as
atmospheric pressure, moisture content, humidity, rainfall etc. The
values of these parameters are obtained from a nearest weather
station and are used to train various forms of neural networks. The
trained model of neural networks is validated using a similar set of
data. The model is then used to predict the wind speed, using the
same meteorological information. This paper reports an Artificial
Neural Network model for short term wind speed prediction, which
uses back propagation algorithm.
Abstract: For numerical prediction of the NOX in the exhaust of
a compression ignition engine a model was developed by considering
the parameter equivalence ratio. This model was validated by
comparing the predicted results of NOX with experimental ones. The
ultimate aim of the work was to access the applicability, robustness
and performance of the improved NOX model against other NOX
models.
Abstract: The size, complexity and number of databases used
for protein information have caused bioinformatics to lag behind in
adapting to the need to handle this distributed information.
Integrating all the information from different databases into one
database is a challenging problem. Our main research is to develop a
tool which can be used to access and manipulate protein information
from difference databases. In our approach, we have integrated
difference databases such as Swiss-prot, PDB, Interpro, and EMBL
and transformed these databases in flat file format into relational
form using XML and Bioperl. As a result, we showed this tool can
search different sizes of protein information stored in relational
database and the result can be retrieved faster compared to flat file
database. A web based user interface is provided to allow user to
access or search for protein information in the local database.
Abstract: Wireless Sensor Networks (WSNs) are wireless
networks consisting of number of tiny, low cost and low power
sensor nodes to monitor various physical phenomena like
temperature, pressure, vibration, landslide detection, presence of any
object, etc. The major limitation in these networks is the use of nonrechargeable
battery having limited power supply. The main cause of
energy consumption WSN is communication subsystem. This paper
presents an efficient grid formation/clustering strategy known as Grid
based level Clustering and Aggregation of Data (GCAD). The
proposed clustering strategy is simple and scalable that uses low duty
cycle approach to keep non-CH nodes into sleep mode thus reducing
energy consumption. Simulation results demonstrate that our
proposed GCAD protocol performs better in various performance
metrics.
Abstract: Structural behavior of ring stiffened thick walled
cylinders made of functionally graded materials (FGMs) is
investigated in this paper. Functionally graded materials are inhomogeneous composites which are usually made from a mixture
of metal and ceramic. The gradient compositional variation of the
constituents from one surface to the other provides an elegant solution to the problem of high transverse shear stresses that are
induced when two dissimilar materials with large differences in material properties are bonded. FGM formation of the cylinder is
modeled by power-law exponent and the variation of characteristics is supposed to be in radial direction.
A finite element formulation is derived for the analysis. According to the property variation of the constituent materials in the radial
direction of the wall, it is not convenient to use conventional elements to model and analyze the structure of the stiffened FGM
cylinders. In this paper a new cylindrical super-element is used to model the finite element formulation and analyze the static and
modal behavior of stiffened FGM thick walled cylinders. By using
this super-element the number of elements, which are needed for
modeling, will reduce significantly and the process time is less in comparison with conventional finite element formulations. Results for static and modal analysis are evaluated and verified by
comparison to finite element formulation with conventional
elements. Comparison indicates a good conformity between results.
Abstract: This paper applies Bayesian Networks to support
information extraction from unstructured, ungrammatical, and
incoherent data sources for semantic annotation. A tool has been
developed that combines ontologies, machine learning, and
information extraction and probabilistic reasoning techniques to
support the extraction process. Data acquisition is performed with the
aid of knowledge specified in the form of ontology. Due to the
variable size of information available on different data sources, it is
often the case that the extracted data contains missing values for
certain variables of interest. It is desirable in such situations to
predict the missing values. The methodology, presented in this paper,
first learns a Bayesian network from the training data and then uses it
to predict missing data and to resolve conflicts. Experiments have
been conducted to analyze the performance of the presented
methodology. The results look promising as the methodology
achieves high degree of precision and recall for information
extraction and reasonably good accuracy for predicting missing
values.
Abstract: The join dependency provides the basis for obtaining
lossless join decomposition in a classical relational schema. The
existence of Join dependency shows that that the tables always
represent the correct data after being joined. Since the classical
relational databases cannot handle imprecise data, they were
extended to fuzzy relational databases so that uncertain, ambiguous,
imprecise and partially known information can also be stored in
databases in a formal way. However like classical databases, the
fuzzy relational databases also undergoes decomposition during
normalization, the issue of joining the decomposed fuzzy relations
remains intact. Our effort in the present paper is to emphasize on this
issue. In this paper we define fuzzy join dependency in the
framework of type-1 fuzzy relational databases & type-2 fuzzy
relational databases using the concept of fuzzy equality which is
defined using fuzzy functions. We use the fuzzy equi-join operator
for computing the fuzzy equality of two attribute values. We also
discuss the dependency preservation property on execution of this
fuzzy equi- join and derive the necessary condition for the fuzzy
functional dependencies to be preserved on joining the decomposed
fuzzy relations. We also derive the conditions for fuzzy join
dependency to exist in context of both type-1 and type-2 fuzzy
relational databases. We find that unlike the classical relational
databases even the existence of a trivial join dependency does not
ensure lossless join decomposition in type-2 fuzzy relational
databases. Finally we derive the conditions for the fuzzy equality to
be non zero and the qualification of an attribute for fuzzy key.
Abstract: Rice husk is one of the alternative fuels for Thailand because of its high potential and environmental benefits. Nonetheless, the environmental profile of the electricity production from rice husk must be assessed to ensure reduced environmental damage. A 10 MW pilot plant using rice husk as feedstock is the study site. The environmental impacts from rice husk power plant are evaluated by using the Life Cycle Assessment (LCA) methodology. Energy, material and carbon balances have been determined for tracing the system flow. Carbon closure has been used for describing of the net amount of CO2 released from the system in relation to the amount being recycled between the power plant and the CO2 adsorbed by rice husk. The transportation of rice husk to the power plant has significant on global warming, but not on acidification and photo-oxidant formation. The results showed that the impact potentials from rice husk power plant are lesser than the conventional plants for most of the categories considered; except the photo-oxidant formation potential from CO. The high CO from rice husk power plant may be due to low boiler efficiency and high moisture content in rice husk. The performance of the study site can be enhanced by improving the combustion efficiency.
Abstract: A systems approach model for prostate cancer in prostate duct, as a sub-system of the organism is developed. It is accomplished in two steps. First this research work starts with a nonlinear system of coupled Fokker-Plank equations which models continuous process of the system like motion of cells. Then extended to PDEs that include discontinuous processes like cell mutations, proliferation and deaths. The discontinuous processes is modeled by using intensity poisson processes. The model incorporates the features of the prostate duct. The system of PDEs spatial coordinate is along the proximal distal axis. Its parameters depend on features of the prostate duct. The movement of cells is biased towards distal region and mutations of prostate cancer cells is localized in the proximal region. Numerical solutions of the full system of equations are provided, and are exhibit traveling wave fronts phenomena. This motivates the use of the standard transformation to derive a canonically related system of ODEs for traveling wave solutions. The results obtained show persistence of prostate cancer by showing that the non-negative cone for the traveling wave system is time invariant. The traveling waves have a unique global attractor is proved also. Biologically, the global attractor verifies that evolution of prostate cancer stem cells exhibit the avascular tumor growth. These numerical solutions show that altering prostate stem cell movement or mutation of prostate cancer cells lead to avascular tumor. Conclusion with comments on clinical implications of the model is discussed.
Abstract: Following the loss of NASA's Space Shuttle
Columbia in 2003, it was determined that problems in the agency's
organization created an environment that led to the accident. One
component of the proposed solution resulted in the formation of the
NASA Engineering Network (NEN), a suite of information retrieval
and knowledge-sharing tools. This paper describes the
implementation of communities of practice, which are formed along
engineering disciplines. Communities of practice enable engineers to
leverage their knowledge and best practices to collaborate and take
information learning back to their jobs and embed it into the
procedures of the agency. This case study offers insight into using
traditional engineering disciplines for virtual collaboration, including
lessons learned during the creation and establishment of NASA-s
communities.
Abstract: Current image-based individual human recognition
methods, such as fingerprints, face, or iris biometric modalities
generally require a cooperative subject, views from certain aspects,
and physical contact or close proximity. These methods cannot
reliably recognize non-cooperating individuals at a distance in the
real world under changing environmental conditions. Gait, which
concerns recognizing individuals by the way they walk, is a relatively
new biometric without these disadvantages. The inherent gait
characteristic of an individual makes it irreplaceable and useful in
visual surveillance.
In this paper, an efficient gait recognition system for human
identification by extracting two features namely width vector of
the binary silhouette and the MPEG-7-based region-based shape
descriptors is proposed. In the proposed method, foreground objects
i.e., human and other moving objects are extracted by estimating
background information by a Gaussian Mixture Model (GMM) and
subsequently, median filtering operation is performed for removing
noises in the background subtracted image. A moving target classification
algorithm is used to separate human being (i.e., pedestrian)
from other foreground objects (viz., vehicles). Shape and boundary
information is used in the moving target classification algorithm.
Subsequently, width vector of the outer contour of binary silhouette
and the MPEG-7 Angular Radial Transform coefficients are taken as
the feature vector. Next, the Principal Component Analysis (PCA)
is applied to the selected feature vector to reduce its dimensionality.
These extracted feature vectors are used to train an Hidden Markov
Model (HMM) for identification of some individuals. The proposed
system is evaluated using some gait sequences and the experimental
results show the efficacy of the proposed algorithm.
Abstract: Cancers could normally be marked by a number of
differentially expressed genes which show enormous potential as
biomarkers for a certain disease. Recent years, cancer classification
based on the investigation of gene expression profiles derived by
high-throughput microarrays has widely been used. The selection of
discriminative genes is, therefore, an essential preprocess step in
carcinogenesis studies. In this paper, we have proposed a novel gene
selector using information-theoretic measures for biological
discovery. This multivariate filter is a four-stage framework through
the analyses of feature relevance, feature interdependence, feature
redundancy-dependence and subset rankings, and having been
examined on the colon cancer data set. Our experimental result show
that the proposed method outperformed other information theorem
based filters in all aspect of classification errors and classification
performance.