Abstract: This paper demonstrates the bus location system for
the route bus through the experiment in the real environment. A
bus location system is a system that provides information such as
the bus delay and positions. This system uses actual services and
positions data of buses, and those information should match data
on the database. The system has two possible problems. One, the
system could cost high in preparing devices to get bus positions.
Two, it could be difficult to match services data of buses. To avoid
these problems, we have developed this system at low cost and short
time by using the smart phone with GPS and the bus route system.
This system realizes the path planning considering bus delay and
displaying position of buses on the map. The bus location system
was demonstrated on route buses with smart phones for two months.
Abstract: A data cutting and sorting method (DCSM) is proposed
to optimize the performance of data mining. DCSM reduces the
calculation time by getting rid of redundant data during the data
mining process. In addition, DCSM minimizes the computational units
by splitting the database and by sorting data with support counts. In the
process of searching for the relationship between metabolic syndrome
and lifestyles with the health examination database of an electronics
manufacturing company, DCSM demonstrates higher search
efficiency than the traditional Apriori algorithm in tests with different
support counts.
Abstract: This paper presented the potential of smart phone to
provide support on mapping the indoor asset. The advantage of using
the smart phone to generate the indoor map is that it has the ability to
capture, store and reproduces still or video images; indeed most of us
do have this powerful gadget. The captured images usually used by
maintenance team to save a record for future reference. Here, these
images are used to generate 3D models of an object precisely and
accurately for efficient and effective solution in data gathering. Thus,
it could be a resource for an informative database in asset
management.
Abstract: Spatial trends are one of the valuable patterns in geo
databases. They play an important role in data analysis and
knowledge discovery from spatial data. A spatial trend is a regular
change of one or more non spatial attributes when spatially moving
away from a start object. Spatial trend detection is a graph search
problem therefore heuristic methods can be good solution. Artificial
immune system (AIS) is a special method for searching and
optimizing. AIS is a novel evolutionary paradigm inspired by the
biological immune system. The models based on immune system
principles, such as the clonal selection theory, the immune network
model or the negative selection algorithm, have been finding
increasing applications in fields of science and engineering.
In this paper, we develop a novel immunological algorithm based
on clonal selection algorithm (CSA) for spatial trend detection. We
are created neighborhood graph and neighborhood path, then select
spatial trends that their affinity is high for antibody. In an
evolutionary process with artificial immune algorithm, affinity of
low trends is increased with mutation until stop condition is satisfied.
Abstract: The size, complexity and number of databases used
for protein information have caused bioinformatics to lag behind in
adapting to the need to handle this distributed information.
Integrating all the information from different databases into one
database is a challenging problem. Our main research is to develop a
tool which can be used to access and manipulate protein information
from difference databases. In our approach, we have integrated
difference databases such as Swiss-prot, PDB, Interpro, and EMBL
and transformed these databases in flat file format into relational
form using XML and Bioperl. As a result, we showed this tool can
search different sizes of protein information stored in relational
database and the result can be retrieved faster compared to flat file
database. A web based user interface is provided to allow user to
access or search for protein information in the local database.
Abstract: The join dependency provides the basis for obtaining
lossless join decomposition in a classical relational schema. The
existence of Join dependency shows that that the tables always
represent the correct data after being joined. Since the classical
relational databases cannot handle imprecise data, they were
extended to fuzzy relational databases so that uncertain, ambiguous,
imprecise and partially known information can also be stored in
databases in a formal way. However like classical databases, the
fuzzy relational databases also undergoes decomposition during
normalization, the issue of joining the decomposed fuzzy relations
remains intact. Our effort in the present paper is to emphasize on this
issue. In this paper we define fuzzy join dependency in the
framework of type-1 fuzzy relational databases & type-2 fuzzy
relational databases using the concept of fuzzy equality which is
defined using fuzzy functions. We use the fuzzy equi-join operator
for computing the fuzzy equality of two attribute values. We also
discuss the dependency preservation property on execution of this
fuzzy equi- join and derive the necessary condition for the fuzzy
functional dependencies to be preserved on joining the decomposed
fuzzy relations. We also derive the conditions for fuzzy join
dependency to exist in context of both type-1 and type-2 fuzzy
relational databases. We find that unlike the classical relational
databases even the existence of a trivial join dependency does not
ensure lossless join decomposition in type-2 fuzzy relational
databases. Finally we derive the conditions for the fuzzy equality to
be non zero and the qualification of an attribute for fuzzy key.
Abstract: This paper presents a technical speaker adaptation
method called WMLLR, which is based on maximum likelihood linear
regression (MLLR). In MLLR, a linear regression-based transform
which adapted the HMM mean vectors was calculated to maximize the
likelihood of adaptation data. In this paper, the prior knowledge of the
initial model is adequately incorporated into the adaptation. A series of
speaker adaptation experiments are carried out at a 30 famous city
names database to investigate the efficiency of the proposed method.
Experimental results show that the WMLLR method outperforms the
conventional MLLR method, especially when only few utterances
from a new speaker are available for adaptation.
Abstract: This article outlines conceptualization and
implementation of an intelligent system capable of extracting
knowledge from databases. Use of hybridized features of both the
Rough and Fuzzy Set theory render the developed system flexibility
in dealing with discreet as well as continuous datasets. A raw data set
provided to the system, is initially transformed in a computer legible
format followed by pruning of the data set. The refined data set is
then processed through various Rough Set operators which enable
discovery of parameter relationships and interdependencies. The
discovered knowledge is automatically transformed into a rule base
expressed in Fuzzy terms. Two exemplary cancer repository datasets
(for Breast and Lung Cancer) have been used to test and implement
the proposed framework.
Abstract: The aim of this paper is to present a new method
which can be used for progressive transmission of electrocardiogram
(ECG). The idea consists in transforming any ECG signal to an
image, containing one beat in each row. In the first step, the beats are
synchronized in order to reduce the high frequencies due to inter-beat
transitions. The obtained image is then transformed using a discrete
version of Radon Transform (DRT). Hence, transmitting the ECG,
leads to transmit the most significant energy of the transformed
image in Radon domain. For decoding purpose, the receptor needs to
use the inverse Radon Transform as well as the two synchronization
frames.
The presented protocol can be adapted for lossy to lossless
compression systems. In lossy mode we show that the compression
ratio can be multiplied by an average factor of 2 for an acceptable
quality of reconstructed signal. These results have been obtained on
real signals from MIT database.
Abstract: This paper proposes a specialized Web robot to automatically collect objectionable Web contents for use in an objectionable Web content classification system, which creates the URL database of objectionable Web contents. It aims at shortening the update period of the DB, increasing the number of URLs in the DB, and enhancing the accuracy of the information in the DB.
Abstract: Researchers of drug-drug interaction alert systems
have often suggested that there were high overridden rate for alerts and
also too false alerts. However, research about decreasing false alerts is
scant. Therefore, the aim of this article attempts to proactive
identification of false alert for drug-drug interaction and provide
solution to decrease false alerts. This research involved retrospective
analysis prescribing database and calculated false alert rate by using
MYSQL and JAVA. Results of this study showed 17% of false alerts
and the false alert rate in the hospitals (37%) was more than in the
clinics. To conclude, this study described the importance that
drug-drug interaction alert system should not only detect drug name
but also detect frequency or route, as well as in providing solution to
decrease false alerts.
Abstract: In this study, workplace environmental monitoring
systems were established using USN(Ubiquitous Sensor Networks)
and LabVIEW. Although existing direct sampling methods enable
finding accurate values as of the time points of measurement, those
methods are disadvantageous in that continuous management and
supervision are difficult and costs for are high when those methods are
used. Therefore, the efficiency and reliability of workplace
management by supervisors are relatively low when those methods are
used. In this study, systems were established so that information on
workplace environmental factors such as temperatures, humidity and
noises is measured and transmitted to the PC in real time to enable
supervisors to monitor workplaces through LabVIEW on the PC.
When any accidents have occurred in workplaces, supervisors can
immediately respond through the monitoring system and this system
enables integrated workplace management and the prevention of
safety accidents. By introducing these monitoring systems, safety
accidents due to harmful environmental factors in workplaces can be
prevented and these monitoring systems will be also helpful in finding
out the correlation between safety accidents and occupational diseases
by comparing and linking databases established by this monitoring
system with existing statistical data.
Abstract: The present study presents a new approach to automatic
data clustering and classification problems in large and complex
databases and, at the same time, derives specific types of explicit rules
describing each cluster. The method works well in both sparse and
dense multidimensional data spaces. The members of the data space
can be of the same nature or represent different classes. A number
of N-dimensional ellipsoids are used for enclosing the data clouds.
Due to the geometry of an ellipsoid and its free rotation in space
the detection of clusters becomes very efficient. The method is based
on genetic algorithms that are used for the optimization of location,
orientation and geometric characteristics of the hyper-ellipsoids. The
proposed approach can serve as a basis for the development of
general knowledge systems for discovering hidden knowledge and
unexpected patterns and rules in various large databases.
Abstract: Face Recognition has always been a fascinating research area. It has drawn the attention of many researchers because of its various potential applications such as security systems, entertainment, criminal identification etc. Many supervised and unsupervised learning techniques have been reported so far. Principal Component Analysis (PCA), Self Organizing Maps (SOM) and Independent Component Analysis (ICA) are the three techniques among many others as proposed by different researchers for Face Recognition, known as the unsupervised techniques. This paper proposes integration of the two techniques, SOM and PCA, for dimensionality reduction and feature selection. Simulation results show that, though, the individual techniques SOM and PCA itself give excellent performance but the combination of these two can also be utilized for face recognition. Experimental results also indicate that for the given face database and the classifier used, SOM performs better as compared to other unsupervised learning techniques. A comparison of two proposed methodologies of SOM, Local and Global processing, shows the superiority of the later but at the cost of more computational time.
Abstract: Nowadays, organizing a repository of documents and
resources for learning on a special field as Information Technology
(IT), together with search techniques based on domain knowledge or
document-s content is an urgent need in practice of teaching, learning
and researching. There have been several works related to methods of
organization and search by content. However, the results are still
limited and insufficient to meet user-s demand for semantic
document retrieval. This paper presents a solution for the
organization of a repository that supports semantic representation and
processing in search. The proposed solution is a model which
integrates components such as an ontology describing domain
knowledge, a database of document repository, semantic
representation for documents and a file system; with problems,
semantic processing techniques and advanced search techniques
based on measuring semantic similarity. The solution is applied to
build a IT learning materials management system of a university with
semantic search function serving students, teachers, and manager as
well. The application has been implemented, tested at the University
of Information Technology, Ho Chi Minh City, Vietnam and has
achieved good results.
Abstract: Images are important in disease research, education,
and clinical medicine. This paper presents a Web Service Platform
(WSP) for support multiple programming languages to access image
from biomedical databases. The main function WSP is to allow web
users access image from biomedical databases. The WSP will
receive web user-s queries. After that, it will send to Querying
Server (QS) and the QS will search and retrieve data from
biomedical databases. Finally, the information will display to the
web users. Simple application is developed and tested for
experiment purpose. Result from experiment indicated WSP can be
used in biomedical environment.
Abstract: Knowledge Discovery in Databases (KDD) has
evolved into an important and active area of research because of
theoretical challenges and practical applications associated with the
problem of discovering (or extracting) interesting and previously
unknown knowledge from very large real-world databases. Rough
Set Theory (RST) is a mathematical formalism for representing
uncertainty that can be considered an extension of the classical set
theory. It has been used in many different research areas, including
those related to inductive machine learning and reduction of
knowledge in knowledge-based systems. One important concept
related to RST is that of a rough relation. In this paper we presented
the current status of research on applying rough set theory to KDD,
which will be helpful for handle the characteristics of real-world
databases. The main aim is to show how rough set and rough set
analysis can be effectively used to extract knowledge from large
databases.
Abstract: The increasing interest on processing data created by
sensor networks has evolved into approaches to implement sensor
networks as databases. The aggregation operator, which calculates a
value from a large group of data such as computing averages or sums,
etc. is an essential function that needs to be provided when
implementing such sensor network databases. This work proposes to
add the DURING clause into TinySQL to calculate values during a
specific long period and suggests a way to implement the aggregation
service in sensor networks by applying materialized view and
incremental view maintenance techniques that is used in data
warehouses. In sensor networks, data values are passed from child
nodes to parent nodes and an aggregation value is computed at the root
node. As such root nodes need to be memory efficient and low
powered, it becomes a problem to recompute aggregate values from all
past and current data. Therefore, applying incremental view
maintenance techniques can reduce the memory consumption and
support fast computation of aggregate values.
Abstract: Electrocardiogram (ECG) is considered to be the
backbone of cardiology. ECG is composed of P, QRS & T waves and
information related to cardiac diseases can be extracted from the
intervals and amplitudes of these waves. The first step in extracting
ECG features starts from the accurate detection of R peaks in the
QRS complex. We have developed a robust R wave detector using
wavelets. The wavelets used for detection are Daubechies and
Symmetric. The method does not require any preprocessing therefore,
only needs the ECG correct recordings while implementing the
detection. The database has been collected from MIT-BIH arrhythmia
database and the signals from Lead-II have been analyzed. MatLab
7.0 has been used to develop the algorithm. The ECG signal under
test has been decomposed to the required level using the selected
wavelet and the selection of detail coefficient d4 has been done based
on energy, frequency and cross-correlation analysis of decomposition
structure of ECG signal. The robustness of the method is apparent
from the obtained results.
Abstract: An empirical study of web applications that use
software frameworks is presented here. The analysis is based on two
approaches. In the first, developers using such frameworks are
required, based on their experience, to assign weights to parameters
such as database connection. In the second approach, a performance
testing tool, OpenSTA, is used to compute start time and other such
measures. From such an analysis, it is concluded that open source
software is superior to proprietary software. The motivation behind
this research is to examine ways in which a quantitative assessment
can be made of software in general and frameworks in particular.
Concepts such as metrics and architectural styles are discussed along
with previously published research.