Abstract: In this paper we are to find the optimum
multiwavelet for compression of electrocardiogram (ECG)
signals. At present, it is not well known which multiwavelet is
the best choice for optimum compression of ECG. In this
work, we examine different multiwavelets on 24 sets of ECG
data with entirely different characteristics, selected from MITBIH
database. For assessing the functionality of the different
multiwavelets in compressing ECG signals, in addition to
known factors such as Compression Ratio (CR), Percent Root
Difference (PRD), Distortion (D), Root Mean Square Error
(RMSE) in compression literature, we also employed the
Cross Correlation (CC) criterion for studying the
morphological relations between the reconstructed and the
original ECG signal and Signal to reconstruction Noise Ratio
(SNR). The simulation results show that the cardbal2 by the
means of identity (Id) prefiltering method to be the best
effective transformation.
Abstract: There have been numerous implementations of
security system using biometric, especially for identification and
verification cases. An example of pattern used in biometric is the iris
pattern in human eye. The iris pattern is considered unique for each
person. The use of iris pattern poses problems in encoding the human
iris.
In this research, an efficient iris recognition method is proposed.
In the proposed method the iris segmentation is based on the
observation that the pupil has lower intensity than the iris, and the
iris has lower intensity than the sclera. By detecting the boundary
between the pupil and the iris and the boundary between the iris and
the sclera, the iris area can be separated from pupil and sclera. A step
is taken to reduce the effect of eyelashes and specular reflection of
pupil. Then the four levels Coiflet wavelet transform is applied to the
extracted iris image. The modified Hamming distance is employed to
measure the similarity between two irises.
This research yields the identification success rate of 84.25% for
the CASIA version 1.0 database. The method gives an accuracy of
77.78% for the left eyes of MMU 1 database and 86.67% for the
right eyes. The time required for the encoding process, from the
segmentation until the iris code is generated, is 0.7096 seconds.
These results show that the accuracy and speed of the method is
better than many other methods.
Abstract: The presented work is motivated by a french law regarding nuclear waste management. In order to avoid the limitation coming with the usage of the existing scenario codes, as COSI, VISION or FAMILY, the Core Library for Advance Scenario Simulation (CLASS) is being develop. CLASS is an open source tool, which allows any user to simulate an electronuclear scenario. The main CLASS asset, is the possibility to include any type of reactor, even a complitely new concept, through the generation of its ACSII evolution database. In the present article, the CLASS working basis will be presented as well as a simple exemple in order to show his potentiel. In the considered exemple, the effect of the transmutation will be assessed on Minor Actinide Inventory produced by PWR reactors.
Abstract: XML is an important standard of data exchange and
representation. As a mature database system, using relational database
to support XML data may bring some advantages. But storing XML in
relational database has obvious redundancy that wastes disk space,
bandwidth and disk I/O when querying XML data. For the efficiency
of storage and query XML, it is necessary to use compressed XML
data in relational database. In this paper, a compressed relational
database technology supporting XML data is presented. Original
relational storage structure is adaptive to XPath query process. The
compression method keeps this feature. Besides traditional relational
database techniques, additional query process technologies on
compressed relations and for special structure for XML are presented.
In this paper, technologies for XQuery process in compressed
relational database are presented..
Abstract: A multimedia presentation system refers to the integration of a multimedia database with a presentation manager which has the functionality of content selection, organization and playout of multimedia presentations. It requires high performance of involved system components. Starting from multimedia information capture until the presentation delivery, high performance tools are required for accessing, manipulating, storing and retrieving these segments, for transferring and delivering them in a presentation terminal according to a playout order. The organization of presentations is a complex task in that the display order of presentation contents (in time and space) must be specified. A multimedia presentation contains audio, video, images and text media types. The critical decisions for presentation construction include what the contents are, how the contents are organized, and once the decision is made on the organization of the contents of the presentation, it must be conveyed to the end user in the correct organizational order and in a timely fashion. This paper introduces a framework for specification of multimedia presentations and describes the design of sample presentations using this framework from a multimedia database.
Abstract: In this paper a new approach to face recognition is presented that achieves double dimension reduction making the system computationally efficient with better recognition results. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results improve with increase in face image resolution and levels off when arriving at a certain resolution level. In the proposed model of face recognition, first image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to better computational speed and feature extraction potential of Discrete Cosine Transform (DCT) it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A trade of between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL database, Yale database and a color database. The proposed technique has performed much better compared to other techniques. The significance of the model is two fold: (1) dimension reduction up to an effective and suitable face image resolution (2) appropriate DCT coefficients are retained to achieve best recognition results with varying image poses, intensity and illumination level.
Abstract: The problem of frequent pattern discovery is defined
as the process of searching for patterns such as sets of features or items that appear in data frequently. Finding such frequent patterns
has become an important data mining task because it reveals associations, correlations, and many other interesting relationships
hidden in a database. Most of the proposed frequent pattern mining
algorithms have been implemented with imperative programming
languages. Such paradigm is inefficient when set of patterns is large
and the frequent pattern is long. We suggest a high-level declarative
style of programming apply to the problem of frequent pattern
discovery. We consider two languages: Haskell and Prolog. Our
intuitive idea is that the problem of finding frequent patterns should
be efficiently and concisely implemented via a declarative paradigm
since pattern matching is a fundamental feature supported by most
functional languages and Prolog. Our frequent pattern mining
implementation using the Haskell and Prolog languages confirms our
hypothesis about conciseness of the program. The comparative
performance studies on line-of-code, speed and memory usage of
declarative versus imperative programming have been reported in the
paper.
Abstract: This article deals with the conceptual modeling under uncertainty. First, the division of information systems with their definition will be described, focusing on those where the construction of a conceptual model is suitable for the design of future information system database. Furthermore, the disadvantages of the traditional approach in creating a conceptual model and database design will be analyzed. A comprehensive methodology for the creation of a conceptual model based on analysis of client requirements and the selection of a suitable domain model is proposed here. This article presents the expert system used for the construction of a conceptual model and is a suitable tool for database designers to create a conceptual model.
Abstract: In data mining, the association rules are used to find
for the associations between the different items of the transactions
database. As the data collected and stored, rules of value can be found
through association rules, which can be applied to help managers
execute marketing strategies and establish sound market frameworks.
This paper aims to use Fuzzy Frequent Pattern growth (FFP-growth)
to derive from fuzzy association rules. At first, we apply fuzzy
partition methods and decide a membership function of quantitative
value for each transaction item. Next, we implement FFP-growth
to deal with the process of data mining. In addition, in order to
understand the impact of Apriori algorithm and FFP-growth algorithm
on the execution time and the number of generated association
rules, the experiment will be performed by using different sizes of
databases and thresholds. Lastly, the experiment results show FFPgrowth
algorithm is more efficient than other existing methods.
Abstract: The given work is devoted to the description of
Information Technologies NAS of Azerbaijan created and
successfully maintained in Institute. On the basis of the decision of
board of the Supreme Certifying commission at the President of the
Azerbaijan Republic and Presidium of National Academy of
Sciences of the Azerbaijan Republic, the organization of training
courses on Computer Sciences for all post-graduate students and
dissertators of the republic, taking of examinations of candidate
minima, it was on-line entrusted to Institute of Information
Technologies of the National Academy of Sciences of Azerbaijan.
Therefore, teaching the computer sciences to post-graduate
students and dissertators a scientific - methodological manual on
effective application of new information technologies for research
works by post-graduate students and dissertators and taking of
candidate minima is carried out in the Educational Center.
Information and communication technologies offer new
opportunities and prospects of their application for teaching and
training. The new level of literacy demands creation of essentially
new technology of obtaining of scientific knowledge. Methods of
training and development, social and professional requirements,
globalization of the communicative economic and political projects
connected with construction of a new society, depends on a level of
application of information and communication technologies in the
educational process. Computer technologies develop ideas of
programmed training, open completely new, not investigated
technological ways of training connected to unique opportunities of
modern computers and telecommunications. Computer technologies
of training are processes of preparation and transfer of the
information to the trainee by means of computer. Scientific and
technical progress as well as global spread of the technologies
created in the most developed countries of the world is the main
proof of the leading role of education in XXI century. Information
society needs individuals having modern knowledge. In practice, all
technologies, using special technical information means (computer,
audio, video) are called information technologies of education.
Abstract: Matching algorithms have significant importance in
speaker recognition. Feature vectors of the unknown utterance are
compared to feature vectors of the modeled speakers as a last step in
speaker recognition. A similarity score is found for every model in
the speaker database. Depending on the type of speaker recognition,
these scores are used to determine the author of unknown speech
samples. For speaker verification, similarity score is tested against a
predefined threshold and either acceptance or rejection result is
obtained. In the case of speaker identification, the result depends on
whether the identification is open set or closed set. In closed set
identification, the model that yields the best similarity score is
accepted. In open set identification, the best score is tested against a
threshold, so there is one more possible output satisfying the
condition that the speaker is not one of the registered speakers in
existing database. This paper focuses on closed set speaker
identification using a modified version of a well known matching
algorithm. The results of new matching algorithm indicated better
performance on YOHO international speaker recognition database.
Abstract: Modern spatial database management systems require a unique Spatial Access Method (SAM) in order solve complex spatial quires efficiently. In this case the spatial data structure takes a prominent place in the SAM. Inadequate data structure leads forming poor algorithmic choices and forging deficient understandings of algorithm behavior on the spatial database. A key step in developing a better semantic spatial object data structure is to quantify the performance effects of semantic and outlier detections that are not reflected in the previous tree structures (R-Tree and its variants). This paper explores a novel SSRO-Tree on SAM to the Topo-Semantic approach. The paper shows how to identify and handle the semantic spatial objects with outlier objects during page overflow/underflow, using gain/loss metrics. We introduce a new SSRO-Tree algorithm which facilitates the achievement of better performance in practice over algorithms that are superior in the R*-Tree and RO-Tree by considering selection queries.
Abstract: Matrix metalloproteinases (MMP) are a class of
structural and functional related enzymes involved in altering the
natural elements of the extracellular matrix. Most of the MMP
structures are cristalographycally determined and published in
WorldWide ProteinDataBank, isolated, in full structure or bound to
natural or synthetic inhibitors. This study proposes an algorithm to
replace missing crystallographic structures in PDB database. We
have compared the results of a chosen docking algorithm with a
known crystallographic structure in order to validate enzyme sites
reconstruction there where crystallographic data are missing.
Abstract: In this paper, we propose an efficient hierarchical DNA
sequence search method to improve the search speed while the
accuracy is being kept constant. For a given query DNA sequence,
firstly, a fast local search method using histogram features is used as a
filtering mechanism before scanning the sequences in the database.
An overlapping processing is newly added to improve the robustness
of the algorithm. A large number of DNA sequences with low
similarity will be excluded for latter searching. The Smith-Waterman
algorithm is then applied to each remainder sequences. Experimental
results using GenBank sequence data show the proposed method
combining histogram information and Smith-Waterman algorithm is
more efficient for DNA sequence search.
Abstract: Number of documents being created increases at an
increasing pace while most of them being in already known topics
and little of them introducing new concepts. This fact has started a
new era in information retrieval discipline where the requirements
have their own specialties. That is digging into topics and concepts
and finding out subtopics or relations between topics. Up to now IR
researches were interested in retrieving documents about a general
topic or clustering documents under generic subjects. However these
conventional approaches can-t go deep into content of documents
which makes it difficult for people to reach to right documents they
were searching. So we need new ways of mining document sets
where the critic point is to know much about the contents of the
documents. As a solution we are proposing to enhance LSI, one of
the proven IR techniques by supporting its vector space with n-gram
forms of words. Positive results we have obtained are shown in two
different application area of IR domain; querying a document
database, clustering documents in the document database.
Abstract: This paper presents data annotation models at
five levels of granularity (database, relation, column, tuple, and cell) of relational data to address the problem of unsuitability of most relational databases to express annotations. These models
do not require any structural and schematic changes to the
underlying database. These models are also flexible, extensible,
customizable, database-neutral, and platform-independent. This paper also presents an SQL-like query language, named Annotation Query Language (AnQL), to query annotation documents. AnQL is simple to understand and exploits the already-existent wide knowledge and skill set of SQL.
Abstract: The development of distributed systems has been affected by the need to accommodate an increasing degree of flexibility, adaptability, and autonomy. The Mobile Agent technology is emerging as an alternative to build a smart generation of highly distributed systems. In this work, we investigate the performance aspect of agent-based technologies for information retrieval. We present a comparative performance evaluation model of Mobile Agents versus Remote Method Invocation by means of an analytical approach. We demonstrate the effectiveness of mobile agents for dynamic code deployment and remote data processing by reducing total latency and at the same time producing minimum network traffic. We argue that exploiting agent-based technologies significantly enhances the performance of distributed systems in the domain of information retrieval.
Abstract: A new automatic system for the recognition and re¬construction of resealed and/or rotated partially occluded objects is presented. The objects to be recognized are described by 2D views and each view is occluded by several half-planes. The whole object views and their visible parts (linear cuts) are then stored in a database. To establish if a region R of an input image represents an object possibly occluded, the system generates a set of linear cuts of R and compare them with the elements in the database. Each linear cut of R is associated to the most similar database linear cut. R is recognized as an instance of the object 0 if the majority of the linear cuts of R are associated to a linear cut of views of 0. In the case of recognition, the system reconstructs the occluded part of R and determines the scale factor and the orientation in the image plane of the recognized object view. The system has been tested on two different datasets of objects, showing good performance both in terms of recognition and reconstruction accuracy.
Abstract: This paper presents an automatic feature recognition
method based on center-surround difference detecting and fuzzy logic
that can be applied in ground-penetrating radar (GPR) image
processing. Adopted center-surround difference method, the salient
local image regions are extracted from the GPR images as features of
detected objects. And fuzzy logic strategy is used to match the
detected features and features in template database. This way, the
problem of objects detecting, which is the key problem in GPR image
processing, can be converted into two steps, feature extracting and
matching. The contributions of these skills make the system have the
ability to deal with changes in scale, antenna and noises. The results of
experiments also prove that the system has higher ratio of features
sensing in using GPR to image the subsurface structures.
Abstract: The latest Geographic Information System (GIS)
technology makes it possible to administer the spatial components of
daily “business object," in the corporate database, and apply suitable
geographic analysis efficiently in a desktop-focused application. We
can use wireless internet technology for transfer process in spatial
data from server to client or vice versa. However, the problem in
wireless Internet is system bottlenecks that can make the process of
transferring data not efficient. The reason is large amount of spatial
data. Optimization in the process of transferring and retrieving data,
however, is an essential issue that must be considered. Appropriate
decision to choose between R-tree and Quadtree spatial data indexing
method can optimize the process. With the rapid proliferation of
these databases in the past decade, extensive research has been
conducted on the design of efficient data structures to enable fast
spatial searching. Commercial database vendors like Oracle have also
started implementing these spatial indexing to cater to the large and
diverse GIS. This paper focuses on the decisions to choose R-tree
and quadtree spatial indexing using Oracle spatial database in mobile
GIS application. From our research condition, the result of using
Quadtree and R-tree spatial data indexing method in one single
spatial database can save the time until 42.5%.