Abstract: XML is an important standard of data exchange and
representation. As a mature database system, using relational database
to support XML data may bring some advantages. But storing XML in
relational database has obvious redundancy that wastes disk space,
bandwidth and disk I/O when querying XML data. For the efficiency
of storage and query XML, it is necessary to use compressed XML
data in relational database. In this paper, a compressed relational
database technology supporting XML data is presented. Original
relational storage structure is adaptive to XPath query process. The
compression method keeps this feature. Besides traditional relational
database techniques, additional query process technologies on
compressed relations and for special structure for XML are presented.
In this paper, technologies for XQuery process in compressed
relational database are presented..
Abstract: Decision support systems are usually based on
multidimensional structures which use the concept of hypercube.
Dimensions are the axes on which facts are analyzed and form a
space where a fact is located by a set of coordinates at the
intersections of members of dimensions. Conventional
multidimensional structures deal with discrete facts linked to discrete
dimensions. However, when dealing with natural continuous
phenomena the discrete representation is not adequate. There is a
need to integrate spatiotemporal continuity within multidimensional
structures to enable analysis and exploration of continuous field data.
Research issues that lead to the integration of spatiotemporal
continuity in multidimensional structures are numerous. In this paper,
we discuss research issues related to the integration of continuity in
multidimensional structures, present briefly a multidimensional
model for continuous field data. We also define new aggregation
operations. The model and the associated operations and measures
are validated by a prototype.
Abstract: Feature-based registration is an effective technique for clinical use, because it can greatly reduce computational costs. However, this technique, which estimates the transformation by using feature points extracted from two images, may cause misalignments. To handle with this limitation, we propose to extract the salient edges and extracted control points (CP) of medical images by using efficiency of multiresolution representation of data nonsubsampled contourlet transform (NSCT) that finds the best feature points. The MR images were first decomposed using the NSCT, and then Edge and CP were extracted from bandpass directional subband of NSCT coefficients and some proposed rules. After edge and CP extraction, mutual information was adopted for the registration of feature points and translation parameters are calculated by using particle swarm optimization (PSO). The experimental results showed that the proposed method produces totally accurate performance for registration medical CT-MR images.
Abstract: With the tremendous growth of World Wide Web
(WWW) data, there is an emerging need for effective information
retrieval at the document level. Several query languages such as
XML-QL, XPath, XQL, Quilt and XQuery are proposed in recent
years to provide faster way of querying XML data, but they still lack of
generality and efficiency. Our approach towards evolving a framework
for querying semistructured documents is based on formal query
algebra. Two elements are introduced in the proposed framework:
first, a generic and flexible data model for logical representation of
semistructured data and second, a set of operators for the manipulation
of objects defined in the data model. In additional to accommodating
several peculiarities of semistructured data, our model offers novel
features such as bidirectional paths for navigational querying and
partitions for data transformation that are not available in other
proposals.
Abstract: In this paper, we propose an approach for the classification of fingerprint databases. It is based on the fact that a fingerprint image is composed of regular texture regions that can be successfully represented by co-occurrence matrices. So, we first extract the features based on certain characteristics of the cooccurrence matrix and then we use these features to train a neural network for classifying fingerprints into four common classes. The obtained results compared with the existing approaches demonstrate the superior performance of our proposed approach.
Abstract: Cellular automata have been used for design of cryptosystems. Recently some secret sharing schemes based on linear memory cellular automata have been introduced which are used for both text and image. In this paper, we illustrate that these secret sharing schemes are vulnerable to dishonest participants- collusion. We propose a cheating model for the secret sharing schemes based on linear memory cellular automata. For this purpose we present a novel uniform model for representation of all secret sharing schemes based on cellular automata. Participants can cheat by means of sending bogus shares or bogus transition rules. Cheaters can cooperate to corrupt a shared secret and compute a cheating value added to it. Honest participants are not aware of cheating and suppose the incorrect secret as the valid one. We prove that cheaters can recover valid secret by removing the cheating value form the corrupted secret. We provide methods of calculating the cheating value.
Abstract: XML is becoming a de facto standard for online data exchange. Existing XML filtering techniques based on a publish/subscribe model are focused on the highly structured data marked up with XML tags. These techniques are efficient in filtering the documents of data-centric XML but are not effective in filtering the element contents of the document-centric XML. In this paper, we propose an extended XPath specification which includes a special matching character '%' used in the LIKE operation of SQL in order to solve the difficulty of writing some queries to adequately filter element contents using the previous XPath specification. We also present a novel technique for filtering a collection of document-centric XMLs, called Pfilter, which is able to exploit the extended XPath specification. We show several performance studies, efficiency and scalability using the multi-query processing time (MQPT).
Abstract: A multimedia presentation system refers to the integration of a multimedia database with a presentation manager which has the functionality of content selection, organization and playout of multimedia presentations. It requires high performance of involved system components. Starting from multimedia information capture until the presentation delivery, high performance tools are required for accessing, manipulating, storing and retrieving these segments, for transferring and delivering them in a presentation terminal according to a playout order. The organization of presentations is a complex task in that the display order of presentation contents (in time and space) must be specified. A multimedia presentation contains audio, video, images and text media types. The critical decisions for presentation construction include what the contents are, how the contents are organized, and once the decision is made on the organization of the contents of the presentation, it must be conveyed to the end user in the correct organizational order and in a timely fashion. This paper introduces a framework for specification of multimedia presentations and describes the design of sample presentations using this framework from a multimedia database.
Abstract: Source code retrieval is of immense importance in the software engineering field. The complex tasks of retrieving and extracting information from source code documents is vital in the development cycle of the large software systems. The two main subtasks which result from these activities are code duplication prevention and plagiarism detection. In this paper, we propose a Mohamed Amine Ouddan, and Hassane Essafi source code retrieval system based on two-level fingerprint representation, respectively the structural and the semantic information within a source code. A sequence alignment technique is applied on these fingerprints in order to quantify the similarity between source code portions. The specific purpose of the system is to detect plagiarism and duplicated code between programs written in different programming languages belonging to the same class, such as C, Cµ, Java and CSharp. These four languages are supported by the actual version of the system which is designed such that it may be easily adapted for any programming language.
Abstract: This paper presents unified theory for local (Savitzky-
Golay) and global polynomial smoothing. The algebraic framework
can represent any polynomial approximation and is seamless from
low degree local, to high degree global approximations. The representation
of the smoothing operator as a projection onto orthonormal
basis functions enables the computation of: the covariance matrix
for noise propagation through the filter; the noise gain and; the
frequency response of the polynomial filters. A virtually perfect Gram
polynomial basis is synthesized, whereby polynomials of degree
d = 1000 can be synthesized without significant errors. The perfect
basis ensures that the filters are strictly polynomial preserving. Given
n points and a support length ls = 2m + 1 then the smoothing
operator is strictly linear phase for the points xi, i = m+1. . . n-m.
The method is demonstrated on geometric surfaces data lying on an
invariant 2D lattice.
Abstract: Representation and description of object shapes by the
slopes of their contours or borders are proposed. The idea is to capture
the essence of the features that make it easier for a shape to be
stored, transmitted, compared and recognized. These features must
be independent of translation, rotation and scaling of the shape. A
approach is proposed to obtain high performance, efficiency and to
merge the boundaries into sequence of straight line segments with
the fewest possible segments. Evaluation on the performance of the
proposed method is based on its comparison with established method
of object shape description.
Abstract: Traditional principal components analysis (PCA)
techniques for face recognition are based on batch-mode training
using a pre-available image set. Real world applications require that
the training set be dynamic of evolving nature where within the
framework of continuous learning, new training images are
continuously added to the original set; this would trigger a costly
continuous re-computation of the eigen space representation via
repeating an entire batch-based training that includes the old and new
images. Incremental PCA methods allow adding new images and
updating the PCA representation. In this paper, two incremental
PCA approaches, CCIPCA and IPCA, are examined and compared.
Besides, different learning and testing strategies are proposed and
applied to the two algorithms. The results suggest that batch PCA is
inferior to both incremental approaches, and that all CCIPCAs are
practically equivalent.
Abstract: Railway Stations are prone to emergency due to
various reasons and proper monitor of railway stations are of
immense importance from various angles. A Petri-net representation
of a web-service-based Emergency management system has been
proposed in this paper which will help in monitoring situation of
train, track, signal etc. and in case of any emergency, necessary
resources can be dispatched.
Abstract: A different concept for designing and detailing of
reinforced concrete precast frame structures is analyzed in this paper.
The new detailing of the joints derives from the special hybrid
moment frame joints. The special reinforcements of this alternative
detailing, named modified special hybrid joint, are bondless with
respect to both column and beams. Full scale tests were performed on
a plan model, which represents a part of 5 story structure, cropped in
the middle of the beams and columns spans. Theoretical approach
was developed, based on testing results on twice repaired model,
subjected to lateral seismic type loading. Discussion regarding the
modified special hybrid joint behavior and further on widening
research needed concludes the presentation.
Abstract: Bioinformatics and Cheminformatics use computer as disciplines providing tools for acquisition, storage, processing, analysis, integrate data and for the development of potential applications of biological and chemical data. A chemical database is one of the databases that exclusively designed to store chemical information. NMRShiftDB is one of the main databases that used to represent the chemical structures in 2D or 3D structures. SMILES format is one of many ways to write a chemical structure in a linear format. In this study we extracted Antimicrobial Structures in SMILES format from NMRShiftDB and stored it in our Local Data Warehouse with its corresponding information. Additionally, we developed a searching tool that would response to user-s query using the JME Editor tool that allows user to draw or edit molecules and converts the drawn structure into SMILES format. We applied Quick Search algorithm to search for Antimicrobial Structures in our Local Data Ware House.
Abstract: Mostly the real life signals are time varying in nature. For proper characterization of such signals, time-frequency representation is required. The STFT (short-time Fourier transform) is a classical tool used for this purpose. The limitation of the STFT is its fixed time-frequency resolution. Thus, an enhanced version of the STFT, which is based on the cross-level sampling, is devised. It can adapt the sampling frequency and the window function length by following the input signal local variations. Therefore, it provides an adaptive resolution time-frequency representation of the input. The computational complexity of the proposed STFT is deduced and compared to the classical one. The results show a significant gain of the computational efficiency and hence of the processing power. The processing error of the proposed technique is also discussed.
Abstract: Ontology-based modelling of multi-formatted
software application content is a challenging area in content
management. When the number of software content unit is huge and
in continuous process of change, content change management is
important. The management of content in this context requires
targeted access and manipulation methods. We present a novel
approach to deal with model-driven content-centric information
systems and access to their content. At the core of our approach is an
ontology-based semantic annotation technique for diversely
formatted content that can improve the accuracy of access and
systems evolution. Domain ontologies represent domain-specific
concepts and conform to metamodels. Different ontologies - from
application domain ontologies to software ontologies - capture and
model the different properties and perspectives on a software content
unit. Interdependencies between domain ontologies, the artifacts and
the content are captured through a trace model. The annotation traces
are formalised and a graph-based system is selected for the
representation of the annotation traces.
Abstract: Nowadays there are many methods for representing
knowledge such as semantic network, neural network, and conceptual
graphs. Nonetheless, these methods are not sufficiently efficient
when applied to perform and deduce on knowledge domains about
supporting in general education such as algebra, analysis or plane
geometry. This leads to the introduction of computational network
which is a useful tool for representation knowledge base, especially
for computational knowledge, especially knowledge domain about
general education. However, when dealing with a practical problem,
we often do not immediately find a new solution, but we search
related problems which have been solved before and then proposing
an appropriate solution for the problem. Besides that, when finding
related problems, we have to determine whether the result of them
can be used to solve the practical problem or not. In this paper, the
extension model of computational network has been presented. In this
model, Sample Problems, which are related problems, will be used
like the experience of human about practical problem, simulate the
way of human thinking, and give the good solution for the practical
problem faster and more effectively. This extension model is applied
to construct an automatic system for solving algebraic problems in
middle school.
Abstract: Conceptualization strengthens intelligent systems in generalization skill, effective knowledge representation, real-time inference, and managing uncertain and indefinite situations in addition to facilitating knowledge communication for learning agents situated in real world. Concept learning introduces a way of abstraction by which the continuous state is formed as entities called concepts which are connected to the action space and thus, they illustrate somehow the complex action space. Of computational concept learning approaches, action-based conceptualization is favored because of its simplicity and mirror neuron foundations in neuroscience. In this paper, a new biologically inspired concept learning approach based on the probabilistic framework is proposed. This approach exploits and extends the mirror neuron-s role in conceptualization for a reinforcement learning agent in nondeterministic environments. In the proposed method, instead of building a huge numerical knowledge, the concepts are learnt gradually from rewards through interaction with the environment. Moreover the probabilistic formation of the concepts is employed to deal with uncertain and dynamic nature of real problems in addition to the ability of generalization. These characteristics as a whole distinguish the proposed learning algorithm from both a pure classification algorithm and typical reinforcement learning. Simulation results show advantages of the proposed framework in terms of convergence speed as well as generalization and asymptotic behavior because of utilizing both success and failures attempts through received rewards. Experimental results, on the other hand, show the applicability and effectiveness of the proposed method in continuous and noisy environments for a real robotic task such as maze as well as the benefits of implementing an incremental learning scenario in artificial agents.
Abstract: Color image segmentation plays an important role in
computer vision and image processing areas. In this paper, the
features of Volterra filter are utilized for color image segmentation.
The discrete Volterra filter exhibits both linear and nonlinear
characteristics. The linear part smoothes the image features in
uniform gray zones and is used for getting a gross representation of
objects of interest. The nonlinear term compensates for the blurring
due to the linear term and preserves the edges which are mainly used
to distinguish the various objects. The truncated quadratic Volterra
filters are mainly used for edge preserving along with Gaussian noise
cancellation. In our approach, the segmentation is based on K-means
clustering algorithm in HSI space. Both the hue and the intensity
components are fully utilized. For hue clustering, the special cyclic
property of the hue component is taken into consideration. The
experimental results show that the proposed technique segments the
color image while preserving significant features and removing noise
effects.