Abstract: The evolutionary tree is an important topic in bioinformation. In 2006, Chen and Lindsay proposed a new method to build the mixture tree from DNA sequences. Mixture tree is a new type evolutionary tree, and it has two additional information besides the information of ordinary evolutionary tree. One of the information is time parameter, and the other is the set of mutated sites. In 2008, Lin and Juan proposed an algorithm to compute the distance between two mixture trees. Their algorithm computes the distance with only considering the time parameter between two mixture trees. In this paper, we proposes a method to measure the similarity of two mixture trees with considering the set of mutated sites and develops two algorithm to compute the distance between two mixture trees. The time complexity of these two proposed algorithms are O(n2 × max{h(T1), h(T2)}) and O(n2), respectively
Abstract: This article presents a computationally tractable probabilistic model for the relation between the complex wavelet coefficients of two images of the same scene. The two images are acquisitioned at distinct moments of times, or from distinct viewpoints, or by distinct sensors. By means of the introduced probabilistic model, we argue that the similarity between the two images is controlled not by the values of the wavelet coefficients, which can be altered by many factors, but by the nature of the wavelet coefficients, that we model with the help of hidden state variables. We integrate this probabilistic framework in the construction of a new image registration algorithm. This algorithm has sub-pixel accuracy and is robust to noise and to other variations like local illumination changes. We present the performance of our algorithm on various image types.
Abstract: An appropriate project delivery system (PDS) is crucial
to the success of a construction projects. Case-based Reasoning (CBR)
is a useful support for PDS selection. However, the traditional CBR
approach represents cases as attribute-value vectors without taking
relations among attributes into consideration, and could not calculate
the similarity when the structures of cases are not strictly same.
Therefore, this paper solves this problem by adopting the Relational
Case-based Reasoning (RCBR) approach for PDS selection,
considering both the structural similarity and feature similarity. To
develop the feature terms of the construction projects, the criteria and
factors governing PDS selection process are first identified. Then
feature terms for the construction projects are developed. Finally, the
mechanism of similarity calculation and a case study indicate how
RCBR works for PDS selection. The adoption of RCBR in PDS
selection expands the scope of application of traditional CBR method
and improves the accuracy of the PDS selection system.
Abstract: In this paper, a method for matching image segments
using triangle-based (geometrical) regions is proposed. Triangular
regions are formed from triples of vertex points obtained from a
keypoint detector (SIFT). However, triangle regions are subject to
noise and distortion around the edges and vertices (especially acute
angles). Therefore, these triangles are expanded into parallelogramshaped
regions. The extracted image segments inherit an important
triangle property; the invariance to affine distortion. Given two
images, matching corresponding regions is conducted by computing
the relative affine matrix, rectifying one of the regions w.r.t. the other
one, then calculating the similarity between the reference and
rectified region. The experimental tests show the efficiency and
robustness of the proposed algorithm against geometrical distortion.
Abstract: A new technique of topological multi-scale analysis is
introduced. By performing a clustering recursively to build a
hierarchy, and analyzing the co-scale and intra-scale similarities, an
Iterated Function System can be extracted from any data set. The study
of fractals shows that this method is efficient to extract
self-similarities, and can find elegant solutions the inverse problem of
building fractals. The theoretical aspects and practical
implementations are discussed, together with examples of analyses of
simple fractals.
Abstract: To offer a large variety of products while maintaining
low costs, high speed, and high quality in a mass customization
product development environment, platform based product
development has much benefit and usefulness in many industry fields.
This paper proposes a product configuration strategy by similarity
measure, incorporating the knowledge engineering principles such as
product information model, ontology engineering, and formal concept
analysis.
Abstract: In order to compare vertical stratification, floristic composition, and woody species diversity of subtropical evergreen broadleaf forests between the Ryukyu Archipelago, Japan, and South China, tree censuses in a 400 m2 plot in Ishigaki Island and a 1225 m2 plot in Dinghushan Nature Reserve were performed. Both of the subtropical forests consisted of five vertical strata. The floristic composition of the Ishigaki forest was quite different from that of the Dinghushan forest in terms of similarity on a species level (Kuno-s similarity index r0 = 0.05). The values of Shannon-s index H' and Pielou-s index J ' tended to increase from the bottom stratum upward in both forests, except H' for the top stratum in the Ishigaki forest and the upper two strata in the Dinghushan forest. The woody species diversity in the Dinghushan forest (H'= 3.01 bit) was much lower than that in the Ishigaki forest (H'= 4.36 bit).
Abstract: Iris pattern is an important biological feature of human body; it becomes very hot topic in both research and practical applications. In this paper, an algorithm is proposed for iris recognition and a simple, efficient and fast method is introduced to extract a set of discriminatory features using first order gradient operator applied on grayscale images. The gradient based features are robust, up to certain extents, against the variations may occur in contrast or brightness of iris image samples; the variations are mostly occur due lightening differences and camera changes. At first, the iris region is located, after that it is remapped to a rectangular area of size 360x60 pixels. Also, a new method is proposed for detecting eyelash and eyelid points; it depends on making image statistical analysis, to mark the eyelash and eyelid as a noise points. In order to cover the features localization (variation), the rectangular iris image is partitioned into N overlapped sub-images (blocks); then from each block a set of different average directional gradient densities values is calculated to be used as texture features vector. The applied gradient operators are taken along the horizontal, vertical and diagonal directions. The low order norms of gradient components were used to establish the feature vector. Euclidean distance based classifier was used as a matching metric for determining the degree of similarity between the features vector extracted from the tested iris image and template features vectors stored in the database. Experimental tests were performed using 2639 iris images from CASIA V4-Interival database, the attained recognition accuracy has reached up to 99.92%.
Abstract: Research has suggested that implicit learning tasks
may rely on episodic processing to generate above chance
performance on the standard classification tasks. The current
research examines the invariant features task (McGeorge and Burton,
1990) and argues that such episodic processing is indeed important.
The results of the experiment suggest that both rejection and
similarity strategies are used by participants in this task to
simultaneously reject unfamiliar items and to accept (falsely) familiar
items. Primarily these decisions are based on the presence of low or
high frequency goal based features of the stimuli presented in the
incidental learning phase. It is proposed that a goal based analysis of
the incidental learning task provides a simple step in understanding
which features of the episodic processing are most important for
explaining the match between incidental, implicit learning and test
performance.
Abstract: This paper presents the application of a signal intensity
independent similarity criterion for rigid and non-rigid body
registration of binary objects. The criterion is defined as the
weighted ratio image of two images. The ratio is computed on a
voxel per voxel basis and weighting is performed by setting the raios
between signal and background voxels to a standard high value. The
mean squared value of the weighted ratio is computed over the union
of the signal areas of the two images and it is minimized using the
Chebyshev polynomial approximation.
Abstract: Design for cost (DFC) is a method that reduces life
cycle cost (LCC) from the angle of designers. Multiple domain
features mapping (MDFM) methodology was given in DFC. Using
MDFM, we can use design features to estimate the LCC. From the
angle of DFC, the design features of family cars were obtained, such
as all dimensions, engine power and emission volume. At the
conceptual design stage, cars- LCC were estimated using back
propagation (BP) artificial neural networks (ANN) method and
case-based reasoning (CBR). Hamming space was used to measure the
similarity among cases in CBR method. Levenberg-Marquardt (LM)
algorithm and genetic algorithm (GA) were used in ANN. The
differences of LCC estimation model between CBR and artificial
neural networks (ANN) were provided. ANN and CBR separately
each method has its shortcomings. By combining ANN and CBR
improved results accuracy was obtained. Firstly, using ANN selected
some design features that affect LCC. Then using LCC estimation
results of ANN could raise the accuracy of LCC estimation in CBR
method. Thirdly, using ANN estimate LCC errors and correct errors in
CBR-s estimation results if the accuracy is not enough accurate.
Finally, economically family cars and sport utility vehicle (SUV) was
given as LCC estimation cases using this hybrid approach combining
ANN and CBR.
Abstract: Network security attacks are the violation of
information security policy that received much attention to the
computational intelligence society in the last decades. Data mining
has become a very useful technique for detecting network intrusions
by extracting useful knowledge from large number of network data
or logs. Naïve Bayesian classifier is one of the most popular data
mining algorithm for classification, which provides an optimal way
to predict the class of an unknown example. It has been tested that
one set of probability derived from data is not good enough to have
good classification rate. In this paper, we proposed a new learning
algorithm for mining network logs to detect network intrusions
through naïve Bayesian classifier, which first clusters the network
logs into several groups based on similarity of logs, and then
calculates the prior and conditional probabilities for each group of
logs. For classifying a new log, the algorithm checks in which cluster
the log belongs and then use that cluster-s probability set to classify
the new log. We tested the performance of our proposed algorithm by
employing KDD99 benchmark network intrusion detection dataset,
and the experimental results proved that it improves detection rates
as well as reduces false positives for different types of network
intrusions.
Abstract: Classifying biomedical literature is a difficult and
challenging task, especially when a large number of biomedical
articles should be organized into a hierarchical structure. In this paper,
we present an approach for classifying a collection of biomedical text
abstracts downloaded from Medline database with the help of
ontology alignment. To accomplish our goal, we construct two types
of hierarchies, the OHSUMED disease hierarchy and the Medline
abstract disease hierarchies from the OHSUMED dataset and the
Medline abstracts, respectively. Then, we enrich the OHSUMED
disease hierarchy before adapting it to ontology alignment process for
finding probable concepts or categories. Subsequently, we compute
the cosine similarity between the vector in probable concepts (in the
“enriched" OHSUMED disease hierarchy) and the vector in Medline
abstract disease hierarchies. Finally, we assign category to the new
Medline abstracts based on the similarity score. The results obtained
from the experiments show the performance of our proposed approach
for hierarchical classification is slightly better than the performance of
the multi-class flat classification.
Abstract: In this paper, we propose effective system for digital music retrieval. We divided proposed system into Client and Server. Client part consists of pre-processing and Content-based feature extraction stages. In pre-processing stage, we minimized Time code Gap that is occurred among same music contents. As content-based feature, first-order differentiated MFCC were used. These presented approximately envelop of music feature sequences. Server part included Music Server and Music Matching stage. Extracted features from 1,000 digital music files were stored in Music Server. In Music Matching stage, we found retrieval result through similarity measure by DTW. In experiment, we used 450 queries. These were made by mixing different compression standards and sound qualities from 50 digital music files. Retrieval accurate indicated 97% and retrieval time was average 15ms in every single query. Out experiment proved that proposed system is effective in retrieve digital music and robust at various user environments of web.
Abstract: Image retrieval is a topic where scientific interest is currently high. The important steps associated with image retrieval system are the extraction of discriminative features and a feasible similarity metric for retrieving the database images that are similar in content with the search image. Gabor filtering is a widely adopted technique for feature extraction from the texture images. The recently proposed sparsity promoting l1-norm minimization technique finds the sparsest solution of an under-determined system of linear equations. In the present paper, the l1-norm minimization technique as a similarity metric is used in image retrieval. It is demonstrated through simulation results that the l1-norm minimization technique provides a promising alternative to existing similarity metrics. In particular, the cases where the l1-norm minimization technique works better than the Euclidean distance metric are singled out.
Abstract: The group invariant solution for Prandtl-s boundary layer equations for an incompressible fluid governing the flow in radial free, wall and liquid jets having finite fluid velocity at the orifice are investigated. For each jet a symmetry is associated with the conserved vector that was used to derive the conserved quantity for the jet elsewhere. This symmetry is then used to construct the group invariant solution for the third-order partial differential equation for the stream function. The general form of the group invariant solution for radial jet flows is derived. The general form of group invariant solution and the general form of the similarity solution which was obtained elsewhere are the same.
Abstract: Multiple criteria decision making (MCDM) is an approach to ranking the solutions and finding the best one when two or more solutions are provided. In this study, MCDM approach is proposed to select the most suitable scheduling rule of robotic flexible assembly cells (RFACs). Two MCDM approaches, Analytic Hierarchy Process (AHP) and Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) are proposed for solving the scheduling rule selection problem. The AHP method is employed to determine the weights of the evaluation criteria, while the TOPSIS method is employed to obtain final ranking order of scheduling rules. Four criteria are used to evaluate the scheduling rules. Also, four scheduling policies of RFAC are examined to choose the most appropriate one for this purpose. A numerical example illustrates applications of the suggested methodology. The results show that the methodology is practical and works in RFAC settings.
Abstract: This paper deals with the application of a fuzzy set in
measuring teachers- beliefs about mathematics. The vagueness of
beliefs was transformed into standard mathematical values using a
fuzzy preferences model. The study employed a fuzzy approach
questionnaire which consists of six attributes for measuring
mathematics teachers- beliefs about mathematics. The fuzzy conjoint
analysis approach based on fuzzy set theory was used to analyze the
data from twenty three mathematics teachers from four secondary
schools in Terengganu, Malaysia. Teachers- beliefs were recorded in
form of degrees of similarity and its levels of agreement. The
attribute 'Drills and practice is one of the best ways of learning
mathematics' scored the highest degree of similarity at 0. 79860 with
level of 'strongly agree'. The results showed that the teachers- beliefs
about mathematics were varied. This is shown by different levels of
agreement and degrees of similarity of the measured attributes.
Abstract: Many real-world data sets consist of a very high dimensional feature space. Most clustering techniques use the distance or similarity between objects as a measure to build clusters. But in high dimensional spaces, distances between points become relatively uniform. In such cases, density based approaches may give better results. Subspace Clustering algorithms automatically identify lower dimensional subspaces of the higher dimensional feature space in which clusters exist. In this paper, we propose a new clustering algorithm, ISC – Intelligent Subspace Clustering, which tries to overcome three major limitations of the existing state-of-art techniques. ISC determines the input parameter such as є – distance at various levels of Subspace Clustering which helps in finding meaningful clusters. The uniform parameters approach is not suitable for different kind of databases. ISC implements dynamic and adaptive determination of Meaningful clustering parameters based on hierarchical filtering approach. Third and most important feature of ISC is the ability of incremental learning and dynamic inclusion and exclusions of subspaces which lead to better cluster formation.
Abstract: Paced Auditory Serial Addition Test (PASAT) has
been used as a common research tool for different neurological
disorders like Multiple Sclerosis. Recently, technology let
researchers to introduce a new versions of the visual test, the paced
visual serial addition test (PVSAT). In this paper, the computerized
version of these two tests is introduced. Beside the number of true
responses are interpreted, the reaction time of subjects are calculated
by the software. We hypothesize that paying attention to the reaction
time may be valuable. For this purpose, sixty eight female normal
subjects and fifty eight male normal subjects are enrolled in the
study. We investigate the similarity between the PASAT3 and
PVSAT3 in number of true responses and the new criterion (the
average reaction time of each subject). The similarity between two
tests were rejected (p-value = 0.000) which means that these two test
differ. The effect of sex in the tests were not approved since the pvalues
of different between PASAT3 and PVSAT3 in both sex is the
same (p-value = 0.000) which means that male and female subjects
performed the tests at no different level of performance. The new
criterion shows a negative correlation with the age which offers aged
normal subjects may have the same number of true responses as the
young subjects but they have latent responses. This will give prove
for the importance of reaction time.