The Mutated Distance between Two Mixture Trees

The evolutionary tree is an important topic in bioinformation. In 2006, Chen and Lindsay proposed a new method to build the mixture tree from DNA sequences. Mixture tree is a new type evolutionary tree, and it has two additional information besides the information of ordinary evolutionary tree. One of the information is time parameter, and the other is the set of mutated sites. In 2008, Lin and Juan proposed an algorithm to compute the distance between two mixture trees. Their algorithm computes the distance with only considering the time parameter between two mixture trees. In this paper, we proposes a method to measure the similarity of two mixture trees with considering the set of mutated sites and develops two algorithm to compute the distance between two mixture trees. The time complexity of these two proposed algorithms are O(n2 × max{h(T1), h(T2)}) and O(n2), respectively

Hidden State Probabilistic Modeling for Complex Wavelet Based Image Registration

This article presents a computationally tractable probabilistic model for the relation between the complex wavelet coefficients of two images of the same scene. The two images are acquisitioned at distinct moments of times, or from distinct viewpoints, or by distinct sensors. By means of the introduced probabilistic model, we argue that the similarity between the two images is controlled not by the values of the wavelet coefficients, which can be altered by many factors, but by the nature of the wavelet coefficients, that we model with the help of hidden state variables. We integrate this probabilistic framework in the construction of a new image registration algorithm. This algorithm has sub-pixel accuracy and is robust to noise and to other variations like local illumination changes. We present the performance of our algorithm on various image types.

A Relational Case-Based Reasoning Framework for Project Delivery System Selection

An appropriate project delivery system (PDS) is crucial to the success of a construction projects. Case-based Reasoning (CBR) is a useful support for PDS selection. However, the traditional CBR approach represents cases as attribute-value vectors without taking relations among attributes into consideration, and could not calculate the similarity when the structures of cases are not strictly same. Therefore, this paper solves this problem by adopting the Relational Case-based Reasoning (RCBR) approach for PDS selection, considering both the structural similarity and feature similarity. To develop the feature terms of the construction projects, the criteria and factors governing PDS selection process are first identified. Then feature terms for the construction projects are developed. Finally, the mechanism of similarity calculation and a case study indicate how RCBR works for PDS selection. The adoption of RCBR in PDS selection expands the scope of application of traditional CBR method and improves the accuracy of the PDS selection system.

Image Segment Matching Using Affine- Invariant Regions

In this paper, a method for matching image segments using triangle-based (geometrical) regions is proposed. Triangular regions are formed from triples of vertex points obtained from a keypoint detector (SIFT). However, triangle regions are subject to noise and distortion around the edges and vertices (especially acute angles). Therefore, these triangles are expanded into parallelogramshaped regions. The extracted image segments inherit an important triangle property; the invariance to affine distortion. Given two images, matching corresponding regions is conducted by computing the relative affine matrix, rectifying one of the regions w.r.t. the other one, then calculating the similarity between the reference and rectified region. The experimental tests show the efficiency and robustness of the proposed algorithm against geometrical distortion.

Recursive Similarity Hashing of Fractal Geometry

A new technique of topological multi-scale analysis is introduced. By performing a clustering recursively to build a hierarchy, and analyzing the co-scale and intra-scale similarities, an Iterated Function System can be extracted from any data set. The study of fractals shows that this method is efficient to extract self-similarities, and can find elegant solutions the inverse problem of building fractals. The theoretical aspects and practical implementations are discussed, together with examples of analyses of simple fractals.

Product Configuration Strategy Based On Product Family Similarity

To offer a large variety of products while maintaining low costs, high speed, and high quality in a mass customization product development environment, platform based product development has much benefit and usefulness in many industry fields. This paper proposes a product configuration strategy by similarity measure, incorporating the knowledge engineering principles such as product information model, ontology engineering, and formal concept analysis.

Comparative Studies on Vertical Stratification,Floristic Composition, and Woody Species Diversity of Subtropical Evergreen Broadleaf Forests Between the Ryukyu Archipelago, Japan, and South China

In order to compare vertical stratification, floristic composition, and woody species diversity of subtropical evergreen broadleaf forests between the Ryukyu Archipelago, Japan, and South China, tree censuses in a 400 m2 plot in Ishigaki Island and a 1225 m2 plot in Dinghushan Nature Reserve were performed. Both of the subtropical forests consisted of five vertical strata. The floristic composition of the Ishigaki forest was quite different from that of the Dinghushan forest in terms of similarity on a species level (Kuno-s similarity index r0 = 0.05). The values of Shannon-s index H' and Pielou-s index J ' tended to increase from the bottom stratum upward in both forests, except H' for the top stratum in the Ishigaki forest and the upper two strata in the Dinghushan forest. The woody species diversity in the Dinghushan forest (H'= 3.01 bit) was much lower than that in the Ishigaki forest (H'= 4.36 bit).

Iris Recognition Based On the Low Order Norms of Gradient Components

Iris pattern is an important biological feature of human body; it becomes very hot topic in both research and practical applications. In this paper, an algorithm is proposed for iris recognition and a simple, efficient and fast method is introduced to extract a set of discriminatory features using first order gradient operator applied on grayscale images. The gradient based features are robust, up to certain extents, against the variations may occur in contrast or brightness of iris image samples; the variations are mostly occur due lightening differences and camera changes. At first, the iris region is located, after that it is remapped to a rectangular area of size 360x60 pixels. Also, a new method is proposed for detecting eyelash and eyelid points; it depends on making image statistical analysis, to mark the eyelash and eyelid as a noise points. In order to cover the features localization (variation), the rectangular iris image is partitioned into N overlapped sub-images (blocks); then from each block a set of different average directional gradient densities values is calculated to be used as texture features vector. The applied gradient operators are taken along the horizontal, vertical and diagonal directions. The low order norms of gradient components were used to establish the feature vector. Euclidean distance based classifier was used as a matching metric for determining the degree of similarity between the features vector extracted from the tested iris image and template features vectors stored in the database. Experimental tests were performed using 2639 iris images from CASIA V4-Interival database, the attained recognition accuracy has reached up to 99.92%.

Goal Based Episodic Processing in Implicit Learning

Research has suggested that implicit learning tasks may rely on episodic processing to generate above chance performance on the standard classification tasks. The current research examines the invariant features task (McGeorge and Burton, 1990) and argues that such episodic processing is indeed important. The results of the experiment suggest that both rejection and similarity strategies are used by participants in this task to simultaneously reject unfamiliar items and to accept (falsely) familiar items. Primarily these decisions are based on the presence of low or high frequency goal based features of the stimuli presented in the incidental learning phase. It is proposed that a goal based analysis of the incidental learning task provides a simple step in understanding which features of the episodic processing are most important for explaining the match between incidental, implicit learning and test performance.

Rigid and Non-rigid Registration of Binary Objects using the Weighted Ratio Image

This paper presents the application of a signal intensity independent similarity criterion for rigid and non-rigid body registration of binary objects. The criterion is defined as the weighted ratio image of two images. The ratio is computed on a voxel per voxel basis and weighting is performed by setting the raios between signal and background voxels to a standard high value. The mean squared value of the weighted ratio is computed over the union of the signal areas of the two images and it is minimized using the Chebyshev polynomial approximation.

A Family Cars- Life Cycle Cost (LCC)-Oriented Hybrid Modelling Approach Combining ANN and CBR

Design for cost (DFC) is a method that reduces life cycle cost (LCC) from the angle of designers. Multiple domain features mapping (MDFM) methodology was given in DFC. Using MDFM, we can use design features to estimate the LCC. From the angle of DFC, the design features of family cars were obtained, such as all dimensions, engine power and emission volume. At the conceptual design stage, cars- LCC were estimated using back propagation (BP) artificial neural networks (ANN) method and case-based reasoning (CBR). Hamming space was used to measure the similarity among cases in CBR method. Levenberg-Marquardt (LM) algorithm and genetic algorithm (GA) were used in ANN. The differences of LCC estimation model between CBR and artificial neural networks (ANN) were provided. ANN and CBR separately each method has its shortcomings. By combining ANN and CBR improved results accuracy was obtained. Firstly, using ANN selected some design features that affect LCC. Then using LCC estimation results of ANN could raise the accuracy of LCC estimation in CBR method. Thirdly, using ANN estimate LCC errors and correct errors in CBR-s estimation results if the accuracy is not enough accurate. Finally, economically family cars and sport utility vehicle (SUV) was given as LCC estimation cases using this hybrid approach combining ANN and CBR.

Mining Network Data for Intrusion Detection through Naïve Bayesian with Clustering

Network security attacks are the violation of information security policy that received much attention to the computational intelligence society in the last decades. Data mining has become a very useful technique for detecting network intrusions by extracting useful knowledge from large number of network data or logs. Naïve Bayesian classifier is one of the most popular data mining algorithm for classification, which provides an optimal way to predict the class of an unknown example. It has been tested that one set of probability derived from data is not good enough to have good classification rate. In this paper, we proposed a new learning algorithm for mining network logs to detect network intrusions through naïve Bayesian classifier, which first clusters the network logs into several groups based on similarity of logs, and then calculates the prior and conditional probabilities for each group of logs. For classifying a new log, the algorithm checks in which cluster the log belongs and then use that cluster-s probability set to classify the new log. We tested the performance of our proposed algorithm by employing KDD99 benchmark network intrusion detection dataset, and the experimental results proved that it improves detection rates as well as reduces false positives for different types of network intrusions.

Classifying Biomedical Text Abstracts based on Hierarchical 'Concept' Structure

Classifying biomedical literature is a difficult and challenging task, especially when a large number of biomedical articles should be organized into a hierarchical structure. In this paper, we present an approach for classifying a collection of biomedical text abstracts downloaded from Medline database with the help of ontology alignment. To accomplish our goal, we construct two types of hierarchies, the OHSUMED disease hierarchy and the Medline abstract disease hierarchies from the OHSUMED dataset and the Medline abstracts, respectively. Then, we enrich the OHSUMED disease hierarchy before adapting it to ontology alignment process for finding probable concepts or categories. Subsequently, we compute the cosine similarity between the vector in probable concepts (in the “enriched" OHSUMED disease hierarchy) and the vector in Medline abstract disease hierarchies. Finally, we assign category to the new Medline abstracts based on the similarity score. The results obtained from the experiments show the performance of our proposed approach for hierarchical classification is slightly better than the performance of the multi-class flat classification.

Effective Digital Music Retrieval System through Content-based Features

In this paper, we propose effective system for digital music retrieval. We divided proposed system into Client and Server. Client part consists of pre-processing and Content-based feature extraction stages. In pre-processing stage, we minimized Time code Gap that is occurred among same music contents. As content-based feature, first-order differentiated MFCC were used. These presented approximately envelop of music feature sequences. Server part included Music Server and Music Matching stage. Extracted features from 1,000 digital music files were stored in Music Server. In Music Matching stage, we found retrieval result through similarity measure by DTW. In experiment, we used 450 queries. These were made by mixing different compression standards and sound qualities from 50 digital music files. Retrieval accurate indicated 97% and retrieval time was average 15ms in every single query. Out experiment proved that proposed system is effective in retrieve digital music and robust at various user environments of web.

Application of l1-Norm Minimization Technique to Image Retrieval

Image retrieval is a topic where scientific interest is currently high. The important steps associated with image retrieval system are the extraction of discriminative features and a feasible similarity metric for retrieving the database images that are similar in content with the search image. Gabor filtering is a widely adopted technique for feature extraction from the texture images. The recently proposed sparsity promoting l1-norm minimization technique finds the sparsest solution of an under-determined system of linear equations. In the present paper, the l1-norm minimization technique as a similarity metric is used in image retrieval. It is demonstrated through simulation results that the l1-norm minimization technique provides a promising alternative to existing similarity metrics. In particular, the cases where the l1-norm minimization technique works better than the Euclidean distance metric are singled out.

Group Invariant Solutions for Radial Jet Having Finite Fluid Velocity at Orifice

The group invariant solution for Prandtl-s boundary layer equations for an incompressible fluid governing the flow in radial free, wall and liquid jets having finite fluid velocity at the orifice are investigated. For each jet a symmetry is associated with the conserved vector that was used to derive the conserved quantity for the jet elsewhere. This symmetry is then used to construct the group invariant solution for the third-order partial differential equation for the stream function. The general form of the group invariant solution for radial jet flows is derived. The general form of group invariant solution and the general form of the similarity solution which was obtained elsewhere are the same.

An MCDM Approach to Selection Scheduling Rule in Robotic Flexibe Assembly Cells

Multiple criteria decision making (MCDM) is an approach to ranking the solutions and finding the best one when two or more solutions are provided. In this study, MCDM approach is proposed to select the most suitable scheduling rule of robotic flexible assembly cells (RFACs). Two MCDM approaches, Analytic Hierarchy Process (AHP) and Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) are proposed for solving the scheduling rule selection problem. The AHP method is employed to determine the weights of the evaluation criteria, while the TOPSIS method is employed to obtain final ranking order of scheduling rules. Four criteria are used to evaluate the scheduling rules. Also, four scheduling policies of RFAC are examined to choose the most appropriate one for this purpose. A numerical example illustrates applications of the suggested methodology. The results show that the methodology is practical and works in RFAC settings.

Measuring Teachers- Beliefs about Mathematics: A Fuzzy Set Approach

This paper deals with the application of a fuzzy set in measuring teachers- beliefs about mathematics. The vagueness of beliefs was transformed into standard mathematical values using a fuzzy preferences model. The study employed a fuzzy approach questionnaire which consists of six attributes for measuring mathematics teachers- beliefs about mathematics. The fuzzy conjoint analysis approach based on fuzzy set theory was used to analyze the data from twenty three mathematics teachers from four secondary schools in Terengganu, Malaysia. Teachers- beliefs were recorded in form of degrees of similarity and its levels of agreement. The attribute 'Drills and practice is one of the best ways of learning mathematics' scored the highest degree of similarity at 0. 79860 with level of 'strongly agree'. The results showed that the teachers- beliefs about mathematics were varied. This is shown by different levels of agreement and degrees of similarity of the measured attributes.

ISC–Intelligent Subspace Clustering, A Density Based Clustering Approach for High Dimensional Dataset

Many real-world data sets consist of a very high dimensional feature space. Most clustering techniques use the distance or similarity between objects as a measure to build clusters. But in high dimensional spaces, distances between points become relatively uniform. In such cases, density based approaches may give better results. Subspace Clustering algorithms automatically identify lower dimensional subspaces of the higher dimensional feature space in which clusters exist. In this paper, we propose a new clustering algorithm, ISC – Intelligent Subspace Clustering, which tries to overcome three major limitations of the existing state-of-art techniques. ISC determines the input parameter such as є – distance at various levels of Subspace Clustering which helps in finding meaningful clusters. The uniform parameters approach is not suitable for different kind of databases. ISC implements dynamic and adaptive determination of Meaningful clustering parameters based on hierarchical filtering approach. Third and most important feature of ISC is the ability of incremental learning and dynamic inclusion and exclusions of subspaces which lead to better cluster formation.

A Revisited View to the Paced Auditory Serial Addition Test (PASAT) in Female and Male Normal Subjects

Paced Auditory Serial Addition Test (PASAT) has been used as a common research tool for different neurological disorders like Multiple Sclerosis. Recently, technology let researchers to introduce a new versions of the visual test, the paced visual serial addition test (PVSAT). In this paper, the computerized version of these two tests is introduced. Beside the number of true responses are interpreted, the reaction time of subjects are calculated by the software. We hypothesize that paying attention to the reaction time may be valuable. For this purpose, sixty eight female normal subjects and fifty eight male normal subjects are enrolled in the study. We investigate the similarity between the PASAT3 and PVSAT3 in number of true responses and the new criterion (the average reaction time of each subject). The similarity between two tests were rejected (p-value = 0.000) which means that these two test differ. The effect of sex in the tests were not approved since the pvalues of different between PASAT3 and PVSAT3 in both sex is the same (p-value = 0.000) which means that male and female subjects performed the tests at no different level of performance. The new criterion shows a negative correlation with the age which offers aged normal subjects may have the same number of true responses as the young subjects but they have latent responses. This will give prove for the importance of reaction time.