Abstract: In order to reduce the number of deaths due to heart
problems, we propose the use of Hierarchical Temporal Memory
Algorithm (HTM) which is a real time anomaly detection algorithm.
HTM is a cortical learning algorithm based on neocortex used for
anomaly detection. In other words, it is based on a conceptual theory
of how the human brain can work. It is powerful in predicting unusual
patterns, anomaly detection and classification. In this paper, HTM
have been implemented and tested on ECG datasets in order to detect
cardiac anomalies. Experiments showed good performance in terms
of specificity, sensitivity and execution time.
Abstract: Character recognition is the process of converting a text image file into editable and searchable text file. Feature Extraction is the heart of any character recognition system. The character recognition rate may be low or high depending on the extracted features. In the proposed paper, 25 features for one character are used in character recognition. Basically, there are three steps of character recognition such as character segmentation, feature extraction and classification. In segmentation step, horizontal cropping method is used for line segmentation and vertical cropping method is used for character segmentation. In the Feature extraction step, features are extracted in two ways. The first way is that the 8 features are extracted from the entire input character using eight direction chain code frequency extraction. The second way is that the input character is divided into 16 blocks. For each block, although 8 feature values are obtained through eight-direction chain code frequency extraction method, we define the sum of these 8 feature values as a feature for one block. Therefore, 16 features are extracted from that 16 blocks in the second way. We use the number of holes feature to cluster the similar characters. We can recognize the almost Myanmar common characters with various font sizes by using these features. All these 25 features are used in both training part and testing part. In the classification step, the characters are classified by matching the all features of input character with already trained features of characters.
Abstract: This paper presents a road vehicle detection approach for the intelligent transportation system. This approach mainly uses low-cost magnetic sensor and associated data collection system to collect magnetic signals. This system can measure the magnetic field changing, and it also can detect and count vehicles. We extend Mel Frequency Cepstral Coefficients to analyze vehicle magnetic signals. Vehicle type features are extracted using representation of cepstrum, frame energy, and gap cepstrum of magnetic signals. We design a 2-dimensional map algorithm using Vector Quantization to classify vehicle magnetic features to four typical types of vehicles in Australian suburbs: sedan, VAN, truck, and bus. Experiments results show that our approach achieves a high level of accuracy for vehicle detection and classification.
Abstract: An artificial neural network is a mathematical model
inspired by biological neural networks. There are several kinds of
neural networks and they are widely used in many areas, such as:
prediction, detection, and classification. Meanwhile, in day to day life,
people always have to make many difficult decisions. For example,
the coach of a soccer club has to decide which offensive player
to be selected to play in a certain game. This work describes a
novel Neural Network using a combination of the General Regression
Neural Network and the Probabilistic Neural Networks to help a
soccer coach make an informed decision.
Abstract: The aim of this work is to build a model based on
tissue characterization that is able to discriminate pathological and
non-pathological regions from three-phasic CT images. With our
research and based on a feature selection in different phases, we are
trying to design a neural network system with an optimal neuron
number in a hidden layer. Our approach consists of three steps:
feature selection, feature reduction, and classification. For each
region of interest (ROI), 6 distinct sets of texture features are
extracted such as: first order histogram parameters, absolute gradient,
run-length matrix, co-occurrence matrix, autoregressive model, and
wavelet, for a total of 270 texture features. When analyzing more
phases, we show that the injection of liquid cause changes to the high
relevant features in each region. Our results demonstrate that for
detecting HCC tumor phase 3 is the best one in most of the features
that we apply to the classification algorithm. The percentage of
detection between pathology and healthy classes, according to our
method, relates to first order histogram parameters with accuracy of
85% in phase 1, 95% in phase 2, and 95% in phase 3.
Abstract: In remote sensing, shadow causes problems in many
applications such as change detection and classification. It is caused
by objects which are elevated, thus can directly affect the accuracy of
information. For these reasons, it is very important to detect shadows
particularly in urban high spatial resolution imagery which created a
significant problem. This paper focuses on automatic shadow
detection based on a new spectral index for multispectral imagery
known as Shadow Detection Index (SDI). The new spectral index
was tested on different areas of WorldView-2 images and the results
demonstrated that the new spectral index has a massive potential to
extract shadows with accuracy of 94% effectively and automatically.
Furthermore, the new shadow detection index improved road
extraction from 82% to 93%.
Abstract: BCI (Brain Computer Interface) is a communication machine that translates brain massages to computer commands. These machines with the help of computer programs can recognize the tasks that are imagined. Feature extraction is an important stage of the process in EEG classification that can effect in accuracy and the computation time of processing the signals. In this study we process the signal in three steps of active segment selection, fractal feature extraction, and classification. One of the great challenges in BCI applications is to improve classification accuracy and computation time together. In this paper, we have used student’s 2D sample t-statistics on continuous wavelet transforms for active segment selection to reduce the computation time. In the next level, the features are extracted from some famous fractal dimension estimation of the signal. These fractal features are Katz and Higuchi. In the classification stage we used ANFIS (Adaptive Neuro-Fuzzy Inference System) classifier, FKNN (Fuzzy K-Nearest Neighbors), LDA (Linear Discriminate Analysis), and SVM (Support Vector Machines). We resulted that active segment selection method would reduce the computation time and Fractal dimension features with ANFIS analysis on selected active segments is the best among investigated methods in EEG classification.
Abstract: The exponential increase in the volume of medical image database has imposed new challenges to clinical routine in maintaining patient history, diagnosis, treatment and monitoring. With the advent of data mining and machine learning techniques it is possible to automate and/or assist physicians in clinical diagnosis. In this research a medical image classification framework using data mining techniques is proposed. It involves feature extraction, feature selection, feature discretization and classification. In the classification phase, the performance of the traditional kNN k nearest neighbor classifier is improved using a feature weighting scheme and a distance weighted voting instead of simple majority voting. Feature weights are calculated using the interestingness measures used in association rule mining. Experiments on the retinal fundus images show that the proposed framework improves the classification accuracy of traditional kNN from 78.57 % to 92.85 %.
Abstract: In this paper, we investigated the characteristic of a
clinical dataseton the feature selection and classification
measurements which deal with missing values problem.And also
posed the appropriated techniques to achieve the aim of the activity;
in this research aims to find features that have high effect to mortality
and mortality time frame. We quantify the complexity of a clinical
dataset. According to the complexity of the dataset, we proposed the
data mining processto cope their complexity; missing values, high
dimensionality, and the prediction problem by using the methods of
missing value replacement, feature selection, and classification.The
experimental results will extend to develop the prediction model for
cardiology.
Abstract: Text categorization is the problem of classifying text
documents into a set of predefined classes. After a preprocessing
step, the documents are typically represented as large sparse vectors.
When training classifiers on large collections of documents, both the
time and memory restrictions can be quite prohibitive. This justifies
the application of feature selection methods to reduce the
dimensionality of the document-representation vector. In this paper,
we present three feature selection methods: Information Gain,
Support Vector Machine feature selection called (SVM_FS) and
Genetic Algorithm with SVM (called GA_SVM). We show that the
best results were obtained with GA_SVM method for a relatively
small dimension of the feature vector.
Abstract: This paper presents the effectiveness of artificial
intelligent technique to apply for pattern recognition and
classification of Partial Discharge (PD). Characteristics of PD signal
for pattern recognition and classification are computed from the
relation of the voltage phase angle, the discharge magnitude and the
repeated existing of partial discharges by using statistical and fractal
methods. The simplified fuzzy ARTMAP (SFAM) is used for pattern
recognition and classification as artificial intelligent technique. PDs
quantities, 13 parameters from statistical method and fractal method
results, are inputted to Simplified Fuzzy ARTMAP to train system
for pattern recognition and classification. The results confirm the
effectiveness of purpose technique.
Abstract: Text similarity measurement is a fundamental issue in
many textual applications such as document clustering, classification,
summarization and question answering. However, prevailing approaches
based on Vector Space Model (VSM) more or less suffer
from the limitation of Bag of Words (BOW), which ignores the semantic
relationship among words. Enriching document representation
with background knowledge from Wikipedia is proven to be an effective
way to solve this problem, but most existing methods still
cannot avoid similar flaws of BOW in a new vector space. In this
paper, we propose a novel text similarity measurement which goes
beyond VSM and can find semantic affinity between documents.
Specifically, it is a unified graph model that exploits Wikipedia as
background knowledge and synthesizes both document representation
and similarity computation. The experimental results on two different
datasets show that our approach significantly improves VSM-based
methods in both text clustering and classification.
Abstract: It has often been said that the strength of any country
resides in the strength of its industrial sector, and Progress in
industrial society has been accomplished by the creation of new
technologies. Developments have been facilitated by the increasing
availability of advanced manufacturing technology (AMT), in
addition the implementation of advanced manufacturing technology
(AMT) requires careful planning at all levels of the organization to
ensure that the implementation will achieve the intended goals.
Justification and implementation of advanced manufacturing
technology (AMT) involves decisions that are crucial for the
practitioners regarding the survival of business in the present days of
uncertain manufacturing world. This paper assists the industrial
managers to consider all the important criteria for success AMT
implementation, when purchasing new technology. Concurrently,
this paper classifies the tangible benefits of a technology that are
evaluated by addressing both cost and time dimensions, and the
intangible benefits are evaluated by addressing technological,
strategic, social and human issues to identify and create awareness of
the essential elements in the AMT implementation process and
identify the necessary actions before implementing AMT.
Abstract: The recognition of human faces, especially those with
different orientations is a challenging and important problem in image
analysis and classification. This paper proposes an effective scheme
for rotation invariant face recognition using Log-Polar Transform and
Discrete Cosine Transform combined features. The rotation invariant
feature extraction for a given face image involves applying the logpolar
transform to eliminate the rotation effect and to produce a row
shifted log-polar image. The discrete cosine transform is then applied
to eliminate the row shift effect and to generate the low-dimensional
feature vector. A PSO-based feature selection algorithm is utilized to
search the feature vector space for the optimal feature subset.
Evolution is driven by a fitness function defined in terms of
maximizing the between-class separation (scatter index).
Experimental results, based on the ORL face database using testing
data sets for images with different orientations; show that the
proposed system outperforms other face recognition methods. The
overall recognition rate for the rotated test images being 97%,
demonstrating that the extracted feature vector is an effective rotation
invariant feature set with minimal set of selected features.
Abstract: An evolutionary method whose selection and recombination
operations are based on generalization error-bounds of
support vector machine (SVM) can select a subset of potentially
informative genes for SVM classifier very efficiently [7]. In this
paper, we will use the derivative of error-bound (first-order criteria)
to select and recombine gene features in the evolutionary process,
and compare the performance of the derivative of error-bound with
the error-bound itself (zero-order) in the evolutionary process. We
also investigate several error-bounds and their derivatives to compare
the performance, and find the best criteria for gene selection
and classification. We use 7 cancer-related human gene expression
datasets to evaluate the performance of the zero-order and first-order
criteria of error-bounds. Though both criteria have the same strategy
in theoretically, experimental results demonstrate the best criterion
for microarray gene expression data.
Abstract: The distinction among urban, periurban and rural areas represents a classical example of uncertainty in land classification. Satellite images, geostatistical analysis and all kinds of spatial data are very useful in urban sprawl studies, but it is important to define precise rules in combining great amounts of data to build complex knowledge about territory. Rough Set theory may be a useful method to employ in this field. It represents a different mathematical approach to uncertainty by capturing the indiscernibility. Two different phenomena can be indiscernible in some contexts and classified in the same way when combining available information about them. This approach has been applied in a case of study, comparing the results achieved with both Map Algebra technique and Spatial Rough Set. The study case area, Potenza Province, is particularly suitable for the application of this theory, because it includes 100 municipalities with different number of inhabitants and morphologic features.
Abstract: The design of a pattern classifier includes an attempt
to select, among a set of possible features, a minimum subset of
weakly correlated features that better discriminate the pattern classes.
This is usually a difficult task in practice, normally requiring the
application of heuristic knowledge about the specific problem
domain. The selection and quality of the features representing each
pattern have a considerable bearing on the success of subsequent
pattern classification. Feature extraction is the process of deriving
new features from the original features in order to reduce the cost of
feature measurement, increase classifier efficiency, and allow higher
classification accuracy. Many current feature extraction techniques
involve linear transformations of the original pattern vectors to new
vectors of lower dimensionality. While this is useful for data
visualization and increasing classification efficiency, it does not
necessarily reduce the number of features that must be measured
since each new feature may be a linear combination of all of the
features in the original pattern vector. In this paper a new approach is
presented to feature extraction in which feature selection, feature
extraction, and classifier training are performed simultaneously using
a genetic algorithm. In this approach each feature value is first
normalized by a linear equation, then scaled by the associated weight
prior to training, testing, and classification. A knn classifier is used to
evaluate each set of feature weights. The genetic algorithm optimizes
a vector of feature weights, which are used to scale the individual
features in the original pattern vectors in either a linear or a nonlinear
fashion. By this approach, the number of features used in classifying
can be finely reduced.
Abstract: This paper proposes a novel feature extraction method,
based on Discrete Wavelet Transform (DWT) and K-L Seperability
(KLS), for the classification of Functional Data (FD). This method
combines the decorrelation and reduction property of DWT and the
additive independence property of KLS, which is helpful to extraction
classification features of FD. It is an advanced approach of the
popular wavelet based shrinkage method for functional data reduction
and classification. A theory analysis is given in the paper to prove the
consistent convergence property, and a simulation study is also done
to compare the proposed method with the former shrinkage ones. The
experiment results show that this method has advantages in improving
classification efficiency, precision and robustness.
Abstract: Shadow detection is still considered as one of the
potential challenges for intelligent automated video surveillance
systems. A pre requisite for reliable and accurate detection and
tracking is the correct shadow detection and classification. In such a
landscape of conditions, privacy issues add more and more
complexity and require reliable shadow detection.
In this work the intertwining between security, accuracy,
reliability and privacy is analyzed and, accordingly, a novel
architecture for Privacy Enhancing Video Surveillance (PEVS) is
introduced. Shadow detection and masking are dealt with through the
combination of two different approaches simultaneously. This results
in a unique privacy enhancement, without affecting security.
Subsequently, the methodology was employed successfully in a
large-scale wireless video surveillance system; privacy relevant
information was stored and encrypted on the unit, without
transferring it over an un-trusted network.
Abstract: Mammography is the most effective procedure for an
early diagnosis of the breast cancer. Nowadays, people are trying to
find a way or method to support as much as possible to the
radiologists in diagnosis process. The most popular way is now being
developed is using Computer-Aided Detection (CAD) system to
process the digital mammograms and prompt the suspicious region to
radiologist. In this paper, an automated CAD system for detection
and classification of massive lesions in mammographic images is
presented. The system consists of three processing steps: Regions-Of-
Interest detection, feature extraction and classification. Our CAD
system was evaluated on Mini-MIAS database consisting 322
digitalized mammograms. The CAD system-s performance is
evaluated using Receiver Operating Characteristics (ROC) and Freeresponse
ROC (FROC) curves. The archived results are 3.47 false
positives per image (FPpI) and sensitivity of 85%.