Abstract: Breast cancer is one of the most frequent occurring cancers in women throughout the world including U.K. The grading of this cancer plays a vital role in the prognosis of the disease. In this paper we present an overview of the use of advanced computational method of fuzzy inference system as a tool for the automation of breast cancer grading. A new spectral data set obtained from Fourier Transform Infrared Spectroscopy (FTIR) of cancer patients has been used for this study. The future work outlines the potential areas of fuzzy systems that can be used for the automation of breast cancer grading.
Abstract: The issue of classifying objects into one of predefined
groups when the measured variables are mixed with different types
of variables has been part of interest among statisticians in many
years. Some methods for dealing with such situation have been
introduced that include parametric, semi-parametric and nonparametric
approaches. This paper attempts to discuss on a problem
in classifying a data when the number of measured mixed variables is
larger than the size of the sample. A propose idea that integrates a
dimensionality reduction technique via principal component analysis
and a discriminant function based on the location model is discussed.
The study aims in offering practitioners another potential tool in a
classification problem that is possible to be considered when the
observed variables are mixed and too large.
Abstract: A novel feature selection strategy to improve the recognition accuracy on the faces that are affected due to nonuniform illumination, partial occlusions and varying expressions is proposed in this paper. This technique is applicable especially in scenarios where the possibility of obtaining a reliable intra-class probability distribution is minimal due to fewer numbers of training samples. Phase congruency features in an image are defined as the points where the Fourier components of that image are maximally inphase. These features are invariant to brightness and contrast of the image under consideration. This property allows to achieve the goal of lighting invariant face recognition. Phase congruency maps of the training samples are generated and a novel modular feature selection strategy is implemented. Smaller sub regions from a predefined neighborhood within the phase congruency images of the training samples are merged to obtain a large set of features. These features are arranged in the order of increasing distance between the sub regions involved in merging. The assumption behind the proposed implementation of the region merging and arrangement strategy is that, local dependencies among the pixels are more important than global dependencies. The obtained feature sets are then arranged in the decreasing order of discriminating capability using a criterion function, which is the ratio of the between class variance to the within class variance of the sample set, in the PCA domain. The results indicate high improvement in the classification performance compared to baseline algorithms.
Abstract: The frontal area in the brain is known to be involved in
behavioral judgement. Because a Kanji character can be discriminated
visually and linguistically from other characters, in Kanji character
discrimination, we hypothesized that frontal event-related potential
(ERP) waveforms reflect two discrimination processes in separate
time periods: one based on visual analysis and the other based
on lexcical access. To examine this hypothesis, we recorded ERPs
while performing a Kanji lexical decision task. In this task, either a
known Kanji character, an unknown Kanji character or a symbol was
presented and the subject had to report if the presented character was
a known Kanji character for the subject or not. The same response
was required for unknown Kanji trials and symbol trials. As a preprocessing
of signals, we examined the performance of a method
using independent component analysis for artifact rejection and found
it was effective. Therefore we used it. In the ERP results, there
were two time periods in which the frontal ERP wavefoms were
significantly different betweeen the unknown Kanji trials and the
symbol trials: around 170ms and around 300ms after stimulus onset.
This result supported our hypothesis. In addition, the result suggests
that Kanji character lexical access may be fully completed by around
260ms after stimulus onset.
Abstract: Principle component analysis is often combined with
the state-of-art classification algorithms to recognize human faces.
However, principle component analysis can only capture these
features contributing to the global characteristics of data because it is a
global feature selection algorithm. It misses those features
contributing to the local characteristics of data because each principal
component only contains some levels of global characteristics of data.
In this study, we present a novel face recognition approach using
non-negative principal component analysis which is added with the
constraint of non-negative to improve data locality and contribute to
elucidating latent data structures. Experiments are performed on the
Cambridge ORL face database. We demonstrate the strong
performances of the algorithm in recognizing human faces in
comparison with PCA and NREMF approaches.
Abstract: Rotation or tilt present in an image capture by digital
means can be detected and corrected using Artificial Neural Network
(ANN) for application with a Face Recognition System (FRS). Principal
Component Analysis (PCA) features of faces at different angles
are used to train an ANN which detects the rotation for an input image
and corrected using a set of operations implemented using another
system based on ANN. The work also deals with the recognition
of human faces with features from the foreheads, eyes, nose and
mouths as decision support entities of the system configured using
a Generalized Feed Forward Artificial Neural Network (GFFANN).
These features are combined to provide a reinforced decision for
verification of a person-s identity despite illumination variations. The
complete system performing facial image rotation detection, correction
and recognition using re-enforced decision support provides a
success rate in the higher 90s.
Abstract: This paper presents a comparative analysis of a new
unsupervised PCA-based technique for steel plates texture segmentation
towards defect detection. The proposed scheme called Variance
Based Component Analysis or VBCA employs PCA for feature
extraction, applies a feature reduction algorithm based on variance of
eigenpictures and classifies the pixels as defective and normal. While
the classic PCA uses a clusterer like Kmeans for pixel clustering,
VBCA employs thresholding and some post processing operations to
label pixels as defective and normal. The experimental results show
that proposed algorithm called VBCA is 12.46% more accurate and
78.85% faster than the classic PCA.
Abstract: A clustering is process to identify a homogeneous
groups of object called as cluster. Clustering is one interesting topic
on data mining. A group or class behaves similarly characteristics.
This paper discusses a robust clustering process for data images with
two reduction dimension approaches; i.e. the two dimensional
principal component analysis (2DPCA) and principal component
analysis (PCA). A standard approach to overcome this problem is
dimension reduction, which transforms a high-dimensional data into
a lower-dimensional space with limited loss of information. One of
the most common forms of dimensionality reduction is the principal
components analysis (PCA). The 2DPCA is often called a variant of
principal component (PCA), the image matrices were directly treated
as 2D matrices; they do not need to be transformed into a vector so
that the covariance matrix of image can be constructed directly using
the original image matrices. The decomposed classical covariance
matrix is very sensitive to outlying observations. The objective of
paper is to compare the performance of robust minimizing vector
variance (MVV) in the two dimensional projection PCA (2DPCA)
and the PCA for clustering on an arbitrary data image when outliers
are hiden in the data set. The simulation aspects of robustness and
the illustration of clustering images are discussed in the end of
paper
Abstract: In this paper, taking Chinese Nanjing Metro ALSTOM vehicle synthesis brake as an example, the subway with synthetic brake shoe formula components of final product performance, has done a lot of research and performance test, final is drawn with hybrid fiber as reinforcing material, modified phenolic resin as matrix material, and then filling friction modifier performance, by the hot pressing process made a new type of domestic subway brake shoe. The product of the test performance indicators that can replace the similar foreign products.
Abstract: The effects of irrigation with dairy factory wastewater on soil properties were investigated at two sites that had received irrigation for > 60 years. Two adjoining paired sites that had never received DFE were also sampled as well as another seven fields from a wider area around the factory. In comparison with paired sites that had not received effluent, long-term wastewater irrigation resulted in an increase in pH, EC, extractable P, exchangeable Na and K and ESP. These changes were related to the use of phosphoric acid, NaOH and KOH as cleaning agents in the factory. Soil organic C content was unaffected by DFE irrigation but the size (microbial biomass C and N) and activity (basal respiration) of the soil microbial community were increased. These increases were attributed to regular inputs of soluble C (e.g. lactose) present as milk residues in the wastewater. Principal component analysis (PCA) of the soils data from all 11sites confirmed that the main effects of DFE irrigation were an increase in exchangeable Na, extractable P and microbial biomass C, an accumulation of soluble salts and a liming effect. PCA analysis of soil bacterial community structure, using PCR-DGGE of 16S rDNA fragments, generally separated individual sites from one another but did not group them according to irrigation history. Thus, whilst the size and activity of the soil microbial community were increased, the structure and diversity of the bacterial community remained unaffected.
Abstract: Diabetes Mellitus is a chronic metabolic disorder, where the improper management of the blood glucose level in the diabetic patients will lead to the risk of heart attack, kidney disease and renal failure. This paper attempts to enhance the diagnostic accuracy of the advancing blood glucose levels of the diabetic patients, by combining principal component analysis and wavelet neural network. The proposed system makes separate blood glucose prediction in the morning, afternoon, evening and night intervals, using dataset from one patient covering a period of 77 days. Comparisons of the diagnostic accuracy with other neural network models, which use the same dataset are made. The comparison results showed overall improved accuracy, which indicates the effectiveness of this proposed system.
Abstract: The automatic classification of non stationary signals is an important practical goal in several domains. An essential classification task is to allocate the incoming signal to a group associated with the kind of physical phenomena producing it. In this paper, we present a modular system composed by three blocs: 1) Representation, 2) Dimensionality reduction and 3) Classification. The originality of our work consists in the use of a new wavelet called "Ben wavelet" in the representation stage. For the dimensionality reduction, we propose a new algorithm based on the random projection and the principal component analysis.
Abstract: Based on the one-bit-matching principle and by turning the de-mixing matrix into an orthogonal matrix via certain normalization, Ma et al proposed a one-bit-matching learning algorithm on the Stiefel manifold for independent component analysis [8]. But this algorithm is not adaptive. In this paper, an algorithm which can extract kurtosis and its sign of each independent source component directly from observation data is firstly introduced.With the algorithm , the one-bit-matching learning algorithm is revised, so that it can make the blind separation on the Stiefel manifold implemented completely in the adaptive mode in the framework of natural gradient.
Abstract: In this paper, a new face recognition method based on
PCA (principal Component Analysis), LDA (Linear Discriminant
Analysis) and neural networks is proposed. This method consists of
four steps: i) Preprocessing, ii) Dimension reduction using PCA, iii)
feature extraction using LDA and iv) classification using neural
network. Combination of PCA and LDA is used for improving the
capability of LDA when a few samples of images are available and
neural classifier is used to reduce number misclassification caused by
not-linearly separable classes. The proposed method was tested on
Yale face database. Experimental results on this database
demonstrated the effectiveness of the proposed method for face
recognition with less misclassification in comparison with previous
methods.
Abstract: Analysis and visualization of microarraydata is veryassistantfor biologists and clinicians in the field of diagnosis and treatment of patients. It allows Clinicians to better understand the structure of microarray and facilitates understanding gene expression in cells. However, microarray dataset is a complex data set and has thousands of features and a very small number of observations. This very high dimensional data set often contains some noise, non-useful information and a small number of relevant features for disease or genotype. This paper proposes a non-linear dimensionality reduction algorithm Local Principal Component (LPC) which aims to maps high dimensional data to a lower dimensional space. The reduced data represents the most important variables underlying the original data. Experimental results and comparisons are presented to show the quality of the proposed algorithm. Moreover, experiments also show how this algorithm reduces high dimensional data whilst preserving the neighbourhoods of the points in the low dimensional space as in the high dimensional space.
Abstract: Frequency domain independent component analysis has
a scaling indeterminacy and a permutation problem. The scaling
indeterminacy can be solved by use of a decomposed spectrum. For
the permutation problem, we have proposed the rules in terms of gain
ratio and phase difference derived from the decomposed spectra and
the source-s coarse directions.
The present paper experimentally clarifies that the gain ratio and
the phase difference work effectively in a real environment but their
performance depends on frequency bands, a microphone-space and
a source-microphone distance. From these facts it is seen that it is
difficult to attain a perfect solution for the permutation problem in a
real environment only by either the gain ratio or the phase difference.
For the perfect solution, this paper gives a solution to the problems
in a real environment. The proposed method is simple, the amount of
calculation is small. And the method has high correction performance
without depending on the frequency bands and distances from source
signals to microphones. Furthermore, it can be applied under the real
environment. From several experiments in a real room, it clarifies
that the proposed method has been verified.
Abstract: The Principal component regression (PCR) is a
combination of principal component analysis (PCA) and multiple linear regression (MLR). The objective of this paper is to revise the
use of PCR in shortwave near infrared (SWNIR) (750-1000nm) spectral analysis. The idea of PCR was explained mathematically and
implemented in the non-destructive assessment of the soluble solid
content (SSC) of pineapple based on SWNIR spectral data. PCR achieved satisfactory results in this application with root mean
squared error of calibration (RMSEC) of 0.7611 Brix°, coefficient of determination (R2) of 0.5865 and root mean squared error of crossvalidation
(RMSECV) of 0.8323 Brix° with principal components
(PCs) of 14.
Abstract: The performance results of the athletes competed in
the 1988-2008 Olympic Games were analyzed (n = 166). The data
were obtained from the IAAF official protocols. In the principal
component analysis, the first three principal components explained
70% of the total variance. In the 1st principal component (with
43.1% of total variance explained) the largest factor loadings were
for 100m (0.89), 400m (0.81), 110m hurdle run (0.76), and long jump
(–0.72). This factor can be interpreted as the 'sprinting performance'.
The loadings on the 2nd factor (15.3% of the total variance)
presented a counter-intuitive throwing-jumping combination: the
highest loadings were for throwing events (javelin throwing 0.76;
shot put 0.74; and discus throwing 0.73) and also for jumping events
(high jump 0.62; pole vaulting 0.58). On the 3rd factor (11.6% of
total variance), the largest loading was for 1500 m running (0.88); all
other loadings were below 0.4.
Abstract: ''Cocktail party problem'' is well known as one of the human auditory abilities. We can recognize the specific sound that we want to listen by this ability even if a lot of undesirable sounds or noises are mixed. Blind source separation (BSS) based on independent component analysis (ICA) is one of the methods by which we can separate only a special signal from their mixed signals with simple hypothesis. In this paper, we propose an online approach for blind source separation using the sliding DFT and the time domain independent component analysis. The proposed method can reduce calculation complexity in comparison with conventional methods, and can be applied to parallel processing by using digital signal processors (DSPs) and so on. We evaluate this method and show its availability.
Abstract: Independent component analysis can estimate unknown
source signals from their mixtures under the assumption that the
source signals are statistically independent. However, in a real environment,
the separation performance is often deteriorated because
the number of the source signals is different from that of the sensors.
In this paper, we propose an estimation method for the number of
the sources based on the joint distribution of the observed signals
under two-sensor configuration. From several simulation results, it
is found that the number of the sources is coincident to that of
peaks in the histogram of the distribution. The proposed method can
estimate the number of the sources even if it is larger than that of
the observed signals. The proposed methods have been verified by
several experiments.