Abstract: Since primary school trips usually start from home,
attention by many scholars have been focused on the home end for
data gathering. Thereafter category analysis has often been relied
upon when predicting school travel demands. In this paper, school
end was relied on for data gathering and multivariate regression for
future travel demand prediction. 9859 pupils were surveyed by way
of questionnaires at 21 primary schools. The town was divided into 5
zones. The study was carried out in Skudai Town, Malaysia. Based
on the hypothesis that the number of primary school trip ends are
expected to be the same because school trips are fixed, the choice of
trip end would have inconsequential effect on the outcome. The
study compared empirical data for home and school trip end
productions and attractions. Variance from both data results was
insignificant, although some claims from home based family survey
were found to be grossly exaggerated. Data from the school trip ends
was relied on for travel demand prediction because of its
completeness. Accessibility, trip attraction and trip production were
then related to school trip rates under daylight and dry weather
conditions. The paper concluded that, accessibility is an important
parameter when predicting demand for future school trip rates.
Abstract: The design of a pattern classifier includes an attempt
to select, among a set of possible features, a minimum subset of
weakly correlated features that better discriminate the pattern classes.
This is usually a difficult task in practice, normally requiring the
application of heuristic knowledge about the specific problem
domain. The selection and quality of the features representing each
pattern have a considerable bearing on the success of subsequent
pattern classification. Feature extraction is the process of deriving
new features from the original features in order to reduce the cost of
feature measurement, increase classifier efficiency, and allow higher
classification accuracy. Many current feature extraction techniques
involve linear transformations of the original pattern vectors to new
vectors of lower dimensionality. While this is useful for data
visualization and increasing classification efficiency, it does not
necessarily reduce the number of features that must be measured
since each new feature may be a linear combination of all of the
features in the original pattern vector. In this paper a new approach is
presented to feature extraction in which feature selection, feature
extraction, and classifier training are performed simultaneously using
a genetic algorithm. In this approach each feature value is first
normalized by a linear equation, then scaled by the associated weight
prior to training, testing, and classification. A knn classifier is used to
evaluate each set of feature weights. The genetic algorithm optimizes
a vector of feature weights, which are used to scale the individual
features in the original pattern vectors in either a linear or a nonlinear
fashion. By this approach, the number of features used in classifying
can be finely reduced.
Abstract: A new approach based on the consideration that electroencephalogram (EEG) signals are chaotic signals was presented for automated diagnosis of electroencephalographic changes. This consideration was tested successfully using the nonlinear dynamics tools, like the computation of Lyapunov exponents. This paper presented the usage of statistics over the set of the Lyapunov exponents in order to reduce the dimensionality of the extracted feature vectors. Since classification is more accurate when the pattern is simplified through representation by important features, feature extraction and selection play an important role in classifying systems such as neural networks. Multilayer perceptron neural network (MLPNN) architectures were formulated and used as basis for detection of electroencephalographic changes. Three types of EEG signals (EEG signals recorded from healthy volunteers with eyes open, epilepsy patients in the epileptogenic zone during a seizure-free interval, and epilepsy patients during epileptic seizures) were classified. The selected Lyapunov exponents of the EEG signals were used as inputs of the MLPNN trained with Levenberg- Marquardt algorithm. The classification results confirmed that the proposed MLPNN has potential in detecting the electroencephalographic changes.
Abstract: The recovery of metal values and safe disposal of
spent catalyst is gaining interest due to both its hazardous nature and
increased regulation associated with disposal methods. Prior to the
recovery of the valuable metals, removal of entrained deposits limit
the diffusion of lixiviate resulting in low recovery of metals must be
taken into consideration. Therefore, petroleum refinery spent catalyst
was subjected to acetone washing and roasting at 500oC. The treated
samples were investigated for metals bioleaching using
Acidithiobacillus ferrooxidans in batch reactors and the leaching
efficiencies were compared. It was found out that acetone washed
spent catalysts results in better metal recovery compare to roasted
spent. About 83% Ni, 20% Al, 50% Mo and 73% V were leached
using the acetone washed spent catalyst. In both the cases, Ni, V and
Mo was high compared to Al.
Abstract: This paper makes an attempt to solve the problem of
searching and retrieving of similar MRI photos via Internet services
using morphological features which are sourced via the original
image. This study is aiming to be considered as an additional tool of
searching and retrieve methods. Until now the main way of the
searching mechanism is based on the syntactic way using keywords.
The technique it proposes aims to serve the new requirements of
libraries. One of these is the development of computational tools for
the control and preservation of the intellectual property of digital
objects, and especially of digital images. For this purpose, this paper
proposes the use of a serial number extracted by using a previously
tested semantic properties method. This method, with its center being
the multi-layers of a set of arithmetic points, assures the following
two properties: the uniqueness of the final extracted number and the
semantic dependence of this number on the image used as the
method-s input. The major advantage of this method is that it can
control the authentication of a published image or its partial
modification to a reliable degree. Also, it acquires the better of the
known Hash functions that the digital signature schemes use and
produces alphanumeric strings for cases of authentication checking,
and the degree of similarity between an unknown image and an
original image.
Abstract: Using a force balanced translational-radial dynamics,
phase space of the moving single bubble sonoluminescence (m-
SBSL) in 85% wt sulfuric acid has been numerically calculated. This
phase space is compared with that of single bubble sonoluminescence
(SBSL) in pure water which has been calculated by using the mere
radial dynamics. It is shown that in 85% wt sulfuric acid, in a general
agreement with experiment, the bubble-s positional instability
threshold lays under the shape instability threshold. At the onset of
spatial instability of moving sonoluminescing (SL) bubble in 85%
wt sulfuric acid, temporal effects of the hydrodynamic force on the
bubble translational-radial dynamics have been investigated. The
appearance of non-zero history force on the moving SL bubble is
because of proper condition which was produced by high viscosity of
acid. Around the moving bubble collapse due to the rapid contraction
of the bubble wall, the inertial based added mass force overcomes the
viscous based history force and induces acceleration on the bubble
translational motion.
Abstract: Subcritical water extraction was investigated as a
novel and alternative technology in the food and pharmaceutical
industry for the separation of Mannitol from olive leaves and its
results was compared with those of Soxhlet extraction. The effects of
temperature, pressure, and flow rate of water and also momentum
and mass transfer dimensionless variables such as Reynolds and
Peclet Numbers on extraction yield and equilibrium partition
coefficient were investigated. The 30-110 bars, 60-150°C, and flow
rates of 0.2-2 mL/min were the water operating conditions. The
results revealed that the highest Mannitol yield was obtained at
100°C and 50 bars. However, extraction of Mannitol was not
influenced by the variations of flow rate. The mathematical modeling
of experimental measurements was also investigated and the model is
capable of predicting the experimental measurements very well. In
addition, the results indicated higher extraction yield for the
subcritical water extraction in contrast to Soxhlet method.
Abstract: An image texture analysis and target recognition approach of using an improved image texture feature coding method (TFCM) and Support Vector Machine (SVM) for target detection is presented. With our proposed target detection framework, targets of interest can be detected accurately. Cascade-Sliding-Window technique was also developed for automated target localization. Application to mammogram showed that over 88% of normal mammograms and 80% of abnormal mammograms can be correctly identified. The approach was also successfully applied to Synthetic Aperture Radar (SAR) and Ground Penetrating Radar (GPR) images for target detection.
Abstract: Text Mining is an important step of Knowledge
Discovery process. It is used to extract hidden information from notstructured
o semi-structured data. This aspect is fundamental because
much of the Web information is semi-structured due to the nested
structure of HTML code, much of the Web information is linked,
much of the Web information is redundant. Web Text Mining helps
whole knowledge mining process to mining, extraction and
integration of useful data, information and knowledge from Web
page contents.
In this paper, we present a Web Text Mining process able to
discover knowledge in a distributed and heterogeneous multiorganization
environment. The Web Text Mining process is based on
flexible architecture and is implemented by four steps able to
examine web content and to extract useful hidden information
through mining techniques. Our Web Text Mining prototype starts
from the recovery of Web job offers in which, through a Text Mining
process, useful information for fast classification of the same are
drawn out, these information are, essentially, job offer place and
skills.
Abstract: This work is focused on the numerical prediction of the fracture resistance of a flat stiffened panel made of the aluminium alloy 2024 T3 under a monotonic traction condition. The performed numerical simulations have been based on the micromechanical Gurson-Tvergaard (GT) model for ductile damage. The applicability of the GT model to this kind of structural problems has been studied and assessed by comparing numerical results, obtained by using the WARP 3D finite element code, with experimental data available in literature. In the sequel a home-made procedure is presented, which aims to increase the residual strength of a cracked stiffened aluminum panel and which is based on the stochastic design improvement (SDI) technique; a whole application example is then given to illustrate the said technique.
Abstract: This paper describes a novel and effective approach to content-based image retrieval (CBIR) that represents each image in the database by a vector of feature values called “Standard deviation of mean vectors of color distribution of rows and columns of images for CBIR". In many areas of commerce, government, academia, and hospitals, large collections of digital images are being created. This paper describes the approach that uses contents as feature vector for retrieval of similar images. There are several classes of features that are used to specify queries: colour, texture, shape, spatial layout. Colour features are often easily obtained directly from the pixel intensities. In this paper feature extraction is done for the texture descriptor that is 'variance' and 'Variance of Variances'. First standard deviation of each row and column mean is calculated for R, G, and B planes. These six values are obtained for one image which acts as a feature vector. Secondly we calculate variance of the row and column of R, G and B planes of an image. Then six standard deviations of these variance sequences are calculated to form a feature vector of dimension six. We applied our approach to a database of 300 BMP images. We have determined the capability of automatic indexing by analyzing image content: color and texture as features and by applying a similarity measure Euclidean distance.
Abstract: Data mining is an extraordinarily demanding field referring to extraction of implicit knowledge and relationships, which are not explicitly stored in databases. A wide variety of methods of data mining have been introduced (classification, characterization, generalization...). Each one of these methods includes more than algorithm. A system of data mining implies different user categories,, which mean that the user-s behavior must be a component of the system. The problem at this level is to know which algorithm of which method to employ for an exploratory end, which one for a decisional end, and how can they collaborate and communicate. Agent paradigm presents a new way of conception and realizing of data mining system. The purpose is to combine different algorithms of data mining to prepare elements for decision-makers, benefiting from the possibilities offered by the multi-agent systems. In this paper the agent framework for data mining is introduced, and its overall architecture and functionality are presented. The validation is made on spatial data. Principal results will be presented.
Abstract: This paper presents a novel method for data hiding based on neighborhood pixels information to calculate the number of bits that can be used for substitution and modified Least Significant Bits technique for data embedding. The modified solution is independent of the nature of the data to be hidden and gives correct results along with un-noticeable image degradation. The technique, to find the number of bits that can be used for data hiding, uses the green component of the image as it is less sensitive to human eye and thus it is totally impossible for human eye to predict whether the image is encrypted or not. The application further encrypts the data using a custom designed algorithm before embedding bits into image for further security. The overall process consists of three main modules namely embedding, encryption and extraction cm.
Abstract: Intelligence tests are series of tasks designed to measure the capacity to make abstractions, to learn, and to deal with novel situations. Testing of the visual abilities of the shape understanding system (SUS) is performed based on the visual intelligence tests. In this paper the progressive matrices tests are formulated as tasks given to SUS. These tests require good visual problem solving abilities of the human subject. SUS solves these tests by performing complex visual reasoning transforming the visual forms (tests) into the string forms. The experiment proved that the proposed method, which is part of the SUS visual understanding abilities, can solve a test that is very difficult for human subject.
Abstract: The Automatic Speech Recognition (ASR) applied to
Arabic language is a challenging task. This is mainly related to the
language specificities which make the researchers facing multiple
difficulties such as the insufficient linguistic resources and the very
limited number of available transcribed Arabic speech corpora. In
this paper, we are interested in the development of a HMM-based
ASR system for Standard Arabic (SA) language. Our fundamental
research goal is to select the most appropriate acoustic parameters
describing each audio frame, acoustic models and speech recognition
unit. To achieve this purpose, we analyze the effect of varying frame
windowing (size and period), acoustic parameter number resulting
from features extraction methods traditionally used in ASR, speech
recognition unit, Gaussian number per HMM state and number of
embedded re-estimations of the Baum-Welch Algorithm. To evaluate
the proposed ASR system, a multi-speaker SA connected-digits
corpus is collected, transcribed and used throughout all experiments.
A further evaluation is conducted on a speaker-independent continue
SA speech corpus. The phonemes recognition rate is 94.02% which is
relatively high when comparing it with another ASR system
evaluated on the same corpus.
Abstract: This paper proposes a novel feature extraction method,
based on Discrete Wavelet Transform (DWT) and K-L Seperability
(KLS), for the classification of Functional Data (FD). This method
combines the decorrelation and reduction property of DWT and the
additive independence property of KLS, which is helpful to extraction
classification features of FD. It is an advanced approach of the
popular wavelet based shrinkage method for functional data reduction
and classification. A theory analysis is given in the paper to prove the
consistent convergence property, and a simulation study is also done
to compare the proposed method with the former shrinkage ones. The
experiment results show that this method has advantages in improving
classification efficiency, precision and robustness.
Abstract: Utilization of bagasse ash for silica sources is one of
the most common application for agricultural wastes and valuable
biomass byproducts in sugar milling. The high percentage silica
content from bagasse ash was used as silica source for sodium
silicate solution. Different heating temperature, time and acid
treatment were studies for silica extraction. The silica was
characterized using various techniques including X-ray fluorescence,
X-ray diffraction, Scanning electron microscopy, and Fourier
Transform Infrared Spectroscopy method,. The synthesis conditions
were optimized to obtain the bagasse ash with the maximum silica
content. The silica content of 91.57 percent was achieved from
heating of bagasse ash at 600°C for 3 hours under oxygen feeding
and HCl treatment. The result can be used as value added for bagasse
ash utilization and minimize the environmental impact of disposal
problems.
Abstract: In historical science and social science, the influence
of natural disaster upon society is a matter of great interest. In
recent years, some archives are made through many hands for natural
disasters, however it is inefficiency and waste. So, we suppose a
computer system to create a historical natural disaster archive. As
the target of this analysis, we consider newspaper articles. The news
articles are considered to be typical examples that prescribe the
temporal relations of affairs for natural disaster. In order to do this
analysis, we identify the occurrences in newspaper articles by some
index entries, considering the affairs which are specific to natural
disasters, and show the temporal relation between natural disasters.
We designed and implemented the automatic system of “extraction
of the occurrences of natural disaster" and “temporal relation table
for natural disaster."
Abstract: We introduce an effective approach for automatic offline au- thentication of handwritten samples where the forgeries are skillfully done, i.e., the true and forgery sample appearances are almost alike. Subtle details of temporal information used in online verification are not available offline and are also hard to recover robustly. Thus the spatial dynamic information like the pen-tip pressure characteristics are considered, emphasizing on the extraction of low density pixels. The points result from the ballistic rhythm of a genuine signature which a forgery, however skillful that may be, always lacks. Ten effective features, including these low density points and den- sity ratio, are proposed to make the distinction between a true and a forgery sample. An adaptive decision criteria is also derived for better verification judgements.
Abstract: In this paper we present an enhanced noise reduction method for robust speech recognition using Adaptive Gain Equalizer with Non linear Spectral Subtraction. In Adaptive Gain Equalizer method (AGE), the input signal is divided into a number of subbands that are individually weighed in time domain, in accordance to the short time Signal-to-Noise Ratio (SNR) in each subband estimation at every time instant. Instead of focusing on suppression the noise on speech enhancement is focused. When analysis was done under various noise conditions for speech recognition, it was found that Adaptive Gain Equalizer method algorithm has an obvious failing point for a SNR of -5 dB, with inadequate levels of noise suppression for SNR less than this point. This work proposes the implementation of AGE when coupled with Non linear Spectral Subtraction (AGE-NSS) for robust speech recognition. The experimental result shows that out AGE-NSS performs the AGE when SNR drops below -5db level.