Abstract: The problem of spam has been seriously troubling the Internet community during the last few years and currently reached an alarming scale. Observations made at CERN (European Organization for Nuclear Research located in Geneva, Switzerland) show that spam mails can constitute up to 75% of daily SMTP traffic. A naïve Bayesian classifier based on a Bag Of Words representation of an email is widely used to stop this unwanted flood as it combines good performance with simplicity of the training and classification processes. However, facing the constantly changing patterns of spam, it is necessary to assure online adaptability of the classifier. This work proposes combining such a classifier with another NBC (naïve Bayesian classifier) based on pairs of adjacent words. Only the latter will be retrained with examples of spam reported by users. Tests are performed on considerable sets of mails both from public spam archives and CERN mailboxes. They suggest that this architecture can increase spam recall without affecting the classifier precision as it happens when only the NBC based on single words is retrained.
Abstract: With high speed vessels getting ever more sophisti-cated, travelling at higher and higher speeds and operating in With high speed vessels getting ever more sophisticated,
travelling at higher and higher speeds and operating in areas of
high maritime traffic density, training becomes of the highest priority
to ensure that safety levels are maintained, and risks are adequately
mitigated. Training onboard the actual craft on the actual route still
remains the most effective way for crews to gain experience. However,
operational experience and incidents during the last 10 years
demonstrate the need for supplementary training whether in the area
of simulation or man to man, man/ machine interaction. Training and
familiarisation of the crew is the most important aspect in preventing
incidents. The use of simulator, computer and web based training
systems in conjunction with onboard training focusing on critical
situations will improve the man machine interaction and thereby
reduce the risk of accidents. Today, both ship simulator and bridge
teamwork courses are now becoming the norm in order to improve
further emergency response and crisis management skills. One of the
main causes of accidents is the human factor. An efficient way to
reduce human errors is to provide high-quality training to the personnel
and to select the navigators carefully.areas of high maritime traffic density, training becomes of the highest priority to ensure that safety levels are maintained, and risks are adequately mitigated. Training onboard the actual craft on the actual route still remains the most effective way for crews to gain experience. How-ever, operational experience and incidents during the last 10 years demonstrate the need for supplementary training whether in the area of simulation or man to man, man/ machine interaction. Training and familiarisation of the crew is the most important aspect in preventing incidents. The use of simulator, computer and web based training systems in conjunction with onboard training focusing on critical situations will improve the man machine interaction and thereby reduce the risk of accidents. Today, both ship simulator and bridge teamwork courses are now becoming the norm in order to improve further emergency response and crisis management skills. One of the main causes of accidents is the human factor. An efficient way to reduce human errors is to provide high-quality training to the person-nel and to select the navigators carefully. KeywordsCBT - WBT systems, Human factors.
Abstract: Hybrid algorithm is the hot issue in Computational
Intelligence (CI) study. From in-depth discussion on Simulation
Mechanism Based (SMB) classification method and composite patterns,
this paper presents the Mamdani model based Adaptive Neural
Fuzzy Inference System (M-ANFIS) and weight updating formula in
consideration with qualitative representation of inference consequent
parts in fuzzy neural networks. M-ANFIS model adopts Mamdani
fuzzy inference system which has advantages in consequent part.
Experiment results of applying M-ANFIS to evaluate traffic Level
of service show that M-ANFIS, as a new hybrid algorithm in computational
intelligence, has great advantages in non-linear modeling,
membership functions in consequent parts, scale of training data and
amount of adjusted parameters.
Abstract: In this paper we present an approach for 3D face
recognition based on extracting principal components of range
images by utilizing modified PCA methods namely 2DPCA and
bidirectional 2DPCA also known as (2D) 2 PCA.A preprocessing
stage was implemented on the images to smooth them using median
and Gaussian filtering. In the normalization stage we locate the nose
tip to lay it at the center of images then crop each image to a standard
size of 100*100. In the face recognition stage we extract the principal
component of each image using both 2DPCA and (2D) 2 PCA.
Finally, we use Euclidean distance to measure the minimum distance
between a given test image to the training images in the database. We
also compare the result of using both methods. The best result
achieved by experiments on a public face database shows that 83.3
percent is the rate of face recognition for a random facial expression.
Abstract: According to the governmental data, the cases of oral
cancers doubled in the past 10 years. This had brought heavy burden to
the patients- family, the society, and the country. The literature
generally evidenced the betel nut contained particular chemicals that
can cause oral cancers. Research in Taiwan had also proofed that 90
percent of oral cancer patients had experience of betel nut chewing. It
is thus important to educate the betel-nut hobbyists to cease such a
hazardous behavior. A program was then organized to establish
several training classes across different areas specific to help ceasing
this particular habit. Purpose of this research was to explore the
attitude and intention toward ceasing betel-nut chewing before and
after attending the training classes. 50 samples were taken from a
ceasing class with average age at 45 years old with high school
education (54%). 74% of the respondents were male in service or
agricultural industries. Experiences in betel-nut chewing were 5-20
years with a dose of 1-20 pieces per day. The data had shown that 60%
of the respondents had cigarette smoking habit, and 30% of the
respondents were concurrently alcoholic dependent. Research results
indicated that the attitude, intentions, and the knowledge on oral
cancers were found significant different between before and after
attendance. This provided evidence for the effectiveness of the training
class. However, we do not perform follow-up after the class.
Noteworthy is the test result also shown that participants who were
drivers as occupation, or habitual smokers or alcoholic dependents
would be less willing to quit the betel-nut chewing. The test results
indicated as well that the educational levels and the type of occupation
may have significant impacts on an individual-s decisions in taking
betel-nut or substance abuse.
Abstract: Conventionally the selection of parameters depends
intensely on the operator-s experience or conservative technological
data provided by the EDM equipment manufacturers that assign
inconsistent machining performance. The parameter settings given by
the manufacturers are only relevant with common steel grades. A
single parameter change influences the process in a complex way.
Hence, the present research proposes artificial neural network (ANN)
models for the prediction of surface roughness on first commenced
Ti-15-3 alloy in electrical discharge machining (EDM) process. The
proposed models use peak current, pulse on time, pulse off time and
servo voltage as input parameters. Multilayer perceptron (MLP) with
three hidden layer feedforward networks are applied. An assessment
is carried out with the models of distinct hidden layer. Training of the
models is performed with data from an extensive series of
experiments utilizing copper electrode as positive polarity. The
predictions based on the above developed models have been verified
with another set of experiments and are found to be in good
agreement with the experimental results. Beside this they can be
exercised as precious tools for the process planning for EDM.
Abstract: In this paper back-propagation artificial neural network
(BPANN) is employed to predict the deformation of the upsetting
process. To prepare a training set for BPANN, some finite element
simulations were carried out. The input data for the artificial neural
network are a set of parameters generated randomly (aspect ratio d/h,
material properties, temperature and coefficient of friction). The
output data are the coefficient of polynomial that fitted on barreling
curves. Neural network was trained using barreling curves generated
by finite element simulations of the upsetting and the corresponding
material parameters. This technique was tested for three different
specimens and can be successfully employed to predict the
deformation of the upsetting process
Abstract: Heterogeneity of solid waste characteristics as well as the complex processes taking place within the landfill ecosystem motivated the implementation of soft computing methodologies such as artificial neural networks (ANN), fuzzy logic (FL), and their combination. The present work uses a hybrid ANN-FL model that employs knowledge-based FL to describe the process qualitatively and implements the learning algorithm of ANN to optimize model parameters. The model was developed to simulate and predict the landfill gas production at a given time based on operational parameters. The experimental data used were compiled from lab-scale experiment that involved various operating scenarios. The developed model was validated and statistically analyzed using F-test, linear regression between actual and predicted data, and mean squared error measures. Overall, the simulated landfill gas production rates demonstrated reasonable agreement with actual data. The discussion focused on the effect of the size of training datasets and number of training epochs.
Abstract: In this paper we designed and implemented a new
ensemble of classifiers based on a sequence of classifiers which were
specialized in regions of the training dataset where errors of its
trained homologous are concentrated. In order to separate this
regions, and to determine the aptitude of each classifier to properly
respond to a new case, it was used another set of classifiers built
hierarchically. We explored a selection based variant to combine the
base classifiers. We validated this model with different base
classifiers using 37 training datasets. It was carried out a statistical
comparison of these models with the well known Bagging and
Boosting, obtaining significantly superior results with the
hierarchical ensemble using Multilayer Perceptron as base classifier.
Therefore, we demonstrated the efficacy of the proposed ensemble,
as well as its applicability to general problems.
Abstract: The present study focuses on the discussion over the
parameter of Artificial Neural Network (ANN). Sensitivity analysis is
applied to assess the effect of the parameters of ANN on the prediction
of turbidity of raw water in the water treatment plant. The result shows
that transfer function of hidden layer is a critical parameter of ANN.
When the transfer function changes, the reliability of prediction of
water turbidity is greatly different. Moreover, the estimated water
turbidity is less sensitive to training times and learning velocity than
the number of neurons in the hidden layer. Therefore, it is important to
select an appropriate transfer function and suitable number of neurons
in the hidden layer in the process of parameter training and validation.
Abstract: Distant-talking voice-based HCI system suffers from
performance degradation due to mismatch between the acoustic
speech (runtime) and the acoustic model (training). Mismatch is
caused by the change in the power of the speech signal as observed at
the microphones. This change is greatly influenced by the change in
distance, affecting speech dynamics inside the room before reaching
the microphones. Moreover, as the speech signal is reflected, its
acoustical characteristic is also altered by the room properties. In
general, power mismatch due to distance is a complex problem. This
paper presents a novel approach in dealing with distance-induced
mismatch by intelligently sensing instantaneous voice power variation
and compensating model parameters. First, the distant-talking speech
signal is processed through microphone array processing, and the
corresponding distance information is extracted. Distance-sensitive
Gaussian Mixture Models (GMMs), pre-trained to capture both
speech power and room property are used to predict the optimal
distance of the speech source. Consequently, pre-computed statistic
priors corresponding to the optimal distance is selected to correct
the statistics of the generic model which was frozen during training.
Thus, model combinatorics are post-conditioned to match the power
of instantaneous speech acoustics at runtime. This results to an
improved likelihood in predicting the correct speech command at
farther distances. We experiment using real data recorded inside two
rooms. Experimental evaluation shows voice recognition performance
using our method is more robust to the change in distance compared
to the conventional approach. In our experiment, under the most
acoustically challenging environment (i.e., Room 2: 2.5 meters), our
method achieved 24.2% improvement in recognition performance
against the best-performing conventional method.
Abstract: One of the essential requirements of a realistic
surgical simulator is to reproduce haptic sensations due to the
interactions in the virtual environment. However, the interaction need
to be performed in real-time, since a delay between the user action
and the system reaction reduces the immersion sensation. In this
paper, a prototype of a coronary stent implant simulator is present;
this system allows real-time interactions with an artery by means of a
specific haptic device. To improve the realism of the simulation, the
building of the virtual environment is based on real patients- images
and a Web Portal is used to search in the geographically remote
medical centres a virtual environment with specific features in terms
of pathology or anatomy. The functional architecture of the system
defines several Medical Centres in which virtual environments built
from the real patients- images and related metadata with specific
features in terms of pathology or anatomy are stored. The searched
data are downloaded from the Medical Centre to the Training Centre
provided with a specific haptic device and with the software
necessary both to manage the interaction in the virtual environment.
After the integration of the virtual environment in the simulation
system it is possible to perform training on the specific surgical
procedure.
Abstract: Optimization is often a critical issue for most system
design problems. Evolutionary Algorithms are population-based,
stochastic search techniques, widely used as efficient global
optimizers. However, finding optimal solution to complex high
dimensional, multimodal problems often require highly
computationally expensive function evaluations and hence are
practically prohibitive. The Dynamic Approximate Fitness based
Hybrid EA (DAFHEA) model presented in our earlier work [14]
reduced computation time by controlled use of meta-models to
partially replace the actual function evaluation by approximate
function evaluation. However, the underlying assumption in
DAFHEA is that the training samples for the meta-model are
generated from a single uniform model. Situations like model
formation involving variable input dimensions and noisy data
certainly can not be covered by this assumption. In this paper we
present an enhanced version of DAFHEA that incorporates a
multiple-model based learning approach for the SVM approximator.
DAFHEA-II (the enhanced version of the DAFHEA framework) also
overcomes the high computational expense involved with additional
clustering requirements of the original DAFHEA framework. The
proposed framework has been tested on several benchmark functions
and the empirical results illustrate the advantages of the proposed
technique.
Abstract: This study was carried out to determine the feedback from the primary school headmasters on in-service training activities. 384 primary school headmasters were asked to give feedback on the in-service training facilities organized by the Ministry of Education through an open –ended question. The study group was formed by 359 primary school headmasters who attended “Management Skills" seminars which were organized in March, April and May in the extent of 2006 In-service Training Facilities Plan by the Ministry of Education and were also volunteers to give feedback on the inservice training activities. The qualitative research method was used in the analysis of the data since the primary school headmasters gave written feedback on the in-service training activities. Having analyzed the feedback, certain categories were formed by coding it. Certain frequencies and percentages were determined according to the codes. It was made obvious that the primary school headmasters benefitted from those facilities by sharing their experiences, problems and their possible solutions mutually
Abstract: The purpose of this research is the analysis of the
impact of ICT-related training in the adoption of a learning
management systems (LMS) for teaching practicesby faculties in a
higher education institution. Based on comparative analyses the
impact will be obtained by the number of LMS courses created and
managed by participants in ICT for teaching workshops and those
who have not attended to any workshops. Involving near 1320 LMS
courses and 265 faculties, the results evidence that(i) faculties who
have not attend any workshop present a larger distribution of empty
courses and (ii) faculties who have attended three or more workshops
managed a higher distribution of courses with a considerable level of
use intensity, when compared to the others groups. These findings
supportthe idea that faculty training is a crucial factor in the process
of LMS integration in higher education institutions and that faculties
who have been enrolled in three or more workshops develop a higher
level of technical and pedagogical proficiency in LMS.
Abstract: In this paper, a new algorithm for generating codebook is proposed for vector quantization (VQ) in image coding. The significant features of the training image vectors are extracted by using the proposed Orthogonal Polynomials based transformation. We propose to generate the codebook by partitioning these feature vectors into a binary tree. Each feature vector at a non-terminal node of the binary tree is directed to one of the two descendants by comparing a single feature associated with that node to a threshold. The binary tree codebook is used for encoding and decoding the feature vectors. In the decoding process the feature vectors are subjected to inverse transformation with the help of basis functions of the proposed Orthogonal Polynomials based transformation to get back the approximated input image training vectors. The results of the proposed coding are compared with the VQ using Discrete Cosine Transform (DCT) and Pairwise Nearest Neighbor (PNN) algorithm. The new algorithm results in a considerable reduction in computation time and provides better reconstructed picture quality.
Abstract: Real-time object tracking is a problem which involves extraction of critical information from complex and uncertain imagedata. In this paper, we present a comprehensive methodology to design an artificial neural network (ANN) for a real-time object tracking application. The object, which is tracked for the purpose of demonstration, is a specific airplane. However, the proposed ANN can be trained to track any other object of interest. The ANN has been simulated and tested on the training and testing datasets, as well as on a real-time streaming video. The tracking error is analyzed with post-regression analysis tool, which finds the correlation among the calculated coordinates and the correct coordinates of the object in the image. The encouraging results from the computer simulation and analysis show that the proposed ANN architecture is a good candidate solution to a real-time object tracking problem.
Abstract: There have been significant improvements in automatic
voice recognition technology. However, existing systems still face difficulties,
particularly when used by non-native speakers with accents.
In this paper we address a problem of identifying the English accented
speech of speakers from different backgrounds. Once an accent is
identified the speech recognition software can utilise training set from
appropriate accent and therefore improve the efficiency and accuracy
of the speech recognition system. We introduced the Q factor, which
is defined by the sum of relationships between frequencies of the
formants. Four different accents were considered and experimented
for this research. A scoring method was introduced in order to
effectively analyse accents. The proposed concept indicates that the
accent could be identified by analysing their formants.
Abstract: Text categorization is the problem of classifying text
documents into a set of predefined classes. After a preprocessing
step, the documents are typically represented as large sparse vectors.
When training classifiers on large collections of documents, both the
time and memory restrictions can be quite prohibitive. This justifies
the application of feature selection methods to reduce the
dimensionality of the document-representation vector. In this paper,
three feature selection methods are evaluated: Random Selection,
Information Gain (IG) and Support Vector Machine feature selection
(called SVM_FS). We show that the best results were obtained with
SVM_FS method for a relatively small dimension of the feature
vector. Also we present a novel method to better correlate SVM
kernel-s parameters (Polynomial or Gaussian kernel).
Abstract: Surface roughness (Ra) is one of the most important requirements in machining process. In order to obtain better surface roughness, the proper setting of cutting parameters is crucial before the process take place. This research presents the development of mathematical model for surface roughness prediction before milling process in order to evaluate the fitness of machining parameters; spindle speed, feed rate and depth of cut. 84 samples were run in this study by using FANUC CNC Milling α-Τ14ιE. Those samples were randomly divided into two data sets- the training sets (m=60) and testing sets(m=24). ANOVA analysis showed that at least one of the population regression coefficients was not zero. Multiple Regression Method was used to determine the correlation between a criterion variable and a combination of predictor variables. It was established that the surface roughness is most influenced by the feed rate. By using Multiple Regression Method equation, the average percentage deviation of the testing set was 9.8% and 9.7% for training data set. This showed that the statistical model could predict the surface roughness with about 90.2% accuracy of the testing data set and 90.3% accuracy of the training data set.