Abstract: This paper proposes a specialized Web robot to automatically collect objectionable Web contents for use in an objectionable Web content classification system, which creates the URL database of objectionable Web contents. It aims at shortening the update period of the DB, increasing the number of URLs in the DB, and enhancing the accuracy of the information in the DB.
Abstract: The wind resource in the Italian site of Lendinara
(RO) is analyzed through a systematic anemometric campaign
performed on the top of the bell tower, at an altitude of over 100 m
above the ground. Both the average wind speed and the Weibull
distribution are computed. The resulting average wind velocity is in
accordance with the numerical predictions of the Italian Wind Atlas,
confirming the accuracy of the extrapolation of wind data adopted for
the evaluation of wind potential at higher altitudes with respect to the
commonly placed measurement stations.
Abstract: Parsing is important in Linguistics and Natural
Language Processing to understand the syntax and semantics of a
natural language grammar. Parsing natural language text is
challenging because of the problems like ambiguity and inefficiency.
Also the interpretation of natural language text depends on context
based techniques. A probabilistic component is essential to resolve
ambiguity in both syntax and semantics thereby increasing accuracy
and efficiency of the parser. Tamil language has some inherent
features which are more challenging. In order to obtain the solutions,
lexicalized and statistical approach is to be applied in the parsing
with the aid of a language model. Statistical models mainly focus on
semantics of the language which are suitable for large vocabulary
tasks where as structural methods focus on syntax which models
small vocabulary tasks. A statistical language model based on Trigram
for Tamil language with medium vocabulary of 5000 words has
been built. Though statistical parsing gives better performance
through tri-gram probabilities and large vocabulary size, it has some
disadvantages like focus on semantics rather than syntax, lack of
support in free ordering of words and long term relationship. To
overcome the disadvantages a structural component is to be
incorporated in statistical language models which leads to the
implementation of hybrid language models. This paper has attempted
to build phrase structured hybrid language model which resolves
above mentioned disadvantages. In the development of hybrid
language model, new part of speech tag set for Tamil language has
been developed with more than 500 tags which have the wider
coverage. A phrase structured Treebank has been developed with 326
Tamil sentences which covers more than 5000 words. A hybrid
language model has been trained with the phrase structured Treebank
using immediate head parsing technique. Lexicalized and statistical
parser which employs this hybrid language model and immediate
head parsing technique gives better results than pure grammar and
trigram based model.
Abstract: Recently, the advanced technologies that offer high
precision product, relative easy, economical process and also rapid
production are needed to realize the high demand of ultra precision
micro part. In our research, micromanufacturing based on soft
lithography and nanopowder injection molding was investigated. The
silicone metal pattern with ultra thick and high aspect ratio succeeds to
fabricate Polydimethylsiloxane (PDMS) micro mold. The process
followed by nanopowder injection molding (PIM) by a simple vacuum
hot press. The 17-4ph nanopowder with diameter of 100 nm, succeed
to be injected and it forms green sample microbearing with thickness,
microchannel and aspect ratio is 700μm, 60μm and 12, respectively.
Sintering process was done in 1200 C for 2 hours and heating rate
0.83oC/min. Since low powder load (45% PL) was applied to achieve
green sample fabrication, ~15% shrinkage happen in the 86% relative
density. Several improvements should be done to produce high
accuracy and full density sintered part.
Abstract: The stochastic nature of tool life using conventional discrete-wear data from experimental tests usually exists due to many individual and interacting parameters. It is a common practice in batch production to continually use the same tool to machine different parts, using disparate machining parameters. In such an environment, the optimal points at which tools have to be changed, while achieving minimum production cost and maximum production rate within the surface roughness specifications, have not been adequately studied. In the current study, two relevant aspects are investigated using coated and uncoated inserts in turning operations: (i) the accuracy of using machinability information, from fixed parameters testing procedures, when variable parameters situations are emerged, and (ii) the credibility of tool life machinability data from prior discrete testing procedures in a non-stop machining. A novel technique is proposed and verified to normalize the conventional fixed parameters machinability data to suit the cases when parameters have to be changed for the same tool. Also, an experimental investigation has been established to evaluate the error in the tool life assessment when machinability from discrete testing procedures is employed in uninterrupted practical machining.
Abstract: Maintenance is one of the most important activities in
the shipyard industry. However, sometimes it is not supported by
adequate services from the shipyard, where inaccuracy in estimating
the duration of the ship maintenance is still common. This makes
estimation of ship maintenance duration is crucial. This study uses
Data Mining approach, i.e., CART (Classification and Regression
Tree) to estimate the duration of ship maintenance that is limited to
dock works or which is known as dry docking. By using the volume
of dock works as an input to estimate the maintenance duration, 4
classes of dry docking duration were obtained with different linear
model and job criteria for each class. These linear models can then be
used to estimate the duration of dry docking based on job criteria.
Abstract: The prediction of Software quality during development life cycle of software project helps the development organization to make efficient use of available resource to produce the product of highest quality. “Whether a module is faulty or not" approach can be used to predict quality of a software module. There are numbers of software quality prediction models described in the literature based upon genetic algorithms, artificial neural network and other data mining algorithms. One of the promising aspects for quality prediction is based on clustering techniques. Most quality prediction models that are based on clustering techniques make use of K-means, Mixture-of-Guassians, Self-Organizing Map, Neural Gas and fuzzy K-means algorithm for prediction. In all these techniques a predefined structure is required that is number of neurons or clusters should be known before we start clustering process. But in case of Growing Neural Gas there is no need of predetermining the quantity of neurons and the topology of the structure to be used and it starts with a minimal neurons structure that is incremented during training until it reaches a maximum number user defined limits for clusters. Hence, in this work we have used Growing Neural Gas as underlying cluster algorithm that produces the initial set of labeled cluster from training data set and thereafter this set of clusters is used to predict the quality of test data set of software modules. The best testing results shows 80% accuracy in evaluating the quality of software modules. Hence, the proposed technique can be used by programmers in evaluating the quality of modules during software development.
Abstract: The ultimate goal of this article is to develop a robust and accurate numerical method for solving hyperbolic conservation laws in one and two dimensions. A hybrid numerical method, coupling a cheap fourth order total variation diminishing (TVD) scheme [1] for smooth region and a Robust seventh-order weighted non-oscillatory (WENO) scheme [2] near discontinuities, is considered. High order multi-resolution analysis is used to detect the high gradients regions of the numerical solution in order to capture the shocks with the WENO scheme, while the smooth regions are computed with fourth order total variation diminishing (TVD). For time integration, we use the third order TVD Runge-Kutta scheme. The accuracy of the resulting hybrid high order scheme is comparable with these of WENO, but with significant decrease of the CPU cost. Numerical demonstrates that the proposed scheme is comparable to the high order WENO scheme and superior to the fourth order TVD scheme. Our scheme has the added advantage of simplicity and computational efficiency. Numerical tests are presented which show the robustness and effectiveness of the proposed scheme.
Abstract: In this study, we experiment on precise control outlet
temperature of water from the water cooler with hot-gas bypass
method based on PI control logic for machine tool. Recently, technical
trend for machine tools is focused on enhancement of speed and
accuracy. High speedy processing causes thermal and structural
deformation of objects from the machine tools. Water cooler has to be
applied to machine tools to reduce the thermal negative influence with
accurate temperature controlling system. The goal of this study is to
minimize temperature error in steady state. In addition, control period
of an electronic expansion valve were considered to increment of
lifetime of the machine tools and quality of product with a water
cooler.
Abstract: Nowadays, OCR systems have got several
applications and are increasingly employed in daily life. Much
research has been done regarding the identification of Latin,
Japanese, and Chinese characters. However, very little investigation
has been performed regarding Farsi/Arabic characters recognition.
Probably the reason is difficulty and complexity of those characters
identification compared to the others and limitation of IT activities in
Farsi and Arabic speaking countries. In this paper, a technique has
been employed to identify isolated Farsi/Arabic characters. A chain
code based algorithm along with other significant peculiarities such
as number and location of dots and auxiliary parts, and the number of
holes existing in the isolated character has been used in this study to
identify Farsi/Arabic characters. Experimental results show the
relatively high accuracy of the method developed when it is tested on
several standard Farsi fonts.
Abstract: In this work, the plate bending formulation of the boundary element method - BEM, based on the Reissner?s hypothesis, is extended to the analysis of plates reinforced by beams taking into account the membrane effects. The formulation is derived by assuming a zoned body where each sub-region defines a beam or a slab and all of them are represented by a chosen reference surface. Equilibrium and compatibility conditions are automatically imposed by the integral equations, which treat this composed structure as a single body. In order to reduce the number of degrees of freedom, the problem values defined on the interfaces are written in terms of their values on the beam axis. Initially are derived separated equations for the bending and stretching problems, but in the final system of equations the two problems are coupled and can not be treated separately. Finally are presented some numerical examples whose analytical results are known to show the accuracy of the proposed model.
Abstract: Cancer classification to their corresponding cohorts has been key area of research in bioinformatics aiming better prognosis of the disease. High dimensionality of gene data has been makes it a complex task and requires significance data identification technique in order to reducing the dimensionality and identification of significant information. In this paper, we have proposed a novel approach for classification of oral cancer into metastasis positive and negative patients. We have used significance analysis of microarrays (SAM) for identifying significant genes which constitutes gene signature. 3 different gene signatures were identified using SAM from 3 different combination of training datasets and their classification accuracy was calculated on corresponding testing datasets using k-Nearest Neighbour (kNN), Fuzzy C-Means Clustering (FCM), Support Vector Machine (SVM) and Backpropagation Neural Network (BPNN). A final gene signature of only 9 genes was obtained from above 3 individual gene signatures. 9 gene signature-s classification capability was compared using same classifiers on same testing datasets. Results obtained from experimentation shows that 9 gene signature classified all samples in testing dataset accurately while individual genes could not classify all accurately.
Abstract: In this paper, an efficient structural approach for
recognizing on-line handwritten digits is proposed. After reading
the digit from the user, the slope is estimated and normalized for
adjacent nodes. Based on the changing of signs of the slope values,
the primitives are identified and extracted. The names of these
primitives are represented by strings, and then a finite state
machine, which contains the grammars of the digits, is traced to
identify the digit. Finally, if there is any ambiguity, it will be
resolved. Experiments showed that this technique is flexible and
can achieve high recognition accuracy for the shapes of the digits
represented in this work.
Abstract: In this article, the phenomenon of nonlinear
consolidation in saturated and homogeneous clay layer is studied.
Considering time-varied drainage model, the excess pore water
pressure in the layer depth is calculated. The Generalized Differential
Quadrature (GDQ) method is used for the modeling and numerical
analysis. For the purpose of analysis, first the domain of independent
variables (i.e., time and clay layer depth) is discretized by the
Chebyshev-Gauss-Lobatto series and then the nonlinear system of
equations obtained from the GDQ method is solved by means of the
Newton-Raphson approach. The obtained results indicate that the
Generalized Differential Quadrature method, in addition to being
simple to apply, enjoys a very high accuracy in the calculation of
excess pore water pressure.
Abstract: Researches show that probability-statistical methods application, especially at the early stage of the aviation Gas Turbine Engine (GTE) technical condition diagnosing, when the flight information has property of the fuzzy, limitation and uncertainty is unfounded. Hence the efficiency of application of new technology Soft Computing at these diagnosing stages with the using of the Fuzzy Logic and Neural Networks methods is considered. According to the purpose of this problem training with high accuracy of fuzzy multiple linear and non-linear models (fuzzy regression equations) which received on the statistical fuzzy data basis is made. For GTE technical condition more adequate model making dynamics of skewness and kurtosis coefficients- changes are analysed. Researches of skewness and kurtosis coefficients values- changes show that, distributions of GTE workand output parameters of the multiple linear and non-linear generalised models at presence of noise measured (the new recursive Least Squares Method (LSM)). The developed GTE condition monitoring system provides stage-by-stage estimation of engine technical conditions. As application of the given technique the estimation of the new operating aviation engine technical condition was made.
Abstract: The counting and analysis of blood cells allows the
evaluation and diagnosis of a vast number of diseases. In particular,
the analysis of white blood cells (WBCs) is a topic of great interest to
hematologists. Nowadays the morphological analysis of blood cells is
performed manually by skilled operators. This involves numerous
drawbacks, such as slowness of the analysis and a nonstandard
accuracy, dependent on the operator skills. In literature there are only
few examples of automated systems in order to analyze the white
blood cells, most of which only partial. This paper presents a
complete and fully automatic method for white blood cells
identification from microscopic images. The proposed method firstly
individuates white blood cells from which, subsequently, nucleus and
cytoplasm are extracted. The whole work has been developed using
MATLAB environment, in particular the Image Processing Toolbox.
Abstract: The paper presents the modeling of nonlinear
longitudinal aerodynamics using flight data of Hansa-3 aircraft at
high angles of attack near stall. The Kirchhoff-s quasi-steady stall
model has been used to incorporate nonlinear aerodynamic effects in
the aerodynamic model used to estimate the parameters, thereby,
making the aerodynamic model nonlinear. The Maximum Likelihood
method has been applied to the flight data (at high angles of attack)
for the estimation of parameters (aerodynamic and stall
characteristics) using the nonlinear aerodynamic model. To improve
the accuracy level of the estimates, an approach of fixing the strong
parameters has also been presented.
Abstract: This paper presents an algorithm for the recognition
and tracking of moving objects, 1/10 scale model car is used to verify
performance of the algorithm. Presented algorithm for the recognition
and tracking of moving objects in the paper is as follows. SURF
algorithm is merged with Lucas-Kanade algorithm. SURF algorithm
has strong performance on contrast, size, rotation changes and it
recognizes objects but it is slow due to many computational
complexities. Processing speed of Lucas-Kanade algorithm is fast but
the recognition of objects is impossible. Its optical flow compares the
previous and current frames so that can track the movement of a pixel.
The fusion algorithm is created in order to solve problems which
occurred using the Kalman Filter to estimate the position and the
accumulated error compensation algorithm was implemented. Kalman
filter is used to create presented algorithm to complement problems
that is occurred when fusion two algorithms. Kalman filter is used to
estimate next location, compensate for the accumulated error. The
resolution of the camera (Vision Sensor) is fixed to be 640x480. To
verify the performance of the fusion algorithm, test is compared to
SURF algorithm under three situations, driving straight, curve, and
recognizing cars behind the obstacles. Situation similar to the actual is
possible using a model vehicle. Proposed fusion algorithm showed
superior performance and accuracy than the existing object
recognition and tracking algorithms. We will improve the performance
of the algorithm, so that you can experiment with the images of the
actual road environment.
Abstract: This paper proposed a novel model for short term load
forecast (STLF) in the electricity market. The prior electricity
demand data are treated as time series. The model is composed of
several neural networks whose data are processed using a wavelet
technique. The model is created in the form of a simulation program
written with MATLAB. The load data are treated as time series data.
They are decomposed into several wavelet coefficient series using
the wavelet transform technique known as Non-decimated Wavelet
Transform (NWT). The reason for using this technique is the belief
in the possibility of extracting hidden patterns from the time series
data. The wavelet coefficient series are used to train the neural
networks (NNs) and used as the inputs to the NNs for electricity load
prediction. The Scale Conjugate Gradient (SCG) algorithm is used as
the learning algorithm for the NNs. To get the final forecast data, the
outputs from the NNs are recombined using the same wavelet
technique. The model was evaluated with the electricity load data of
Electronic Engineering Department in Mandalay Technological
University in Myanmar. The simulation results showed that the
model was capable of producing a reasonable forecasting accuracy in
STLF.
Abstract: Information on weed distribution within the field is necessary to implement spatially variable herbicide application. Since hand labor is costly, an automated weed control system could be feasible. This paper deals with the development of an algorithm for real time specific weed recognition system based on Histogram Maxima with threshold of an image that is used for the weed classification. This algorithm is specifically developed to classify images into broad and narrow class for real-time selective herbicide application. The developed system has been tested on weeds in the lab, which have shown that the system to be very effectiveness in weed identification. Further the results show a very reliable performance on images of weeds taken under varying field conditions. The analysis of the results shows over 95 percent classification accuracy over 140 sample images (broad and narrow) with 70 samples from each category of weeds.