Abstract: Opinion extraction about products from customer
reviews is becoming an interesting area of research. Customer
reviews about products are nowadays available from blogs and
review sites. Also tools are being developed for extraction of opinion
from these reviews to help the user as well merchants to track the
most suitable choice of product. Therefore efficient method and
techniques are needed to extract opinions from review and blogs. As
reviews of products mostly contains discussion about the features,
functions and services, therefore, efficient techniques are required to
extract user comments about the desired features, functions and
services. In this paper we have proposed a novel idea to find features
of product from user review in an efficient way. Our focus in this
paper is to get the features and opinion-oriented words about
products from text through auxiliary verbs (AV) {is, was, are, were,
has, have, had}. From the results of our experiments we found that
82% of features and 85% of opinion-oriented sentences include AVs.
Thus these AVs are good indicators of features and opinion
orientation in customer reviews.
Abstract: A new target detection technique is presented in this
paper for the identification of small boats in coastal surveillance. The
proposed technique employs an adaptive progressive thresholding (APT) scheme to first process the given input scene to separate any
objects present in the scene from the background. The preprocessing
step results in an image having only the foreground objects, such as
boats, trees and other cluttered regions, and hence reduces the search
region for the correlation step significantly. The processed image is then fed to the shifted phase-encoded fringe-adjusted joint transform
correlator (SPFJTC) technique which produces single and delta-like
correlation peak for a potential target present in the input scene. A
post-processing step involves using a peak-to-clutter ratio (PCR) to determine whether the boat in the input scene is authorized or unauthorized. Simulation results are presented to show that the
proposed technique can successfully determine the presence of an authorized boat and identify any intruding boat present in the given input scene.
Abstract: This paper presents a part of research on the
rheological properties of bitumen modified by thermoplastic namely
linear low density polyethylene (LLDPE), high density polyethylene
(HDPE) and polypropylene (PP) and its interaction with 80 pen base
bitumen. As it is known that the modification of bitumen by the use
of polymers enhances its performance characteristics but at the same
time significantly alters its rheological properties. The rheological
study of polymer modified bitumen (PMB) was made through
penetration, ring & ball softening point and viscosity test. The results
were then related to the changes in the rheological properties of
polymer modified bitumen. It was observed that thermoplastic
copolymer shows profound effect on penetration rather than
softening point. The viscoelastic behavior of polymer modified
bitumen depend on the concentration of polymer, mixing
temperature, mixing technique, solvating power of base bitumen and
molecular structure of polymer used. PP offer better blend in
comparison to HDPE and LLDPE. The viscosity of base bitumen was
also enhanced with the addition of polymer. The pseudoplastic
behavior was more prominent for HDPE and LLDPE than PP. Best
results were obtained when polymer concentration was kept below
3%
Abstract: Fundamental sensor-motor couplings form the backbone
of most mobile robot control tasks, and often need to be implemented
fast, efficiently and nevertheless reliably. Machine learning
techniques are therefore often used to obtain the desired sensor-motor
competences.
In this paper we present an alternative to established machine
learning methods such as artificial neural networks, that is very fast,
easy to implement, and has the distinct advantage that it generates
transparent, analysable sensor-motor couplings: system identification
through nonlinear polynomial mapping.
This work, which is part of the RobotMODIC project at the
universities of Essex and Sheffield, aims to develop a theoretical understanding
of the interaction between the robot and its environment.
One of the purposes of this research is to enable the principled design
of robot control programs.
As a first step towards this aim we model the behaviour of the
robot, as this emerges from its interaction with the environment, with
the NARMAX modelling method (Nonlinear, Auto-Regressive, Moving
Average models with eXogenous inputs). This method produces
explicit polynomial functions that can be subsequently analysed using
established mathematical methods.
In this paper we demonstrate the fidelity of the obtained NARMAX
models in the challenging task of robot route learning; we present a
set of experiments in which a Magellan Pro mobile robot was taught
to follow four different routes, always using the same mechanism to
obtain the required control law.
Abstract: Concerning the measurement of friction properties of
textiles and fabrics using Kawabata Evaluation System (KES), whose
output is constrained to the surface friction factor of fabric, and no
other data would be generated; this research has been conducted to
gain information about surface roughness regarding its surface
friction factor. To assess roughness properties of light nonwovens, a
3-dimensional model of a surface has been simulated with regular
sinuous waves through it as an ideal surface. A new factor was
defined, namely Surface Roughness Factor, through comparing
roughness properties of simulated surface and real specimens. The
relation between the proposed factor and friction factor of specimens
has been analyzed by regression, and results showed a meaningful
correlation between them. It can be inferred that the new presented
factor can be used as an acceptable criterion for evaluating the
roughness properties of light nonwoven fabrics.
Abstract: The use of new technologies such internet (e-mail, chat
rooms) and cell phones has steeply increased in recent years.
Especially among children and young people, use of technological
tools and equipments is widespread. Although many teachers and
administrators now recognize the problem of school bullying, few are
aware that students are being harassed through electronic
communication. Referred to as electronic bullying, cyber bullying, or
online social cruelty, this phenomenon includes bullying through email,
instant messaging, in a chat room, on a website, or through
digital messages or images sent to a cell phone. Cyber bullying is
defined as causing deliberate/intentional harm to others using internet
or other digital technologies. It has a quantitative research design nd
uses relational survey as its method. The participants consisted of
300 secondary school students in the city of Konya, Turkey. 195
(64.8%) participants were female and 105 (35.2%) were male. 39
(13%) students were at grade 1, 187 (62.1%) were at grade 2 and 74
(24.6%) were at grade 3. The “Cyber Bullying Question List"
developed by Ar─▒cak (2009) was given to students. Following
questions about demographics, a functional definition of cyber
bullying was provided. In order to specify students- human values,
“Human Values Scale (HVS)" developed by Dilmaç (2007) for
secondary school students was administered. The scale consists of 42
items in six dimensions. Data analysis was conducted by the primary
investigator of the study using SPSS 14.00 statistical analysis
software. Descriptive statistics were calculated for the analysis of
students- cyber bullying behaviour and simple regression analysis was
conducted in order to test whether each value in the scale could
explain cyber bullying behaviour.
Abstract: The objective of this research was to identify the
vegetation-soil relationships in Nodushan arid rangelands of Yazd. 5
sites were selected for measuring the cover of plant species and soil
attributes. Soil samples were taken in 0-10 and 10-80 cm layers. The
species studied were Salsola tomentosa, Salsola arbuscula, Peganum
harmala, Zygophylum eurypterum and Eurotia ceratoides. Canonical
correspondence analysis (CCA) was used to analyze the data. Based
on the CCA results, 74.9 % of vegetation-soil variation was explained
by axis 1-3. Axis 1, 2 and 3 accounted for 27.2%, 24.9 % and 22.8%
of variance respectively. Correlation between axis 1, 2, 3 and speciesedaphic
variables were 0.995, 0.989, 0.981 respectively. Soil texture,
lime, salinity and organic matter significantly influenced the
distribution of these plant species. Determination of soil-vegetation
relationships will be useful for managing and improving rangelands
in arid and semi arid environments.
Abstract: The aim of this study is to find out and analyze the
role of gender and age on the perceptions of students to the distant
online program offered by Vocational High School in Sakarya
University. The research is based on a questionnaire as a mean of
data collection method to find out the role of age and gender on the
student-s perceptions toward online education, and the study
progressed through finding relationships between the variables used
in the data collection instrument. The findings of the analysis
revealed that although the students registered to the online program
by will, they preferred the traditional face-to-face education due to
the difficulty of the nonverbal communication, their incompetence of
using the technology required, and their belief in traditional face-toface
learning more than online education.
Regarding gender, the results showed that the female students
have a better perception of the online education as opposed to the
male students. Regarding age, the results showed that the older the
students are the more is their preference towards attending face-toface
classes.
Abstract: Breast skin-line estimation and breast segmentation is an important pre-process in mammogram image processing and computer-aided diagnosis of breast cancer. Limiting the area to be processed into a specific target region in an image would increase the accuracy and efficiency of processing algorithms. In this paper we are presenting a new algorithm for estimating skin-line and breast segmentation using fast marching algorithm. Fast marching is a partial-differential equation based numerical technique to track evolution of interfaces. We have introduced some modifications to the traditional fast marching method, specifically to improve the accuracy of skin-line estimation and breast tissue segmentation. Proposed modifications ensure that the evolving front stops near the desired boundary. We have evaluated the performance of the algorithm by using 100 mammogram images taken from mini-MIAS database. The results obtained from the experimental evaluation indicate that this algorithm explains 98.6% of the ground truth breast region and accuracy of the segmentation is 99.1%. Also this algorithm is capable of partially-extracting nipple when it is available in the profile.
Abstract: In blended learning environments, the Internet can be combined with other technologies. The aim of this research was to design, introduce and validate a model to support synchronous and asynchronous activities by managing content domains in an Adaptive Hypermedia System (AHS). The application is based on information recovery techniques, clustering algorithms and adaptation rules to adjust the user's model to contents and objects of study. This system was applied to blended learning in higher education. The research strategy used was the case study method. Empirical studies were carried out on courses at two universities to validate the model. The results of this research show that the model had a positive effect on the learning process. The students indicated that the synchronous and asynchronous scenario is a good option, as it involves a combination of work with the lecturer and the AHS. In addition, they gave positive ratings to the system and stated that the contents were adapted to each user profile.
Abstract: Restoration research has become important on principle recently in Czech Republic. The reason is simple. More than 70 % of mined brown coal comes from the North Bohemian Basin these days. Open cast brown coal mining has lead to large damage on the landscape. Reclamation of phytotoxic areas is one of the serious problems in the North Bohemian Basin. It mainly concerns the areas with the occurrence of overburden rocks from the coal bed enriched with coal. The presented paper includes the characteristics of the important phytotoxic areas and the methodology of their reclamation. The results are documented with the long term monitoring of physical, mineralogical, chemical and pedological parameters of rocks in the testing areas.
Abstract: Knowledge bases are basic components of expert
systems or intelligent computational programs. Knowledge bases
provide knowledge, events that serve deduction activity,
computation and control. Therefore, researching and developing of
models for knowledge representation play an important role in
computer science, especially in Artificial Intelligence Science and
intelligent educational software. In this paper, the extensive
deduction computational model is proposed to design knowledge
bases whose attributes are able to be real values or functional values.
The system can also solve problems based on knowledge bases.
Moreover, the models and algorithms are applied to produce the
educational software for solving alternating current problems or
solving set of equations automatically.
Abstract: This paper reports on an effort to address the issue of
inequality in girls- and women-s access to science, engineering and
technology (SET) education and careers through raising awareness on
SET among secondary school girls in South Africa. Girls participated
in hands-on high-tech rapid prototyping environment of a fabrication
laboratory that was aimed at stimulating creativity and innovation as
part of a Fab Kids initiative. The Fab Kids intervention is about
creating a SET pipeline as part of the Young Engineers and Scientists
of Africa Initiative.The methodology was based on a real world
situation and a hands-on approach. In the process, participants
acquired a number of skills including computer-aided design,
research skills, communication skills, teamwork skills, technical
drawing skills, writing skills and problem-solving skills. Exposure to
technology enhanced the girls- confidence in being able to handle
technology-related tasks.
Abstract: Quantitative trait loci (QTL) experiments have yielded
important biological and biochemical information necessary for
understanding the relationship between genetic markers and
quantitative traits. For many years, most QTL algorithms only
allowed one observation per genotype. Recently, there has been an
increasing demand for QTL algorithms that can accommodate more
than one observation per genotypic distribution. The Bayesian
hierarchical model is very flexible and can easily incorporate this
information into the model. Herein a methodology is presented that
uses a Bayesian hierarchical model to capture the complexity of the
data. Furthermore, the Markov chain Monte Carlo model composition
(MC3) algorithm is used to search and identify important markers. An
extensive simulation study illustrates that the method captures the
true QTL, even under nonnormal noise and up to 6 QTL.
Abstract: The overall objective of this research is a strain
improvement technology for efficient pectinase production. A novel
cells cultivation technology by immobilization of fungal cells has
been studied in long time continuous fermentations. Immobilization
was achieved by using of new material for absorption of stores of
immobilized cultures which was for the first time used for
immobilization of microorganisms. Effects of various conditions of
nitrogen and carbon nutrition on the biosynthesis of pectolytic
enzymes in Aspergillus awamori 1-8 strain were studied. Proposed
cultivation technology along with optimization of media components
for pectinase overproduction led to increased pectinase productivity
in Aspergillus awamori 1-8 from 7 to 8 times. Proposed technology
can be applied successfully for production of major industrial
enzymes such as α-amylase, protease, collagenase etc.
Abstract: Video-on-demand (VOD) is designed by using content delivery networks (CDN) to minimize the overall operational cost and to maximize scalability. Estimation of the viewing pattern (i.e., the relationship between the number of viewings and the ranking of VOD contents) plays an important role in minimizing the total operational cost and maximizing the performance of the VOD systems. In this paper, we have analyzed a large body of commercial VOD viewing data and found that the viewing rank distribution fits well with the parabolic fractal distribution. The weighted linear model fitting function is used to estimate the parameters (coefficients) of the parabolic fractal distribution. This paper presents an analytical basis for designing an optimal hierarchical VOD contents distribution system in terms of its cost and performance.
Abstract: Natural organic matter (NOM) is heterogeneous
mixture of organic compounds that enter the water media from
animal and plant remains, domestic and industrial wastes.
Researches showed that NOM is likely precursor material for
disinfection by products (DBPs). Chlorine very commenly used for
disinfection purposes and NOM and chlorine reacts then
Trihalomethane (THM) and Haloacetic acids (HAAs) which are
cancerogenics for human health are produced. The aim of the study is
to search NOM removal by enhanced coagulation from drinking
water source of Eskisehir which is supplied from Porsuk Dam.
Recently, Porsuk dam water is getting highly polluted and therefore
NOM concentration is increasing. Enhanced coagulation studies were
evaluated by measurement of Dissolved Organic Carbon (DOC), UV
absorbance at 254 nm (UV254), and different trihalomethane
formation potential (THMFP) tests. Results of jar test experiments
showed that NOM can be removed from water about 40-50 % of
efficiency by enhanced coagulation. Optimum coagulant type and
coagulant dosages were determined using FeCl3 and Alum.
Abstract: The ability of information systems to operate in conjunction with each other encompassing communication protocols, hardware, software, application, and data compatibility layers. There has been considerable work in industry on the development of component interoperability models, such as CORBA, (D)COM and JavaBeans. These models are intended to reduce the complexity of software development and to facilitate reuse of off-the-shelf components. The focus of these models is syntactic interface specification, component packaging, inter-component communications, and bindings to a runtime environment. What these models lack is a consideration of architectural concerns – specifying systems of communicating components, explicitly representing loci of component interaction, and exploiting architectural styles that provide well-understood global design solutions. The development of complex business applications is now focused on an assembly of components available on a local area network or on the net. These components must be localized and identified in terms of available services and communication protocol before any request. The first part of the article introduces the base concepts of components and middleware while the following sections describe the different up-todate models of communication and interaction and the last section shows how different models can communicate among themselves.
Abstract: Anchovy (Engraulis Encrasicholus) and sardine
(Sardina Pilchardus) are blue fishes linked to our alimentary tradition
of Mediterranean. In our work, particularly, we tested for the first
time physical and enzymatic methods to verify the freshness of
species of blue fish, anchovy and sardine of Mediterranean. In
connection with to the lowering of the pH after post-mortem stage we
assisted to a increase in proteolytic activity of calpaine and catpsine.
Already after 2 h in post-mortem there was a significant increase.
Abstract: This paper presents a novel genetic algorithm, termed
the Optimum Individual Monogenetic Algorithm (OIMGA) and
describes its hardware implementation. As the monogenetic strategy
retains only the optimum individual, the memory requirement is
dramatically reduced and no crossover circuitry is needed, thereby
ensuring the requisite silicon area is kept to a minimum.
Consequently, depending on application requirements, OIMGA
allows the investigation of solutions that warrant either larger GA
populations or individuals of greater length. The results given in this
paper demonstrate that both the performance of OIMGA and its
convergence time are superior to those of existing hardware GA
implementations. Local convergence is achieved in OIMGA by
retaining elite individuals, while population diversity is ensured by
continually searching for the best individuals in fresh regions of the
search space.