Abstract: The future of business intelligence (BI) is to integrate
intelligence into operational systems that works in real-time
analyzing small chunks of data based on requirements on continuous
basis. This is moving away from traditional approach of doing
analysis on ad-hoc basis or sporadically in passive and off-line mode
analyzing huge amount data. Various AI techniques such as expert
systems, case-based reasoning, neural-networks play important role
in building business intelligent systems. Since BI involves various
tasks and models various types of problems, hybrid intelligent
techniques can be better choice. Intelligent systems accessible
through web services make it easier to integrate them into existing
operational systems to add intelligence in every business processes.
These can be built to be invoked in modular and distributed way to
work in real time. Functionality of such systems can be extended to
get external inputs compatible with formats like RSS. In this paper,
we describe a framework that use effective combinations of these
techniques, accessible through web services and work in real-time.
We have successfully developed various prototype systems and done
few commercial deployments in the area of personalization and
recommendation on mobile and websites.
Abstract: The public sector holds large amounts of data of
various areas such as social affairs, economy, or tourism. Various
initiatives such as Open Government Data or the EU Directive on
public sector information aim to make these data available for public
and private service providers. Requirements for the provision of
public sector data are defined by legal and organizational
frameworks. Surprisingly, the defined requirements hardly cover
security aspects such as integrity or authenticity.
In this paper we discuss the importance of these missing
requirements and present a concept to assure the integrity and
authenticity of provided data based on electronic signatures. We
show that our concept is perfectly suitable for the provisioning of
unaltered data. We also show that our concept can also be extended
to data that needs to be anonymized before provisioning by
incorporating redactable signatures. Our proposed concept enhances
trust and reliability of provided public sector data.
Abstract: This paper presents a new strategy of identification
and classification of pathological voices using the hybrid method
based on wavelet transform and neural networks. After speech
acquisition from a patient, the speech signal is analysed in order to
extract the acoustic parameters such as the pitch, the formants, Jitter,
and shimmer. Obtained results will be compared to those normal and
standard values thanks to a programmable database. Sounds are
collected from normal people and patients, and then classified into
two different categories. Speech data base is consists of several
pathological and normal voices collected from the national hospital
“Rabta-Tunis". Speech processing algorithm is conducted in a
supervised mode for discrimination of normal and pathology voices
and then for classification between neural and vocal pathologies
(Parkinson, Alzheimer, laryngeal, dyslexia...). Several simulation
results will be presented in function of the disease and will be
compared with the clinical diagnosis in order to have an objective
evaluation of the developed tool.
Abstract: Analysis and visualization of microarraydata is veryassistantfor biologists and clinicians in the field of diagnosis and treatment of patients. It allows Clinicians to better understand the structure of microarray and facilitates understanding gene expression in cells. However, microarray dataset is a complex data set and has thousands of features and a very small number of observations. This very high dimensional data set often contains some noise, non-useful information and a small number of relevant features for disease or genotype. This paper proposes a non-linear dimensionality reduction algorithm Local Principal Component (LPC) which aims to maps high dimensional data to a lower dimensional space. The reduced data represents the most important variables underlying the original data. Experimental results and comparisons are presented to show the quality of the proposed algorithm. Moreover, experiments also show how this algorithm reduces high dimensional data whilst preserving the neighbourhoods of the points in the low dimensional space as in the high dimensional space.
Abstract: This paper deals with econometric analysis of real
retail trade turnover. It is a part of an extensive scientific research
about modern trends in Croatian national economy. At the end of the
period of transition economy, Croatia confronts with challenges and
problems of high consumption society. In such environment as
crucial economic variables: real retail trade turnover, average
monthly real wages and household loans are chosen for consequence
analysis. For the purpose of complete procedure of multiple
econometric analysis data base adjustment has been provided.
Namely, it has been necessary to deflate original national statistics
data of retail trade turnover using consumer price indices, as well as
provide process of seasonally adjustment of its contemporary
behavior. In model establishment it has been necessary to involve the
overcoming procedure for the autocorrelation and colinearity
problems. Moreover, for case of time-series shift a specific
appropriate econometric instrument has been applied. It would be
emphasize that the whole methodology procedure is based on the real
Croatian national economy time-series.
Abstract: Competing risks survival data that comprises of more
than one type of event has been used in many applications, and one
of these is in clinical study (e.g. in breast cancer study). The
decision tree method can be extended to competing risks survival
data by modifying the split function so as to accommodate two or
more risks which might be dependent on each other. Recently,
researchers have constructed some decision trees for recurrent
survival time data using frailty and marginal modelling. We further
extended the method for the case of competing risks. In this paper,
we developed the decision tree method for competing risks survival
time data based on proportional hazards for subdistribution of
competing risks. In particular, we grow a tree by using deviance
statistic. The application of breast cancer data is presented. Finally,
to investigate the performance of the proposed method, simulation
studies on identification of true group of observations were executed.
Abstract: since in natural accidents, facilities that relate to this vita element are underground so, it is difficult to find quickly some right, exact and definite information about water utilities. There fore, this article has done operationally in Boukan city in Western Azarbaijan of Iran and it tries to represent operation and capabilities of Geographical Information system (GIS) in urban water management at the time of natural accidents. Structure of this article is that firstly it has established a comprehensive data base related to water utilities by collecting, entering, saving and data management, then by modeling water utilities we have practically considered its operational aspects related to water utility problems in urban regions.
Abstract: Young patients suffering from Cerebral Palsy are
facing difficult choices concerning heavy surgeries. Diagnosis settled
by surgeons can be complex and on the other hand decision for
patient about getting or not such a surgery involves important
reflection effort. Proposed software combining prediction for
surgeries and post surgery kinematic values, and from 3D model
representing the patient is an innovative tool helpful for both patients
and medicine professionals. Beginning with analysis and
classification of kinematics values from Data Base extracted from
gait analysis in 3 separated clusters, it is possible to determine close
similarity between patients. Prediction surgery best adapted to
improve a patient gait is then determined by operating a suitable
preconditioned neural network. Finally, patient 3D modeling based
on kinematic values analysis, is animated thanks to post surgery
kinematic vectors characterizing the closest patient selected from
patients clustering.
Abstract: In this paper an algorithm based on the adaptive
neuro-fuzzy controller is provided to enhance the tipover stability of
mobile manipulators when they are subjected to predefined
trajectories for the end-effector and the vehicle. The controller
creates proper configurations for the manipulator to prevent the robot
from being overturned. The optimal configuration and thus the most
favorable control are obtained through soft computing approaches
including a combination of genetic algorithm, neural networks, and
fuzzy logic. The proposed algorithm, in this paper, is that a look-up
table is designed by employing the obtained values from the genetic
algorithm in order to minimize the performance index and by using
this data base, rule bases are designed for the ANFIS controller and
will be exerted on the actuators to enhance the tipover stability of the
mobile manipulator. A numerical example is presented to
demonstrate the effectiveness of the proposed algorithm.
Abstract: This paper investigates several factors affecting the
cost of capital for listed Romanian companies. Although there is a
large amount of literature investigating the drivers of the cost of
capital internationally, there is currently little evidence from
emergent markets. Based on a sample of 19 Romanian listed
companies followed by financial analysts for the years 2008-2010,
according to Thomson Reuters- I/B/E/S data base, the paper confirms
the international trends, showing that size, corporate governance
policies, and growth are negatively correlated with the cost of capital.
Abstract: This paper presents an adaptive differentiator
of sequential data based on the adaptive control theory. The
algorithm is applied to detect moving objects by estimating a
temporal gradient of sequential data at a specified pixel. We
adopt two nonlinear intensity functions to reduce the influence
of noises. The derivatives of the nonlinear intensity functions
are estimated by an adaptive observer with σ-modification
update law.
Abstract: All Text processing systems allow their users to
search a pattern of string from a given text. String matching is
fundamental to database and text processing applications. Every text
editor must contain a mechanism to search the current document for
arbitrary strings. Spelling checkers scan an input text for words in the
dictionary and reject any strings that do not match. We store our
information in data bases so that later on we can retrieve the same
and this retrieval can be done by using various string matching
algorithms. This paper is describing a new string matching algorithm
for various applications. A new algorithm has been designed with the
help of Rabin Karp Matcher, to improve string matching process.
Abstract: Classifying data hierarchically is an efficient approach
to analyze data. Data is usually classified into multiple categories, or
annotated with a set of labels. To analyze multi-labeled data, such
data must be specified by giving a set of labels as a semantic range.
There are some certain purposes to analyze data. This paper shows
which multi-labeled data should be the target to be analyzed for
those purposes, and discusses the role of a label against a set of
labels by investigating the change when a label is added to the set of
labels. These discussions give the methods for the advanced analysis
of multi-labeled data, which are based on the role of a label against
a semantic range.
Abstract: The sensitivity of UAVs to the atmospheric effects are
apparent. All the same the meteorological support for the UAVs
missions is often non-adequate or partly missing.
In our paper we show a new complex meteorological support
system for different types of UAVs pilots, specialists and decision
makers, too. The mentioned system has two important parts with
different forecasts approach such as the statistical and dynamical
ones.
The statistical prediction approach is based on a large
climatological data base and the special analog method which is able
to select similar weather situations from the mentioned data base to
apply them during the forecasting procedure.
The applied dynamic approach uses the specific WRF model runs
twice a day and produces 96 hours, high resolution weather forecast
for the UAV users over the Hungary. An easy to use web-based
system can give important weather information over the Carpathian
basin in Central-Europe. The mentioned products can be reached via
internet connection.
Abstract: The purpose of this paper is to study Database Models
to use them efficiently in E-commerce websites. In this paper we are
going to find a method which can save and retrieve information in Ecommerce
websites. Thus, semantic web applications can work with,
and we are also going to study different technologies of E-commerce
databases and we know that one of the most important deficits in
semantic web is the shortage of semantic data, since most of the
information is still stored in relational databases, we present an
approach to map legacy data stored in relational databases into the
Semantic Web using virtually any modern RDF query language, as
long as it is closed within RDF. To achieve this goal we study XML
structures for relational data bases of old websites and eventually we
will come up one level over XML and look for a map from relational
model (RDM) to RDF. Noting that a large number of semantic webs
get advantage of relational model, opening the ways which can be
converted to XML and RDF in modern systems (semantic web) is
important.
Abstract: We have developed a computer program consisting of
6 subtests assessing the children hand dexterity applicable in the
rehabilitation medicine. We have carried out a normative study on a
representative sample of 285 children aged from 7 to 15 (mean age
11.3) and we have proposed clinical standards for three age groups
(7-9, 9-11, 12-15 years). We have shown statistical significance of
differences among the corresponding mean values of the task time
completion. We have also found a strong correlation between the task
time completion and the age of the subjects, as well as we have
performed the test-retest reliability checks in the sample of 84
children, giving the high values of the Pearson coefficients for the
dominant and non-dominant hand in the range 0.740.97 and
0.620.93, respectively.
A new MATLAB-based programming tool aiming at analysis of
cardiologic RR intervals and blood pressure descriptors, is worked
out, too. For each set of data, ten different parameters are extracted: 2
in time domain, 4 in frequency domain and 4 in Poincaré plot
analysis. In addition twelve different parameters of baroreflex
sensitivity are calculated. All these data sets can be visualized in time
domain together with their power spectra and Poincaré plots. If
available, the respiratory oscillation curves can be also plotted for
comparison. Another application processes biological data obtained
from BLAST analysis.
Abstract: The purpose of this study is to introduce a new
interface program to calculate a dose distribution with Monte Carlo method in complex heterogeneous systems such as organs or tissues
in proton therapy. This interface program was developed under
MATLAB software and includes a friendly graphical user interface
with several tools such as image properties adjustment or results display. Quadtree decomposition technique was used as an image
segmentation algorithm to create optimum geometries from Computed Tomography (CT) images for dose calculations of proton
beam. The result of the mentioned technique is a number of nonoverlapped
squares with different sizes in every image. By this way
the resolution of image segmentation is high enough in and near
heterogeneous areas to preserve the precision of dose calculations
and is low enough in homogeneous areas to reduce the number of
cells directly. Furthermore a cell reduction algorithm can be used to combine neighboring cells with the same material. The validation of this method has been done in two ways; first, in comparison with experimental data obtained with 80 MeV proton beam in Cyclotron
and Radioisotope Center (CYRIC) in Tohoku University and second, in comparison with data based on polybinary tissue calibration method, performed in CYRIC. These results are presented in this paper. This program can read the output file of Monte Carlo code while region of interest is selected manually, and give a plot of dose distribution of proton beam superimposed onto the CT images.
Abstract: In this paper a new method is suggested for
distributed data-mining by the probability patterns. These patterns
use decision trees and decision graphs. The patterns are cared to be
valid, novel, useful, and understandable. Considering a set of
functions, the system reaches to a good pattern or better objectives.
By using the suggested method we will be able to extract the useful
information from massive and multi-relational data bases.
Abstract: Recently the use of data mining to scientific bibliographic data bases has been implemented to analyze the pathways of the knowledge or the core scientific relevances of a laureated novel or a country. This specific case of data mining has been named citation mining, and it is the integration of citation bibliometrics and text mining. In this paper we present an improved WEB implementation of statistical physics algorithms to perform the text mining component of citation mining. In particular we use an entropic like distance between the compression of text as an indicator of the similarity between them. Finally, we have included the recently proposed index h to characterize the scientific production. We have used this web implementation to identify users, applications and impact of the Mexican scientific institutions located in the State of Morelos.
Abstract: PARIS (Personal Archiving and Retrieving Image
System) is an experiment personal photograph library, which includes
more than 80,000 of consumer photographs accumulated within a
duration of approximately five years, metadata based on our proposed
MPEG-7 annotation architecture, Dozen Dimensional Digital Content
(DDDC), and a relational database structure. The DDDC architecture
is specially designed for facilitating the managing, browsing and
retrieving of personal digital photograph collections. In annotating
process, we also utilize a proposed Spatial and Temporal Ontology
(STO) designed based on the general characteristic of personal
photograph collections. This paper explains PRAIS system.