Abstract: Mammography is the most effective procedure for an
early diagnosis of the breast cancer. Nowadays, people are trying to
find a way or method to support as much as possible to the
radiologists in diagnosis process. The most popular way is now being
developed is using Computer-Aided Detection (CAD) system to
process the digital mammograms and prompt the suspicious region to
radiologist. In this paper, an automated CAD system for detection
and classification of massive lesions in mammographic images is
presented. The system consists of three processing steps: Regions-Of-
Interest detection, feature extraction and classification. Our CAD
system was evaluated on Mini-MIAS database consisting 322
digitalized mammograms. The CAD system-s performance is
evaluated using Receiver Operating Characteristics (ROC) and Freeresponse
ROC (FROC) curves. The archived results are 3.47 false
positives per image (FPpI) and sensitivity of 85%.
Abstract: This paper deals with the application for contentbased
image retrieval to extract color feature from natural images
stored in the image database by segmenting the image through
clustering. We employ a class of nonparametric techniques in which
the data points are regarded as samples from an unknown probability
density. Explicit computation of the density is avoided by using the
mean shift procedure, a robust clustering technique, which does not
require prior knowledge of the number of clusters, and does not
constrain the shape of the clusters. A non-parametric technique for
the recovery of significant image features is presented and
segmentation module is developed using the mean shift algorithm to
segment each image. In these algorithms, the only user set parameter
is the resolution of the analysis and either gray level or color images
are accepted as inputs. Extensive experimental results illustrate
excellent performance.
Abstract: Automatic keyphrase extraction is useful in efficiently
locating specific documents in online databases. While several
techniques have been introduced over the years, improvement on
accuracy rate is minimal. This research examines attribute scores for
author-supplied keyphrases to better understand how the scores affect
the accuracy rate of automatic keyphrase extraction. Five attributes
are chosen for examination: Term Frequency, First Occurrence, Last
Occurrence, Phrase Position in Sentences, and Term Cohesion
Degree. The results show that First Occurrence is the most reliable
attribute. Term Frequency, Last Occurrence and Term Cohesion
Degree display a wide range of variation but are still usable with
suggested tweaks. Only Phrase Position in Sentences shows a totally
unpredictable pattern. The results imply that the commonly used
ranking approach which directly extracts top ranked potential phrases
from candidate keyphrase list as the keyphrases may not be reliable.
Abstract: The main objective of our study is to collect data
about the profile of the asthmatic patients in Assam and thereby have
a comprehensive knowledge of the factors influencing the asthmatic
patients of the state and their medication pattern. We developed a
search strategy to find any publication about the community based
survey asthma related and used. These to search the MEDLINE
(1996 to current literature) CINAHL DOAJ pubmed databases using
the key phrases, Asthma, Respiratory disorders, Drug therapy of
Asthma, database decision support system and asthma. The
appropriate literature was printed out from the online source and
library (Journal) source. The study was conducted through a set of
structured and non-structured questionnaires targeted on the
asthmatic patients belonging to the rural and urban areas of Assam,
during the month of Dec 2006 to July 2007, 138 cases were studied
in Gauwathi Medical College & Hospital located in Bhangagarh,
Assam in India. The demographic characteristics a factor in 138
patients with asthma with allergic rhinitis (cases) gives the detail
profile of asthmatic patient-s distribution of Assam as classified on
the basis of age and sex. It is evident from the study that male
populations (66%) are more prone to asthma as compared to the
females (34%).Another striking features that emerged from this
survey is the maximum prevalence of asthma in the age group of 20-
30 years followed by infants belonging to the age group of 7 (0.05%)
0-10years among both male and female populations of Assam. The
high incidence of asthma in the age group of 20-30 years may
probably be due to the allergy arising out of sudden exposure to dust
and pollen which the children face while playing and going to the
school. The rural females in the age group of 30-40 years are more
prone to asthma than urban females in the same age group may be
due to sex differentiation among the tribal population of the state.
Pharmacists should educate the asthmatics how to use inhalers
considering growing menace of asthma in the state. Safer drugs
should be produced in the form of aerosol so that easy administration
by the asthmatic patients and physicians of the state is possible for
curing asthma. The health centers should be more equipped with the
medicines to cure asthma in the state like Assam.
Abstract: The aim of this paper is to investigate the effect of
mean size of industry on survival of new firms in East-Azarbaijan
province through 1981-2006 using hazard function. So the effect of
two variables including mean employment of industry and mean
capital of industry are investigated on firm's survival. The Industry &
Mine Ministry database has used for data gathering and the data are
analyzed using the semi-parametric cox regression model. The results
of this study shows that there is a meaningful negative relationship
between mean capital of industry and firm's survival, but the mean
employment of industry has no meaningful effect on survival of new
firms.
Abstract: On a such wide-area environment as a Grid, data
placement is an important aspect of distributed database systems. In
this paper, we address the problem of initial placement of database
no-replicated fragments in Grid architecture. We propose a graph
based approach that considers resource restrictions. The goal is to
optimize the use of computing, storage and communication
resources. The proposed approach is developed in two phases: in the
first phase, we perform fragment grouping using knowledge about
fragments dependency and, in the second phase, we determine an
efficient placement of the fragment groups on the Grid. We also
show, via experimental analysis that our approach gives solutions
that are close to being optimal for different databases and Grid
configurations.
Abstract: This research paper presents a framework on how to
build up malware dataset.Many researchers took longer time to
clean the dataset from any noise or to transform the dataset into a
format that can be used straight away for testing. Therefore, this
research is proposing a framework to help researchers to speed up
the malware dataset cleaningprocesses which later can be used for
testing. It is believed, an efficient malware dataset cleaning
processes, can improved the quality of the data, thus help to improve
the accuracy and the efficiency of the subsequent analysis. Apart
from that, an in-depth understanding of the malware taxonomy is
also important prior and during the dataset cleaning processes. A
new Trojan classification has been proposed to complement this
framework.This experiment has been conducted in a controlled lab
environment and using the dataset from VxHeavens dataset. This
framework is built based on the integration of static and dynamic
analyses, incident response method and knowledge database
discovery (KDD) processes.This framework can be used as the basis
guideline for malware researchers in building malware dataset.
Abstract: There are many issues that affect modeling and designing real-time databases. One of those issues is maintaining consistency between the actual state of the real-time object of the external environment and its images as reflected by all its replicas distributed over multiple nodes. The need to improve the scalability is another important issue. In this paper, we present a general framework to design a replicated real-time database for small to medium scale systems and maintain all timing constrains. In order to extend the idea for modeling a large scale database, we present a general outline that consider improving the scalability by using an existing static segmentation algorithm applied on the whole database, with the intent to lower the degree of replication, enables segments to have individual degrees of replication with the purpose of avoiding excessive resource usage, which all together contribute in solving the scalability problem for DRTDBS.
Abstract: This manuscript presents, palmprint recognition by
combining different texture extraction approaches with high accuracy.
The Region of Interest (ROI) is decomposed into different frequencytime
sub-bands by wavelet transform up-to two levels and only the
approximate image of two levels is selected, which is known as
Approximate Image ROI (AIROI). This AIROI has information of
principal lines of the palm. The Competitive Index is used as the
features of the palmprint, in which six Gabor filters of different
orientations convolve with the palmprint image to extract the orientation
information from the image. The winner-take-all strategy
is used to select dominant orientation for each pixel, which is
known as Competitive Index. Further, PCA is applied to select highly
uncorrelated Competitive Index features, to reduce the dimensions of
the feature vector, and to project the features on Eigen space. The
similarity of two palmprints is measured by the Euclidean distance
metrics. The algorithm is tested on Hong Kong PolyU palmprint
database. Different AIROI of different wavelet filter families are also
tested with the Competitive Index and PCA. AIROI of db7 wavelet
filter achievs Equal Error Rate (EER) of 0.0152% and Genuine
Acceptance Rate (GAR) of 99.67% on the palm database of Hong
Kong PolyU.
Abstract: The medical studies often require different methods
for parameters selection, as a second step of processing, after the
database-s designing and filling with information. One common
task is the selection of fields that act as risk factors using wellknown
methods, in order to find the most relevant risk factors and
to establish a possible hierarchy between them. Different methods
are available in this purpose, one of the most known being the
binary logistic regression. We will present the mathematical
principles of this method and a practical example of using it in the
analysis of the influence of 10 different psychiatric diagnostics
over 4 different types of offences (in a database made from 289
psychiatric patients involved in different types of offences).
Finally, we will make some observations about the relation
between the risk factors hierarchy established through binary
logistic regression and the individual risks, as well as the results of
Chi-squared test. We will show that the hierarchy built using the
binary logistic regression doesn-t agree with the direct order of risk
factors, even if it was naturally to assume this hypothesis as being
always true.
Abstract: A Ground Control System (GCS), which controls Unmanned Aerial Vehicles (UAVs) and monitors their missionrelated data, is one of the major components of UAVs. In fact, some traditional GCSs were built on an expensive, complicated hardware infrastructure with workstations and PCs. In contrast, a GCS on a portable device – such as an Android phone or tablet – takes advantage of its light-weight hardware and the rich User Interface supported by the Android Operating System. We implemented that kind of GCS and called it Ground System Software (GSS) in this paper. In operation, our GSS communicates with UAVs or other GSS via TCP/IP connection to get mission-related data, visualizes it on the device-s screen, and saves the data in its own database. Our study showed that this kind of system will become a potential instrument in UAV-related systems and this kind of topic will appear in many research studies in the near future.
Abstract: In this paper a combined feature selection method is
proposed which takes advantages of sample domain filtering,
resampling and feature subset evaluation methods to reduce
dimensions of huge datasets and select reliable features. This method
utilizes both feature space and sample domain to improve the process
of feature selection and uses a combination of Chi squared with
Consistency attribute evaluation methods to seek reliable features.
This method consists of two phases. The first phase filters and
resamples the sample domain and the second phase adopts a hybrid
procedure to find the optimal feature space by applying Chi squared,
Consistency subset evaluation methods and genetic search.
Experiments on various sized datasets from UCI Repository of
Machine Learning databases show that the performance of five
classifiers (Naïve Bayes, Logistic, Multilayer Perceptron, Best First
Decision Tree and JRIP) improves simultaneously and the
classification error for these classifiers decreases considerably. The
experiments also show that this method outperforms other feature
selection methods.
Abstract: In this paper a novel algorithm is proposed that integrates the process of fuzzy hierarchy generation and rule discovery for automated discovery of Production Rules with Fuzzy Hierarchy (PRFH) in large databases.A concept of frequency matrix (Freq) introduced to summarize large database that helps in minimizing the number of database accesses, identification and removal of irrelevant attribute values and weak classes during the fuzzy hierarchy generation.Experimental results have established the effectiveness of the proposed algorithm.
Abstract: This paper presents a method to detect multiple cracks
based on frequency information. When a structure is subjected to
dynamic or static loads, cracks may develop and the modal
frequencies of the cracked structure may change. To detect cracks in a
structure, we construct a high precision wavelet finite element (EF)
model of a certain structure using the B-spline wavelet on the interval
(BSWI). Cracks can be modeled by rotational springs and added to the
FE model. The crack detection database will be obtained by solving
that model. Then the crack locations and depths can be determined
based on the frequency information from the database. The
performance of the proposed method has been numerically verified by
a rotor example.
Abstract: This paper proposes a new approach to perform the
problem of real-time face detection. The proposed method combines
primitive Haar-Like feature and variance value to construct a new
feature, so-called Variance based Haar-Like feature. Face in image
can be represented with a small quantity of features using this
new feature. We used SVM instead of AdaBoost for training and
classification. We made a database containing 5,000 face samples
and 10,000 non-face samples extracted from real images for learning
purposed. The 5,000 face samples contain many images which have
many differences of light conditions. And experiments showed that
face detection system using Variance based Haar-Like feature and
SVM can be much more efficient than face detection system using
primitive Haar-Like feature and AdaBoost. We tested our method on
two Face databases and one Non-Face database. We have obtained
96.17% of correct detection rate on YaleB face database, which is
higher 4.21% than that of using primitive Haar-Like feature and
AdaBoost.
Abstract: The huge development of new technologies and the
apparition of open communication system more and more
sophisticated create a new challenge to protect digital content from
piracy. Digital watermarking is a recent research axis and a new
technique suggested as a solution to these problems. This technique
consists in inserting identification information (watermark) into
digital data (audio, video, image, databases...) in an invisible and
indelible manner and in such a way not to degrade original medium-s
quality. Moreover, we must be able to correctly extract the
watermark despite the deterioration of the watermarked medium (i.e
attacks). In this paper we propose a system for watermarking satellite
images. We chose to embed the watermark into frequency domain,
precisely the discrete wavelet transform (DWT). We applied our
algorithm on satellite images of Tunisian center. The experiments
show satisfying results. In addition, our algorithm showed an
important resistance facing different attacks, notably the compression
(JEPG, JPEG2000), the filtering, the histogram-s manipulation and
geometric distortions such as rotation, cropping, scaling.
Abstract: The speculative locking (SL) protocol extends the twophase locking (2PL) protocol to allow for parallelism among conflicting transactions. The adaptive speculative locking (ASL) protocol provided further enhancements and outperformed SL protocols under most conditions. Neither of these protocols consider the impact of network latency on the performance of the distributed database systems. We have studied the performance of ASL protocol taking into account the communication overhead. The results indicate that though system load can counter network latency, it can still become a bottleneck in many situations. The impact of latency on performance depends on many factors including the system resources. A flexible discrete event simulator was used as the testbed for this study.
Abstract: A comparison between the performance of Latin and
Arabic handwritten digits recognition problems is presented. The
performance of ten different classifiers is tested on two similar
Arabic and Latin handwritten digits databases. The analysis shows
that Arabic handwritten digits recognition problem is easier than that
of Latin digits. This is because the interclass difference in case of
Latin digits is smaller than in Arabic digits and variances in writing
Latin digits are larger. Consequently, weaker yet fast classifiers are
expected to play more prominent role in Arabic handwritten digits
recognition.
Abstract: In this paper, the noise maps for the area encircled by
the Second Ring Road in Riyadh city are developed based on real
measured data. Sound level meters, GPS receivers to determine
measurement position, a database program to manage the measured
data, and a program to develop the maps are used. A baseline noise
level has been established at each short-term site so subsequent
monitoring may be conducted to describe changes in Riyadh-s noise
environment. Short-term sites are used to show typical daytime and
nighttime noise levels at specific locations by short duration grab
sampling.
Abstract: This paper introduces and studies new indexing techniques for content-based queries in images databases. Indexing is the key to providing sophisticated, accurate and fast searches for queries in image data. This research describes a new indexing approach, which depends on linear modeling of signals, using bases for modeling. A basis is a set of chosen images, and modeling an image is a least-squares approximation of the image as a linear combination of the basis images. The coefficients of the basis images are taken together to serve as index for that image. The paper describes the implementation of the indexing scheme, and presents the findings of our extensive evaluation that was conducted to optimize (1) the choice of the basis matrix (B), and (2) the size of the index A (N). Furthermore, we compare the performance of our indexing scheme with other schemes. Our results show that our scheme has significantly higher performance.