Abstract: Introduction: There are multiple social, individual and
cultural factors that influence an individual’s decision to adopt family
planning methods especially among non-users in patriarchal societies
like Pakistan. Non-users, if targeted efficiently, can contribute
significantly to country’s CPR. A research study showed that nonusers
if convinced to adopt lactational amenorrhea method can shift
to long term methods in future. Research shows that if non users are
targeted efficiently a 59% reduction in unintended pregnancies in
Saharan Africa and South-Central and South-East Asia is anticipated.
Methods: We did secondary data analysis on Pakistan
Demographic Heath Survey (2012-13) dataset. Use of contraception
(never-use/ever-use) was the outcome variable. At univariate level
Chi-square/Fisher Exact test was used to assess relationship of
baseline covariates with contraception use. Then variables to be
incorporated in the model were checked for multicollinearity,
confounding and interaction. Then binary logistic regression (with an
urban-rural stratification) was done to find relationship between
contraception use and baseline demographic and social variables.
Results: The multivariate analyses of the study showed that
younger women (≤ 29 years)were more prone to be never users as
compared to those who were >30 years and this trend was seen in
urban areas (AOR 1.92, CI 1.453-2.536) as well as rural areas (AOR
1.809, CI 1.421-2.303). While looking at regional variation, women
from urban Sindh (AOR 1.548, CI 1.142-2.099) and urban
Balochistan (AOR 2.403, CI 1.504-3.839) had more never users as
compared to other urban regions. Women in the rich wealth quintile
were more never users and this was seen both in urban and rural
localities (urban (AOR 1.106 CI .753-1.624); rural areas (AOR 1.162,
CI .887-1.524)) even though these were not statistically significant.
Women idealizing more children (>4) are more never users as
compared to those idealizing less children in both urban (AOR 1.854,
CI 1.275-2.697) and rural areas (AOR 2.101, CI 1.514-2.916).
Women who never lost a pregnancy were more inclined to be nonusers
in rural areas (AOR 1.394, CI 1.127-1.723) .Women familiar
with only traditional or no method had more never users in rural areas
(AOR 1.717, CI 1.127-1.723) but in urban areas it wasn’t significant.
Women unaware of Lady Health Worker’s presence in their area
were more never users especially in rural areas (AOR 1.276, CI
1.014-1.607). Women who did not visit any care provider were more
never users (urban (AOR 11.738, CI 9.112-15.121) rural areas (AOR
7.832, CI 6.243-9.826)).
Discussion/Conclusion: This study concluded that government,
policy makers and private sector family planning programs should
focus on the untapped pool of never users (younger women from underserved provinces, in higher wealth quintiles, who desire more
children.). We need to make sure to cover catchment areas where
there are less LHWs and less providers as ignorance to modern
methods and never been visited by an LHW are important
determinants of never use. This all is in sync with previous literate
from similar developing countries.
Abstract: In the past few years, the amount of malicious software
increased exponentially and, therefore, machine learning algorithms
became instrumental in identifying clean and malware files through
(semi)-automated classification. When working with very large
datasets, the major challenge is to reach both a very high malware
detection rate and a very low false positive rate. Another challenge
is to minimize the time needed for the machine learning algorithm to
do so. This paper presents a comparative study between different
machine learning techniques such as linear classifiers, ensembles,
decision trees or various hybrids thereof. The training dataset consists
of approximately 2 million clean files and 200.000 infected files,
which is a realistic quantitative mixture. The paper investigates the
above mentioned methods with respect to both their performance
(detection rate and false positive rate) and their practicability.
Abstract: Introduction: Researchers globally have strived to explore diverse factors that augment the continuation and uptake of family planning methods. Clients’ satisfaction is one of the core determinants facilitating continuation of family planning methods. There is a major debate yet scanty evidence to contrast public and private sectors with respect to client satisfaction. The objective of this study is to compare quality-of-care provided by public and private sectors of Pakistan through a client satisfaction lens. Methods: We used Pakistan Demographic Heath Survey 2012-13 dataset on 3133 women. Ten different multivariate models were made. to explore the relationship between client satisfaction and dependent outcome after adjusting for all known confounding factors and results are presented as OR and AOR (95% CI). Results: Multivariate analyses showed that clients were less satisfied in contraceptive provision from private sector as compared to public sector (AOR 0.92, 95% CI 0.63-1.68) even though the result was not statistically significant. Clients were more satisfied from private sector as compared to the public sector with respect to other determinants of quality-of-care follow-up care (AOR 3.29, 95% CI 1.95-5.55), infection prevention (AOR 2.41, 95% CI 1.60-3.62), counseling services (AOR 2.01, 95% CI 1.27-3.18, timely treatment (AOR 3.37, 95% CI 2.20-5.15), attitude of staff (AOR 2.23, 95% CI 1.50-3.33), punctuality of staff (AOR 2.28, 95% CI 1.92-4.13), timely referring (AOR 2.34, 95% CI 1.63-3.35), staff cooperation (AOR 1.75, 95% CI 1.22-2.51) and complications handling (AOR 2.27, 95% CI 1.56-3.29). Discussion: Public sector has successfully attained substantial satisfaction levels with respect to provision of contraceptives, but it contrasts previous literature from a multi country studies. Our study though in is concordance with a study from Tanzania where public sector was more likely to offer family planning services to clients as compared to private facilities. Conclusion: In majority of the developing countries, public sector is more involved in FP service provision; however, in Pakistan clients’ satisfaction in private sector is more, which opens doors for public-private partnerships and collaboration in the near future.
Abstract: This study analyzes the innovative orientation of the
Croatian entrepreneurs. Innovative orientation is represented by the
perceived extent to which an entrepreneur’s product or service or
technology is new, and no other businesses offer the same product.
The sample is extracted from the GEM Croatia Adult Population
Survey dataset for the years 2003-2013. We apply descriptive
statistics, t-test, Chi-square test and logistic regression. Findings
indicate that innovative orientations vary with personal, firm, meso
and macro level variables, and between different stages in
entrepreneurship process. Significant predictors are occupation of the
entrepreneurs, size of the firm and export aspiration for both early
stage and established entrepreneurs. In addition, fear of failure,
expecting to start a new business and seeing an entrepreneurial career
as a desirable choice are predictors of innovative orientation among
early stage entrepreneurs.
Abstract: The effects of hypertension are often lethal thus its
early detection and prevention is very important for everybody. In
this paper, a neural network (NN) model was developed and trained
based on a dataset of hypertension causative parameters in order to
forecast the likelihood of occurrence of hypertension in patients. Our
research goal was to analyze the potential of the presented NN to
predict, for a period of time, the risk of hypertension or the risk of
developing this disease for patients that are or not currently
hypertensive. The results of the analysis for a given patient can
support doctors in taking pro-active measures for averting the
occurrence of hypertension such as recommendations regarding the
patient behavior in order to lower his hypertension risk. Moreover,
the paper envisages a set of three example scenarios in order to
determine the age when the patient becomes hypertensive, i.e.
determine the threshold for hypertensive age, to analyze what
happens if the threshold hypertensive age is set to a certain age and
the weight of the patient if being varied, and, to set the ideal weight
for the patient and analyze what happens with the threshold of
hypertensive age.
Abstract: The detection of moving objects from a video image
sequences is very important for object tracking, activity recognition,
and behavior understanding in video surveillance.
The most used approach for moving objects detection / tracking is
background subtraction algorithms. Many approaches have been
suggested for background subtraction. But, these are illumination
change sensitive and the solutions proposed to bypass this problem
are time consuming.
In this paper, we propose a robust yet computationally efficient
background subtraction approach and, mainly, focus on the ability to
detect moving objects on dynamic scenes, for possible applications in
complex and restricted access areas monitoring, where moving and
motionless persons must be reliably detected. It consists of three
main phases, establishing illumination changes invariance,
background/foreground modeling and morphological analysis for
noise removing.
We handle illumination changes using Contrast Limited Histogram
Equalization (CLAHE), which limits the intensity of each pixel to
user determined maximum. Thus, it mitigates the degradation due to
scene illumination changes and improves the visibility of the video
signal. Initially, the background and foreground images are extracted
from the video sequence. Then, the background and foreground
images are separately enhanced by applying CLAHE.
In order to form multi-modal backgrounds we model each channel
of a pixel as a mixture of K Gaussians (K=5) using Gaussian Mixture
Model (GMM). Finally, we post process the resulting binary
foreground mask using morphological erosion and dilation
transformations to remove possible noise.
For experimental test, we used a standard dataset to challenge the
efficiency and accuracy of the proposed method on a diverse set of
dynamic scenes.
Abstract: This paper presents two techniques, local feature
extraction using image spectrum and low frequency spectrum
modelling using GMM to capture the underlying statistical
information to improve the performance of face recognition
system. Local spectrum features are extracted using overlap sub
block window that are mapped on the face image. For each of this
block, spatial domain is transformed to frequency domain using
DFT. A low frequency coefficient is preserved by discarding high
frequency coefficients by applying rectangular mask on the
spectrum of the facial image. Low frequency information is non-
Gaussian in the feature space and by using combination of several
Gaussian functions that has different statistical properties, the best
feature representation can be modelled using probability density
function. The recognition process is performed using maximum
likelihood value computed using pre-calculated GMM components.
The method is tested using FERET datasets and is able to achieved
92% recognition rates.
Abstract: Governments collect and produce large amounts of
data. Increasingly, governments worldwide have started to implement
open data initiatives and also launch open data portals to enable the
release of these data in open and reusable formats. Therefore, a large
number of open data repositories, catalogues and portals have been
emerging in the world. The greater availability of interoperable and
linkable open government data catalyzes secondary use of such data,
so they can be used for building useful applications which leverage
their value, allow insight, provide access to government services, and
support transparency. The efficient development of successful open
data portals makes it necessary to evaluate them systematic, in order
to understand them better and assess the various types of value they
generate, and identify the required improvements for increasing this
value. Thus, the attention of this paper is directed particularly to the
field of open data portals. The main aim of this paper is to compare
the selected open data portals on the national level using content
analysis and propose a new evaluation framework, which further
improves the quality of these portals. It also establishes a set of
considerations for involving businesses and citizens to create eservices
and applications that leverage on the datasets available from
these portals.
Abstract: The web’s increased popularity has included a huge
amount of information, due to which automated web page
classification systems are essential to improve search engines’
performance. Web pages have many features like HTML or XML
tags, hyperlinks, URLs and text contents which can be considered
during an automated classification process. It is known that Webpage
classification is enhanced by hyperlinks as it reflects Web page
linkages. The aim of this study is to reduce the number of features to
be used to improve the accuracy of the classification of web pages. In
this paper, a novel feature selection method using an improved
Particle Swarm Optimization (PSO) using principle of evolution is
proposed. The extracted features were tested on the WebKB dataset
using a parallel Neural Network to reduce the computational cost.
Abstract: The study of the electrical signals produced by neural
activities of human brain is called Electroencephalography. In this
paper, we propose an automatic and efficient EEG signal
classification approach. The proposed approach is used to classify the
EEG signal into two classes: epileptic seizure or not. In the proposed
approach, we start with extracting the features by applying Discrete
Wavelet Transform (DWT) in order to decompose the EEG signals
into sub-bands. These features, extracted from details and
approximation coefficients of DWT sub-bands, are used as input to
Principal Component Analysis (PCA). The classification is based on
reducing the feature dimension using PCA and deriving the supportvectors
using Support Vector Machine (SVM). The experimental are
performed on real and standard dataset. A very high level of
classification accuracy is obtained in the result of classification.
Abstract: Image search engines rely on the surrounding textual
keywords for the retrieval of images. It is a tedious work for the
search engines like Google and Bing to interpret the user’s search
intention and to provide the desired results. The recent researches
also state that the Google image search engines do not work well on
all the images. Consequently, this leads to the emergence of efficient
image retrieval technique, which interprets the user’s search intention
and shows the desired results. In order to accomplish this task, an
efficient image re-ranking framework is required. Sequentially, to
provide best image retrieval, the new image re-ranking framework is
experimented in this paper. The implemented new image re-ranking
framework provides best image retrieval from the image dataset by
making use of re-ranking of retrieved images that is based on the
user’s desired images. This is experimented in two sections. One is
offline section and other is online section. In offline section, the reranking
framework studies differently (reference classes or Semantic
Spaces) for diverse user query keywords. The semantic signatures get
generated by combining the textual and visual features of the images.
In the online section, images are re-ranked by comparing the
semantic signatures that are obtained from the reference classes with
the user specified image query keywords. This re-ranking
methodology will increases the retrieval image efficiency and the
result will be effective to the user.
Abstract: The use of eXtensible Markup Language (XML) in
web, business and scientific databases lead to the development of
methods, techniques and systems to manage and analyze XML data.
Semi-structured documents suffer due to its heterogeneity and
dimensionality. XML structure and content mining represent
convergence for research in semi-structured data and text mining. As
the information available on the internet grows drastically, extracting
knowledge from XML documents becomes a harder task. Certainly,
documents are often so large that the data set returned as answer to a
query may also be very big to convey the required information. To
improve the query answering, a Semantic Tree Based Association
Rule (STAR) mining method is proposed. This method provides
intentional information by considering the structure, content and the
semantics of the content. The method is applied on Reuter’s dataset
and the results show that the proposed method outperforms well.
Abstract: Red blood cells (RBC) are the most common types of
blood cells and are the most intensively studied in cell biology. The
lack of RBCs is a condition in which the amount of hemoglobin level
is lower than normal and is referred to as “anemia”. Abnormalities in
RBCs will affect the exchange of oxygen. This paper presents a
comparative study for various techniques for classifying the RBCs as
normal or abnormal (anemic) using WEKA. WEKA is an open
source consists of different machine learning algorithms for data
mining applications. The algorithms tested are Radial Basis Function
neural network, Support vector machine, and K-Nearest Neighbors
algorithm. Two sets of combined features were utilized for
classification of blood cells images. The first set, exclusively consist
of geometrical features, was used to identify whether the tested blood
cell has a spherical shape or non-spherical cells. While the second
set, consist mainly of textural features was used to recognize the
types of the spherical cells. We have provided an evaluation based on
applying these classification methods to our RBCs image dataset
which were obtained from Serdang Hospital - Malaysia, and
measuring the accuracy of test results. The best achieved
classification rates are 97%, 98%, and 79% for Support vector
machines, Radial Basis Function neural network, and K-Nearest
Neighbors algorithm respectively.
Abstract: Artificial Immune Systems (AIS), inspired by the
human immune system, are algorithms and mechanisms which are
self-adaptive and self-learning classifiers capable of recognizing and
classifying by learning, long-term memory and association. Unlike
other human system inspired techniques like genetic algorithms and
neural networks, AIS includes a range of algorithms modeling on
different immune mechanism of the body. In this paper, a mechanism
of a human immune system based on apoptosis is adopted to build an
Intrusion Detection System (IDS) to protect computer networks.
Features are selected from network traffic using Fisher Score. Based
on the selected features, the record/connection is classified as either
an attack or normal traffic by the proposed methodology. Simulation
results demonstrates that the proposed AIS based on apoptosis
performs better than existing AIS for intrusion detection.
Abstract: Thanks to informational technologies development
every sphere of economics is becoming more and more datacentralized
as people are generating huge datasets containing
information on any aspect of their life. Applying research of such
data to human resources management allows getting scarce statistics
on labor market state including salary expectations and potential
employees’ typical career behavior, and this information can become
a reliable basis for management decisions.
The following article presents results of career behavior research
based on freely accessible resume data. Information used for study is
much wider than one usually uses in human resources surveys. That
is why there is enough data for statistically significant results even
for subgroups analysis.
Abstract: Microarray technology is universally used in the study
of disease diagnosis using gene expression levels. The main
shortcoming of gene expression data is that it includes thousands of
genes and a small number of samples. Abundant methods and
techniques have been proposed for tumor classification using
microarray gene expression data. Feature or gene selection methods
can be used to mine the genes that directly involve in the
classification and to eliminate irrelevant genes. In this paper
statistical measures like T-Statistics, Signal-to-Noise Ratio (SNR)
and F-Statistics are used to rank the genes. The ranked genes are used
for further classification. Particle Swarm Optimization (PSO)
algorithm and Shuffled Frog Leaping (SFL) algorithm are used to
find the significant genes from the top-m ranked genes. The Naïve
Bayes Classifier (NBC) is used to classify the samples based on the
significant genes. The proposed work is applied on Lung and Ovarian
datasets. The experimental results show that the proposed method
achieves 100% accuracy in all the three datasets and the results are
compared with previous works.
Abstract: This paper presents a real-time visualization technique
and filtering of classified LiDAR point clouds. The visualization is
capable of displaying filtered information organized in layers by the
classification attribute saved within LiDAR datasets. We explain the
used data structure and data management, which enables real-time
presentation of layered LiDAR data. Real-time visualization is
achieved with LOD optimization based on the distance from the
observer without loss of quality. The filtering process is done in two
steps and is entirely executed on the GPU and implemented using
programmable shaders.
Abstract: Different strategies and tools are available at the oil
and gas industry for detecting and analyzing tension and possible
fractures in borehole walls. Most of these techniques are based on
manual observation of the captured borehole images. While this
strategy may be possible and convenient with small images and few
data, it may become difficult and suitable to errors when big
databases of images must be treated. While the patterns may differ
among the image area, depending on many characteristics (drilling
strategy, rock components, rock strength, etc.). In this work we
propose the inclusion of data-mining classification strategies in order
to create a knowledge database of the segmented curves. These
classifiers allow that, after some time using and manually pointing
parts of borehole images that correspond to tension regions and
breakout areas, the system will indicate and suggest automatically
new candidate regions, with higher accuracy. We suggest the use of
different classifiers methods, in order to achieve different knowledge
dataset configurations.
Abstract: Over the past era, there have been a lot of efforts and
studies are carried out in growing proficient tools for performing
various tasks in big data. Recently big data have gotten a lot of
publicity for their good reasons. Due to the large and complex
collection of datasets it is difficult to process on traditional data
processing applications. This concern turns to be further mandatory
for producing various tools in big data. Moreover, the main aim of
big data analytics is to utilize the advanced analytic techniques
besides very huge, different datasets which contain diverse sizes from
terabytes to zettabytes and diverse types such as structured or
unstructured and batch or streaming. Big data is useful for data sets
where their size or type is away from the capability of traditional
relational databases for capturing, managing and processing the data
with low-latency. Thus the out coming challenges tend to the
occurrence of powerful big data tools. In this survey, a various
collection of big data tools are illustrated and also compared with the
salient features.
Abstract: Consumer-to-Consumer (C2C) E-commerce has been
growing at a very high speed in recent years. Since identical or
nearly-same kinds of products compete one another by relying on
keyword search in C2C E-commerce, some sellers describe their
products with spam keywords that are popular but are not related to
their products. Though such products get more chances to be retrieved
and selected by consumers than those without spam keywords,
the spam keywords mislead the consumers and waste their time.
This problem has been reported in many commercial services like
ebay and taobao, but there have been little research to solve this
problem. As a solution to this problem, this paper proposes a method
to classify whether keywords of a product are spam or not. The
proposed method assumes that a keyword for a given product is
more reliable if the keyword is observed commonly in specifications
of products which are the same or the same kind as the given
product. This is because that a hierarchical category of a product
in general determined precisely by a seller of the product and so is
the specification of the product. Since higher layers of the hierarchical
category represent more general kinds of products, a reliable degree
is differently determined according to the layers. Hence, reliable
degrees from different layers of a hierarchical category become
features for keywords and they are used together with features only
from specifications for classification of the keywords. Support Vector
Machines are adopted as a basic classifier using the features, since
it is powerful, and widely used in many classification tasks. In
the experiments, the proposed method is evaluated with a golden
standard dataset from Yi-han-wang, a Chinese C2C E-commerce,
and is compared with a baseline method that does not consider
the hierarchical category. The experimental results show that the
proposed method outperforms the baseline in F1-measure, which
proves that spam keywords are effectively identified by a hierarchical
category in C2C E-commerce.