Abstract: The theory of Groebner Bases, which has recently been
honored with the ACM Paris Kanellakis Theory and Practice Award,
has become a crucial building block to computer algebra, and is
widely used in science, engineering, and computer science. It is wellknown
that Groebner bases computation is EXP-SPACE in a general
setting. In this paper, we give an algorithm to show that Groebner
bases computation is P-SPACE in Boolean rings. We also show that
with this discovery, the Groebner bases method can theoretically be
as efficient as other methods for automated verification of hardware
and software. Additionally, many useful and interesting properties of
Groebner bases including the ability to efficiently convert the bases
for different orders of variables making Groebner bases a promising
method in automated verification.
Abstract: In this paper we investigated a number of the Internet
congestion control algorithms that has been developed in the last few
years. It was obviously found that many of these algorithms were
designed to deal with the Internet traffic merely as a train of
consequent packets. Other few algorithms were specifically tailored
to handle the Internet congestion caused by running media traffic that
represents audiovisual content. This later set of algorithms is
considered to be aware of the nature of this media content. In this
context we briefly explained a number of congestion control
algorithms and hence categorized them into the two following
categories: i) Media congestion control algorithms. ii) Common
congestion control algorithms. We hereby recommend the usage of
the media congestion control algorithms for the reason of being
media content-aware rather than the other common type of
algorithms that blindly manipulates such traffic. We showed that the
spread of such media content-aware algorithms over Internet will
lead to better congestion control status in the coming years. This is
due to the observed emergence of the era of digital convergence
where the media traffic type will form the majority of the Internet
traffic.
Abstract: In this paper, a reliable cooperative multipath routing
algorithm is proposed for data forwarding in wireless sensor networks
(WSNs). In this algorithm, data packets are forwarded towards the
base station (BS) through a number of paths, using a set of relay
nodes. In addition, the Rayleigh fading model is used to calculate
the evaluation metric of links. Here, the quality of reliability is
guaranteed by selecting optimal relay set with which the probability
of correct packet reception at the BS will exceed a predefined
threshold. Therefore, the proposed scheme ensures reliable packet
transmission to the BS. Furthermore, in the proposed algorithm,
energy efficiency is achieved by energy balancing (i.e. minimizing
the energy consumption of the bottleneck node of the routing path)
at the same time. This work also demonstrates that the proposed
algorithm outperforms existing algorithms in extending longevity of
the network, with respect to the quality of reliability. Given this, the
obtained results make possible reliable path selection with minimum
energy consumption in real time.
Abstract: Neighborhood Rough Sets (NRS) has been proven to
be an efficient tool for heterogeneous attribute reduction. However,
most of researches are focused on dealing with complete and noiseless
data. Factually, most of the information systems are noisy, namely,
filled with incomplete data and inconsistent data. In this paper, we
introduce a generalized neighborhood rough sets model, called
VPTNRS, to deal with the problem of heterogeneous attribute
reduction in noisy system. We generalize classical NRS model with
tolerance neighborhood relation and the probabilistic theory.
Furthermore, we use the neighborhood dependency to evaluate the
significance of a subset of heterogeneous attributes and construct a
forward greedy algorithm for attribute reduction based on it.
Experimental results show that the model is efficient to deal with noisy
data.
Abstract: Clustering is a very well known technique in data mining. One of the most widely used clustering techniques is the k-means algorithm. Solutions obtained from this technique are dependent on the initialization of cluster centers. In this article we propose a new algorithm to initialize the clusters. The proposed algorithm is based on finding a set of medians extracted from a dimension with maximum variance. The algorithm has been applied to different data sets and good results are obtained.
Abstract: A challenging problem in radar signal processing is to
achieve reliable target detection in the presence of interferences. In
this paper, we propose a novel algorithm for automatic censoring of
radar interfering targets in log-normal clutter. The proposed
algorithm, termed the forward automatic censored cell averaging
detector (F-ACCAD), consists of two steps: removing the corrupted
reference cells (censoring) and the actual detection. Both steps are
performed dynamically by using a suitable set of ranked cells to
estimate the unknown background level and set the adaptive
thresholds accordingly. The F-ACCAD algorithm does not require
any prior information about the clutter parameters nor does it require
the number of interfering targets. The effectiveness of the F-ACCAD
algorithm is assessed by computing, using Monte Carlo simulations,
the probability of censoring and the probability of detection in
different background environments.
Abstract: Atlantic herring (Clupea harengus) is an important
commercial fish and shows to be more and more demanded for
human consumption. Therefore, it is very important to find good
methods for monitoring the freshness of the fish in order to keep it in
the best quality for human consumption. In this study, the fish was
stored in ice up to 2 weeks. Quality changes during storage were
assessed by the Quality Index Method (QIM), quantitative
descriptive analysis (QDA) and Torry scheme, by texture
measurements: puncture tests and Texture Profile Analysis (TPA)
tests on texture analyzer TA.XT2i, and by electronic nose (e-nose)
measurements using FreshSense instrument. Storage time of herring
in ice could be estimated by QIM with ± 2 days using 5 herring per
lot. No correlation between instrumental texture parameters and
storage time or between sensory and instrumental texture variables
was found. E-nose measurements could be use to detect the onset of
spoilage.
Abstract: Every commercial bank optimises its asset portfolio
depending on the profitability of assets and chosen or imposed
constraints. This paper proposes and applies a stylized model for
optimising banks' asset and liability structure, reflecting profitability
of different asset categories and their risks as well as costs associated
with different liability categories and reserve requirements. The level
of detail for asset and liability categories is chosen to create a
suitably parsimonious model and to include the most important
categories in the model. It is shown that the most appropriate
optimisation criterion for the model is the maximisation of the ratio
of net interest income to assets. The maximisation of this ratio is
subject to several constraints. Some are accounting identities or
dictated by legislative requirements; others vary depending on the
market objectives for a particular bank. The model predicts variable
amount of assets allocated to loan provision.
Abstract: The demand of the energy management systems (EMS) set forth by modern power systems requires fast energy management systems. Contingency analysis is among the functions in EMS which is time consuming. In order to handle this limitation, this paper introduces agent based technology in the contingency analysis. The main function of agents is to speed up the performance. Negotiations process in decision making is explained and the issue set forth is the minimization of the operating costs. The IEEE 14 bus system and its line outage have been used in the research and simulation results are presented.
Abstract: Heart-s electric field can be measured anywhere on
the surface of the body (ECG). When individuals touch, one person-s
ECG signal can be registered in other person-s EEG and elsewhere
on his body. Now, the aim of this study was to test the hypothesis
that physical contact (hand-holding) of two persons changes their
heart rate variability. Subjects were sixteen healthy female (age: 20-
26) which divided into eight sets. In each sets, we had two friends
that they passed intimacy test of J.sternberg. ECG of two subjects
(each set) acquired for 5 minutes before hand-holding (as control
group) and 5 minutes during they held their hands (as experimental
group). Then heart rate variability signals were extracted from
subjects' ECG and analyzed in linear feature space (time and
frequency domain) and nonlinear feature space. Considering the
results, we conclude that physical contact (hand-holding of two
friends) increases parasympathetic activity, as indicate by increase
SD1, SD1/SD2, HF and MF power (p
Abstract: The paper describes a new approach for fingerprint
classification, based on the distribution of local features (minute
details or minutiae) of the fingerprints. The main advantage is that
fingerprint classification provides an indexing scheme to facilitate
efficient matching in a large fingerprint database. A set of rules based
on heuristic approach has been proposed. The area around the core
point is treated as the area of interest for extracting the minutiae
features as there are substantial variations around the core point as
compared to the areas away from the core point. The core point in a
fingerprint has been located at a point where there is maximum
curvature. The experimental results report an overall average
accuracy of 86.57 % in fingerprint classification.
Abstract: The proposed paper examines strategies whose aim is
to counter the all too often sighted process of abandonment that
characterizes contemporary cities. The city of Nicosia in Cyprus is
used as an indicative case study, whereby several recent projects are
presented as capitalizing on traditional cultural assets to revive the
downtown. The reuse of existing building stock as museums,
performing arts centers and theaters but also as in the form of various
housing typologies is geared to strengthen the ranks of local residents
and to spur economic growth. Unlike the examples from the 1960s,
the architecture of more recent adaptive reuse for urban regeneration
seems to be geared in reinforcing a connection to the city where the
buildings often reflect the characteristics of their urban context.
Abstract: Throughput is an important measure of performance of production system. Analyzing and modeling of production throughput is complex in today-s dynamic production systems due to uncertainties of production system. The main reasons are that uncertainties are materialized when the production line faces changes in setup time, machinery break down, lead time of manufacturing, and scraps. Besides, demand changes are fluctuating from time to time for each product type. These uncertainties affect the production performance. This paper proposes Bayesian inference for throughput modeling under five production uncertainties. Bayesian model utilized prior distributions related to previous information about the uncertainties where likelihood distributions are associated to the observed data. Gibbs sampling algorithm as the robust procedure of Monte Carlo Markov chain was employed for sampling unknown parameters and estimating the posterior mean of uncertainties. The Bayesian model was validated with respect to convergence and efficiency of its outputs. The results presented that the proposed Bayesian models were capable to predict the production throughput with accuracy of 98.3%.
Abstract: Mechanical interaction between endothelial cells (ECs) and the extracellular matrix (or collagen gel) is known to influence the sprouting response of endothelial cells during angiogenesis. This influence is believed to impact on the capability of endothelial cells to sense soluble chemical cues. Quantitative analysis of endothelial-cell-mediated displacement of the collagen gel provides a means to explore this mechanical interaction. Existing analysis in this context is generally limited to 2D settings. In this paper, we investigate the mechanical interaction between endothelial cells and the extracellular matrix in terms of the endothelial-cellmediated displacement of the collagen gel in both 2D and 3D. Digital image correlation and Digital volume correlation are applied on confocal reflectance image stacks to analyze cell-mediated displacement of the gel. The skeleton of the sprout is extracted from phase contrast images and superimposed on the displacement field to further investigate the link between the development of the sprout and the displacement of the gel.
Abstract: Timetabling problems are often hard and timeconsuming
to solve. Most of the methods of solving them concern
only one problem instance or class. This paper describes a universal
method for solving large, highly constrained timetabling problems
from different domains. The solution is based on evolutionary
algorithm-s framework and operates on two levels – first-level
evolutionary algorithm tries to find a solution basing on given set of
operating parameters, second-level algorithm is used to establish
those parameters. Tabu search is employed to speed up the solution
finding process on first level. The method has been used to solve
three different timetabling problems with promising results.
Abstract: Hepatitis C is an infectious disease transmitted by
blood and due to hepatitis C virus (HCV), which attacks the liver.
The infection is characterized by liver inflammation (hepatitis) that is
often asymptomatic but can progress to chronic hepatitis and later
cirrhosis and liver cancer. Our problem tends to highlight on the one
hand the prevalence of infectious disease in the population of the
region of Batna and on other hand the biological characteristics of
this disease by a screening and a specific diagnosis based on
serological tests, liver checkup (measurement of haematological and
biochemical parameters).
The results showed:
The serology of hepatitis C establishes the diagnosis of infection
with hepatitis C. In this study and with the serological test, 24 cases
of the disease of hepatitis C were found in 1000 suspected cases (7
cases with normal transaminases and 17 cases with elevated
transaminases). The prevalence of this disease in this study
population was 2.4%.
The presence of hepatitis C disrupts liver function including the
onset of cytolysis, cholestasis, jaundice, thrombocytopenia, and
coagulation disorders.
Abstract: Human activity is a major concern in a wide variety of
applications, such as video surveillance, human computer interface
and face image database management. Detecting and recognizing
faces is a crucial step in these applications. Furthermore, major
advancements and initiatives in security applications in the past years
have propelled face recognition technology into the spotlight. The
performance of existing face recognition systems declines significantly
if the resolution of the face image falls below a certain level.
This is especially critical in surveillance imagery where often, due to
many reasons, only low-resolution video of faces is available. If these
low-resolution images are passed to a face recognition system, the
performance is usually unacceptable. Hence, resolution plays a key
role in face recognition systems. In this paper we introduce a new
low resolution face recognition system based on mixture of expert
neural networks. In order to produce the low resolution input images
we down-sampled the 48 × 48 ORL images to 12 × 12 ones using
the nearest neighbor interpolation method and after that applying
the bicubic interpolation method yields enhanced images which is
given to the Principal Component Analysis feature extractor system.
Comparison with some of the most related methods indicates that
the proposed novel model yields excellent recognition rate in low
resolution face recognition that is the recognition rate of 100% for
the training set and 96.5% for the test set.
Abstract: Obtaining labeled data in supervised learning is often
difficult and expensive, and thus the trained learning algorithm tends
to be overfitting due to small number of training data. As a result,
some researchers have focused on using unlabeled data which may
not necessary to follow the same generative distribution as the labeled
data to construct a high-level feature for improving performance on
supervised learning tasks. In this paper, we investigate the impact of
the relationship between unlabeled and labeled data for classification
performance. Specifically, we will apply difference unlabeled data
which have different degrees of relation to the labeled data for
handwritten digit classification task based on MNIST dataset. Our
experimental results show that the higher the degree of relation
between unlabeled and labeled data, the better the classification
performance. Although the unlabeled data that is completely from
different generative distribution to the labeled data provides the lowest
classification performance, we still achieve high classification performance.
This leads to expanding the applicability of the supervised
learning algorithms using unsupervised learning.
Abstract: Gene expression profiling is rapidly evolving into a
powerful technique for investigating tumor malignancies. The
researchers are overwhelmed with the microarray-based platforms
and methods that confer them the freedom to conduct large-scale
gene expression profiling measurements. Simultaneously,
investigations into cross-platform integration methods have started
gaining momentum due to their underlying potential to help
comprehend a myriad of broad biological issues in tumor diagnosis,
prognosis, and therapy. However, comparing results from different
platforms remains to be a challenging task as various inherent
technical differences exist between the microarray platforms. In this
paper, we explain a simple ratio-transformation method, which can
provide some common ground for cDNA and Affymetrix platform
towards cross-platform integration. The method is based on the
characteristic data attributes of Affymetrix- and cDNA- platform. In
the work, we considered seven childhood leukemia patients and their
gene expression levels in either platform. With a dataset of 822
differentially expressed genes from both these platforms, we carried
out a specific ratio-treatment to Affymetrix data, which subsequently
showed an improvement in the relationship with the cDNA data.
Abstract: In comparison to the original SVM, which involves a
quadratic programming task; LS–SVM simplifies the required
computation, but unfortunately the sparseness of standard SVM is
lost. Another problem is that LS-SVM is only optimal if the training
samples are corrupted by Gaussian noise. In Least Squares SVM
(LS–SVM), the nonlinear solution is obtained, by first mapping the
input vector to a high dimensional kernel space in a nonlinear
fashion, where the solution is calculated from a linear equation set. In
this paper a geometric view of the kernel space is introduced, which
enables us to develop a new formulation to achieve a sparse and
robust estimate.