Abstract: The overlay approach has been widely used by many service providers for Traffic Engineering (TE) in large Internet backbones. In the overlay approach, logical connections are set up between edge nodes to form a full mesh virtual network on top of the physical topology. IP routing is then run over the virtual network. Traffic engineering objectives are achieved through carefully routing logical connections over the physical links. Although the overlay approach has been implemented in many operational networks, it has a number of well-known scaling issues. This paper proposes a new approach to achieve traffic engineering without full-mesh overlaying with the help of integrated approach and equal subset split method. Traffic engineering needs to determine the optimal routing of traffic over the existing network infrastructure by efficiently allocating resource in order to optimize traffic performance on an IP network. Even though constraint-based routing [1] of Multi-Protocol Label Switching (MPLS) is developed to address this need, since it is not widely tested or debugged, Internet Service Providers (ISPs) resort to TE methods under Open Shortest Path First (OSPF), which is the most commonly used intra-domain routing protocol. Determining OSPF link weights for optimal network performance is an NP-hard problem. As it is not possible to solve this problem, we present a subset split method to improve the efficiency and performance by minimizing the maximum link utilization in the network via a small number of link weight modifications. The results of this method are compared against results of MPLS architecture [9] and other heuristic methods.
Abstract: In this article, we introduce a mechanism by which the same concept of differentiated services used in network transmission can be applied to provide quality of service levels to pervasive systems applications. The classical DiffServ model, including marking and classification, assured forwarding, and expedited forwarding, are all utilized to create quality of service guarantees for various pervasive applications requiring different levels of quality of service. Through a collection of various sensors, personal devices, and data sources, the transmission of contextsensitive data can automatically occur within a pervasive system with a given quality of service level. Triggers, initiators, sources, and receivers are four entities labeled in our mechanism. An explanation of the role of each is provided, and how quality of service is guaranteed.
Abstract: In this paper we address the issue of classifying the fluorescent intensity of a sample in Indirect Immuno-Fluorescence (IIF). Since IIF is a subjective, semi-quantitative test in its very nature, we discuss a strategy to reliably label the image data set by using the diagnoses performed by different physicians. Then, we discuss image pre-processing, feature extraction and selection. Finally, we propose two ANN-based classifiers that can separate intrinsically dubious samples and whose error tolerance can be flexibly set. Measured performance shows error rates less than 1%, which candidates the method to be used in daily medical practice either to perform pre-selection of cases to be examined, or to act as a second reader.
Abstract: The volume of XML data exchange is explosively increasing, and the need for efficient mechanisms of XML data management is vital. Many XML storage models have been proposed for storing XML DTD-independent documents in relational database systems. Benchmarking is the best way to highlight pros and cons of different approaches. In this study, we use a common benchmarking scheme, known as XMark to compare the most cited and newly proposed DTD-independent methods in terms of logical reads, physical I/O, CPU time and duration. We show the effect of Label Path, extracting values and storing in another table and type of join needed for each method's query answering.
Abstract: Collected data must be organized to be utilized efficiently, and hierarchical classification of data is efficient approach to organize data. When data is classified to multiple categories or annotated with a set of labels, users request multi-labeled data by giving a set of labels. There are several interpretations of the data expressed by a set of labels. This paper discusses which data is expressed by a set of labels by introducing orders for sets of labels and shows that there are four types of orders, which are characterized by whether the labels of expressed data includes every label of the given set of labels within the range of the set. Desirable properties of the orders, data is also expressed by the higher set of labels and different sets of labels express different data, are discussed for the orders.
Abstract: A new blind symbol by symbol equalizer is proposed.
The operation of the proposed equalizer is based on the geometric
properties of the two dimensional data constellation. An unsupervised
clustering technique is used to locate the clusters formed by the
received data. The symmetric properties of the clusters labels are
subsequently utilized in order to label the clusters. Following this
step, the received data are compared to clusters and decisions are
made on a symbol by symbol basis, by assigning to each data
the label of the nearest cluster. The operation of the equalizer is
investigated both in linear and nonlinear channels. The performance
of the proposed equalizer is compared to the performance of a CMAbased
blind equalizer.
Abstract: Assume that we have m identical graphs where the
graphs consists of paths with k vertices where k is a positive integer.
In this paper, we discuss certain labelling of the m graphs called
c-Erdösian for some positive integers c. We regard labellings of the
vertices of the graphs by positive integers, which induce the edge
labels for the paths as the sum of the two incident vertex labels.
They have the property that each vertex label and edge label appears
only once in the set of positive integers {c, . . . , c+6m- 1}. Here,
we show how to construct certain c-Erdösian of m paths with 2 and
3 vertices by using Skolem sequences.
Abstract: An adaptive spatial Gaussian mixture model is proposed for clustering based color image segmentation. A new clustering objective function which incorporates the spatial information is introduced in the Bayesian framework. The weighting parameter for controlling the importance of spatial information is made adaptive to the image content to augment the smoothness towards piecewisehomogeneous region and diminish the edge-blurring effect and hence the name adaptive spatial finite mixture model. The proposed approach is compared with the spatially variant finite mixture model for pixel labeling. The experimental results with synthetic and Berkeley dataset demonstrate that the proposed method is effective in improving the segmentation and it can be employed in different practical image content understanding applications.
Abstract: Lighting upgrades involve relatively lower costs which
allow the benefits to be spread more widely than is possible with any
other energy efficiency measure. In order to popularize the adoption of
CFL in Taiwan, the authority proposes to implement a new energy efficient lamp comparative label system. The current study was
accordingly undertaken to investigate the factors affecting the performance and the deviation of actual and labeled performance of
commercially available integrated CFLs. In this paper, standard test
methods to determine the electrical and photometric performances of
CFL were developed based on CIE 84-1989 and CIE 60901-1987,
then 55 selected CFLs from market were tested. The results show that
with higher color temperature of CFLs lower efficacy are achieved. It
was noticed that the most packaging of CFL often lack the information of Color Rendering Index. Also, there was no correlation between
price and performance of the CFLs was indicated in this work. The results of this paper might help consumers to make more informed
CFL-purchasing decisions.
Abstract: Obtaining labeled data in supervised learning is often
difficult and expensive, and thus the trained learning algorithm tends
to be overfitting due to small number of training data. As a result,
some researchers have focused on using unlabeled data which may
not necessary to follow the same generative distribution as the labeled
data to construct a high-level feature for improving performance on
supervised learning tasks. In this paper, we investigate the impact of
the relationship between unlabeled and labeled data for classification
performance. Specifically, we will apply difference unlabeled data
which have different degrees of relation to the labeled data for
handwritten digit classification task based on MNIST dataset. Our
experimental results show that the higher the degree of relation
between unlabeled and labeled data, the better the classification
performance. Although the unlabeled data that is completely from
different generative distribution to the labeled data provides the lowest
classification performance, we still achieve high classification performance.
This leads to expanding the applicability of the supervised
learning algorithms using unsupervised learning.
Abstract: Term Extraction, a key data preparation step in Text
Mining, extracts the terms, i.e. relevant collocation of words,
attached to specific concepts (e.g. genetic-algorithms and decisiontrees
are terms associated to the concept “Machine Learning" ). In
this paper, the task of extracting interesting collocations is achieved
through a supervised learning algorithm, exploiting a few
collocations manually labelled as interesting/not interesting. From
these examples, the ROGER algorithm learns a numerical function,
inducing some ranking on the collocations. This ranking is optimized
using genetic algorithms, maximizing the trade-off between the false
positive and true positive rates (Area Under the ROC curve). This
approach uses a particular representation for the word collocations,
namely the vector of values corresponding to the standard statistical
interestingness measures attached to this collocation. As this
representation is general (over corpora and natural languages),
generality tests were performed by experimenting the ranking
function learned from an English corpus in Biology, onto a French
corpus of Curriculum Vitae, and vice versa, showing a good
robustness of the approaches compared to the state-of-the-art Support
Vector Machine (SVM).
Abstract: Sparse representation which can represent high dimensional
data effectively has been successfully used in computer vision
and pattern recognition problems. However, it doesn-t consider the
label information of data samples. To overcome this limitation,
we develop a novel dimensionality reduction algorithm namely
dscriminatively regularized sparse subspace learning(DR-SSL) in this
paper. The proposed DR-SSL algorithm can not only make use of
the sparse representation to model the data, but also can effective
employ the label information to guide the procedure of dimensionality
reduction. In addition,the presented algorithm can effectively deal
with the out-of-sample problem.The experiments on gene-expression
data sets show that the proposed algorithm is an effective tool for
dimensionality reduction and gene-expression data classification.
Abstract: This paper proposes an improvement method of classification
efficiency in a classification model. The model is used
in a risk search system and extracts specific labels from articles
posted at bulletin board sites. The system can analyze the important
discussions composed of the articles. The improvement method
introduces ensemble learning methods that use multiple classification
models. Also, it introduces expressions related to the specific labels
into generation of word vectors. The paper applies the improvement
method to articles collected from three bulletin board sites selected
by users and verifies the effectiveness of the improvement method.
Abstract: The paper deals with quality labels used in the food products market, especially with labels of quality, labels of origin, and labels of organic farming. The aim of the paper is to identify perception of these labels by consumers in the Czech Republic. The first part refers to the definition and specification of food quality labels that are relevant in the Czech Republic. The second part includes the discussion of marketing research results. Data were collected with personal questioning method. Empirical findings on 150 respondents are related to consumer awareness and perception of national and European food quality labels used in the Czech Republic, attitudes to purchases of labelled products, and interest in information regarding the labels. Statistical methods, in the concrete Pearson´s chi-square test of independence, coefficient of contingency, and coefficient of association are used to determinate if significant differences do exist among selected demographic categories of Czech consumers.
Abstract: The aim of this in vitro study was to evaluate the possible interference of a Nectandra membranacea extract (i) on the labeling of blood cells (BC), (ii) on the labeling process of BC and plasma (P) proteins with technetium-99m (Tc-99m) and (iii) on the morphology of red blood cells (RBC). Blood samples were incubated with a Nectandra membranacea crude extract, stannous chloride, Tc- 99m (sodium pertechnetate) was added, and soluble (SF) and insoluble (IF) fractions were isolated. Morphometry studies were performed with blood samples incubated with Nectandra membranacea extract. The results show that the Nectandra membranacea extract does not promote significant alteration of the labeling of BC, IF-P and IF-BC. The Nectandra membranacea extract was able to alter the erythrocyte membrane morphology, but these morphological changes were not capable to interfere on the labeling of blood constituents with Tc-99m.
Abstract: In general, image-based 3D scenes can now be found in many popular vision systems, computer games and virtual reality tours. So, It is important to segment ROI (region of interest) from input scenes as a preprocessing step for geometric stricture detection in 3D scene. In this paper, we propose a method for segmenting ROI based on tensor voting and Dirichlet process mixture model. In particular, to estimate geometric structure information for 3D scene from a single outdoor image, we apply the tensor voting and Dirichlet process mixture model to a image segmentation. The tensor voting is used based on the fact that homogeneous region in an image are usually close together on a smooth region and therefore the tokens corresponding to centers of these regions have high saliency values. The proposed approach is a novel nonparametric Bayesian segmentation method using Gaussian Dirichlet process mixture model to automatically segment various natural scenes. Finally, our method can label regions of the input image into coarse categories: “ground", “sky", and “vertical" for 3D application. The experimental results show that our method successfully segments coarse regions in many complex natural scene images for 3D.
Abstract: Y chromosome microdeletions are the most common
genetic cause of male infertility and screening for these
microdeletions in azoospermic or severely oligospermic men is now
standard practice. Analysis of the Y chromosome in men with
azoospermia or severe oligozoospermia has resulted in the
identification of three regions in the euchromatic part of the long arm
of the human Y chromosome (Yq11) that are frequently deleted in
men with otherwise unexplained spermatogenic failure. PCR analysis
of microdeletions in the AZFa, AZFb and AZFc regions of the
human Y chromosome is an important screening tool. The aim of this
study was to analyse the type of microdeletions in men with fertility
disorders in Slovakia. We evaluated 227 patients with azoospermia
and with normal karyotype. All patient samples were analyzed
cytogenetically. For PCR amplification of sequence-tagged sites
(STS) of the AZFa, AZFb and AZFc regions of the Y chromosome
was used Devyser AZF set. Fluorescently labeled primers for all
markers in one multiplex PCR reaction were used and for automated
visualization and identification of the STS markers we used genetic
analyzer ABi 3500xl (Life Technologies). We reported 13 cases of
deletions in the AZF region 5,73%. Particular types of deletions were
recorded in each region AZFa,b,c .The presence of microdeletions in
the AZFc region was the most frequent. The study confirmed that
percentage of microdeletions in the AZF region is low in Slovak
azoospermic patients, but important from a prognostic view.
Abstract: In this paper a way of hiding text message (Steganography) in the gray image has been presented. In this method tried to find binary value of each character of text message and then in the next stage, tried to find dark places of gray image (black) by converting the original image to binary image for labeling each object of image by considering on 8 connectivity. Then these images have been converted to RGB image in order to find dark places. Because in this way each sequence of gray color turns into RGB color and dark level of grey image is found by this way if the Gary image is very light the histogram must be changed manually to find just dark places. In the final stage each 8 pixels of dark places has been considered as a byte and binary value of each character has been put in low bit of each byte that was created manually by dark places pixels for increasing security of the main way of steganography (LSB).
Abstract: In this paper, we propose a new image segmentation approach for colour textured images. The proposed method for image segmentation consists of two stages. In the first stage, textural features using gray level co-occurrence matrix(GLCM) are computed for regions of interest (ROI) considered for each class. ROI acts as ground truth for the classes. Ohta model (I1, I2, I3) is the colour model used for segmentation. Statistical mean feature at certain inter pixel distance (IPD) of I2 component was considered to be the optimized textural feature for further segmentation. In the second stage, the feature matrix obtained is assumed to be the degraded version of the image labels and modeled as Markov Random Field (MRF) model to model the unknown image labels. The labels are estimated through maximum a posteriori (MAP) estimation criterion using ICM algorithm. The performance of the proposed approach is compared with that of the existing schemes, JSEG and another scheme which uses GLCM and MRF in RGB colour space. The proposed method is found to be outperforming the existing ones in terms of segmentation accuracy with acceptable rate of convergence. The results are validated with synthetic and real textured images.
Abstract: This paper estimates the economic values of
household preference for enhanced solid waste disposal services in
Malaysia. The contingent valuation (CV) method estimates an
average additional monthly willingness-to-pay (WTP) in solid waste
management charges of Ôé¼0.77 to 0.80 for improved waste disposal
services quality. The finding of a slightly higher WTP from the
generic CV question than that of label-specific, further reveals a
higher WTP for sanitary landfill, at Ôé¼0.90, than incineration, at Ôé¼0.63.
This suggests that sanitary landfill is a more preferred alternative.
The logistic regression estimation procedure reveals that household-s
concern of where their rubbish is disposed, age, ownership of house,
household income and format of CV question are significant factors
in influencing WTP.