Abstract: Increasing growth of information volume in the
internet causes an increasing need to develop new (semi)automatic
methods for retrieval of documents and ranking them according to
their relevance to the user query. In this paper, after a brief review
on ranking models, a new ontology based approach for ranking
HTML documents is proposed and evaluated in various
circumstances. Our approach is a combination of conceptual,
statistical and linguistic methods. This combination reserves the
precision of ranking without loosing the speed. Our approach
exploits natural language processing techniques to extract phrases
from documents and the query and doing stemming on words. Then
an ontology based conceptual method will be used to annotate
documents and expand the query. To expand a query the spread
activation algorithm is improved so that the expansion can be done
flexible and in various aspects. The annotated documents and the
expanded query will be processed to compute the relevance degree
exploiting statistical methods. The outstanding features of our
approach are (1) combining conceptual, statistical and linguistic
features of documents, (2) expanding the query with its related
concepts before comparing to documents, (3) extracting and using
both words and phrases to compute relevance degree, (4) improving
the spread activation algorithm to do the expansion based on
weighted combination of different conceptual relationships and (5)
allowing variable document vector dimensions. A ranking system
called ORank is developed to implement and test the proposed
model. The test results will be included at the end of the paper.
Abstract: The major objective of this paper is to introduce a new method to select genes from DNA microarray data. As criterion to select genes we suggest to measure the local changes in the correlation graph of each gene and to select those genes whose local changes are largest. More precisely, we calculate the correlation networks from DNA microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to tumor progression. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth. This indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.
Abstract: Sorting appears the most attention among all computational tasks over the past years because sorted data is at the heart of many computations. Sorting is of additional importance to parallel computing because of its close relation to the task of routing data among processes, which is an essential part of many parallel algorithms. Many parallel sorting algorithms have been investigated for a variety of parallel computer architectures. In this paper, three parallel sorting algorithms have been implemented and compared in terms of their overall execution time. The algorithms implemented are the odd-even transposition sort, parallel merge sort and parallel rank sort. Cluster of Workstations or Windows Compute Cluster has been used to compare the algorithms implemented. The C# programming language is used to develop the sorting algorithms. The MPI (Message Passing Interface) library has been selected to establish the communication and synchronization between processors. The time complexity for each parallel sorting algorithm will also be mentioned and analyzed.
Abstract: In this paper the General Game problem is described.
In this problem the competition or cooperation dilemma occurs as the
two basic types of strategies. The strategy possibilities have been
analyzed for finding winning strategy in uncertain situations (no
information about the number of players and their strategy types).
The winning strategy is missing, but a good solution can be found by
simulation by varying the ratio of the two types of strategies. This
new method has been used in a real contest with human players,
where the created strategies by simulation have reached very good
ranks. This construction can be applied in other real social games as
well.
Abstract: This study presents the numerical simulation of
optimum pin-fin heat sink with air impinging cooling by using
Taguchi method. 9 L ( 4 3 ) orthogonal array is selected as a plan for
the four design-parameters with three levels. The governing
equations are discretized by using the
control-volume-based-finite-difference method with a power-law
scheme on the non-uniform staggered grid. We solved the coupling of
the velocity and the pressure terms of momentum equations using
SIMPLEC algorithm. We employ the k −ε two-equations
turbulence model to describe the turbulent behavior. The parameters
studied include fin height H (35mm-45mm), inter-fin spacing a , b ,
and c (2 mm-6.4 mm), and Reynolds number ( Re = 10000- 25000).
The objective of this study is to examine the effects of the fin
spacings and fin height on the thermal resistance and to find the
optimum group by using the Taguchi method. We found that the fin
spacings from the center to the edge of the heat sink gradually
extended, and the longer the fin’s height the better the results. The
optimum group is 3 1 2 3 H a b c . In addition, the effects of parameters are
ranked by importance as a , H , c , and b .
Abstract: Fuzzy linear programming is an application of fuzzy set theory in linear decision making problems and most of these problems are related to linear programming with fuzzy variables. A convenient method for solving these problems is based on using of auxiliary problem. In this paper a new method for solving fuzzy variable linear programming problems directly using linear ranking functions is proposed. This method uses simplex tableau which is used for solving linear programming problems in crisp environment before.
Abstract: Recent quasi-experimental evaluation of the Canadian Active Labour Market Policies (ALMP) by Human Resources and Skills Development Canada (HRSDC) has provided an opportunity to examine alternative methods to estimating the incremental effects of Employment Benefits and Support Measures (EBSMs) on program participants. The focus of this paper is to assess the efficiency and robustness of inverse probability weighting (IPW) relative to kernel matching (KM) in the estimation of program effects. To accomplish this objective, the authors compare pairs of 1,080 estimates, along with their associated standard errors, to assess which type of estimate is generally more efficient and robust. In the interest of practicality, the authorsalso document the computationaltime it took to produce the IPW and KM estimates, respectively.
Abstract: This research is to study the types of products and
services that employs 'ambient media and respective techniques in its
advertisement materials. Data collection has been done via analyses of a total of 62 advertisements that employed ambient media
approach in Thailand during the years 2004 to 2011. The 62 advertisement were qualifying advertisements of the Adman Awards
& Symposium under the category of Outdoor & Ambience. Analysis
results reveal that there is a total of 14 products and services that
chooses to utilize ambient media in its advertisement. Amongst all ambient media techniques, 'intrusion' uses the value of a medium in
its representation of content most often. Following intrusion is 'interaction', where consumers are invited to participate and interact
with the advertising materials. 'Illusion' ranks third in its ability to subject the viewers to distortions of reality that makes the division
between reality and fantasy less clear.
Abstract: Median filters with larger windows offer greater smoothing and are more robust than the median filters of smaller windows. However, the larger median smoothers (the median filters with the larger windows) fail to track low order polynomial trends in the signals. Due to this, constant regions are produced at the signal corners, leading to the loss of fine details. In this paper, an algorithm, which combines the ability of the 3-point median smoother in preserving the low order polynomial trends and the superior noise filtering characteristics of the larger median smoother, is introduced. The proposed algorithm (called the combiner algorithm in this paper) is evaluated for its performance on a test image corrupted with different types of noise and the results obtained are included.
Abstract: The paper provides a discussion of the most relevant
aspects of yield curve modeling. Two classes of models are
considered: stochastic and parsimonious function based, through the
approaches developed by Vasicek (1977) and Nelson and Siegel
(1987). Yield curve estimates for Croatia are presented and their
dynamics analyzed and finally, a comparative analysis of models is
conducted.
Abstract: Protein 3D structure prediction has always been an
important research area in bioinformatics. In particular, the
prediction of secondary structure has been a well-studied research
topic. Despite the recent breakthrough of combining multiple
sequence alignment information and artificial intelligence algorithms
to predict protein secondary structure, the Q3 accuracy of various
computational prediction algorithms rarely has exceeded 75%. In a
previous paper [1], this research team presented a rule-based method
called RT-RICO (Relaxed Threshold Rule Induction from Coverings)
to predict protein secondary structure. The average Q3 accuracy on
the sample datasets using RT-RICO was 80.3%, an improvement
over comparable computational methods. Although this demonstrated
that RT-RICO might be a promising approach for predicting
secondary structure, the algorithm-s computational complexity and
program running time limited its use. Herein a parallelized
implementation of a slightly modified RT-RICO approach is
presented. This new version of the algorithm facilitated the testing of
a much larger dataset of 396 protein domains [2]. Parallelized RTRICO
achieved a Q3 score of 74.6%, which is higher than the
consensus prediction accuracy of 72.9% that was achieved for the
same test dataset by a combination of four secondary structure
prediction methods [2].
Abstract: The study was conducted to evaluate the quality
characteristics of cookies produced from sweet potato-fermented
soybean flour. Cookies were subjected to proximate and sensory
analysis to determine the acceptability of the product. Protein, fat and
ash increased as the proportion of soybean flour increased, ranging
from 13.8-21.7, 1.22-5.25 and 2.20-2.57 respectively. The crude fibre
content was within the range of 3.08-4.83%. The moisture content of
the cookies decreased with increase in soybean flour from 3.42-
2.13%. Cookies produced from whole sweet potato flour had the
highest moisture content of 3.42% while 30% substitution had the
lowest moisture content 2.13%. A nine point hedonic scale was used
to evaluate the organoleptic characteristics of the cookies. The
sensory analysis indicated that there was no significant difference
between the cookies produced even when compared to the control
100% sweet potato cookies. The overall acceptance of the cookies
was ranked to 20% soybean flour substitute.
Abstract: This paper gives a novel method for improving
classification performance for cancer classification with very few
microarray Gene expression data. The method employs classification
with individual gene ranking and gene subset ranking. For selection
and classification, the proposed method uses the same classifier. The
method is applied to three publicly available cancer gene expression
datasets from Lymphoma, Liver and Leukaemia datasets. Three
different classifiers namely Support vector machines-one against all
(SVM-OAA), K nearest neighbour (KNN) and Linear Discriminant
analysis (LDA) were tested and the results indicate the improvement
in performance of SVM-OAA classifier with satisfactory results on
all the three datasets when compared with the other two classifiers.
Abstract: This paper presents a computational methodology
based on matrix operations for a computer based solution to the
problem of performance analysis of software reliability models
(SRMs). A set of seven comparison criteria have been formulated to
rank various non-homogenous Poisson process software reliability
models proposed during the past 30 years to estimate software
reliability measures such as the number of remaining faults, software
failure rate, and software reliability. Selection of optimal SRM for
use in a particular case has been an area of interest for researchers in
the field of software reliability. Tools and techniques for software
reliability model selection found in the literature cannot be used with
high level of confidence as they use a limited number of model
selection criteria. A real data set of middle size software project from
published papers has been used for demonstration of matrix method.
The result of this study will be a ranking of SRMs based on the
Permanent value of the criteria matrix formed for each model based
on the comparison criteria. The software reliability model with
highest value of the Permanent is ranked at number – 1 and so on.
Abstract: This paper presents three models which enable the
customisation of Universal Description, Discovery and Integration
(UDDI) query results, based on some pre-defined and/or real-time
changing parameters. These proposed models detail the requirements,
design and techniques which make ranking of Web service discovery
results from a service registry possible. Our contribution is two fold:
First, we present an extension to the UDDI inquiry capabilities. This
enables a private UDDI registry owner to customise or rank the query
results, based on its business requirements. Second, our proposal
utilises existing technologies and standards which require minimal
changes to existing UDDI interfaces or its data structures. We believe
these models will serve as valuable reference for enhancing the
service discovery methods within a private UDDI registry
environment.
Abstract: Application of synthetic antioxidants such as tertbutylhydroquinon
(TBHQ), in spite of their efficiency, is questioned
because of their possible carcinogenic effect. The purpose of this
study was application of mixtures of natural antioxidants that provide
the best oxidative stability for margarine. Antioxidant treatments
included 10 various mixtures (F1- F10) containing 100-500ppm
tocopherol mixture (Toc), 100-200ppm ascorbyl palmitate (AP), 100-
200ppm rosemary extract (Ros) and 1000ppm lecithin(Lec) along
with a control or F0 (with no antioxidant) and F11 with 120ppm
TBHQ. The effect of antioxidant mixtures on the stability of
margarine samples during oven test (60°C), rancimat test at 110°C
and storage at 4°C was evaluated. Final ranking of natural antioxidant
mixtures was as follows: F2,F10>F5,F9>F8>F1,F3,F4>F6, F7.
Considering the results of this research and ranking criteria,
F2(200ppmAp + 200ppmRos) and F10(200ppmRos + 200ppmToc
+1000ppmLec) were recommended as substitutes for TBHQ to
maintain the quality and increase the shelf-life of margarine.
Abstract: Meta-reasoning is essential for multi-agent communication. In this paper we propose a framework of multi-agent communication in which agents employ meta-reasoning to reason with agent and ontology locations in order to communicate semantic information with other agents on the semantic web and also reason with multiple distributed ontologies. We shall argue that multi-agent communication of Semantic Web information cannot be realized without the need to reason with agent and ontology locations. This is because for an agent to be able to communicate with another agent, it must know where and how to send a message to that agent. Similarly, for an agent to be able to reason with an external semantic web ontology, it must know where and how to access to that ontology. The agent framework and its communication mechanism are formulated entirely in meta-logic.
Abstract: Voice over Internet Protocol (VoIP) application or commonly known as softphone has been developing an increasingly large market in today-s telecommunication world and the trend is expected to continue with the enhancement of additional features. This includes leveraging on the existing presence services, location and contextual information to enable more ubiquitous and seamless communications. In this paper, we discuss the concept of seamless session transfer for real-time application such as VoIP and IPTV, and our prototype implementation of such concept on a selected open source VoIP application. The first part of this paper is about conducting performance evaluation and assessments across some commonly found open source VoIP applications that are Ekiga, Kphone, Linphone and Twinkle so as to identify one of them for implementing our design of seamless session transfer. Subjective testing has been carried out to evaluate the audio performance on these VoIP applications and rank them according to their Mean Opinion Score (MOS) results. The second part of this paper is to discuss on the performance evaluations of our prototype implementation of session transfer using Linphone.
Abstract: This paper gives an introduction to Web mining, then
describes Web Structure mining in detail, and explores the data
structure used by the Web. This paper also explores different Page
Rank algorithms and compare those algorithms used for Information
Retrieval. In Web Mining, the basics of Web mining and the Web
mining categories are explained. Different Page Rank based
algorithms like PageRank (PR), WPR (Weighted PageRank), HITS
(Hyperlink-Induced Topic Search), DistanceRank and DirichletRank
algorithms are discussed and compared. PageRanks are calculated for
PageRank and Weighted PageRank algorithms for a given hyperlink
structure. Simulation Program is developed for PageRank algorithm
because PageRank is the only ranking algorithm implemented in the
search engine (Google). The outputs are shown in a table and chart
format.
Abstract: Palladium-catalyzed hydrodechlorination is a
promising alternative for the treatment of environmentally relevant
water bodies, such as groundwater, contaminated with chlorinated
organic compounds (COCs). In the aqueous phase
hydrodechlorination of COCs, Pd-based catalysts were found to have
a very high catalytic activity. However, the full utilization of the
catalyst-s potential is impeded by the sensitivity of the catalyst to
poisoning and deactivation induced by reduced sulfur compounds
(e.g. sulfides). Several regenerants have been tested before to recover
the performance of sulfide-fouled Pd catalyst. But these only
delivered partial success with respect to re-establishment of the
catalyst activity. In this study, the deactivation behaviour of
Pd/Al2O3 in the presence of sulfide was investigated. Subsequent to
total deactivation the catalyst was regenerated in the aqueous phase
using potassium permanganate. Under neutral pH condition,
oxidative regeneration with permanganate delivered a slow recovery
of catalyst activity. However, changing the pH of the bulk solution to
acidic resulted in the complete recovery of catalyst activity within a
regeneration time of about half an hour. These findings suggest the
superiority of permanganate as regenerant in re-activating Pd/Al2O3
by oxidizing Pd-bound sulfide.