Abstract: Network warfare is an emerging concept that focuses on the network and computer based forms through which information is attacked and defended. Various computer and network security concepts thus play a role in network warfare. Due the intricacy of the various interacting components, a model to better understand the complexity in a network warfare environment would be beneficial. Non-quantitative modeling is a useful method to better characterize the field due to the rich ideas that can be generated based on the use of secular associations, chronological origins, linked concepts, categorizations and context specifications. This paper proposes the use of non-quantitative methods through a morphological analysis to better explore and define the influential conditions in a network warfare environment.
Abstract: In text categorization problem the most used method
for documents representation is based on words frequency vectors
called VSM (Vector Space Model). This representation is based only
on words from documents and in this case loses any “word context"
information found in the document. In this article we make a
comparison between the classical method of document representation
and a method called Suffix Tree Document Model (STDM) that is
based on representing documents in the Suffix Tree format. For the
STDM model we proposed a new approach for documents
representation and a new formula for computing the similarity
between two documents. Thus we propose to build the suffix tree
only for any two documents at a time. This approach is faster, it has
lower memory consumption and use entire document representation
without using methods for disposing nodes. Also for this method is
proposed a formula for computing the similarity between documents,
which improves substantially the clustering quality. This
representation method was validated using HAC - Hierarchical
Agglomerative Clustering. In this context we experiment also the
stemming influence in the document preprocessing step and highlight
the difference between similarity or dissimilarity measures to find
“closer" documents.
Abstract: Text categorization (the assignment of texts in natural language into predefined categories) is an important and extensively studied problem in Machine Learning. Currently, popular techniques developed to deal with this task include many preprocessing and learning algorithms, many of which in turn require tuning nontrivial internal parameters. Although partial studies are available, many authors fail to report values of the parameters they use in their experiments, or reasons why these values were used instead of others. The goal of this work then is to create a more thorough comparison of preprocessing parameters and their mutual influence, and report interesting observations and results.
Abstract: This study examined the role of driving experience in hazard perception and categorization using traffic scene pictures. Specifically, young-inexperienced, moderately experienced and very experienced (taxi) drivers observed traffic scene pictures while connected to an eye tracking system and were asked to rate the level of hazardousness of each picture and to mention the three most prominent hazards in it. Target pictures included nine, nearly identical, pairs of pictures where one picture in each pair included an actual hazard as an additional element. Altogether, 22 areas of interest (AOIs) were predefined and included 13 potential hazards and 9 actual hazards. Data analysis included both verbal reports and eye scanning patterns of these AOIs. Generally, both experienced and taxi drivers noted a relatively larger number of potential hazards than young inexperienced drivers Thus, by relating to less salient potential hazards, experienced drivers have demonstrated a better situation model of the traffic environment.
Abstract: Text categorization is the problem of classifying text
documents into a set of predefined classes. In this paper, we
investigated three approaches to build a meta-classifier in order to
increase the classification accuracy. The basic idea is to learn a metaclassifier
to optimally select the best component classifier for each
data point. The experimental results show that combining classifiers
can significantly improve the accuracy of classification and that our
meta-classification strategy gives better results than each individual
classifier. For 7083 Reuters text documents we obtained a
classification accuracies up to 92.04%.
Abstract: Text categorization is the problem of classifying text
documents into a set of predefined classes. After a preprocessing
step, the documents are typically represented as large sparse vectors.
When training classifiers on large collections of documents, both the
time and memory restrictions can be quite prohibitive. This justifies
the application of feature selection methods to reduce the
dimensionality of the document-representation vector. In this paper,
three feature selection methods are evaluated: Random Selection,
Information Gain (IG) and Support Vector Machine feature selection
(called SVM_FS). We show that the best results were obtained with
SVM_FS method for a relatively small dimension of the feature
vector. Also we present a novel method to better correlate SVM
kernel-s parameters (Polynomial or Gaussian kernel).
Abstract: This paper presents a new approach for automatic
document categorization. Exploiting the logical structure of the
document, our approach assigns a HTML document to one or more
categories (thesis, paper, call for papers, email, ...). Using a set of
training documents, our approach generates a set of rules used to
categorize new documents. The approach flexibility is carried out
with rule weight association representing your importance in the
discrimination between possible categories. This weight is
dynamically modified at each new document categorization. The
experimentation of the proposed approach provides satisfactory
results.
Abstract: Apart from geometry, functionality is one of the most
significant hallmarks of a product. The functionality of a product can
be considered as the fundamental justification for a product
existence. Therefore a functional analysis including a complete and
reliable descriptor has a high potential to improve product
development process in various fields especially in knowledge-based
design. One of the important applications of the functional analysis
and indexing is in retrieval and design reuse concept. More than 75%
of design activity for a new product development contains reusing
earlier and existing design know-how. Thus, analysis and
categorization of product functions concluded by functional
indexing, influences directly in design optimization. This paper
elucidates and evaluates major classes for functional analysis by
discussing their major methods. Moreover it is finalized by
presenting a noble hybrid approach for functional analysis.
Abstract: Text categorization is the problem of classifying text documents into a set of predefined classes. After a preprocessing step the documents are typically represented as large sparse vectors. When training classifiers on large collections of documents, both the time and memory restrictions can be quite prohibitive. This justifies the application of features selection methods to reduce the dimensionality of the document-representation vector. Four feature selection methods are evaluated: Random Selection, Information Gain (IG), Support Vector Machine (called SVM_FS) and Genetic Algorithm with SVM (GA_FS). We showed that the best results were obtained with SVM_FS and GA_FS methods for a relatively small dimension of the features vector comparative with the IG method that involves longer vectors, for quite similar classification accuracies. Also we present a novel method to better correlate SVM kernel-s parameters (Polynomial or Gaussian kernel).
Abstract: In this paper, in order to categorize ORL database face
pictures, principle Component Analysis (PCA) and Kernel Principal
Component Analysis (KPCA) methods by using Elman neural
network and Support Vector Machine (SVM) categorization methods
are used. Elman network as a recurrent neural network is proposed
for modeling storage systems and also it is used for reviewing the
effect of using PCA numbers on system categorization precision rate
and database pictures categorization time. Categorization stages are
conducted with various components numbers and the obtained results
of both Elman neural network categorization and support vector
machine are compared. In optimum manner 97.41% recognition
accuracy is obtained.
Abstract: This paper presents a semi-supervised learning algorithm called Iterative-Cross Training (ICT) to solve the Web pages classification problems. We apply Inductive logic programming (ILP) as a strong learner in ICT. The objective of this research is to evaluate the potential of the strong learner in order to boost the performance of the weak learner of ICT. We compare the result with the supervised Naive Bayes, which is the well-known algorithm for the text classification problem. The performance of our learning algorithm is also compare with other semi-supervised learning algorithms which are Co-Training and EM. The experimental results show that ICT algorithm outperforms those algorithms and the performance of the weak learner can be enhanced by ILP system.
Abstract: Color categorization is shared among members in a
society. This allows communication of color, especially when using
natural language such as English. Hence sociable robot, to live
coexist with human in human society, must also have the shared
color categorization. To achieve this, many works have been done
relying on modeling of human color perception and mathematical
complexities. In contrast, in this work, the computer as brain of the
robot learns color categorization through interaction with humans
without much mathematical complexities.
Abstract: The presented article deals with the description of a
numerical model of a corridor at a Central Interim Spent Fuel Storage
Facility (hereinafter CISFSF). The model takes into account the
effect of air flows on the temperature of stored waste. The
computational model was implemented in the ANSYS/CFX
programming environment in the form of a CFD task solution, which
was compared with an approximate analytical calculation. The article
includes a categorization of the individual alternatives for the
ventilation of such underground systems. The aim was to evaluate a
ventilation system for a CISFSF with regard to its stability and
capacity to provide sufficient ventilation for the removal of heat
produced by stored casks with spent nuclear fuel.
Abstract: IT consultants often take over an important role as an
interface between technological, organizational and managerial
structures. As a result, the services offered are in many cases
assigned to different disciplines which can cause a lack of
transparency on the market for consulting services. However, not all
consulting products are suitable for every company because of
different frameworks and business processes. In this context the
questions arises as to what consulting products are currently offered
and how they can be compared as well as how the market for IT
consulting services is structured on the supply side. The presented
study aims to shed light on the IT consulting market by giving an
overview of the current structure of the supply-side for IT consulting
services as well as proposing a categorization of the currently
available consulting services (consulting fields) in order to provide a
theoretical background for the empirical study. Apart from these
theoretical considerations, the empirical results of field surveys on
the Austrian IT consulting market are presented and analyzed.
Abstract: Modern information and communication technologies
offer a variety of support options for the efficient handling of
customer relationships. CRM systems have been developed, which
are designed to support the processes in the areas of marketing, sales
and service. Along with technological progress, CRM systems are
constantly changing, i.e. the systems are continually enhanced by
new functions. However, not all functions are suitable for every
company because of different frameworks and business processes. In
this context the question arises whether or not CRM systems are
widely used in Austrian companies and which business processes are
most frequently supported by CRM systems. This paper aims to shed
light on the popularity of CRM systems in Austrian companies in
general and the use of different functions to support their daily
business. First of all, the paper provides a theoretical overview of the
structure of modern CRM systems and proposes a categorization of
currently available software functionality for collaborative,
operational and analytical CRM processes, which provides the
theoretical background for the empirical study. Apart from these
theoretical considerations, the paper presents the empirical results of
a field survey on the use of CRM systems in Austrian companies and
analyzes its findings.
Abstract: Text categorization techniques are widely used to many Information Retrieval (IR) applications. In this paper, we proposed a simple but efficient method that can automatically find the relationship between any pair of terms and documents, also an indexing matrix is established for text categorization. We call this method Indexing Matrix Categorization Machine (IMCM). Several experiments are conducted to show the efficiency and robust of our algorithm.
Abstract: Most neural network (NN) models of human category learning use a gradient-based learning method, which assumes that locally-optimal changes are made to model parameters on each learning trial. This method tends to under predict variability in individual-level cognitive processes. In addition many recent models of human category learning have been criticized for not being able to replicate rapid changes in categorization accuracy and attention processes observed in empirical studies. In this paper we introduce stochastic learning algorithms for NN models of human category learning and show that use of the algorithms can result in (a) rapid changes in accuracy and attention allocation, and (b) different learning trajectories and more realistic variability at the individual-level.
Abstract: The most common result of analysis of highthroughput
data in molecular biology represents a global list of
genes, ranked accordingly to a certain score. The score can be a
measure of differential expression. Recent work proposed a new
method for selecting a number of genes in a ranked gene list from
microarray gene expression data such that this set forms the
Optimally Functionally Enriched Network (OFTEN), formed by
known physical interactions between genes or their products. Here
we present calculation results of relative connectivity of genes from
META-OFTEN network and tentative biological interpretation of the
most reproducible signal. The relative connectivity and
inbetweenness values of genes from META-OFTEN network were
estimated.