Abstract: Active network was developed to solve the problem of
the current sharing-based network–difficulty in applying new
technology, service or standard, and duplicated operation at several
protocol layers. Active network can transport the packet loaded with
the executable codes, which enables to change the state of the network
node. However, if the network node is placed in the sharing-based
network, security and safety issues should be resolved. To satisfy this
requirement, various security aspects are required such as
authentication, authorization, confidentiality and integrity. Among
these security components, the core factor is the encryption key. As a
result, this study is designed to propose the scheme that manages the
encryption key, which is used to provide security of the
comprehensive active directory, based on the domain.
Abstract: Fuzzy Cognitive Maps (FCMs) have successfully
been applied in numerous domains to show relations between
essential components. In some FCM, there are more nodes, which
related to each other and more nodes means more complex in system
behaviors and analysis. In this paper, a novel learning method used to
construct FCMs based on historical data and by using data mining
and DEMATEL method, a new method defined to reduce nodes
number. This method cluster nodes in FCM based on their cause and
effect behaviors.
Abstract: Nowadays, efficiency, effectiveness and economy are regarded as the main objectives of managers and the secret of the continuity of an organization in competing economy. In such competing settings, it is essential that the management of an organization has not been neglected and been obliged to identify quickly the opportunities for improving the operation of organization and remove the shortcomings of their managed system in order to use the opportunities for development. Operational auditing is a useful tool for system adjustment and leading an organization toward its objectives. Operational auditing is indeed a viewpoint which identifies the causes of insufficiencies, weaknesses and deficiencies of system and plans to eliminate them. Operational auditing is useful in the effectiveness and optimization of executive managers- decisions and increasing the efficiency and economy of their performance in the future and prevents the waste and incorrect use of resources. Evidence shows that operational auditing is used at a limited level in Iran. This matter raises some questions like the following ones in the minds. Why do a limited number of corporations use operational auditing? Which factors can guarantee its full implementation? What obstacles are there in its implementation? The purpose of this article is to determine executive objectives, the operation domain of operational auditing, the components of operational auditing and the executive obstacles to operational auditing in Iran.
Abstract: Discrimination between different classes of environmental
sounds is the goal of our work. The use of a sound recognition
system can offer concrete potentialities for surveillance and
security applications. The first paper contribution to this research
field is represented by a thorough investigation of the applicability
of state-of-the-art audio features in the domain of environmental
sound recognition. Additionally, a set of novel features obtained by
combining the basic parameters is introduced. The quality of the
features investigated is evaluated by a HMM-based classifier to which
a great interest was done. In fact, we propose to use a Multi-Style
training system based on HMMs: one recognizer is trained on a
database including different levels of background noises and is used
as a universal recognizer for every environment. In order to enhance
the system robustness by reducing the environmental variability, we
explore different adaptation algorithms including Maximum Likelihood
Linear Regression (MLLR), Maximum A Posteriori (MAP)
and the MAP/MLLR algorithm that combines MAP and MLLR.
Experimental evaluation shows that a rather good recognition rate
can be reached, even under important noise degradation conditions
when the system is fed by the convenient set of features.
Abstract: In recent years, fast neural networks for object/face detection have been introduced based on cross correlation in the frequency domain between the input matrix and the hidden weights of neural networks. In our previous papers [3,4], fast neural networks for certain code detection was introduced. It was proved in [10] that for fast neural networks to give the same correct results as conventional neural networks, both the weights of neural networks and the input matrix must be symmetric. This condition made those fast neural networks slower than conventional neural networks. Another symmetric form for the input matrix was introduced in [1-9] to speed up the operation of these fast neural networks. Here, corrections for the cross correlation equations (given in [13,15,16]) to compensate for the symmetry condition are presented. After these corrections, it is proved mathematically that the number of computation steps required for fast neural networks is less than that needed by classical neural networks. Furthermore, there is no need for converting the input data into symmetric form. Moreover, such new idea is applied to increase the speed of neural networks in case of processing complex values. Simulation results after these corrections using MATLAB confirm the theoretical computations.
Abstract: The study of piezoelectric material in the past was in
T-Domain form; however, no one has studied piezoelectric material in the S-Domain form. This paper will present the piezoelectric material in the transfer function or S-Domain model. S-Domain is a
well known mathematical model, used for analyzing the stability of the material and determining the stability limits. By using S-Domain
in testing stability of piezoelectric material, it will provide a new tool for the scientific world to study this material in various forms.
Abstract: Domain-specific languages describe specific solutions to problems in the application domain. Traditionally they form a solution composing black-box abstractions together. This, usually, involves non-deep transformations over the target model. In this paper we argue that it is potentially powerful to operate with grey-box abstractions to build a domain-specific software system. We present parametric code templates as grey-box abstractions and conceptual tools to encapsulate and manipulate these templates. Manipulations introduce template-s merging routines and can be defined in a generic way. This involves reasoning mechanisms at the code templates level. We introduce the concept of Neurath Modelling Language (NML) that operates with parametric code templates and specifies a visualisation mapping mechanism for target models. Finally we provide an example of calculating a domain-specific software system with predefined NML elements.
Abstract: This paper explores the scalability issues associated
with solving the Named Entity Recognition (NER) problem using
Support Vector Machines (SVM) and high-dimensional features. The
performance results of a set of experiments conducted using binary
and multi-class SVM with increasing training data sizes are
examined. The NER domain chosen for these experiments is the
biomedical publications domain, especially selected due to its
importance and inherent challenges. A simple machine learning
approach is used that eliminates prior language knowledge such as
part-of-speech or noun phrase tagging thereby allowing for its
applicability across languages. No domain-specific knowledge is
included. The accuracy measures achieved are comparable to those
obtained using more complex approaches, which constitutes a
motivation to investigate ways to improve the scalability of multiclass
SVM in order to make the solution more practical and useable.
Improving training time of multi-class SVM would make support
vector machines a more viable and practical machine learning
solution for real-world problems with large datasets. An initial
prototype results in great improvement of the training time at the
expense of memory requirements.
Abstract: Ontology-based modelling of multi-formatted
software application content is a challenging area in content
management. When the number of software content unit is huge and
in continuous process of change, content change management is
important. The management of content in this context requires
targeted access and manipulation methods. We present a novel
approach to deal with model-driven content-centric information
systems and access to their content. At the core of our approach is an
ontology-based semantic annotation technique for diversely
formatted content that can improve the accuracy of access and
systems evolution. Domain ontologies represent domain-specific
concepts and conform to metamodels. Different ontologies - from
application domain ontologies to software ontologies - capture and
model the different properties and perspectives on a software content
unit. Interdependencies between domain ontologies, the artifacts and
the content are captured through a trace model. The annotation traces
are formalised and a graph-based system is selected for the
representation of the annotation traces.
Abstract: Expert systems are used extensively in many domains.
This paper discusses the use of medical expert systems in Pakistan.
Countries all over the world pay special attention on health facilities.
A country like Pakistan faces a lot of trouble in health sector.
Several attempts have been made in Pakistan to improve the health
conditions of the people but the situation is still not encouraging.
There is a shortage of doctors and other trained personnel in
Pakistan. Expert systems can play a vital role in such cases where the
medical expert is not readily available. The purpose of this paper is
to analyze the role that such systems can play in improving the health
conditions of the people in Pakistan.
Abstract: Assessment for image quality traditionally needs its
original image as a reference. The conventional method for assessment
like Mean Square Error (MSE) or Peak Signal to Noise Ratio (PSNR)
is invalid when there is no reference. In this paper, we present a new
No-Reference (NR) assessment of image quality using blur and noise.
The recent camera applications provide high quality images by help of
digital Image Signal Processor (ISP). Since the images taken by the
high performance of digital camera have few blocking and ringing
artifacts, we only focus on the blur and noise for predicting the
objective image quality. The experimental results show that the
proposed assessment method gives high correlation with subjective
Difference Mean Opinion Score (DMOS). Furthermore, the proposed
method provides very low computational load in spatial domain and
similar extraction of characteristics to human perceptional assessment.
Abstract: Image watermarking has become an important tool for
intellectual property protection and authentication. In this paper a
watermarking technique is suggested that incorporates two
watermarks in a host image for improved protection and robustness.
A watermark, in form of a PN sequence (will be called the secondary
watermark), is embedded in the wavelet domain of a primary
watermark before being embedded in the host image. The technique
has been tested using Lena image as a host and the camera man as
the primary watermark. The embedded PN sequence was detectable
through correlation among other five sequences where a PSNR of
44.1065 dB was measured. Furthermore, to test the robustness of the
technique, the watermarked image was exposed to four types of
attacks, namely compression, low pass filtering, salt and pepper noise
and luminance change. In all cases the secondary watermark was
easy to detect even when the primary one is severely distorted.
Abstract: A submerged horizontal plate type breakwater is
pointed out as an efficient wave protection device for cage culture in
marine fishery. In order to reveal the wave elimination principle of this
type breakwater, boundary element method is utilized to investigate
this problem. The flow field and the trajectory of water particles are
studied carefully. The flow field analysis shows that: the interaction of
incident wave and adverse current above the plate disturbs the water
domain drastically. This can slow down the horizontal velocity and
vertical velocity of the water particles.
Abstract: Nowadays there are many methods for representing
knowledge such as semantic network, neural network, and conceptual
graphs. Nonetheless, these methods are not sufficiently efficient
when applied to perform and deduce on knowledge domains about
supporting in general education such as algebra, analysis or plane
geometry. This leads to the introduction of computational network
which is a useful tool for representation knowledge base, especially
for computational knowledge, especially knowledge domain about
general education. However, when dealing with a practical problem,
we often do not immediately find a new solution, but we search
related problems which have been solved before and then proposing
an appropriate solution for the problem. Besides that, when finding
related problems, we have to determine whether the result of them
can be used to solve the practical problem or not. In this paper, the
extension model of computational network has been presented. In this
model, Sample Problems, which are related problems, will be used
like the experience of human about practical problem, simulate the
way of human thinking, and give the good solution for the practical
problem faster and more effectively. This extension model is applied
to construct an automatic system for solving algebraic problems in
middle school.
Abstract: The Far From Most Strings Problem (FFMSP) is to obtain a string which is far from as many as possible of a given set of strings. All the input and the output strings are of the same length, and two strings are said to be far if their hamming distance is greater than or equal to a given positive integer. FFMSP belongs to the class of sequences consensus problems which have applications in molecular biology. The problem is NP-hard; it does not admit a constant-ratio approximation either, unless P = NP. Therefore, in addition to exact and approximate algorithms, (meta)heuristic algorithms have been proposed for the problem in recent years. On the other hand, in the recent years, hybrid algorithms have been proposed and successfully used for many hard problems in a variety of domains. In this paper, a new metaheuristic algorithm, called Constructive Beam and Local Search (CBLS), is investigated for the problem, which is a hybridization of constructive beam search and local search algorithms. More specifically, the proposed algorithm consists of two phases, the first phase is to obtain several candidate solutions via the constructive beam search and the second phase is to apply local search to the candidate solutions obtained by the first phase. The best solution found is returned as the final solution to the problem. The proposed algorithm is also similar to memetic algorithms in the sense that both use local search to further improve individual solutions. The CBLS algorithm is compared with the most recent published algorithm for the problem, GRASP, with significantly positive results; the improvement is by order of magnitudes in most cases.
Abstract: In large datasets, identifying exceptional or rare cases
with respect to a group of similar cases is considered very significant
problem. The traditional problem (Outlier Mining) is to find
exception or rare cases in a dataset irrespective of the class label of
these cases, they are considered rare events with respect to the whole
dataset. In this research, we pose the problem that is Class Outliers
Mining and a method to find out those outliers. The general
definition of this problem is “given a set of observations with class
labels, find those that arouse suspicions, taking into account the
class labels". We introduce a novel definition of Outlier that is Class
Outlier, and propose the Class Outlier Factor (COF) which measures
the degree of being a Class Outlier for a data object. Our work
includes a proposal of a new algorithm towards mining of the Class
Outliers, presenting experimental results applied on various domains
of real world datasets and finally a comparison study with other
related methods is performed.
Abstract: The mechanical deformation and the electrical conductivity of lanthanum strontium cobalt ferrite oxide under uniaxial compression were investigated at various temperatures up to 1073 K. The material reveals a rather complex mechanical behaviour related to its ferroelasticity and completely different stress-strain curves are obtained during the 1st and 2nd loading cycles. A distinctive ferroelastic creep was observed at 293 K whilst typical ferroelastic stress-strain curve were obtained in the temperature range from 473 K to 873 K. At 1073 K, on the other hand, high-temperature creep deformation was observed instead of ferroelastic deformation. The conductivity increases with increasing compressive stress at all the temperatures. The increase in conductivity is related to both geometrical and piezoelectric effects. From 293 K to 873 K, where the material exhibits ferroelastic behaviour, the variation in the total conductivity decreases with increasing temperature. The contribution of the piezoelectric effect to the total conductivity variation also decreases with increasing temperature and the maximum in piezoconductivity has a value of about 0.75 % at 293 K for a compressive stress of 100 MPa. There is no effect of domain switching on conductivity except for the geometric effect. At 1073 K, the conductivity is simply proportional to the compressive strain.
Abstract: Real world Speaker Identification (SI) application
differs from ideal or laboratory conditions causing perturbations that
leads to a mismatch between the training and testing environment
and degrade the performance drastically. Many strategies have been
adopted to cope with acoustical degradation; wavelet based Bayesian
marginal model is one of them. But Bayesian marginal models
cannot model the inter-scale statistical dependencies of different
wavelet scales. Simple nonlinear estimators for wavelet based
denoising assume that the wavelet coefficients in different scales are
independent in nature. However wavelet coefficients have significant
inter-scale dependency. This paper enhances this inter-scale
dependency property by a Circularly Symmetric Probability Density
Function (CS-PDF) related to the family of Spherically Invariant
Random Processes (SIRPs) in Log Gabor Wavelet (LGW) domain
and corresponding joint shrinkage estimator is derived by Maximum
a Posteriori (MAP) estimator. A framework is proposed based on
these to denoise speech signal for automatic speaker identification
problems. The robustness of the proposed framework is tested for
Text Independent Speaker Identification application on 100 speakers
of POLYCOST and 100 speakers of YOHO speech database in three
different noise environments. Experimental results show that the
proposed estimator yields a higher improvement in identification
accuracy compared to other estimators on popular Gaussian Mixture
Model (GMM) based speaker model and Mel-Frequency Cepstral
Coefficient (MFCC) features.
Abstract: The human friendly interaction is the key function of a human-centered system. Over the years, it has received much attention to develop the convenient interaction through intention recognition. Intention recognition processes multimodal inputs including speech, face images, and body gestures. In this paper, we suggest a novel approach of intention recognition using a graph representation called Intention Graph. A concept of valid intention is proposed, as a target of intention recognition. Our approach has two phases: goal recognition phase and intention recognition phase. In the goal recognition phase, we generate an action graph based on the observed actions, and then the candidate goals and their plans are recognized. In the intention recognition phase, the intention is recognized with relevant goals and user profile. We show that the algorithm has polynomial time complexity. The intention graph is applied to a simple briefcase domain to test our model.
Abstract: A study of the obtainable watermark data rate for information hiding algorithms is presented in this paper. As the perceptual entropy for wideband monophonic audio signals is in the range of four to five bits per sample, a significant amount of additional information can be inserted into signal without causing any perceptual distortion. Experimental results showed that transform domain watermark embedding outperforms considerably watermark embedding in time domain and that signal decompositions with a high gain of transform coding, like the wavelet transform, are the most suitable for high data rate information hiding. Keywords?Digital watermarking, information hiding, audio watermarking, watermark data rate.