Abstract: Game theory could be used to analyze the conflicted
issues in the field of information hiding. In this paper, 2-phase game
can be used to build the embedder-attacker system to analyze the
limits of hiding capacity of embedding algorithms: the embedder
minimizes the expected damage and the attacker maximizes it. In the
system, the embedder first consumes its resource to build embedded
units (EU) and insert the secret information into EU. Then the attacker
distributes its resource evenly to the attacked EU. The expected
equilibrium damage, which is maximum damage in value from the
point of view of the attacker and minimum from the embedder against
the attacker, is evaluated by the case when the attacker attacks a
subset from all the EU. Furthermore, the optimal equilibrium capacity
of hiding information is calculated through the optimal number of EU
with the embedded secret information. Finally, illustrative examples
of the optimal equilibrium capacity are presented.
Abstract: An approach of design of stable of control systems with ultimately wide ranges of uncertainly disturbed parameters is offered. The method relies on using of nonlinear structurally stable functions from catastrophe theory as controllers. Theoretical part presents an analysis of designed nonlinear second-order control systems. As more important the integrators in series, canonical controllable form and Jordan forms are considered. The analysis resumes that due to added controllers systems become stable and insensitive to any disturbance of parameters. Experimental part presents MATLAB simulation of design of control systems of epidemic spread, aircrafts angular motion and submarine depth. The results of simulation confirm the efficiency of offered method of design. KeywordsCatastrophes, robust control, simulation, uncertain parameters.
Abstract: Authentication plays a vital role in many secure
systems. Most of these systems require user to log in with his or her
secret password or pass phrase before entering it. This is to ensure all
the valuables information is kept confidential guaranteeing also its
integrity and availability. However, to achieve this goal, users are
required to memorize high entropy passwords or pass phrases.
Unfortunately, this sometimes causes difficulty for user to remember
meaningless strings of data. This paper presents a new scheme which
assigns a weight to each personal question given to the user in
revealing the encrypted secrets or password. Concentration of this
scheme is to offer fault tolerance to users by allowing them to forget
the specific password to a subset of questions and still recover the
secret and achieve successful authentication. Comparison on level of
security for weight-based and weightless secret recovery scheme is
also discussed. The paper concludes with the few areas that requires
more investigation in this research.
Abstract: Segmentation of a color image composed of different
kinds of regions can be a hard problem, namely to compute for an
exact texture fields. The decision of the optimum number of
segmentation areas in an image when it contains similar and/or un
stationary texture fields. A novel neighborhood-based segmentation
approach is proposed. A genetic algorithm is used in the proposed
segment-pass optimization process. In this pass, an energy function,
which is defined based on Markov Random Fields, is minimized. In
this paper we use an adaptive threshold estimation method for image
thresholding in the wavelet domain based on the generalized
Gaussian distribution (GGD) modeling of sub band coefficients. This
method called Normal Shrink is computationally more efficient and
adaptive because the parameters required for estimating the threshold
depend on sub band data energy that used in the pre-stage of
segmentation. A quad tree is employed to implement the multi
resolution framework, which enables the use of different strategies at
different resolution levels, and hence, the computation can be
accelerated. The experimental results using the proposed
segmentation approach are very encouraging.
Abstract: The efficient use of available licensed spectrum is
becoming more and more critical with increasing demand and usage
of the radio spectrum. This paper shows how the use of spectrum as
well as dynamic spectrum management can be effectively managed
and spectrum allocation schemes in the wireless communication
systems be implemented and used, in future. This paper would be an
attempt towards better utilization of the spectrum. This research will
focus on the decision-making process mainly, with an
assumption that the radio environment has already been sensed and
the QoS requirements for the application have been specified either
by the sensed radio environment or by the secondary user itself. We
identify and study the characteristic parameters of Cognitive Radio
and use Genetic Algorithm for spectrum allocation. Performance
evaluation is done using MATLAB toolboxes.
Abstract: An optimal control of Reverse Osmosis (RO) plant is
studied in this paper utilizing the auto tuning concept in conjunction
with PID controller. A control scheme composing an auto tuning
stochastic technique based on an improved Genetic Algorithm (GA) is
proposed. For better evaluation of the process in GA, objective
function defined newly in sense of root mean square error has been
used. Also in order to achieve better performance of GA, more
pureness and longer period of random number generation in operation
are sought. The main improvement is made by replacing the uniform
distribution random number generator in conventional GA technique
to newly designed hybrid random generator composed of Cauchy
distribution and linear congruential generator, which provides
independent and different random numbers at each individual steps in
Genetic operation. The performance of newly proposed GA tuned
controller is compared with those of conventional ones via simulation.
Abstract: The binary phase-only filter digital watermarking
embeds the phase information of the discrete Fourier transform of the
image into the corresponding magnitudes for better image authentication.
The paper proposed an approach of how to implement watermark
embedding by quantizing the magnitude, with discussing how to
regulate the quantization steps based on the frequencies of the magnitude
coefficients of the embedded watermark, and how to embed the
watermark at low frequency quantization. The theoretical analysis and
simulation results show that algorithm flexibility, security, watermark
imperceptibility and detection performance of the binary phase-only
filter digital watermarking can be effectively improved with quantization
based watermark embedding, and the robustness against JPEG
compression will also be increased to some extent.
Abstract: Modern spatial database management systems require a unique Spatial Access Method (SAM) in order solve complex spatial quires efficiently. In this case the spatial data structure takes a prominent place in the SAM. Inadequate data structure leads forming poor algorithmic choices and forging deficient understandings of algorithm behavior on the spatial database. A key step in developing a better semantic spatial object data structure is to quantify the performance effects of semantic and outlier detections that are not reflected in the previous tree structures (R-Tree and its variants). This paper explores a novel SSRO-Tree on SAM to the Topo-Semantic approach. The paper shows how to identify and handle the semantic spatial objects with outlier objects during page overflow/underflow, using gain/loss metrics. We introduce a new SSRO-Tree algorithm which facilitates the achievement of better performance in practice over algorithms that are superior in the R*-Tree and RO-Tree by considering selection queries.
Abstract: In this paper we present a novel approach for density estimation. The proposed approach is based on using the logistic regression model to get initial density estimation for the given empirical density. The empirical data does not exactly follow the logistic regression model, so, there will be a deviation between the empirical density and the density estimated using logistic regression model. This deviation may be positive and/or negative. In this paper we use a linear combination of Gaussian (LCG) with positive and negative components as a model for this deviation. Also, we will use the expectation maximization (EM) algorithm to estimate the parameters of LCG. Experiments on real images demonstrate the accuracy of our approach.
Abstract: Simultaneous determination of multicomponents of phenol, resorcinol and catechol with a chemometric technique a PCranking artificial neural network (PCranking-ANN) algorithm is reported in this study. Based on the data correlation coefficient method, 3 representative PCs are selected from the scores of original UV spectral data (35 PCs) as the original input patterns for ANN to build a neural network model. The results obtained by iterating 8000 .The RMSEP for phenol, resorcinol and catechol with PCranking- ANN were 0.6680, 0.0766 and 0.1033, respectively. Calibration matrices were 0.50-21.0, 0.50-15.1 and 0.50-20.0 μg ml-1 for phenol, resorcinol and catechol, respectively. The proposed method was successfully applied for the determination of phenol, resorcinol and catechol in synthetic and water samples.
Abstract: As the network based technologies become
omnipresent, demands to secure networks/systems against threat
increase. One of the effective ways to achieve higher security is
through the use of intrusion detection systems (IDS), which are a
software tool to detect anomalous in the computer or network. In this
paper, an IDS has been developed using an improved machine
learning based algorithm, Locally Linear Neuro Fuzzy Model
(LLNF) for classification whereas this model is originally used for
system identification. A key technical challenge in IDS and LLNF
learning is the curse of high dimensionality. Therefore a feature
selection phase is proposed which is applicable to any IDS. While
investigating the use of three feature selection algorithms, in this
model, it is shown that adding feature selection phase reduces
computational complexity of our model. Feature selection algorithms
require the use of a feature goodness measure. The use of both a
linear and a non-linear measure - linear correlation coefficient and
mutual information- is investigated respectively
Abstract: This paper presents a software quality support tool, a
Java source code evaluator and a code profiler based on
computational intelligence techniques. It is Java prototype software
developed by AI Group [1] from the Research Laboratories at
Universidad de Palermo: an Intelligent Java Analyzer (in Spanish:
Analizador Java Inteligente, AJI). It represents a new approach to
evaluate and identify inaccurate source code usage and transitively,
the software product itself.
The aim of this project is to provide the software development
industry with a new tool to increase software quality by extending
the value of source code metrics through computational intelligence.
Abstract: In this paper we describe the recognition process of Greek compound words using the PC-KIMMO software. We try to show certain limitations of the system with respect to the principles of compound formation in Greek. Moreover, we discuss the computational processing of phenomena such as stress and syllabification which are indispensable for the analysis of such constructions and we try to propose linguistically-acceptable solutions within the particular system.
Abstract: CEMTool is a command style design and analyzing
package for scientific and technological algorithm and a matrix based
computation language. In this paper, we present new 2D & 3D
finite element method (FEM) packages for CEMTool. We discuss
the detailed structures and the important features of pre-processor,
solver, and post-processor of CEMTool 2D & 3D FEM packages. In
contrast to the existing MATLAB PDE Toolbox, our proposed FEM
packages can deal with the combination of the reserved words. Also,
we can control the mesh in a very effective way. With the introduction
of new mesh generation algorithm and fast solving technique, our
FEM packages can guarantee the shorter computational time than
MATLAB PDE Toolbox. Consequently, with our new FEM packages,
we can overcome some disadvantages or limitations of the existing
MATLAB PDE Toolbox.
Abstract: Embedded hardware simulator is a valuable computeraided
tool for embedded application development. This paper focuses
on the ARM926EJ-S MMU, builds state transition models and
formally verifies critical properties for the models. The state transition
models include loading instruction model, reading data model, and
writing data model. The properties of the models are described by
CTL specification language, and they are verified in VIS. The results
obtained in VIS demonstrate that the critical properties of MMU are
satisfied in the state transition models. The correct models can be
used to implement the MMU component in our simulator. In the
end of this paper, the experimental results show that the MMU can
successfully accomplish memory access requests from CPU.
Abstract: This paper proposes a new model to support user
queries on postgraduate research information at Universiti Tenaga
Nasional. The ontology to be developed will contribute towards
shareable and reusable domain knowledge that makes knowledge
assets intelligently accessible to both people and software. This work
adapts a methodology for ontology development based on the
framework proposed by Uschold and King. The concepts and
relations in this domain are represented in a class diagram using the
Protégé software. The ontology will be used to support a menudriven
query system for assisting students in searching for
information related to postgraduate research at the university.
Abstract: Using maximal consistent blocks of tolerance relation
on the universe in incomplete decision table, the concepts of join block
and meet block are introduced and studied. Including tolerance class,
other blocks such as tolerant kernel and compatible kernel of an object
are also discussed at the same time. Upper and lower approximations
based on those blocks are also defined. Default definite decision rules
acquired from incomplete decision table are proposed in the paper. An
incremental algorithm to update default definite decision rules is
suggested for effective mining tasks from incomplete decision table
into which data is appended. Through an example, we demonstrate
how default definite decision rules based on maximal consistent
blocks, join blocks and meet blocks are acquired and how optimization
is done in support of discernibility matrix and discernibility function
in the incomplete decision table.
Abstract: Duplicated region detection is a technical method to
expose copy-paste forgeries on digital images. Copy-paste is one
of the common types of forgeries to clone portion of an image
in order to conceal or duplicate special object. In this type of
forgery detection, extracting robust block feature and also high
time complexity of matching step are two main open problems.
This paper concentrates on computational time and proposes a local
block matching algorithm based on block clustering to enhance time
complexity. Time complexity of the proposed algorithm is formulated
and effects of two parameter, block size and number of cluster, on
efficiency of this algorithm are considered. The experimental results
and mathematical analysis demonstrate this algorithm is more costeffective
than lexicographically algorithms in time complexity issue
when the image is complex.
Abstract: This paper aims at improving web server performance
by establishing a middleware layer between web and database
servers, which minimizes the overload on the database server. A
middleware system has been developed as a service mainly to
improve the performance. This system manages connection accesses
in a way that would result in reducing the overload on the database
server. In addition to the connection management, this system acts as
an object-oriented model for best utilization of operating system
resources. A web developer can use this Service Broker to improve
web server performance.
Abstract: Vehicular communications play a substantial role in providing safety in transportation by means of safety message exchange. Researchers have proposed several solutions for securing safety messages. Protocols based on a fixed key infrastructure are more efficient in implementation and maintain stronger security in comparison with dynamic structures. These protocols utilize zone partitioning to establish distinct key infrastructure under Certificate Authority (CA) supervision in different regions. Secure anonymous broadcasting (SAB) is one of these protocols that preserves most of security aspects but it has some deficiencies in practice. A very important issue is region change of a vehicle for its mobility. Changing regions leads to change of CA and necessity of having new key set to resume communication. In this paper, we propose solutions for informing vehicles about region change to obtain new key set before entering next region. This hinders attackers- intrusion, packet loss and lessons time delay. We also make key request messages secure by confirming old CA-s public key to the message, hence stronger security for safety message broadcasting is attained.