Abstract: Intrusion detection systems (IDS)are crucial components
of the security mechanisms of today-s computer systems.
Existing research on intrusion detection has focused on sequential
intrusions. However, intrusions can also be formed by concurrent
interactions of multiple processes. Some of the intrusions caused
by these interactions cannot be detected using sequential intrusion
detection methods. Therefore, there is a need for a mechanism that
views the distributed system as a whole. L-BIDS (Lattice-Based
Intrusion Detection System) is proposed to address this problem. In
the L-BIDS framework, a library of intrusions and distributed traces
are represented as lattices. Then these lattices are compared in order
to detect intrusions in the distributed traces.
Abstract: this paper presents an auto-regressive network called the Auto-Regressive Multi-Context Recurrent Neural Network (ARMCRN), which forecasts the daily peak load for two large power plant systems. The auto-regressive network is a combination of both recurrent and non-recurrent networks. Weather component variables are the key elements in forecasting because any change in these variables affects the demand of energy load. So the AR-MCRN is used to learn the relationship between past, previous, and future exogenous and endogenous variables. Experimental results show that using the change in weather components and the change that occurred in past load as inputs to the AR-MCRN, rather than the basic weather parameters and past load itself as inputs to the same network, produce higher accuracy of predicted load. Experimental results also show that using exogenous and endogenous variables as inputs is better than using only the exogenous variables as inputs to the network.
Abstract: In this paper, we introduce an mobile agent framework
with proactive load balancing for ambient intelligence (AmI) environments.
One of the main obstacles of AmI is the scalability in
which the openness of AmI environment introduces dynamic resource
requirements on agencies. To mediate this scalability problem, our
framework proposes a load balancing module to proactively analyze
the resource consumption of network bandwidth and preferred agencies
to suggest the optimal communication method to its user. The
framework generally formulates an AmI environment that consists
of three main components: (1) mobile devices, (2) hosts or agencies,
and (3) directory service center (DSC). A preliminary implementation
was conducted with NetLogo and the experimental results show that
the proposed approach provides enhanced system performance by
minimizing the network utilization to provide users with responsive
services.
Abstract: one of the significant factors for improving the
accuracy of Land Surface Temperature (LST) retrieval is the correct
understanding of the directional anisotropy for thermal radiance. In
this paper, the multiple scattering effect between heterogeneous
non-isothermal surfaces is described rigorously according to the
concept of configuration factor, based on which a directional thermal
radiance model is built, and the directional radiant character for urban
canopy is analyzed. The model is applied to a simple urban canopy
with row structure to simulate the change of Directional Brightness
Temperature (DBT). The results show that the DBT is aggrandized
because of the multiple scattering effects, whereas the change range of
DBT is smoothed. The temperature difference, spatial distribution,
emissivity of the components can all lead to the change of DBT. The
“hot spot" phenomenon occurs when the proportion of high
temperature component in the vision field came to a head. On the other
hand, the “cool spot" phenomena occur when low temperature
proportion came to the head. The “spot" effect disappears only when
the proportion of every component keeps invariability. The model
built in this paper can be used for the study of directional effect on
emissivity, the LST retrieval over urban areas and the adjacency effect
of thermal remote sensing pixels.
Abstract: Three-dimensional geometric models have been used
to present architectural and engineering works, showing their final
configuration. When the clarification of a detail or the constitution of
a construction step in needed, these models are not appropriate. They
do not allow the observation of the construction progress of a
building. Models that could present dynamically changes of the
building geometry are a good support to the elaboration of projects.
Techniques of geometric modeling and virtual reality were used to
obtain models that could visually simulate the construction activity.
The applications explain the construction work of a cavity wall and a
bridge. These models allow the visualization of the physical
progression of the work following a planned construction sequence,
the observation of details of the form of every component of the
works and support the study of the type and method of operation of
the equipment applied in the construction. These models presented
distinct advantage as educational aids in first-degree courses in Civil
Engineering. The use of Virtual Reality techniques in the
development of educational applications brings new perspectives to
the teaching of subjects related to the field of civil construction.
Abstract: In any trust model, the two information sources that a peer relies on to predict trustworthiness of another peer are direct experience as well as reputation. These two vital components evolve over time. Trust evolution is an important issue, where the objective is to observe a sequence of past values of a trust parameter and determine the future estimates. Unfortunately, trust evolution algorithms received little attention and the proposed algorithms in the literature do not comply with the conditions and the nature of trust. This paper contributes to this important problem in the following ways: (a) presents an algorithm that manages and models trust evolution in a P2P environment, (b) devises new mechanisms for effectively maintaining trust values based on the conditions that influence trust evolution , and (c) introduces a new methodology for incorporating trust-nurture incentives into the trust evolution algorithm. Simulation experiments are carried out to evaluate our trust evolution algorithm.
Abstract: In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.
Abstract: Over the years, there is a growing trend towards
quality-based specifications in highway construction. In many
Quality Control/Quality Assurance (QC/QA) specifications, the
contractor is primarily responsible for quality control of the process,
whereas the highway agency is responsible for testing the acceptance
of the product. A cooperative investigation was conducted in Illinois
over several years to develop a prototype End-Result Specification
(ERS) for asphalt pavement construction. The final characteristics of
the product are stipulated in the ERS and the contractor is given
considerable freedom in achieving those characteristics. The risk for
the contractor or agency depends on how the acceptance limits and
processes are specified. Stochastic simulation models are very useful
in estimating and analyzing payment risk in ERS systems and these
form an integral part of the Illinois-s prototype ERS system. This
paper describes the development of an innovative methodology to
estimate the variability components in in-situ density, air voids and
asphalt content data from ERS projects. The information gained from
this would be crucial in simulating these ERS projects for estimation
and analysis of payment risks associated with asphalt pavement
construction. However, these methods require at least two parties to
conduct tests on all the split samples obtained according to the
sampling scheme prescribed in present ERS implemented in Illinois.
Abstract: In recent years, it has been proposed security
architecture for sensor network.[2][4]. One of these, TinySec by Chris
Kalof, Naveen Sastry, David Wagner had proposed Link layer security
architecture, considering some problems of sensor network. (i.e :
energy, bandwidth, computation capability,etc). The TinySec employs
CBC_mode of encryption and CBC-MAC for authentication based on
SkipJack Block Cipher. Currently, This TinySec is incorporated in the
TinyOS for sensor network security.
This paper introduces TinyHash based on general hash algorithm.
TinyHash is the module in order to replace parts of authentication and
integrity in the TinySec. it implies that apply hash algorithm on
TinySec architecture. For compatibility about TinySec, Components
in TinyHash is constructed as similar structure of TinySec. And
TinyHash implements the HMAC component for authentication and
the Digest component for integrity of messages. Additionally, we
define the some interfaces for service associated with hash algorithm.
Abstract: Abovepresented work deals with the new scope of application of information and communication technologies for the improvement of the election process in the biased environment. We are introducing a new concept of construction of the information-communication system for the election participant. It consists of four main components: Software, Physical Infrastructure, Structured Information and the Trained Stuff. The Structured Information is the bases of the whole system and is the collection of all possible events (irregularities among them) at the polling stations, which are structured in special templates, forms and integrated in mobile devices.The software represents a package of analytic modules, which operates with the dynamic database. The application of modern communication technologies facilities the immediate exchange of information and of relevant documents between the polling stations and the Server of the participant. No less important is the training of the staff for the proper functioning of the system. The e-training system with various modules should be applied in this respect. The presented methodology is primarily focused on the election processes in the countries of emerging democracies.It can be regarded as the tool for the monitoring of elections process by the political organization(s) and as one of the instruments to foster the spread of democracy in these countries.
Abstract: This paper presents design, analysis and comparison of the different rotor type permanent magnet machines. The presented machines are designed as having same geometrical dimensions and same materials for comparison. The main machine parameters of interior and exterior rotor type machines including eddy current effect, torque-speed characteristics and magnetic analysis are investigated using MAXWELL program. With this program, the components of the permanent magnet machines can be calculated with high accuracy. Six types of Permanent machines are compared with respect to their topology, size, magnetic field, air gap flux, voltage, torque, loss and efficiency. The analysis results demonstrate the effectiveness of the proposed machines design methodology. We believe that, this study will be a helpful resource in terms of examination and comparison of the basic structure and magnetic features of the PM (Permanent magnet) machines which have different rotor structure.
Abstract: Most of the Question Answering systems
composed of three main modules: question processing,
document processing and answer processing. Question
processing module plays an important role in QA systems. If
this module doesn't work properly, it will make problems for
other sections. Moreover answer processing module is an
emerging topic in Question Answering, where these systems
are often required to rank and validate candidate answers.
These techniques aiming at finding short and precise answers
are often based on the semantic classification.
This paper discussed about a new model for question
answering which improved two main modules, question
processing and answer processing.
There are two important components which are the bases
of the question processing. First component is question
classification that specifies types of question and answer.
Second one is reformulation which converts the user's
question into an understandable question by QA system in a
specific domain. Answer processing module, consists of
candidate answer filtering, candidate answer ordering
components and also it has a validation section for interacting
with user. This module makes it more suitable to find exact
answer. In this paper we have described question and answer
processing modules with modeling, implementing and
evaluating the system. System implemented in two versions.
Results show that 'Version No.1' gave correct answer to 70%
of questions (30 correct answers to 50 asked questions) and
'version No.2' gave correct answers to 94% of questions (47
correct answers to 50 asked questions).
Abstract: Due to the tremendous amount of information provided
by the World Wide Web (WWW) developing methods for mining
the structure of web-based documents is of considerable interest. In
this paper we present a similarity measure for graphs representing
web-based hypertext structures. Our similarity measure is mainly
based on a novel representation of a graph as linear integer strings,
whose components represent structural properties of the graph. The
similarity of two graphs is then defined as the optimal alignment of
the underlying property strings. In this paper we apply the well known
technique of sequence alignments for solving a novel and challenging
problem: Measuring the structural similarity of generalized trees.
In other words: We first transform our graphs considered as high
dimensional objects in linear structures. Then we derive similarity
values from the alignments of the property strings in order to
measure the structural similarity of generalized trees. Hence, we
transform a graph similarity problem to a string similarity problem for
developing a efficient graph similarity measure. We demonstrate that
our similarity measure captures important structural information by
applying it to two different test sets consisting of graphs representing
web-based document structures.
Abstract: In analyzing large scale nonlinear dynamical systems,
it is often desirable to treat the overall system as a collection of
interconnected subsystems. Solutions properties of the large scale
system are then deduced from the solution properties of the
individual subsystems and the nature of the interconnections. In this
paper a new approach is proposed for the stability analysis of large
scale systems, which is based upon the concept of vector Lyapunov
functions and the decomposition methods. The present results make
use of graph theoretic decomposition techniques in which the overall
system is partitioned into a hierarchy of strongly connected
components. We show then, that under very reasonable assumptions,
the overall system is stable once the strongly connected subsystems
are stables. Finally an example is given to illustrate the constructive
methodology proposed.
Abstract: Accurate modeling of high speed RLC interconnects
has become a necessity to address signal integrity issues in current
VLSI design. To accurately model a dispersive system of interconnects
at higher frequencies; a full-wave analysis is required.
However, conventional circuit simulation of interconnects with full
wave models is extremely CPU expensive. We present an algorithm
for reducing large VLSI circuits to much smaller ones with similar
input-output behavior. A key feature of our method, called Frequency
Shift Technique, is that it is capable of reducing linear time-varying
systems. This enables it to capture frequency-translation and sampling
behavior, important in communication subsystems such as mixers,
RF components and switched-capacitor filters. Reduction is obtained
by projecting the original system described by linear differential
equations into a lower dimension. Experiments have been carried out
using Cadence Design Simulator cwhich indicates that the proposed
technique achieves more % reduction with less CPU time than the
other model order reduction techniques existing in literature. We
also present applications to RF circuit subsystems, obtaining size
reductions and evaluation speedups of orders of magnitude with
insignificant loss of accuracy.
Abstract: Machine tools are improved capacity remarkably during the 20th century. Improving the precision of machine tools are related with precision of products and accurate processing is always associated with the subject of interest. There are a lot of the elements that determine the precision of the machine, as guides, motors, structure, control, etc. In this paper we focused on the phenomenon that vertical movement system has worse precision than horizontal movement system even they were made up with same components. The vertical movement system needs to be studied differently from the horizontal movement system to develop its precision. The vertical movement system has load on its transfer direction and it makes the movement system weak in precision than the horizontal one. Some machines have mechanical counter balance, hydraulic or pneumatic counter balance to compensate the weight of the machine head. And there is several type of compensating the weight. It can push the machine head and also can use chain or wire lope to transfer the compensating force from counter balance to machine head. According to the type of compensating, there could be error from friction, pressure error of hydraulic or pressure control error. Also according to what to use for transferring the compensating force, transfer error of compensating force could be occur.
Abstract: The study of cytokine expression in mice under the
influence of inactivated poliovirus and Imovaks polio vaccine in
combination with derivatives of chitosan shows various kinds of
processes. There is a significant increase in IL-12 in the serum of
immunized animals, which should stimulate the production of IFN-γ
NK-cells and T-cells and polarize the immune response to Th1 type.
Thus, the derivatives of chitosan can promote cell component of the
immune response, providing a full antiviral immunity.
Abstract: An early and accurate detection of Alzheimer's disease (AD) is an important stage in the treatment of individuals suffering from AD. We present an approach based on the use of structural magnetic resonance imaging (sMRI) phase images to distinguish between normal controls (NC), mild cognitive impairment (MCI) and AD patients with clinical dementia rating (CDR) of 1. Independent component analysis (ICA) technique is used for extracting useful features which form the inputs to the support vector machines (SVM), K nearest neighbour (kNN) and multilayer artificial neural network (ANN) classifiers to discriminate between the three classes. The obtained results are encouraging in terms of classification accuracy and effectively ascertain the usefulness of phase images for the classification of different stages of Alzheimer-s disease.
Abstract: The elution process for the removal of Co and Cu from clinoptilolite as an ion-exchanger was investigated using three parameters: bed volume, pH and contact time. The present paper study has shown quantitatively that acid concentration has a significant effect on the elution process. The favorable eluant concentration was found to be 2 M HCl and 2 M H2SO4, respectively. The multi-component equilibrium relationship in the process can be very complex, and perhaps ill-defined. In such circumstances, it is preferable to use a non-parametric technique such as Neural Network to represent such an equilibrium relationship.
Abstract: Occurrence of a multiple-points fault in machine operations could result in exhibiting complex fault signatures, which could result in lowering fault diagnosis accuracy. In this study, a multiple-points defect model (MPDM) is proposed which can simulate fault signature-s dynamics for n-points bearing faults. Furthermore, this study identifies that in case of multiple-points fault in the rotary machine, the location of the dominant component of defect frequency shifts depending upon the relative location of the fault points which could mislead the fault diagnostic model to inaccurate detections. Analytical and experimental results are presented to characterize and validate the variation in the dominant component of defect frequency. Based on envelop detection analysis, a modification is recommended in the existing fault diagnostic models to consider the multiples of defect frequency rather than only considering the frequency spectrum at the defect frequency in order to incorporate the impact of multiple points fault.