Abstract: Introducing survivability into embedded real-time system (ERTS) can improve the survivability power of the system. This paper mainly discusses about the survivability of ERTS. The first is the survivability origin of ERTS. The second is survivability analysis. According to the definition of survivability based on survivability specification and division of the entire survivability analysis process for ERTS, a survivability analysis profile is presented. The quantitative analysis model of this profile is emphasized and illuminated in detail, the quantifying analysis of system was showed helpful to evaluate system survivability more accurate. The third is platform design of survivability analysis. In terms of the profile, the analysis process is encapsulated and assembled into one platform, on which quantification, standardization and simplification of survivability analysis are all achieved. The fourth is survivability design. According to character of ERTS, strengthened design method is selected to realize system survivability design. Through the analysis of embedded mobile video-on-demand system, intrusion tolerant technology is introduced in whole survivability design.
Abstract: In-core memory requirement is a bottleneck in solving
large three dimensional Navier-Stokes finite element problem
formulations using sparse direct solvers. Out-of-core solution
strategy is a viable alternative to reduce the in-core memory
requirements while solving large scale problems. This study
evaluates the performance of various out-of-core sequential solvers
based on multifrontal or supernodal techniques in the context of
finite element formulations for three dimensional problems on a
Windows platform. Here three different solvers, HSL_MA78,
MUMPS and PARDISO are compared. The performance of these
solvers is evaluated on a 64-bit machine with 16GB RAM for finite
element formulation of flow through a rectangular channel. It is
observed that using out-of-core PARDISO solver, relatively large
problems can be solved. The implementation of Newton and
modified Newton's iteration is also discussed.
Abstract: The main goal of this paper is to show a possibility, how to solve numerically elliptic boundary value problems arising in 2D linear elasticity by using the fictitious domain method (FDM) and the Total-FETI domain decomposition method. We briefly mention the theoretical background of these methods and demonstrate their performance on a benchmark.
Abstract: This paper presents a mathematical model and a
methodology to analyze the losses in transmission expansion
planning (TEP) under uncertainty in demand. The methodology is
based on discrete particle swarm optimization (DPSO). DPSO is a
useful and powerful stochastic evolutionary algorithm to solve the
large-scale, discrete and nonlinear optimization problems like TEP.
The effectiveness of the proposed idea is tested on an actual
transmission network of the Azerbaijan regional electric company,
Iran. The simulation results show that considering the losses even for
transmission expansion planning of a network with low load growth
is caused that operational costs decreases considerably and the
network satisfies the requirement of delivering electric power more
reliable to load centers.
Abstract: In this paper, some practical solid transportation models are formulated considering per trip capacity of each type of conveyances with crisp and rough unit transportation costs. This is applicable for the system in which full vehicles, e.g. trucks, rail coaches are to be booked for transportation of products so that transportation cost is determined on the full of the conveyances. The models with unit transportation costs as rough variables are transformed into deterministic forms using rough chance constrained programming with the help of trust measure. Numerical examples are provided to illustrate the proposed models in crisp environment as well as with unit transportation costs as rough variables.
Abstract: A satured liquid is warmed until boiling in a parallelepipedic boiler. The heat is supplied in a liquid through the horizontal bottom of the boiler, the other walls being adiabatic. During the process of boiling, the liquid evaporates through its free surface by deforming it. This surface which subdivides the boiler into two regions occupied on both sides by the boiled liquid (broth) and its vapor which surmounts it. The broth occupying the region and its vapor the superior region. A two- fluids model is used to describe the dynamics of the broth, its vapor and their interface. In this model, the broth is treated as a monophasic fluid (homogeneous model) and form with its vapor adiphasic pseudo fluid (two-fluid model). Furthermore, the interface is treated as a zone of mixture characterized by superficial void fraction noted α* . The aim of this article is to describe the dynamics of the interface between the boiled fluid and its vapor within a boiler. The resolution of the problem allowed us to show the evolution of the broth and the level of the liquid.
Abstract: A new concept for long-term reagent storage for Labon- a-Chip (LoC) devices is described. Here we present a polymer multilayer stack with integrated stick packs for long-term storage of several liquid reagents, which are necessary for many diagnostic applications. Stick packs are widely used in packaging industry for storing solids and liquids for long time. The storage concept fulfills two main requirements: First, a long-term storage of reagents in stick packs without significant losses and interaction with surroundings, second, on demand releasing of liquids, which is realized by pushing a membrane against the stick pack through pneumatic pressure. This concept enables long-term on-chip storage of liquid reagents at room temperature and allows an easy implementation in different LoC devices.
Abstract: In this paper, we propose an algorithm to compute
initial cluster centers for K-means clustering. Data in a cell is
partitioned using a cutting plane that divides cell in two smaller cells.
The plane is perpendicular to the data axis with the highest variance
and is designed to reduce the sum squared errors of the two cells as
much as possible, while at the same time keep the two cells far apart
as possible. Cells are partitioned one at a time until the number of
cells equals to the predefined number of clusters, K. The centers of
the K cells become the initial cluster centers for K-means. The
experimental results suggest that the proposed algorithm is effective,
converge to better clustering results than those of the random
initialization method. The research also indicated the proposed
algorithm would greatly improve the likelihood of every cluster
containing some data in it.
Abstract: In this paper, a generalized form of the Banzhaf-Owen value for cooperative fuzzy games with a coalition structure is proposed. Its axiomatic system is given by extending crisp case. In order to better understand the Banzhaf-Owen value for fuzzy games with a coalition structure, we briefly introduce the Banzhaf-Owen values for two special kinds of fuzzy games with a coalition structure, and give their explicit forms.
Abstract: In this paper back-propagation artificial neural
network (BPANN) with Levenberg–Marquardt algorithm is
employed to predict the limiting drawing ratio (LDR) of the deep
drawing process. To prepare a training set for BPANN, some finite
element simulations were carried out. die and punch radius, die arc
radius, friction coefficient, thickness, yield strength of sheet and
strain hardening exponent were used as the input data and the LDR
as the specified output used in the training of neural network. As a
result of the specified parameters, the program will be able to
estimate the LDR for any new given condition. Comparing FEM and
BPANN results, an acceptable correlation was found.
Abstract: In this work we study the reflection of circularly
polarised light from a nano-structured biological material found in
the exocuticle of scarabus beetles. This material is made of a stack
of ultra-thin (~5 nm) uniaxial layers arranged in a left-handed
helicoidal stack, which resonantly reflects circularly polarized light.
A chirp in the layer thickness combined with a finite absorption
coefficient produce a broad smooth reflectance spectrum. By
comparing model calculations and electron microscopy with
measured spectra we can explain our observations and quantify most
relevant structural parameters.
Abstract: In image processing and visualization, comparing two
bitmapped images needs to be compared from their pixels by matching
pixel-by-pixel. Consequently, it takes a lot of computational time
while the comparison of two vector-based images is significantly
faster. Sometimes these raster graphics images can be approximately
converted into the vector-based images by various techniques. After
conversion, the problem of comparing two raster graphics images
can be reduced to the problem of comparing vector graphics images.
Hence, the problem of comparing pixel-by-pixel can be reduced to
the problem of polynomial comparisons. In computer aided geometric
design (CAGD), the vector graphics images are the composition of
curves and surfaces. Curves are defined by a sequence of control
points and their polynomials. In this paper, the control points will be
considerably used to compare curves. The same curves after relocated
or rotated are treated to be equivalent while two curves after different
scaled are considered to be similar curves. This paper proposed an
algorithm for comparing the polynomial curves by using the control
points for equivalence and similarity. In addition, the geometric
object-oriented database used to keep the curve information has also
been defined in XML format for further used in curve comparisons.
Abstract: In this paper, position vector of a partially null unit speed curve with respect to standard frame of Minkowski space-time is studied. First, it is proven that position vector of every partially null unit speed curve satisfies a vector differential equation of fourth order. In terms of solution of the differential equation, position vector of a partially null unit speed curve is expressed.
Abstract: The main goal of microarray experiments is to quantify the expression of every object on a slide as precisely as possible, with a further goal of clustering the objects. Recently, many studies have discussed clustering issues involving similar patterns of gene expression. This paper presents an application of fuzzy-type methods for clustering DNA microarray data that can be applied to typical comparisons. Clustering and analyses were performed on microarray and simulated data. The results show that fuzzy-possibility c-means clustering substantially improves the findings obtained by others.
Abstract: This paper is introduced a modification to Diffie-
Hellman protocol to be applicable on the decimal numbers, which
they are the numbers between zero and one. For this purpose we
extend the theory of the congruence. The new congruence is over
the set of the real numbers and it is called the “real congruence"
or the “real modulus". We will refer to the existing congruence by
the “integer congruence" or the “integer modulus". This extension
will define new terms and redefine the existing terms. As the
properties and the theorems of the integer modulus are extended as
well. Modified Diffie-Hellman key exchange protocol is produced a
sharing, secure and decimal secret key for the the cryptosystems that
depend on decimal numbers.
Abstract: Prime Factorization based on Quantum approach in
two phases has been performed. The first phase has been achieved at
Quantum computer and the second phase has been achieved at the
classic computer (Post Processing). At the second phase the goal is to
estimate the period r of equation xrN ≡ 1 and to find the prime factors
of the composite integer N in classic computer. In this paper we
present a method based on Randomized Approach for estimation the
period r with a satisfactory probability and the composite integer N
will be factorized therefore with the Randomized Approach even the
gesture of the period is not exactly the real period at least we can find
one of the prime factors of composite N. Finally we present some
important points for designing an Emulator for Quantum Computer
Simulation.
Abstract: Bagging and boosting are among the most popular resampling ensemble methods that generate and combine a diversity of classifiers using the same learning algorithm for the base-classifiers. Boosting algorithms are considered stronger than bagging on noisefree data. However, there are strong empirical indications that bagging is much more robust than boosting in noisy settings. For this reason, in this work we built an ensemble using a voting methodology of bagging and boosting ensembles with 10 subclassifiers in each one. We performed a comparison with simple bagging and boosting ensembles with 25 sub-classifiers, as well as other well known combining methods, on standard benchmark datasets and the proposed technique was the most accurate.
Abstract: This paper deals with the helical flow of a Newtonian
fluid in an infinite circular cylinder, due to both longitudinal and
rotational shear stress. The velocity field and the resulting shear
stress are determined by means of the Laplace and finite Hankel
transforms and satisfy all imposed initial and boundary conditions.
For large times, these solutions reduce to the well-known steady-state
solutions.
Abstract: In this communication an expression for mean
velocity of waste flow via an open channel is proposed which
is an improvement over Manning formula. The discharges,
storages and depths are computed at all locations of the Lyari river
by utilizing proposed expression. The results attained through
proposed expression are in good agreement with the observed data
and better than those acquired using Manning formula.