Abstract: In this paper, a novel method for a biometric system based on the ECG signal is proposed, using spectral coefficients computed through linear predictive coding (LPC). ECG biometric systems have traditionally incorporated characteristics of fiducial points of the ECG signal as the feature set. These systems have been shown to contain loopholes and thus a non-fiducial system allows for tighter security. In the proposed system, incorporating non-fiducial features from the LPC spectrum produced a segment and subject recognition rate of 99.52% and 100% respectively. The recognition rates outperformed the biometric system that is based on the wavelet packet decomposition (WPD) algorithm in terms of recognition rates and computation time. This allows for LPC to be used in a practical ECG biometric system that requires fast, stringent and accurate recognition.
Abstract: In this paper, a two-dimensional mathematical model is developed for estimating the extent of inland inundation due to Indonesian tsunami of 2004 along the coastal belts of Peninsular Malaysia and Thailand. The model consists of the shallow water equations together with open and coastal boundary conditions. In order to route the water wave towards the land, the coastal boundary is treated as a time dependent moving boundary. For computation of tsunami inundation, the initial tsunami wave is generated in the deep ocean with the strength of the Indonesian tsunami of 2004. Several numerical experiments are carried out by changing the slope of the beach to examine the extent of inundation with slope. The simulated inundation is found to decrease with the increase of the slope of the orography. Correlation between inundation / recession and run-up are found to be directly proportional to each other.
Abstract: The small interfering RNA (siRNA) alters the
regulatory role of mRNA during gene expression by translational
inhibition. Recent studies show that upregulation of mRNA because
serious diseases like cancer. So designing effective siRNA with good
knockdown effects plays an important role in gene silencing. Various
siRNA design tools had been developed earlier. In this work, we are
trying to analyze the existing good scoring second generation siRNA
predicting tools and to optimize the efficiency of siRNA prediction
by designing a computational model using Artificial Neural Network
and whole stacking energy (%G), which may help in gene silencing
and drug design in cancer therapy. Our model is trained and tested
against a large data set of siRNA sequences. Validation of our results
is done by finding correlation coefficient of experimental versus
observed inhibition efficacy of siRNA. We achieved a correlation
coefficient of 0.727 in our previous computational model and we
could improve the correlation coefficient up to 0.753 when the
threshold of whole tacking energy is greater than or equal to -32.5
kcal/mol.
Abstract: The incorporation of computational fluid dynamics in the design of modern hydraulic turbines appears to be necessary in order to improve their efficiency and cost-effectiveness beyond the traditional design practices. A numerical optimization methodology is developed and applied in the present work to a Turgo water turbine. The fluid is simulated by a Lagrangian mesh-free approach that can provide detailed information on the energy transfer and enhance the understanding of the complex, unsteady flow field, at very small computing cost. The runner blades are initially shaped according to hydrodynamics theory, and parameterized using Bezier polynomials and interpolation techniques. The use of a limited number of free design variables allows for various modifications of the standard blade shape, while stochastic optimization using evolutionary algorithms is implemented to find the best blade that maximizes the attainable hydraulic efficiency of the runner. The obtained optimal runner design achieves considerably higher efficiency than the standard one, and its numerically predicted performance is comparable to a real Turgo turbine, verifying the reliability and the prospects of the new methodology.
Abstract: Cloud Computing is an approach that provides computation and storage services on-demand to clients over the network, independent of device and location. In the last few years, cloud computing became a trend in information technology with many companies that transfer their business processes and applications in the cloud. Cloud computing with service oriented architecture has contributed to rapid development of Geographic Information Systems. Open Geospatial Consortium with its standards provides the interfaces for hosted spatial data and GIS functionality to integrated GIS applications. Furthermore, with the enormous processing power, clouds provide efficient environment for data intensive applications that can be performed efficiently, with higher precision, and greater reliability. This paper presents our work on the geospatial data services within the cloud computing environment and its technology. A cloud computing environment with the strengths and weaknesses of the geographic information system will be introduced. The OGC standards that solve our application interoperability are highlighted. Finally, we outline our system architecture with utilities for requesting and invoking our developed data intensive applications as a web service.
Abstract: In this paper, we propose a dual version of the first
threshold ring signature scheme based on error-correcting code proposed
by Aguilar et. al in [1]. Our scheme uses an improvement of
Véron zero-knowledge identification scheme, which provide smaller
public and private key sizes and better computation complexity than
the Stern one. This scheme is secure in the random oracle model.
Abstract: In this paper a novel approach for generalized image
retrieval based on semantic contents is presented. A combination of
three feature extraction methods namely color, texture, and edge
histogram descriptor. There is a provision to add new features in
future for better retrieval efficiency. Any combination of these
methods, which is more appropriate for the application, can be used
for retrieval. This is provided through User Interface (UI) in the
form of relevance feedback. The image properties analyzed in this
work are by using computer vision and image processing algorithms.
For color the histogram of images are computed, for texture cooccurrence
matrix based entropy, energy, etc, are calculated and for
edge density it is Edge Histogram Descriptor (EHD) that is found.
For retrieval of images, a novel idea is developed based on greedy
strategy to reduce the computational complexity. The entire system
was developed using AForge.Imaging (an open source product),
MATLAB .NET Builder, C#, and Oracle 10g. The system was tested
with Coral Image database containing 1000 natural images and
achieved better results.
Abstract: Numerical analysis naturally finds applications in all
fields of engineering and the physical sciences, but in the
21st century, the life sciences and even the arts have adopted
elements of scientific computations. The numerical data analysis
became key process in research and development of all the fields [6].
In this paper we have made an attempt to analyze the specified
numerical patterns with reference to the association rule mining
techniques with minimum confidence and minimum support mining
criteria. The extracted rules and analyzed results are graphically
demonstrated. Association rules are a simple but very useful form of
data mining that describe the probabilistic co-occurrence of certain
events within a database [7]. They were originally designed to
analyze market-basket data, in which the likelihood of items being
purchased together within the same transactions are analyzed.
Abstract: The vertex connectivity of a graph is the smallest number of vertices whose deletion separates the graph or makes it trivial. This work is devoted to the problem of vertex connectivity test of graphs in a distributed environment based on a general and a constructive approach. The contribution of this paper is threefold. First, using a preconstructed spanning tree of the considered graph, we present a protocol to test whether a given graph is 2-connected using only local knowledge. Second, we present an encoding of this protocol using graph relabeling systems. The last contribution is the implementation of this protocol in the message passing model. For a given graph G, where M is the number of its edges, N the number of its nodes and Δ is its degree, our algorithms need the following requirements: The first one uses O(Δ×N2) steps and O(Δ×logΔ) bits per node. The second one uses O(Δ×N2) messages, O(N2) time and O(Δ × logΔ) bits per node. Furthermore, the studied network is semi-anonymous: Only the root of the pre-constructed spanning tree needs to be identified.
Abstract: Prospective readers can quickly determine whether a document is relevant to their information need if the significant phrases (or keyphrases) in this document are provided. Although keyphrases are useful, not many documents have keyphrases assigned to them, and manually assigning keyphrases to existing documents is costly. Therefore, there is a need for automatic keyphrase extraction. This paper introduces a new domain independent keyphrase extraction algorithm. The algorithm approaches the problem of keyphrase extraction as a classification task, and uses a combination of statistical and computational linguistics techniques, a new set of attributes, and a new machine learning method to distinguish keyphrases from non-keyphrases. The experiments indicate that this algorithm performs better than other keyphrase extraction tools and that it significantly outperforms Microsoft Word 2000-s AutoSummarize feature. The domain independence of this algorithm has also been confirmed in our experiments.
Abstract: The flow field around a flat plate of infinite span has
been investigated for several values of the angle of attack. Numerical
predictions have been compared to experimental measurements, in
order to examine the effect of turbulence model and grid resolution
on the resultant aerodynamic forces acting on the plate. Also the
influence of the free-stream turbulence intensity, at the entrance of
the computational domain, has been investigated. A full campaign of
simulations has been conducted for three inclination angles (9°, 15°
and 30°), in order to obtain some practical guidelines to be used for
the simulation of the flow field around inclined plates and discs.
Abstract: One of the approaches enabling people with amputated
limbs to establish some sort of interface with the real world includes
the utilization of the myoelectric signal (MES) from the remaining
muscles of those limbs. The MES can be used as a control input to a
multifunction prosthetic device. In this control scheme, known as the
myoelectric control, a pattern recognition approach is usually utilized
to discriminate between the MES signals that belong to different
classes of the forearm movements. Since the MES is recorded using
multiple channels, the feature vector size can become very large. In
order to reduce the computational cost and enhance the generalization
capability of the classifier, a dimensionality reduction method is
needed to identify an informative yet moderate size feature set. This
paper proposes a new fuzzy version of the well known Fisher-s
Linear Discriminant Analysis (LDA) feature projection technique.
Furthermore, based on the fact that certain muscles might contribute
more to the discrimination process, a novel feature weighting scheme
is also presented by employing Particle Swarm Optimization (PSO)
for estimating the weight of each feature. The new method, called
PSOFLDA, is tested on real MES datasets and compared with other
techniques to prove its superiority.
Abstract: Traditional parallel single string matching algorithms
are always based on PRAM computation model. Those algorithms
concentrate on the cost optimal design and the theoretical speed.
Based on the distributed string matching algorithm proposed by
CHEN, a practical distributed string matching algorithm architecture
is proposed in this paper. And also an improved single string matching
algorithm based on a variant Boyer-Moore algorithm is presented. We
implement our algorithm on the above architecture and the
experiments prove that it is really practical and efficient on distributed
memory machine. Its computation complexity is O(n/p + m), where n
is the length of the text, and m is the length of the pattern, and p is the
number of the processors.
Abstract: Preliminary results for a new flat plate test
facility are presented here in the form of Computational Fluid Dynamics (CFD), flow visualisation, pressure measurements and thermal anemometry. The results from the CFD and flow
visualisation show the effectiveness of the plate design, with the trailing edge flap anchoring the stagnation point on the working surface and reducing the extent of the leading edge separation. The flow visualization technique demonstrates the
two-dimensionality of the flow in the location where the
thermal anemometry measurements are obtained.
Measurements of the boundary layer mean velocity profiles compare favourably with the Blasius solution, thereby allowing for comparison of future measurements with the
wealth of data available on zero pressure gradient Blasius
flows. Results for the skin friction, boundary layer thickness,
frictional velocity and wall shear stress are shown to agree well with the Blasius theory, with a maximum experimental deviation from theory of 5%. Two turbulence generating grids
have been designed and characterized and it is shown that the turbulence decay downstream of both grids agrees with established correlations. It is also demonstrated that there is
little dependence of turbulence on the freestream velocity.
Abstract: This paper focuses on cost and profit analysis of
single-server Markovian queuing system with two priority classes. In
this paper, functions of total expected cost, revenue and profit of the
system are constructed and subjected to optimization with respect to
its service rates of lower and higher priority classes. A computing
algorithm has been developed on the basis of fast converging
numerical method to solve the system of non linear equations formed
out of the mathematical analysis. A novel performance measure of
cost and profit analysis in view of its economic interpretation for the
system with priority classes is attempted to discuss in this paper. On
the basis of computed tables observations are also drawn to enlighten
the variational-effect of the model on the parameters involved
therein.
Abstract: The frequency contents of the non-stationary
signals vary with time. For proper characterization of such
signals, a smart time-frequency representation is necessary.
Classically, the STFT (short-time Fourier transform) is
employed for this purpose. Its limitation is the fixed timefrequency
resolution. To overcome this drawback an enhanced
STFT version is devised. It is based on the signal driven
sampling scheme, which is named as the cross-level sampling.
It can adapt the sampling frequency and the window function
(length plus shape) by following the input signal local
variations. This adaptation results into the proposed technique
appealing features, which are the adaptive time-frequency
resolution and the computational efficiency.
Abstract: The purpose of this paper is to propose a framework for constructing correct parallel processing programs based on Equivalent Transformation Framework (ETF). ETF regards computation as In the framework, a problem-s domain knowledge and a query are described in definite clauses, and computation is regarded as transformation of the definite clauses. Its meaning is defined by a model of the set of definite clauses, and the transformation rules generated must preserve meaning. We have proposed a parallel processing method based on “specialization", a part of operation in the transformations, which resembles substitution in logic programming. The method requires “Memo-tree", a history of specialization to maintain correctness. In this paper we proposes the new method for the specialization-base parallel processing without Memo-tree.
Abstract: We consider a typical problem in the assembly of
printed circuit boards (PCBs) in a two-machine flow shop system to
simultaneously minimize the weighted sum of weighted tardiness and
weighted flow time. The investigated problem is a group scheduling
problem in which PCBs are assembled in groups and the interest is to
find the best sequence of groups as well as the boards within each
group to minimize the objective function value. The type of setup
operation between any two board groups is characterized as carryover
sequence-dependent setup time, which exactly matches with the real
application of this problem. As a technical constraint, all of the
boards must be kitted before the assembly operation starts (kitting
operation) and by kitting staff. The main idea developed in this paper
is to completely eliminate the role of kitting staff by assigning the
task of kitting to the machine operator during the time he is idle
which is referred to as integration of internal (machine) and external
(kitting) setup times. Performing the kitting operation, which is a
preparation process of the next set of boards while the other boards
are currently being assembled, results in the boards to continuously
enter the system or have dynamic arrival times. Consequently, a
dynamic PCB assembly system is introduced for the first time in the
assembly of PCBs, which also has characteristics similar to that of
just-in-time manufacturing. The problem investigated is
computationally very complex, meaning that finding the optimal
solutions especially when the problem size gets larger is impossible.
Thus, a heuristic based on Genetic Algorithm (GA) is employed. An
example problem on the application of the GA developed is
demonstrated and also numerical results of applying the GA on
solving several instances are provided.
Abstract: A generalization of the concepts of Feistel Networks (FN), known as Extended Feistel Network (EFN) is examined. EFN splits the input blocks into n > 2 sub-blocks. Like conventional FN, EFN consists of a series of rounds whereby at least one sub-block is subjected to an F function. The function plays a key role in the diffusion process due to its completeness property. It is also important to note that in EFN the F-function is the most computationally expensive operation in a round. The aim of this paper is to determine a suitable type of EFN for a scalable cipher. This is done by analyzing the threshold number of rounds for different types of EFN to achieve the completeness property as well as the number of F-function required in the network. The work focuses on EFN-Type I, Type II and Type III only. In the analysis it is found that EFN-Type II and Type III diffuses at the same rate and both are faster than Type-I EFN. Since EFN-Type-II uses less F functions as compared to EFN-Type III, therefore Type II is the most suitable EFN for use in a scalable cipher.
Abstract: The use of the mechanical simulation (in particular the finite element analysis) requires the management of assumptions in order to analyse a real complex system. In finite element analysis (FEA), two modeling steps require assumptions to be able to carry out the computations and to obtain some results: the building of the physical model and the building of the simulation model. The simplification assumptions made on the analysed system in these two steps can generate two kinds of errors: the physical modeling errors (mathematical model, domain simplifications, materials properties, boundary conditions and loads) and the mesh discretization errors. This paper proposes a mesh adaptive method based on the use of an h-adaptive scheme in combination with an error estimator in order to choose the mesh of the simulation model. This method allows us to choose the mesh of the simulation model in order to control the cost and the quality of the finite element analysis.