Abstract: This paper presents a formant-tracking linear prediction
(FTLP) model for speech processing in noise. The main focus of this
work is the detection of formant trajectory based on Hidden Markov
Models (HMM), for improved formant estimation in noise. The
approach proposed in this paper provides a systematic framework for
modelling and utilization of a time- sequence of peaks which satisfies
continuity constraints on parameter; the within peaks are modelled
by the LP parameters. The formant tracking LP model estimation
is composed of three stages: (1) a pre-cleaning multi-band spectral
subtraction stage to reduce the effect of residue noise on formants
(2) estimation stage where an initial estimate of the LP model of
speech for each frame is obtained (3) a formant classification using
probability models of formants and Viterbi-decoders. The evaluation
results for the estimation of the formant tracking LP model tested
in Gaussian white noise background, demonstrate that the proposed
combination of the initial noise reduction stage with formant tracking
and LPC variable order analysis, results in a significant reduction in
errors and distortions. The performance was evaluated with noisy
natual vowels extracted from international french and English vocabulary
speech signals at SNR value of 10dB. In each case, the
estimated formants are compared to reference formants.
Abstract: Society has grown to rely on Internet services, and the
number of Internet users increases every day. As more and more
users become connected to the network, the window of opportunity
for malicious users to do their damage becomes very great and
lucrative. The objective of this paper is to incorporate different
techniques into classier system to detect and classify intrusion from
normal network packet. Among several techniques, Steady State
Genetic-based Machine Leaning Algorithm (SSGBML) will be used
to detect intrusions. Where Steady State Genetic Algorithm (SSGA),
Simple Genetic Algorithm (SGA), Modified Genetic Algorithm and
Zeroth Level Classifier system are investigated in this research.
SSGA is used as a discovery mechanism instead of SGA. SGA
replaces all old rules with new produced rule preventing old good
rules from participating in the next rule generation. Zeroth Level
Classifier System is used to play the role of detector by matching
incoming environment message with classifiers to determine whether
the current message is normal or intrusion and receiving feedback
from environment. Finally, in order to attain the best results,
Modified SSGA will enhance our discovery engine by using Fuzzy
Logic to optimize crossover and mutation probability. The
experiments and evaluations of the proposed method were performed
with the KDD 99 intrusion detection dataset.
Abstract: The use of the oncologic index ISTER allows for a more effective planning of the radiotherapic facilities in the hospitals. Any change in the radiotherapy treatment, due to unexpected stops, may be adapted by recalculating the doses to the new treatment duration while keeping the optimal prognosis. The results obtained in a simulation model on millions of patients allow the definition of optimal success probability algorithms.
Abstract: This paper presents a reliability-based approach to select appropriate wind turbine types for a wind farm considering site-specific wind speed patterns. An actual wind farm in the northern region of Iran with the wind speed registration of one year is studied in this paper. An analytic approach based on total probability theorem is utilized in this paper to model the probabilistic behavior of both turbines- availability and wind speed. Well-known probabilistic reliability indices such as loss of load expectation (LOLE), expected energy not supplied (EENS) and incremental peak load carrying capability (IPLCC) for wind power integration in the Roy Billinton Test System (RBTS) are examined. The most appropriate turbine type achieving the highest reliability level is chosen for the studied wind farm.
Abstract: Ren et al. presented an efficient carrier frequency offset
(CFO) estimation method for orthogonal frequency division multiplexing
(OFDM), which has an estimation range as large as the
bandwidth of the OFDM signal and achieves high accuracy without
any constraint on the structure of the training sequence. However,
its detection probability of the integer frequency offset (IFO) rapidly
varies according to the fractional frequency offset (FFO) change. In
this paper, we first analyze the Ren-s method and define two criteria
suitable for detection of IFO. Then, we propose a novel method for
the IFO estimation based on the maximum-likelihood (ML) principle
and the detection criteria defined in this paper. The simulation results
demonstrate that the proposed method outperforms the Ren-s method
in terms of the IFO detection probability irrespective of a value of
the FFO.
Abstract: In this paper, we propose a Perceptually Optimized Foveation based Embedded ZeroTree Image Coder (POEFIC) that introduces a perceptual weighting to wavelet coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to a given bit rate a fixation point which determines the region of interest ROI. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEFIC quality assessment. Our POEFIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) foveation masking to remove or reduce considerable high frequencies from peripheral regions 2) luminance and Contrast masking, 3) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.
Abstract: The bit error rate (BER) performance for ultra-wide
band (UWB) indoor communication with impact of metallic furniture
is investigated. The impulse responses of different indoor
environments for any transmitter and receiver location are computed
by shooting and bouncing ray/image and inverse Fourier transform
techniques. By using the impulse responses of these multipath
channels, the BER performance for binary pulse amplitude
modulation (BPAM) impulse radio UWB communication system are
calculated. Numerical results have shown that the multi-path effect
by the metallic cabinets is an important factor for BER performance.
Also the outage probability for the UWB multipath environment with
metallic cabinets is more serious (about 18%) than with wooden
cabinets. Finally, it is worth noting that in these cases the present
work provides not only comparative information but also quantitative
information on the performance reduction.
Abstract: Whereas cellular wireless communication systems are
subject to short-and long-term fading. The effect of wireless channel
has largely been ignored in most of the teletraffic assessment
researches. In this paper, a mathematical teletraffic model is proposed
to estimate blocking and forced termination probabilities of cellular
wireless networks as a result of teletraffic behavior as well as the
outage of the propagation channel. To evaluate the proposed
teletraffic model, gamma inter-arrival and general service time
distributions have been considered based on wireless channel fading
effect. The performance is evaluated and compared with the classical
model. The proposed model is dedicated and investigated in different
operational conditions. These conditions will consider not only the
arrival rate process, but also, the different faded channels models.
Abstract: Prime Factorization based on Quantum approach in
two phases has been performed. The first phase has been achieved at
Quantum computer and the second phase has been achieved at the
classic computer (Post Processing). At the second phase the goal is to
estimate the period r of equation xrN ≡ 1 and to find the prime factors
of the composite integer N in classic computer. In this paper we
present a method based on Randomized Approach for estimation the
period r with a satisfactory probability and the composite integer N
will be factorized therefore with the Randomized Approach even the
gesture of the period is not exactly the real period at least we can find
one of the prime factors of composite N. Finally we present some
important points for designing an Emulator for Quantum Computer
Simulation.
Abstract: Decrease in hardware costs and advances in computer
networking technologies have led to increased interest in the use of
large-scale parallel and distributed computing systems. One of the
biggest issues in such systems is the development of effective
techniques/algorithms for the distribution of the processes/load of a
parallel program on multiple hosts to achieve goal(s) such as
minimizing execution time, minimizing communication delays,
maximizing resource utilization and maximizing throughput.
Substantive research using queuing analysis and assuming job
arrivals following a Poisson pattern, have shown that in a multi-host
system the probability of one of the hosts being idle while other host
has multiple jobs queued up can be very high. Such imbalances in
system load suggest that performance can be improved by either
transferring jobs from the currently heavily loaded hosts to the lightly
loaded ones or distributing load evenly/fairly among the hosts .The
algorithms known as load balancing algorithms, helps to achieve the
above said goal(s). These algorithms come into two basic categories -
static and dynamic. Whereas static load balancing algorithms (SLB)
take decisions regarding assignment of tasks to processors based on
the average estimated values of process execution times and
communication delays at compile time, Dynamic load balancing
algorithms (DLB) are adaptive to changing situations and take
decisions at run time.
The objective of this paper work is to identify qualitative
parameters for the comparison of above said algorithms. In future this
work can be extended to develop an experimental environment to
study these Load balancing algorithms based on comparative
parameters quantitatively.
Abstract: Transportation is of great importance in the current
life of human beings. The transportation system plays many roles,
from economical development to after-catastrophe aids such as
rescue operation in the first hours and days after an earthquake. In
after earthquakes response phase, transportation system acts as a
basis for ground operations including rescue and relief operation,
food providing for victims and etc. It is obvious that partial or
complete obstruction of this system results in the stop of these
operations. Bridges are one of the most important elements of
transportation network. Failure of a bridge, in the most optimistic
case, cuts the relation between two regions and in more developed
countries, cuts the relation of numerous regions. In this paper, to
evaluate the vulnerability and estimate the damage level of Tehran
bridges, HAZUS method, developed by Federal Emergency
Management Agency (FEMA) with the aid of National Institute of
Building Science (NIBS), is used for the first time in Iran. In this
method, to evaluate the collapse probability, fragility curves are
used. Iran is located on seismic belt and thus, it is vulnerable to
earthquakes. Thus, the study of the probability of bridge collapses, as
an important part of transportation system, during earthquakes is of
great importance. The purpose of this study is to provide fragility
curves for Gisha Bridge, one of the longest steel bridges in Tehran,
as an important lifeline element. Besides, the damage probability for
this bridge during a specific earthquake, introduced as scenario
earthquakes, is calculated. The fragility curves show that for the
considered scenario, the probability of occurrence of complete
collapse for the bridge is 8.6%.
Abstract: This research work is concerned with the eigenvalue problem for the integral operators which are obtained by linearization of a nonlocal evolution equation. The purpose of section II.A is to describe the nature of the problem and the objective of the project. The problem is related to the “stable solution" of the evolution equation which is the so-called “instanton" that describe the interface between two stable phases. The analysis of the instanton and its asymptotic behavior are described in section II.C by imposing the Green function and making use of a probability kernel. As a result , a classical Theorem which is important for an instanton is proved. Section III devoted to a study of the integral operators related to interface dynamics which concern the analysis of the Cauchy problem for the evolution equation with initial data close to different phases and different regions of space.
Abstract: Recent quasi-experimental evaluation of the Canadian Active Labour Market Policies (ALMP) by Human Resources and Skills Development Canada (HRSDC) has provided an opportunity to examine alternative methods to estimating the incremental effects of Employment Benefits and Support Measures (EBSMs) on program participants. The focus of this paper is to assess the efficiency and robustness of inverse probability weighting (IPW) relative to kernel matching (KM) in the estimation of program effects. To accomplish this objective, the authors compare pairs of 1,080 estimates, along with their associated standard errors, to assess which type of estimate is generally more efficient and robust. In the interest of practicality, the authorsalso document the computationaltime it took to produce the IPW and KM estimates, respectively.
Abstract: Optical burst switching (OBS) has been proposed to
realize the next generation Internet based on the wavelength division
multiplexing (WDM) network technologies. In the OBS, the burst
contention is one of the major problems. The deflection routing has
been designed for resolving the problem. However, the deflection
routing becomes difficult to prevent from the burst contentions as the
network load becomes high. In this paper, we introduce a flow rate
control methods to reduce burst contentions. We propose new flow
rate control methods based on the leaky bucket algorithm and
deflection routing, i.e. separate leaky bucket deflection method, and
dynamic leaky bucket deflection method. In proposed methods, edge
nodes which generate data bursts carry out the flow rate control
protocols. In order to verify the effectiveness of the flow rate control in
OBS networks, we show that the proposed methods improve the
network utilization and reduce the burst loss probability through
computer simulations.
Abstract: Reactiondiffusion systems are mathematical models that describe how the concentration of one or more substances distributed in space changes under the influence of local chemical reactions in which the substances are converted into each other, and diffusion which causes the substances to spread out in space. The classical representation of a reaction-diffusion system is given by semi-linear parabolic partial differential equations, whose general form is ÔêétX(x, t) = DΔX(x, t), where X(x, t) is the state vector, D is the matrix of the diffusion coefficients and Δ is the Laplace operator. If the solute move in an homogeneous system in thermal equilibrium, the diffusion coefficients are constants that do not depend on the local concentration of solvent and of solutes and on local temperature of the medium. In this paper a new stochastic reaction-diffusion model in which the diffusion coefficients are function of the local concentration, viscosity and frictional forces of solvent and solute is presented. Such a model provides a more realistic description of the molecular kinetics in non-homogenoeus and highly structured media as the intra- and inter-cellular spaces. The movement of a molecule A from a region i to a region j of the space is described as a first order reaction Ai k- → Aj , where the rate constant k depends on the diffusion coefficient. Representing the diffusional motion as a chemical reaction allows to assimilate a reaction-diffusion system to a pure reaction system and to simulate it with Gillespie-inspired stochastic simulation algorithms. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the specific speed of reaction and diffusion events. Redi is the software tool, developed to implement the model of reaction-diffusion kinetics and dynamics. It is a free software, that can be downloaded from http://www.cosbi.eu. To demonstrate the validity of the new reaction-diffusion model, the simulation results of the chaperone-assisted protein folding in cytoplasm obtained with Redi are reported. This case study is redrawing the attention of the scientific community due to current interests on protein aggregation as a potential cause for neurodegenerative diseases.
Abstract: Stochastic resonance (SR) is a phenomenon whereby
the signal transmission or signal processing through certain nonlinear
systems can be improved by adding noise. This paper discusses SR in
nonlinear signal detection by a simple test statistic, which can be
computed from multiple noisy data in a binary decision problem based
on a maximum a posteriori probability criterion. The performance of
detection is assessed by the probability of detection error Per . When
the input signal is subthreshold signal, we establish that benefit from
noise can be gained for different noises and confirm further that the
subthreshold SR exists in nonlinear signal detection. The efficacy of
SR is significantly improved and the minimum of Per can
dramatically approach to zero as the sample number increases. These
results show the robustness of SR in signal detection and extend the
applicability of SR in signal processing.
Abstract: In large Internet backbones, Service Providers
typically have to explicitly manage the traffic flows in order to
optimize the use of network resources. This process is often referred
to as Traffic Engineering (TE). Common objectives of traffic
engineering include balance traffic distribution across the network
and avoiding congestion hot spots. Raj P H and SVK Raja designed
the Bayesian network approach to identify congestion hors pots in
MPLS. In this approach for every node in the network the
Conditional Probability Distribution (CPD) is specified. Based on
the CPD the congestion hot spots are identified. Then the traffic can
be distributed so that no link in the network is either over utilized or
under utilized. Although the Bayesian network approach has been
implemented in operational networks, it has a number of well known
scaling issues.
This paper proposes a new approach, which we call the Pragati
(means Progress) Node Popularity (PNP) approach to identify the
congestion hot spots with the network topology alone. In the new
Pragati Node Popularity approach, IP routing runs natively over the
physical topology rather than depending on the CPD of each node as
in Bayesian network. We first illustrate our approach with a simple
network, then present a formal analysis of the Pragati Node
Popularity approach. Our PNP approach shows that for any given
network of Bayesian approach, it exactly identifies the same result
with minimum efforts. We further extend the result to a more
generic one: for any network topology and even though the network
is loopy. A theoretical insight of our result is that the optimal routing
is always shortest path routing with respect to some considerations of
hot spots in the networks.
Abstract: this paper gives a novel approach towards real-time speed estimation of multiple traffic vehicles using fuzzy logic and image processing techniques with proper arrangement of camera parameters. The described algorithm consists of several important steps. First, the background is estimated by computing median over time window of specific frames. Second, the foreground is extracted using fuzzy similarity approach (FSA) between estimated background pixels and the current frame pixels containing foreground and background. Third, the traffic lanes are divided into two parts for both direction vehicles for parallel processing. Finally, the speeds of vehicles are estimated by Maximum a Posterior Probability (MAP) estimator. True ground speed is determined by utilizing infrared sensors for three different vehicles and the results are compared to the proposed algorithm with an accuracy of ± 0.74 kmph.
Abstract: Probabilistic characteristics of seismic responses of the
Partially Restrained connection rotation (PRCR) and panel zone
deformation (PZD) installed in older steel moment frames were
investigated in accordance with statistical inference in
decision-making process. The 4, 6 and 8 story older steel moment
frames with clip angle and T-stub connections were designed and
analyzed using 2%/50yrs ground motions in four cities of the
Mid-America earthquake region. The probability density function and
cumulative distribution function of PRCR and PZD were determined
by the goodness-of-fit tests based on probabilistic parameters
measured from the results of the nonlinear time-history analyses. The
obtained probabilistic parameters and distributions can be used to find
out what performance level mainly PR connections and panel zones
satisfy and how many PR connections and panel zones experience a
serious damage under the Mid-America ground motions.
Abstract: In high powered dense wavelength division
multiplexed (WDM) systems with low chromatic dispersion,
four-wave mixing (FWM) can prove to be a major source of noise.
The MultiCanonical Monte Carlo Method (MCMC) and the Split
Step Fourier Method (SSFM) are combined to accurately evaluate the
probability density function of the decision variable of a receiver,
limited by FWM. The combination of the two methods leads to more
accurate results, and offers the possibility of adding other optical
noises such as the Amplified Spontaneous Emission (ASE) noise.