Abstract: In this paper, an estimation accuracy of multiple moving
talker tracking using a microphone array is improved. The tracking
can be achieved by the adaptive method in which two algorithms are integrated, namely, the PAST (Projection Approximation Subspace
Tracking) algorithm and the IPLS (Interior Point Least Square) algorithm. When either talker begins to speak again after a silent
period, an appropriate feasible region for an evaluation function of
the IPLS algorithm might not be set. Then, the tracking fails due to the incorrect updating. Therefore, if an increment of the number of
active talkers is detected, the feasible region must be reset. Then, a low cost realization is required for the high speed tracking and a high
accuracy realization is desired for the precise tracking. In this paper,
the directions roughly estimated using the delayed-sum-array method
are used for the resetting. Several results of experiments performed in
an actual room environment show the effectiveness of the proposed method.
Abstract: The world economic crises and budget constraints
have caused authorities, especially those in developing countries, to
rationalize water quality monitoring activities. Rationalization
consists of reducing the number of monitoring sites, the number of
samples, and/or the number of water quality variables measured. The
reduction in water quality variables is usually based on correlation. If
two variables exhibit high correlation, it is an indication that some of
the information produced may be redundant. Consequently, one
variable can be discontinued, and the other continues to be measured.
Later, the ordinary least squares (OLS) regression technique is
employed to reconstitute information about discontinued variable by
using the continuously measured one as an explanatory variable. In
this paper, two record extension techniques are employed to
reconstitute information about discontinued water quality variables,
the OLS and the Line of Organic Correlation (LOC). An empirical
experiment is conducted using water quality records from the Nile
Delta water quality monitoring network in Egypt. The record
extension techniques are compared for their ability to predict
different statistical parameters of the discontinued variables. Results
show that the OLS is better at estimating individual water quality
records. However, results indicate an underestimation of the variance
in the extended records. The LOC technique is superior in preserving
characteristics of the entire distribution and avoids underestimation
of the variance. It is concluded from this study that the OLS can be
used for the substitution of missing values, while LOC is preferable
for inferring statements about the probability distribution.
Abstract: The RR interval series is non-stationary and unevenly
spaced in time. For estimating its power spectral density (PSD) using
traditional techniques like FFT, require resampling at uniform
intervals. The researchers have used different interpolation
techniques as resampling methods. All these resampling methods
introduce the low pass filtering effect in the power spectrum. The
lomb transform is a means of obtaining PSD estimates directly from
irregularly sampled RR interval series, thus avoiding resampling. In
this work, the superiority of Lomb transform method has been
established over FFT based approach, after applying linear and
cubicspline interpolation as resampling methods, in terms of
reproduction of exact frequency locations as well as the relative
magnitudes of each spectral component.
Abstract: The major problem that wireless communication
systems undergo is multipath fading caused by scattering of the
transmitted signal. However, we can treat multipath propagation as
multiple channels between the transmitter and receiver to improve
the signal-to-scattering-noise ratio. While using Single Input
Multiple Output (SIMO) systems, the diversity receivers extract
multiple signal branches or copies of the same signal received from
different channels and apply gain combining schemes such as Root
Mean Square Gain Combining (RMSGC). RMSGC asymptotically
yields an identical performance to that of the theoretically optimal
Maximum Ratio Combining (MRC) for values of mean Signal-to-
Noise-Ratio (SNR) above a certain threshold value without the need
for SNR estimation. This paper introduces an improvement of
RMSGC using two different issues. We found that post-detection and
de-noising the received signals improve the performance of RMSGC
and lower the threshold SNR.
Abstract: In 3D-wavelet video coding framework temporal
filtering is done along the trajectory of motion using Motion
Compensated Temporal Filtering (MCTF). Hence computationally
efficient motion estimation technique is the need of MCTF. In this
paper a predictive technique is proposed in order to reduce the
computational complexity of the MCTF framework, by exploiting
the high correlation among the frames in a Group Of Picture (GOP).
The proposed technique applies coarse and fine searches of any fast
block based motion estimation, only to the first pair of frames in a
GOP. The generated motion vectors are supplied to the next
consecutive frames, even to subsequent temporal levels and only fine
search is carried out around those predicted motion vectors. Hence
coarse search is skipped for all the motion estimation in a GOP
except for the first pair of frames. The technique has been tested for
different fast block based motion estimation algorithms over different
standard test sequences using MC-EZBC, a state-of-the-art scalable
video coder. The simulation result reveals substantial reduction (i.e.
20.75% to 38.24%) in the number of search points during motion
estimation, without compromising the quality of the reconstructed
video compared to non-predictive techniques. Since the motion
vectors of all the pair of frames in a GOP except the first pair will
have value ±1 around the motion vectors of the previous pair of
frames, the number of bits required for motion vectors is also
reduced by 50%.
Abstract: Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.
Abstract: This article concerns the presentation of an integrated
method for detection of steganographic content embedded by new
unknown programs. The method is based on data mining and
aggregated hypothesis testing. The article contains the theoretical
basics used to deploy the proposed detection system and the
description of improvement proposed for the basic system idea.
Further main results of experiments and implementation details are
collected and described. Finally example results of the tests are
presented.
Abstract: The focus of this research is in the area of the soviet period and the mission of the Russian Orthodox Church in Kazakhstan in the XIX century. There was close connection of national customs and traditions with religious practices, outlooks and attitudes. In particular, such an approach has alleged estimation by Kazakh historians of the process of Christianization of the local population. Some of them are inclined to consider the small number of Christening Kazakhs as evidence that the Russian Orthodox Church didn’t achieve its mission. The number of historians who think that the church didn’t achieve its mission has thousand over the last centuries, however our calculations of the number of Kazakhs who became Orthodox Christian is much more than other historians think. Such Christians can be divided into 3 groups: Some remained Christian until their deaths, others had two faiths and the third hid their true religions, having returned to their former belief. Therefore, to define the exact amount of Christening Kazakhs represented a challenge. Some data does not create a clear picture of the level of Christianization, constant and accurate was not collected. The data appearing in reports of spiritual attendants and civil authorities is not always authentic. Article purpose is illumination and the analysis missionary activity of Russian Orthodox Church in Kazakhstan.
Abstract: The innovative fuzzy estimator is used to estimate the
ground motion acceleration of the retaining structure in this study. The
Kalman filter without the input term and the fuzzy weighting recursive
least square estimator are two main portions of this method. The
innovation vector can be produced by the Kalman filter, and be
applied to the fuzzy weighting recursive least square estimator to
estimate the acceleration input over time. The excellent performance
of this estimator is demonstrated by comparing it with the use of
difference weighting function, the distinct levels of the measurement
noise covariance and the initial process noise covariance. The
availability and the precision of the proposed method proposed in this
study can be verified by comparing the actual value and the one
obtained by numerical simulation.
Abstract: In this paper, many techniques for blind identification of moving average (MA) process are presented. These methods utilize third- and fourth-order cumulants of the noisy observations of the system output. The system is driven by an independent and identically distributed (i.i.d) non-Gaussian sequence that is not observed. Two nonlinear optimization algorithms, namely the Gradient Descent and the Gauss-Newton algorithms are exposed. An algorithm based on the joint-diagonalization of the fourth-order cumulant matrices (FOSI) is also considered, as well as an improved version of the classical C(q, 0, k) algorithm based on the choice of the Best 1-D Slice of fourth-order cumulants. To illustrate the effectiveness of our methods, various simulation examples are presented.
Abstract: Wireless capsule Endoscopy (WCE) has rapidly
shown its wide applications in medical domain last ten years
thanks to its noninvasiveness for patients and support for thorough
inspection through a patient-s entire digestive system including
small intestine. However, one of the main barriers to efficient
clinical inspection procedure is that it requires large amount of
effort for clinicians to inspect huge data collected during the
examination, i.e., over 55,000 frames in video. In this paper, we
propose a method to compute meaningful motion changes of
WCE by analyzing the obtained video frames based on regional
optical flow estimations. The computed motion vectors are used to
remove duplicate video frames caused by WCE-s imaging nature,
such as repetitive forward-backward motions from peristaltic
movements. The motion vectors are derived by calculating
directional component vectors in four local regions. Our
experiments are performed on small intestine area, which is of
main interest to clinical experts when using WCEs, and our
experimental results show significant frame reductions comparing
with a simple frame-to-frame similarity-based image reduction
method.
Abstract: Corner detection and optical flow are common techniques for feature-based video stabilization. However, these algorithms are computationally expensive and should be performed at a reasonable rate. This paper presents an algorithm for discarding irrelevant feature points and maintaining them for future use so as to improve the computational cost. The algorithm starts by initializing a maintained set. The feature points in the maintained set are examined against its accuracy for modeling. Corner detection is required only when the feature points are insufficiently accurate for future modeling. Then, optical flows are computed from the maintained feature points toward the consecutive frame. After that, a motion model is estimated based on the simplified affine motion model and least square method, with outliers belonging to moving objects presented. Studentized residuals are used to eliminate such outliers. The model estimation and elimination processes repeat until no more outliers are identified. Finally, the entire algorithm repeats along the video sequence with the points remaining from the previous iteration used as the maintained set. As a practical application, an efficient video stabilization can be achieved by exploiting the computed motion models. Our study shows that the number of times corner detection needs to perform is greatly reduced, thus significantly improving the computational cost. Moreover, optical flow vectors are computed for only the maintained feature points, not for outliers, thus also reducing the computational cost. In addition, the feature points after reduction can sufficiently be used for background objects tracking as demonstrated in the simple video stabilizer based on our proposed algorithm.
Abstract: This paper addresses the problem of blind source separation
(BSS). To recover original signals, from linear instantaneous
mixtures, we propose a new contrast function based on the use of a
double referenced system. Our approach assumes statistical independence
sources. The reference vectors will be incrusted in the cumulant
to evaluate the independence. The estimation of the separating matrix
will be performed in two steps: whitening observations and joint
diagonalization of a set of referenced cumulant matrices. Computer
simulations are presented to demonstrate the effectiveness of the
suggested approach.
Abstract: Since the feasibility study of R&D programs have been
initiated for efficient public R&D investments, year 2008, feasibility
studies have improved in terms of precision. Although experience
related to these studies of R&D programs have increased to a certain
point, still methodological improvement is required. The feasibility
studies of R&D programs are consisted of various viewpoints, such as
technology, policy, and economics. This research is to provide
improvement methods to the economic perspective; especially the cost
estimation process of R&D activities. First of all, the fundamental
concept of cost estimation is reviewed. After the review, a statistical
and econometric analysis method is applied as empirical analysis.
Conclusively, limitations and further research directions are provided.
Abstract: There are two types of drought as conceptual drought
and operational drought. The three parameters as the beginning, the
end and the degree of severity of the drought can be identifying in
operational drought by average precipitation in the whole region. One
of the methods classified to measure drought is Reconnaissance
Drought Index (RDI). Evapotranspiration is calculated using
Penman-Monteith method by analyzing thirty nine years prolong
climatic data. The evapotranspiration is then utilized in RDI to
classify normalized and standardized RDI. These RDI classifications
led to what kind of drought faced in Bhavnagar region on 12 month
time scale basis. The comparison between actual drought conditions
and RDI method used to find out drought are also illustrated. It can
be concluded that the index results of drought in a particular year are
same in both methods but having different index values where as
severity remain same.
Abstract: In this paper, an automated algorithm to estimate and remove the continuous baseline from measured spectra containing both continuous and discontinuous bands is proposed. The algorithm uses previous information contained in a Continuous Database Spectra (CDBS) to obtain a linear basis, with minimum number of sampled vectors, capable of representing a continuous baseline. The proposed algorithm was tested by using a CDBS of flame spectra where Principal Components Analysis and Non-negative Matrix Factorization were used to obtain linear bases. Thus, the radical emissions of natural gas, oil and bio-oil flames spectra at different combustion conditions were obtained. In order to validate the performance in the baseline estimation process, the Goodness-of-fit Coefficient and the Root Mean-squared Error quality metrics were evaluated between the estimated and the real spectra in absence of discontinuous emission. The achieved results make the proposed method a key element in the development of automatic monitoring processes strategies involving discontinuous spectral bands.
Abstract: The adaptive power control of Code Division Multiple
Access (CDMA) communications using Remote Radio Head
(RRH) between multiple Unmanned Aerial Vehicles (UAVs) with
a link-budget based Signal-to-Interference Ratio (SIR) estimate is
applied to four inner loop power control algorithms. It is concluded
that Base Station (BS) can calculate not only UAV distance using
linearity between speed and Consecutive Transmit-Power-Control
Ratio (CTR) of Adaptive Step-size Closed Loop Power Control (ASCLPC),
Consecutive TPC Ratio Step-size Closed Loop Power Control
(CS-CLPC), Fixed Step-size Power Control (FSPC), but also UAV
position with Received Signal Strength Indicator (RSSI) ratio of
RRHs.
Abstract: This paper presents a integer frequency offset (IFO)
estimation scheme for the 3GPP long term evolution (LTE) downlink
system. Firstly, the conventional joint detection method for IFO and
sector cell index (CID) information is introduced. Secondly, an IFO
estimation without explicit sector CID information is proposed, which
can operate jointly with the proposed IFO estimation and reduce
the time delay in comparison with the conventional joint method.
Also, the proposed method is computationally efficient and has almost
similar performance in comparison with the conventional method over
the Pedestrian and Vehicular channel models.
Abstract: The identification and elimination of bad
measurements is one of the basic functions of a robust state estimator
as bad data have the effect of corrupting the results of state
estimation according to the popular weighted least squares method.
However this is a difficult problem to handle especially when dealing
with multiple errors from the interactive conforming type. In this
paper, a self adaptive genetic based algorithm is proposed. The
algorithm utilizes the results of the classical linearized normal
residuals approach to tune the genetic operators thus instead of
making a randomized search throughout the whole search space it is
more likely to be a directed search thus the optimum solution is
obtained at very early stages(maximum of 5 generations). The
algorithm utilizes the accumulating databases of already computed
cases to reduce the computational burden to minimum. Tests are
conducted with reference to the standard IEEE test systems. Test
results are very promising.
Abstract: Most of the real queuing systems include special properties and constraints, which can not be analyzed directly by using the results of solved classical queuing models. Lack of Markov chains features, unexponential patterns and service constraints, are the mentioned conditions. This paper represents an applied general algorithm for analysis and optimizing the queuing systems. The algorithm stages are described through a real case study. It is consisted of an almost completed non-Markov system with limited number of customers and capacities as well as lots of common exception of real queuing networks. Simulation is used for optimizing this system. So introduced stages over the following article include primary modeling, determining queuing system kinds, index defining, statistical analysis and goodness of fit test, validation of model and optimizing methods of system with simulation.