Multiple Moving Talker Tracking by Integration of Two Successive Algorithms

In this paper, an estimation accuracy of multiple moving talker tracking using a microphone array is improved. The tracking can be achieved by the adaptive method in which two algorithms are integrated, namely, the PAST (Projection Approximation Subspace Tracking) algorithm and the IPLS (Interior Point Least Square) algorithm. When either talker begins to speak again after a silent period, an appropriate feasible region for an evaluation function of the IPLS algorithm might not be set. Then, the tracking fails due to the incorrect updating. Therefore, if an increment of the number of active talkers is detected, the feasible region must be reset. Then, a low cost realization is required for the high speed tracking and a high accuracy realization is desired for the precise tracking. In this paper, the directions roughly estimated using the delayed-sum-array method are used for the resetting. Several results of experiments performed in an actual room environment show the effectiveness of the proposed method.

Reconstitute Information about Discontinued Water Quality Variables in the Nile Delta Monitoring Network Using Two Record Extension Techniques

The world economic crises and budget constraints have caused authorities, especially those in developing countries, to rationalize water quality monitoring activities. Rationalization consists of reducing the number of monitoring sites, the number of samples, and/or the number of water quality variables measured. The reduction in water quality variables is usually based on correlation. If two variables exhibit high correlation, it is an indication that some of the information produced may be redundant. Consequently, one variable can be discontinued, and the other continues to be measured. Later, the ordinary least squares (OLS) regression technique is employed to reconstitute information about discontinued variable by using the continuously measured one as an explanatory variable. In this paper, two record extension techniques are employed to reconstitute information about discontinued water quality variables, the OLS and the Line of Organic Correlation (LOC). An empirical experiment is conducted using water quality records from the Nile Delta water quality monitoring network in Egypt. The record extension techniques are compared for their ability to predict different statistical parameters of the discontinued variables. Results show that the OLS is better at estimating individual water quality records. However, results indicate an underestimation of the variance in the extended records. The LOC technique is superior in preserving characteristics of the entire distribution and avoids underestimation of the variance. It is concluded from this study that the OLS can be used for the substitution of missing values, while LOC is preferable for inferring statements about the probability distribution.

Improved Power Spectrum Estimation for RR-Interval Time Series

The RR interval series is non-stationary and unevenly spaced in time. For estimating its power spectral density (PSD) using traditional techniques like FFT, require resampling at uniform intervals. The researchers have used different interpolation techniques as resampling methods. All these resampling methods introduce the low pass filtering effect in the power spectrum. The lomb transform is a means of obtaining PSD estimates directly from irregularly sampled RR interval series, thus avoiding resampling. In this work, the superiority of Lomb transform method has been established over FFT based approach, after applying linear and cubicspline interpolation as resampling methods, in terms of reproduction of exact frequency locations as well as the relative magnitudes of each spectral component.

Improved Root-Mean-Square-Gain-Combining for SIMO Channels

The major problem that wireless communication systems undergo is multipath fading caused by scattering of the transmitted signal. However, we can treat multipath propagation as multiple channels between the transmitter and receiver to improve the signal-to-scattering-noise ratio. While using Single Input Multiple Output (SIMO) systems, the diversity receivers extract multiple signal branches or copies of the same signal received from different channels and apply gain combining schemes such as Root Mean Square Gain Combining (RMSGC). RMSGC asymptotically yields an identical performance to that of the theoretically optimal Maximum Ratio Combining (MRC) for values of mean Signal-to- Noise-Ratio (SNR) above a certain threshold value without the need for SNR estimation. This paper introduces an improvement of RMSGC using two different issues. We found that post-detection and de-noising the received signals improve the performance of RMSGC and lower the threshold SNR.

Motion Prediction and Motion Vector Cost Reduction during Fast Block Motion Estimation in MCTF

In 3D-wavelet video coding framework temporal filtering is done along the trajectory of motion using Motion Compensated Temporal Filtering (MCTF). Hence computationally efficient motion estimation technique is the need of MCTF. In this paper a predictive technique is proposed in order to reduce the computational complexity of the MCTF framework, by exploiting the high correlation among the frames in a Group Of Picture (GOP). The proposed technique applies coarse and fine searches of any fast block based motion estimation, only to the first pair of frames in a GOP. The generated motion vectors are supplied to the next consecutive frames, even to subsequent temporal levels and only fine search is carried out around those predicted motion vectors. Hence coarse search is skipped for all the motion estimation in a GOP except for the first pair of frames. The technique has been tested for different fast block based motion estimation algorithms over different standard test sequences using MC-EZBC, a state-of-the-art scalable video coder. The simulation result reveals substantial reduction (i.e. 20.75% to 38.24%) in the number of search points during motion estimation, without compromising the quality of the reconstructed video compared to non-predictive techniques. Since the motion vectors of all the pair of frames in a GOP except the first pair will have value ±1 around the motion vectors of the previous pair of frames, the number of bits required for motion vectors is also reduced by 50%.

Identification of Aircraft Gas Turbine Engine's Temperature Condition

Groundlessness of application probability-statistic methods are especially shown at an early stage of the aviation GTE technical condition diagnosing, when the volume of the information has property of the fuzzy, limitations, uncertainty and efficiency of application of new technology Soft computing at these diagnosing stages by using the fuzzy logic and neural networks methods. It is made training with high accuracy of multiple linear and nonlinear models (the regression equations) received on the statistical fuzzy data basis. At the information sufficiency it is offered to use recurrent algorithm of aviation GTE technical condition identification on measurements of input and output parameters of the multiple linear and nonlinear generalized models at presence of noise measured (the new recursive least squares method (LSM)). As application of the given technique the estimation of the new operating aviation engine D30KU-154 technical condition at height H=10600 m was made.

Integrated Method for Detection of Unknown Steganographic Content

This article concerns the presentation of an integrated method for detection of steganographic content embedded by new unknown programs. The method is based on data mining and aggregated hypothesis testing. The article contains the theoretical basics used to deploy the proposed detection system and the description of improvement proposed for the basic system idea. Further main results of experiments and implementation details are collected and described. Finally example results of the tests are presented.

Mission of Russian Orthodox Church in Kazakhstan in the XIX Century: Activity, Expectations and Results

The focus of this research is in the area of the soviet period and the mission of the Russian Orthodox Church in Kazakhstan in the XIX century. There was close connection of national customs and traditions with religious practices, outlooks and attitudes. In particular, such an approach has alleged estimation by Kazakh historians of the process of Christianization of the local population. Some of them are inclined to consider the small number of Christening Kazakhs as evidence that the Russian Orthodox Church didn’t achieve its mission. The number of historians who think that the church didn’t achieve its mission has thousand over the last centuries, however our calculations of the number of Kazakhs who became Orthodox Christian is much more than other historians think. Such Christians can be divided into 3 groups: Some remained Christian until their deaths, others had two faiths and the third hid their true religions, having returned to their former belief. Therefore, to define the exact amount of Christening Kazakhs represented a challenge. Some data does not create a clear picture of the level of Christianization, constant and accurate was not collected. The data appearing in reports of spiritual attendants and civil authorities is not always authentic. Article purpose is illumination and the analysis missionary activity of Russian Orthodox Church in Kazakhstan. 

Inverse Dynamic Active Ground Motion Acceleration Inputs Estimation of the Retaining Structure

The innovative fuzzy estimator is used to estimate the ground motion acceleration of the retaining structure in this study. The Kalman filter without the input term and the fuzzy weighting recursive least square estimator are two main portions of this method. The innovation vector can be produced by the Kalman filter, and be applied to the fuzzy weighting recursive least square estimator to estimate the acceleration input over time. The excellent performance of this estimator is demonstrated by comparing it with the use of difference weighting function, the distinct levels of the measurement noise covariance and the initial process noise covariance. The availability and the precision of the proposed method proposed in this study can be verified by comparing the actual value and the one obtained by numerical simulation.

Blind Identification of MA Models Using Cumulants

In this paper, many techniques for blind identification of moving average (MA) process are presented. These methods utilize third- and fourth-order cumulants of the noisy observations of the system output. The system is driven by an independent and identically distributed (i.i.d) non-Gaussian sequence that is not observed. Two nonlinear optimization algorithms, namely the Gradient Descent and the Gauss-Newton algorithms are exposed. An algorithm based on the joint-diagonalization of the fourth-order cumulant matrices (FOSI) is also considered, as well as an improved version of the classical C(q, 0, k) algorithm based on the choice of the Best 1-D Slice of fourth-order cumulants. To illustrate the effectiveness of our methods, various simulation examples are presented.

Motion Analysis for Duplicate Frame Removal in Wireless Capsule Endoscope Video

Wireless capsule Endoscopy (WCE) has rapidly shown its wide applications in medical domain last ten years thanks to its noninvasiveness for patients and support for thorough inspection through a patient-s entire digestive system including small intestine. However, one of the main barriers to efficient clinical inspection procedure is that it requires large amount of effort for clinicians to inspect huge data collected during the examination, i.e., over 55,000 frames in video. In this paper, we propose a method to compute meaningful motion changes of WCE by analyzing the obtained video frames based on regional optical flow estimations. The computed motion vectors are used to remove duplicate video frames caused by WCE-s imaging nature, such as repetitive forward-backward motions from peristaltic movements. The motion vectors are derived by calculating directional component vectors in four local regions. Our experiments are performed on small intestine area, which is of main interest to clinical experts when using WCEs, and our experimental results show significant frame reductions comparing with a simple frame-to-frame similarity-based image reduction method.

Feature Point Reduction for Video Stabilization

Corner detection and optical flow are common techniques for feature-based video stabilization. However, these algorithms are computationally expensive and should be performed at a reasonable rate. This paper presents an algorithm for discarding irrelevant feature points and maintaining them for future use so as to improve the computational cost. The algorithm starts by initializing a maintained set. The feature points in the maintained set are examined against its accuracy for modeling. Corner detection is required only when the feature points are insufficiently accurate for future modeling. Then, optical flows are computed from the maintained feature points toward the consecutive frame. After that, a motion model is estimated based on the simplified affine motion model and least square method, with outliers belonging to moving objects presented. Studentized residuals are used to eliminate such outliers. The model estimation and elimination processes repeat until no more outliers are identified. Finally, the entire algorithm repeats along the video sequence with the points remaining from the previous iteration used as the maintained set. As a practical application, an efficient video stabilization can be achieved by exploiting the computed motion models. Our study shows that the number of times corner detection needs to perform is greatly reduced, thus significantly improving the computational cost. Moreover, optical flow vectors are computed for only the maintained feature points, not for outliers, thus also reducing the computational cost. In addition, the feature points after reduction can sufficiently be used for background objects tracking as demonstrated in the simple video stabilizer based on our proposed algorithm.

A Double Referenced Contrast for Blind Source Separation

This paper addresses the problem of blind source separation (BSS). To recover original signals, from linear instantaneous mixtures, we propose a new contrast function based on the use of a double referenced system. Our approach assumes statistical independence sources. The reference vectors will be incrusted in the cumulant to evaluate the independence. The estimation of the separating matrix will be performed in two steps: whitening observations and joint diagonalization of a set of referenced cumulant matrices. Computer simulations are presented to demonstrate the effectiveness of the suggested approach.

Estimating Cost of R&D Activities for Feasibility Study of Public R&D Investment

Since the feasibility study of R&D programs have been initiated for efficient public R&D investments, year 2008, feasibility studies have improved in terms of precision. Although experience related to these studies of R&D programs have increased to a certain point, still methodological improvement is required. The feasibility studies of R&D programs are consisted of various viewpoints, such as technology, policy, and economics. This research is to provide improvement methods to the economic perspective; especially the cost estimation process of R&D activities. First of all, the fundamental concept of cost estimation is reviewed. After the review, a statistical and econometric analysis method is applied as empirical analysis. Conclusively, limitations and further research directions are provided.

Estimation of Reconnaissance Drought Index (RDI) for Bhavnagar District, Gujarat, India

There are two types of drought as conceptual drought and operational drought. The three parameters as the beginning, the end and the degree of severity of the drought can be identifying in operational drought by average precipitation in the whole region. One of the methods classified to measure drought is Reconnaissance Drought Index (RDI). Evapotranspiration is calculated using Penman-Monteith method by analyzing thirty nine years prolong climatic data. The evapotranspiration is then utilized in RDI to classify normalized and standardized RDI. These RDI classifications led to what kind of drought faced in Bhavnagar region on 12 month time scale basis. The comparison between actual drought conditions and RDI method used to find out drought are also illustrated. It can be concluded that the index results of drought in a particular year are same in both methods but having different index values where as severity remain same.

Automated Algorithm for Removing Continuous Flame Spectrum Based On Sampled Linear Bases

In this paper, an automated algorithm to estimate and remove the continuous baseline from measured spectra containing both continuous and discontinuous bands is proposed. The algorithm uses previous information contained in a Continuous Database Spectra (CDBS) to obtain a linear basis, with minimum number of sampled vectors, capable of representing a continuous baseline. The proposed algorithm was tested by using a CDBS of flame spectra where Principal Components Analysis and Non-negative Matrix Factorization were used to obtain linear bases. Thus, the radical emissions of natural gas, oil and bio-oil flames spectra at different combustion conditions were obtained. In order to validate the performance in the baseline estimation process, the Goodness-of-fit Coefficient and the Root Mean-squared Error quality metrics were evaluated between the estimated and the real spectra in absence of discontinuous emission. The achieved results make the proposed method a key element in the development of automatic monitoring processes strategies involving discontinuous spectral bands.

UAV Position Estimation Using Remote Radio Head With Adaptive Power Control

The adaptive power control of Code Division Multiple Access (CDMA) communications using Remote Radio Head (RRH) between multiple Unmanned Aerial Vehicles (UAVs) with a link-budget based Signal-to-Interference Ratio (SIR) estimate is applied to four inner loop power control algorithms. It is concluded that Base Station (BS) can calculate not only UAV distance using linearity between speed and Consecutive Transmit-Power-Control Ratio (CTR) of Adaptive Step-size Closed Loop Power Control (ASCLPC), Consecutive TPC Ratio Step-size Closed Loop Power Control (CS-CLPC), Fixed Step-size Power Control (FSPC), but also UAV position with Received Signal Strength Indicator (RSSI) ratio of RRHs.

Low-complexity Integer Frequency Offset Synchronization for OFDMA System

This paper presents a integer frequency offset (IFO) estimation scheme for the 3GPP long term evolution (LTE) downlink system. Firstly, the conventional joint detection method for IFO and sector cell index (CID) information is introduced. Secondly, an IFO estimation without explicit sector CID information is proposed, which can operate jointly with the proposed IFO estimation and reduce the time delay in comparison with the conventional joint method. Also, the proposed method is computationally efficient and has almost similar performance in comparison with the conventional method over the Pedestrian and Vehicular channel models.

A Self Adaptive Genetic Based Algorithm for the Identification and Elimination of Bad Data

The identification and elimination of bad measurements is one of the basic functions of a robust state estimator as bad data have the effect of corrupting the results of state estimation according to the popular weighted least squares method. However this is a difficult problem to handle especially when dealing with multiple errors from the interactive conforming type. In this paper, a self adaptive genetic based algorithm is proposed. The algorithm utilizes the results of the classical linearized normal residuals approach to tune the genetic operators thus instead of making a randomized search throughout the whole search space it is more likely to be a directed search thus the optimum solution is obtained at very early stages(maximum of 5 generations). The algorithm utilizes the accumulating databases of already computed cases to reduce the computational burden to minimum. Tests are conducted with reference to the standard IEEE test systems. Test results are very promising.

A New Heuristic Statistical Methodology for Optimizing Queuing Networks Using Discreet Event Simulation

Most of the real queuing systems include special properties and constraints, which can not be analyzed directly by using the results of solved classical queuing models. Lack of Markov chains features, unexponential patterns and service constraints, are the mentioned conditions. This paper represents an applied general algorithm for analysis and optimizing the queuing systems. The algorithm stages are described through a real case study. It is consisted of an almost completed non-Markov system with limited number of customers and capacities as well as lots of common exception of real queuing networks. Simulation is used for optimizing this system. So introduced stages over the following article include primary modeling, determining queuing system kinds, index defining, statistical analysis and goodness of fit test, validation of model and optimizing methods of system with simulation.