Radar Task Schedulers based on Multiple Queue

There are very complex communication systems, as the multifunction radar, MFAR (Multi-Function Array Radar), where functions are integrated all together, and simultaneously are performed the classic functions of tracking and surveillance, as all the functions related to the communication, countermeasures, and calibration. All these functions are divided into the tasks to execute. The task scheduler is a key element of the radar, since it does the planning and distribution of energy and time resources to be shared and used by all tasks. This paper presents schedulers based on the use of multiple queue. Several schedulers have been designed and studied, and it has been made a comparative analysis of different performed schedulers. The tests and experiments have been done by means of system software simulation. Finally a suitable set of radar characteristics has been selected to evaluate the behavior of the task scheduler working.

Recognition by Online Modeling – a New Approach of Recognizing Voice Signals in Linear Time

This work presents a novel means of extracting fixedlength parameters from voice signals, such that words can be recognized in linear time. The power and the zero crossing rate are first calculated segment by segment from a voice signal; by doing so, two feature sequences are generated. We then construct an FIR system across these two sequences. The parameters of this FIR system, used as the input of a multilayer proceptron recognizer, can be derived by recursive LSE (least-square estimation), implying that the complexity of overall process is linear to the signal size. In the second part of this work, we introduce a weighting factor λ to emphasize recent input; therefore, we can further recognize continuous speech signals. Experiments employ the voice signals of numbers, from zero to nine, spoken in Mandarin Chinese. The proposed method is verified to recognize voice signals efficiently and accurately.

Patents as Indicators of Innovative Environment

The main problem is that there is a very low innovation performance in Latvia. Since Latvia is a Member State of European Union, it also shall have to fulfill the set targets and to improve innovative results.Universities are one of the main performers to provide innovative capacity of country. University, industry and government need to cooperate for getting best results.The intellectual property is one of the indicators to determine innovation level in the country or organization, and patents are one of the characteristics of intellectual property.The objective of the article is to determine indicators characterizing innovative environment in Latvia and influence of the development of universities on them.The methods that will be used in the article to achieve the objectives are quantitative and qualitative analysis of the literature, statistical data analysis and graphical analysis methods.

Binary Mixture of Copper-Cobalt Ions Uptake by Zeolite using Neural Network

In this study a neural network (NN) was proposed to predict the sorption of binary mixture of copper-cobalt ions into clinoptilolite as ion-exchanger. The configuration of the backpropagation neural network giving the smallest mean square error was three-layer NN with tangent sigmoid transfer function at hidden layer with 10 neurons, linear transfer function at output layer and Levenberg-Marquardt backpropagation training algorithm. Experiments have been carried out in the batch reactor to obtain equilibrium data of the individual sorption and the mixture of coppercobalt ions. The obtained modeling results have shown that the used of neural network has better adjusted the equilibrium data of the binary system when compared with the conventional sorption isotherm models.

Optimization of Communication Protocols by stochastic Delay Mechanisms

The paper is concerned with developing stochastic delay mechanisms for efficient multicast protocols and for smooth mobile handover processes which are capable of preserving a given Quality of Service (QoS). In both applications the participating entities (receiver nodes or subscribers) sample a stochastic timer and generate load after a random delay. In this way, the load on the networking resources is evenly distributed which helps to maintain QoS communication. The optimal timer distributions have been sought in different p.d.f. families (e.g. exponential, power law and radial basis function) and the optimal parameter have been found in a recursive manner. Detailed simulations have demonstrated the improvement in performance both in the case of multicast and mobile handover applications.

Role of Director's Philosophical Approach in Cinematographic Expression

The original idea for a feature film may come from a writer, director or a producer. Director is the person responsible for the creative aspects, both interpretive and technical, of a motion picture production in a film. Director may be shot discussing his project with his or her cowriters, members of production staff, and producer, and director may be shown selecting locales or constructing sets. All these activities provide, of course, ways of externalizing director-s ideas about the film. A director sometimes pushes both the film image and techniques of narration to new artistic limits, but main responsibility of director is take the spectator to an original opinion in his philosophical approach. Director tries to find an artistic angle in every scene and change screenplay into an effective story and sets his film on a spiritual and philosophical base.

Image Clustering Framework for BAVM Segmentation in 3DRA Images: Performance Analysis

Brain ArterioVenous Malformation (BAVM) is an abnormal tangle of brain blood vessels where arteries shunt directly into veins with no intervening capillary bed which causes high pressure and hemorrhage risk. The success of treatment by embolization in interventional neuroradiology is highly dependent on the accuracy of the vessels visualization. In this paper the performance of clustering techniques on vessel segmentation from 3- D rotational angiography (3DRA) images is investigated and a new technique of segmentation is proposed. This method consists in: preprocessing step of image enhancement, then K-Means (KM), Fuzzy C-Means (FCM) and Expectation Maximization (EM) clustering are used to separate vessel pixels from background and artery pixels from vein pixels when possible. A post processing step of removing false-alarm components is applied before constructing a three-dimensional volume of the vessels. The proposed method was tested on six datasets along with a medical assessment of an expert. Obtained results showed encouraging segmentations.

Multi-threshold Approach for License Plate Recognition System

The objective of this paper is to propose an adaptive multi threshold for image segmentation precisely in object detection. Due to the different types of license plates being used, the requirement of an automatic LPR is rather different for each country. The proposed technique is applied on Malaysian LPR application. It is based on Multi Layer Perceptron trained by back propagation. The proposed adaptive threshold is introduced to find the optimum threshold values. The technique relies on the peak value from the graph of the number object versus specific range of threshold values. The proposed approach has improved the overall performance compared to current optimal threshold techniques. Further improvement on this method is in progress to accommodate real time system specification.

An Overview of Some High Order and Multi-Level Finite Difference Schemes in Computational Aeroacoustics

In this paper, we have combined some spatial derivatives with the optimised time derivative proposed by Tam and Webb in order to approximate the linear advection equation which is given by = 0. Ôêé Ôêé + Ôêé Ôêé x f t u These spatial derivatives are as follows: a standard 7-point 6 th -order central difference scheme (ST7), a standard 9-point 8 th -order central difference scheme (ST9) and optimised schemes designed by Tam and Webb, Lockard et al., Zingg et al., Zhuang and Chen, Bogey and Bailly. Thus, these seven different spatial derivatives have been coupled with the optimised time derivative to obtain seven different finite-difference schemes to approximate the linear advection equation. We have analysed the variation of the modified wavenumber and group velocity, both with respect to the exact wavenumber for each spatial derivative. The problems considered are the 1-D propagation of a Boxcar function, propagation of an initial disturbance consisting of a sine and Gaussian function and the propagation of a Gaussian profile. It is known that the choice of the cfl number affects the quality of results in terms of dissipation and dispersion characteristics. Based on the numerical experiments solved and numerical methods used to approximate the linear advection equation, it is observed in this work, that the quality of results is dependent on the choice of the cfl number, even for optimised numerical methods. The errors from the numerical results have been quantified into dispersion and dissipation using a technique devised by Takacs. Also, the quantity, Exponential Error for Low Dispersion and Low Dissipation, eeldld has been computed from the numerical results. Moreover, based on this work, it has been found that when the quantity, eeldld can be used as a measure of the total error. In particular, the total error is a minimum when the eeldld is a minimum.

Optimal Control of Viscoelastic Melt Spinning Processes

The optimal control problem for the viscoelastic melt spinning process has not been reported yet in the literature. In this study, an optimal control problem for a mathematical model of a viscoelastic melt spinning process is considered. Maxwell-Oldroyd model is used to describe the rheology of the polymeric material, the fiber is made of. The extrusion velocity of the polymer at the spinneret as well as the velocity and the temperature of the quench air and the fiber length serve as control variables. A constrained optimization problem is derived and the first–order optimality system is set up to obtain the adjoint equations. Numerical solutions are carried out using a steepest descent algorithm. A computer program in MATLAB is developed for simulations.

Combination of Different Classifiers for Cardiac Arrhythmia Recognition

This paper describes a new supervised fusion (hybrid) electrocardiogram (ECG) classification solution consisting of a new QRS complex geometrical feature extraction as well as a new version of the learning vector quantization (LVQ) classification algorithm aimed for overcoming the stability-plasticity dilemma. Toward this objective, after detection and delineation of the major events of ECG signal via an appropriate algorithm, each QRS region and also its corresponding discrete wavelet transform (DWT) are supposed as virtual images and each of them is divided into eight polar sectors. Then, the curve length of each excerpted segment is calculated and is used as the element of the feature space. To increase the robustness of the proposed classification algorithm versus noise, artifacts and arrhythmic outliers, a fusion structure consisting of five different classifiers namely as Support Vector Machine (SVM), Modified Learning Vector Quantization (MLVQ) and three Multi Layer Perceptron-Back Propagation (MLP–BP) neural networks with different topologies were designed and implemented. The new proposed algorithm was applied to all 48 MIT–BIH Arrhythmia Database records (within–record analysis) and the discrimination power of the classifier in isolation of different beat types of each record was assessed and as the result, the average accuracy value Acc=98.51% was obtained. Also, the proposed method was applied to 6 number of arrhythmias (Normal, LBBB, RBBB, PVC, APB, PB) belonging to 20 different records of the aforementioned database (between– record analysis) and the average value of Acc=95.6% was achieved. To evaluate performance quality of the new proposed hybrid learning machine, the obtained results were compared with similar peer– reviewed studies in this area.

A Reconfigurable Processing Element for Cholesky Decomposition and Matrix Inversion

Fixed-point simulation results are used for the performance measure of inverting matrices by Cholesky decomposition. The fixed-point Cholesky decomposition algorithm is implemented using a fixed-point reconfigurable processing element. The reconfigurable processing element provides all mathematical operations required by Cholesky decomposition. The fixed-point word length analysis is based on simulations using different condition numbers and different matrix sizes. Simulation results show that 16 bits word length gives sufficient performance for small matrices with low condition number. Larger matrices and higher condition numbers require more dynamic range for a fixedpoint implementation.

Influence of Heat Transfer on Stability of Newtonian and Non-Newtonian Extending Films

The stability of Newtonian and Non-Newtonian extending films under local or global heating or cooling conditions are considered. The thickness-averaged mass, momentum and energy equations with convective and radiative heat transfer are derived, both for Newtonian and non-Newtonian fluids (Maxwell, PTT and Giesekus models considered). The stability of the system is explored using either eigenvalue analysis or transient simulations. The results showed that the influence of heating and cooling on stability strongly depends on the magnitude of the Peclet number. Examples of stabilization or destabilization of heating or cooling are shown for Pe

The Link between Unemployment and Inflation Using Johansen’s Co-Integration Approach and Vector Error Correction Modelling

In this paper bi-annual time series data on unemployment rates (from the Labour Force Survey) are expanded to quarterly rates and linked to quarterly unemployment rates (from the Quarterly Labour Force Survey). The resultant linked series and the consumer price index (CPI) series are examined using Johansen’s cointegration approach and vector error correction modeling. The study finds that both the series are integrated of order one and are cointegrated. A statistically significant co-integrating relationship is found to exist between the time series of unemployment rates and the CPI. Given this significant relationship, the study models this relationship using Vector Error Correction Models (VECM), one with a restriction on the deterministic term and the other with no restriction. A formal statistical confirmation of the existence of a unique linear and lagged relationship between inflation and unemployment for the period between September 2000 and June 2011 is presented. For the given period, the CPI was found to be an unbiased predictor of the unemployment rate. This relationship can be explored further for the development of appropriate forecasting models incorporating other study variables.

Drug Combinations with Steroid Dispensing in Drugstores: A Study in the Center Area of Bangkok, Thailand

The purposes of this research were 1) to survey the number of drugstores that unlawful dispense of asthma prescription drugs, in form of drug combinations in the Phaya Thai district of Bangkok, 2) to find the steroids contained in that drug combinations, 3) to find a means for informing general public about the dangers of drugs and for a campaign to stop dispensing them. Researcher collected drug combinations from 69 drugstores in Phaya Thai district from Feb 15, 2012 to Mar 15, 2012. The survey found 30.43%, 21, drug stores, sold asthma drug combinations to customers without a prescription. These collected samples were tested for steroid contamination by using Immunochromatography kits. Eleven samples, 52.38%, were found contaminated with steroids. In short, there should be control and inspection of drugstores in the distribution of steroid medications. To improve the knowledge of self health maintenance and drug usage among public, Thai Government and Department of Public Health should educate people about the side effects of using drug combinations and steroids.

The Impact of Subsequent Stock Market Liberalization on the Integration of Stock Markets in ASEAN-4 + South Korea

To strengthen the capital market, there is a need to integrate the capital markets within the region by removing legal or informal restriction, specifically, stock market liberalization. Thus the paper is to investigate the effects of the subsequent stock market liberalization on stock market integration in 4 ASEAN countries (Malaysia, Indonesia, Thailand, Singapore) and Korea from 1997 to 2007. The correlation between stock market liberalization and stock market integration are to be examined by analyzing the stock prices and returns within the region and in comparison with the world MSCI index. Event study method is to be used with windows of ±12 months and T-7 + T. The results show that the subsequent stock market liberalization generally, gives minor positive effects to stock returns, except for one or two countries. The subsequent liberalization also integrates the markets short-run and long-run.

An Intelligent System for Phish Detection, using Dynamic Analysis and Template Matching

Phishing, or stealing of sensitive information on the web, has dealt a major blow to Internet Security in recent times. Most of the existing anti-phishing solutions fail to handle the fuzziness involved in phish detection, thus leading to a large number of false positives. This fuzziness is attributed to the use of highly flexible and at the same time, highly ambiguous HTML language. We introduce a new perspective against phishing, that tries to systematically prove, whether a given page is phished or not, using the corresponding original page as the basis of the comparison. It analyzes the layout of the pages under consideration to determine the percentage distortion between them, indicative of any form of malicious alteration. The system design represents an intelligent system, employing dynamic assessment which accurately identifies brand new phishing attacks and will prove effective in reducing the number of false positives. This framework could potentially be used as a knowledge base, in educating the internet users against phishing.

Virulent-GO: Prediction of Virulent Proteins in Bacterial Pathogens Utilizing Gene Ontology Terms

Prediction of bacterial virulent protein sequences can give assistance to identification and characterization of novel virulence-associated factors and discover drug/vaccine targets against proteins indispensable to pathogenicity. Gene Ontology (GO) annotation which describes functions of genes and gene products as a controlled vocabulary of terms has been shown effectively for a variety of tasks such as gene expression study, GO annotation prediction, protein subcellular localization, etc. In this study, we propose a sequence-based method Virulent-GO by mining informative GO terms as features for predicting bacterial virulent proteins. Each protein in the datasets used by the existing method VirulentPred is annotated by using BLAST to obtain its homologies with known accession numbers for retrieving GO terms. After investigating various popular classifiers using the same five-fold cross-validation scheme, Virulent-GO using the single kind of GO term features with an accuracy of 82.5% is slightly better than VirulentPred with 81.8% using five kinds of sequence-based features. For the evaluation of independent test, Virulent-GO also yields better results (82.0%) than VirulentPred (80.7%). When evaluating single kind of feature with SVM, the GO term feature performs much well, compared with each of the five kinds of features.