Mathematical Approach towards Fault Detection and Isolation of Linear Dynamical Systems

The main objective of this work is to provide a fault detection and isolation based on Markov parameters for residual generation and a neural network for fault classification. The diagnostic approach is accomplished in two steps: In step 1, the system is identified using a series of input / output variables through an identification algorithm. In step 2, the fault is diagnosed comparing the Markov parameters of faulty and non faulty systems. The Artificial Neural Network is trained using predetermined faulty conditions serves to classify the unknown fault. In step 1, the identification is done by first formulating a Hankel matrix out of Input/ output variables and then decomposing the matrix via singular value decomposition technique. For identifying the system online sliding window approach is adopted wherein an open slit slides over a subset of 'n' input/output variables. The faults are introduced at arbitrary instances and the identification is carried out in online. Fault residues are extracted making a comparison of the first five Markov parameters of faulty and non faulty systems. The proposed diagnostic approach is illustrated on benchmark problems with encouraging results.

CFD Analysis of Natural Ventilation Behaviour in Four Sided Wind Catcher

Wind catchers are traditional natural ventilation systems attached to buildings in order to ventilate the indoor air. The most common type of wind catcher is four sided one which is capable to catch wind in all directions. CFD simulation is the perfect way to evaluate the wind catcher performance. The accuracy of CFD results is the issue of concern, so sensitivity analyses is crucial to find out the effect of different settings of CFD on results. This paper presents a series of 3D steady RANS simulations for a generic isolated four-sided wind catcher attached to a room subjected to wind direction ranging from 0º to 180º with an interval of 45º. The CFD simulations are validated with detailed wind tunnel experiments. The influence of an extensive range of computational parameters is explored in this paper, including the resolution of the computational grid, the size of the computational domain and the turbulence model. This study found that CFD simulation is a reliable method for wind catcher study, but it is less accurate in prediction of models with non perpendicular wind directions.

Comparison of Artificial Neural Network Architectures in the Task of Tourism Time Series Forecast

The authors have been developing several models based on artificial neural networks, linear regression models, Box- Jenkins methodology and ARIMA models to predict the time series of tourism. The time series consist in the “Monthly Number of Guest Nights in the Hotels" of one region. Several comparisons between the different type models have been experimented as well as the features used at the entrance of the models. The Artificial Neural Network (ANN) models have always had their performance at the top of the best models. Usually the feed-forward architecture was used due to their huge application and results. In this paper the author made a comparison between different architectures of the ANNs using simply the same input. Therefore, the traditional feed-forward architecture, the cascade forwards, a recurrent Elman architecture and a radial based architecture were discussed and compared based on the task of predicting the mentioned time series.

A Universal Model for Content-Based Image Retrieval

In this paper a novel approach for generalized image retrieval based on semantic contents is presented. A combination of three feature extraction methods namely color, texture, and edge histogram descriptor. There is a provision to add new features in future for better retrieval efficiency. Any combination of these methods, which is more appropriate for the application, can be used for retrieval. This is provided through User Interface (UI) in the form of relevance feedback. The image properties analyzed in this work are by using computer vision and image processing algorithms. For color the histogram of images are computed, for texture cooccurrence matrix based entropy, energy, etc, are calculated and for edge density it is Edge Histogram Descriptor (EHD) that is found. For retrieval of images, a novel idea is developed based on greedy strategy to reduce the computational complexity. The entire system was developed using AForge.Imaging (an open source product), MATLAB .NET Builder, C#, and Oracle 10g. The system was tested with Coral Image database containing 1000 natural images and achieved better results.

Maximum Common Substructure Extraction in RNA Secondary Structures Using Clique Detection Approach

The similarity comparison of RNA secondary structures is important in studying the functions of RNAs. In recent years, most existing tools represent the secondary structures by tree-based presentation and calculate the similarity by tree alignment distance. Different to previous approaches, we propose a new method based on maximum clique detection algorithm to extract the maximum common structural elements in compared RNA secondary structures. A new graph-based similarity measurement and maximum common subgraph detection procedures for comparing purely RNA secondary structures is introduced. Given two RNA secondary structures, the proposed algorithm consists of a process to determine the score of the structural similarity, followed by comparing vertices labelling, the labelled edges and the exact degree of each vertex. The proposed algorithm also consists of a process to extract the common structural elements between compared secondary structures based on a proposed maximum clique detection of the problem. This graph-based model also can work with NC-IUB code to perform the pattern-based searching. Therefore, it can be used to identify functional RNA motifs from database or to extract common substructures between complex RNA secondary structures. We have proved the performance of this proposed algorithm by experimental results. It provides a new idea of comparing RNA secondary structures. This tool is helpful to those who are interested in structural bioinformatics.

Study of Unsteady Swirling Flow in a Hydrodynamic Vortex Chamber

The paper reports on the results of experimental and numerical study of nonstationary swirling flow in an isothermal model of vortex burner. It has been identified that main source of the instability is related to a precessing vortex core (PVC) phenomenon. The PVC induced flow pulsation characteristics such as precession frequency and its variation as a function of flowrate and swirl number have been explored making use of acoustic probes. Additionally pressure transducers were used to measure the pressure drops on the working chamber and across the vortex flow. The experiments have been included also the mean velocity measurements making use of a laser-Doppler anemometry. The features of instantaneous flowfield generated by the PVC were analyzed employing a commercial CFD code (Star-CCM+) based on Detached Eddy Simulation (DES) approach. Validity of the numerical code has been checked by comparison calculated flowfield data with the obtained experimental results. It has been confirmed particularly that the CFD code applied correctly reproduces the flow features.

Comprehensive Evaluation on Land Supply System Performance: In Terms of System Transformation

This evaluation of land supply system performance in China shall examine the combination of government functions and national goals in order to perform a cost-benefit analysis of system results. From the author's point of view, it is most productive to evaluate land supply system performance at moments of system transformation for the following reasons. The behavior and input-output change of beneficial results at different times can be observed when the system or policy changes and system performance can be evaluated through a cost-benefit analysis during the process of system transformation. Moreover, this evaluation method can avoid the influence of land resource endowment. Different land resource endowment methods and different economy development periods result in different systems. This essay studies the contents, principles and methods of land supply system performance evaluation. Taking Beijing as an example, this essay optimizes and classifies the land supply index, makes a quantitative evaluation of land supply system performance through principal component analysis (PCA), and finally analyzes the factors that influence land supply system performance at times of system transformation.

Approximate Frequent Pattern Discovery Over Data Stream

Frequent pattern discovery over data stream is a hard problem because a continuously generated nature of stream does not allow a revisit on each data element. Furthermore, pattern discovery process must be fast to produce timely results. Based on these requirements, we propose an approximate approach to tackle the problem of discovering frequent patterns over continuous stream. Our approximation algorithm is intended to be applied to process a stream prior to the pattern discovery process. The results of approximate frequent pattern discovery have been reported in the paper.

A Combination of Similarity Ranking and Time for Social Research Paper Searching

Nowadays social media are important tools for web resource discovery. The performance and capabilities of web searches are vital, especially search results from social research paper bookmarking. This paper proposes a new algorithm for ranking method that is a combination of similarity ranking with paper posted time or CSTRank. The paper posted time is static ranking for improving search results. For this particular study, the paper posted time is combined with similarity ranking to produce a better ranking than other methods such as similarity ranking or SimRank. The retrieval performance of combination rankings is evaluated using mean values of NDCG. The evaluation in the experiments implies that the chosen CSTRank ranking by using weight score at ratio 90:10 can improve the efficiency of research paper searching on social bookmarking websites.

Defect Cause Modeling with Decision Tree and Regression Analysis

The main aim of this study is to identify the most influential variables that cause defects on the items produced by a casting company located in Turkey. To this end, one of the items produced by the company with high defective percentage rates is selected. Two approaches-the regression analysis and decision treesare used to model the relationship between process parameters and defect types. Although logistic regression models failed, decision tree model gives meaningful results. Based on these results, it can be claimed that the decision tree approach is a promising technique for determining the most important process variables.

Developing and Implementing Successful Key Performance Indicators

Measurement and the following evaluation of performance represent important part of management. The paper focuses on indicators as the basic elements of performance measurement system. It emphasizes a necessity of searching requirements for quality indicators so that they can become part of the useful system. It introduces standpoints for a systematic dividing of indicators so that they have as high as possible informative value of background sources for searching, analysis, designing and using of indicators. It draws attention to requirements for indicators' quality and at the same it deals with some dangers decreasing indicator's informative value. It submits a draft of questions that should be answered at the construction of indicator. It is obvious that particular indicators need to be defined exactly to stimulate the desired behavior in order to attain expected results. In the enclosure a concrete example of the defined indicator in the concrete conditions of a small firm is given. The authors of the paper pay attention to the fact that a quality indicator makes it possible to get to the basic causes of the problem and include the established facts into the company information system. At the same time they emphasize that developing of a quality indicator is a prerequisite for the utilization of the system of measurement in management.

Comparison of Imputation Techniques for Efficient Prediction of Software Fault Proneness in Classes

Missing data is a persistent problem in almost all areas of empirical research. The missing data must be treated very carefully, as data plays a fundamental role in every analysis. Improper treatment can distort the analysis or generate biased results. In this paper, we compare and contrast various imputation techniques on missing data sets and make an empirical evaluation of these methods so as to construct quality software models. Our empirical study is based on NASA-s two public dataset. KC4 and KC1. The actual data sets of 125 cases and 2107 cases respectively, without any missing values were considered. The data set is used to create Missing at Random (MAR) data Listwise Deletion(LD), Mean Substitution(MS), Interpolation, Regression with an error term and Expectation-Maximization (EM) approaches were used to compare the effects of the various techniques.

Design, Analysis and Modeling of Dual Band Microstrip Loop Antenna Using Defective Ground Plane

Present wireless communication demands compact and intelligent devices with multitasking capabilities at affordable cost. The focus in the presented paper is on a dual band antenna for wireless communication with the capability of operating at two frequency bands with same structure. Two resonance frequencies are observed with the second operation band at 4.2GHz approximately three times the first resonance frequency at 1.5GHz. Structure is simple loop of microstrip line with characteristic impedance 50 ohms. The proposed antenna is designed using defective ground structure (DGS) and shows the nearly one third reductions in size as compared to without DGS. This antenna was simulated on electromagnetic (EM) simulation software and fabricated using microwave integrated circuit technique on RT-Duroid dielectric substrate (εr= 2.22) of thickness (H=15 mils). The designed antenna was tested on automatic network analyzer and shows the good agreement with simulated results. The proposed structure is modeled into an equivalent electrical circuit and simulated on circuit simulator. Subsequently, theoretical analysis was carried out and simulated. The simulated, measured, equivalent circuit response, and theoretical results shows good resemblance. The bands of operation draw many potential applications in today’s wireless communication.

Feature Subset Selection Using Ant Colony Optimization

Feature selection is an important step in many pattern classification problems. It is applied to select a subset of features, from a much larger set, such that the selected subset is sufficient to perform the classification task. Due to its importance, the problem of feature selection has been investigated by many researchers. In this paper, a novel feature subset search procedure that utilizes the Ant Colony Optimization (ACO) is presented. The ACO is a metaheuristic inspired by the behavior of real ants in their search for the shortest paths to food sources. It looks for optimal solutions by considering both local heuristics and previous knowledge. When applied to two different classification problems, the proposed algorithm achieved very promising results.

Non-equilibrium Statistical Mechanics of a Driven Lattice Gas Model: Probability Function, FDT-violation, and Monte Carlo Simulations

The study of non-equilibrium systems has attracted increasing interest in recent years, mainly due to the lack of theoretical frameworks, unlike their equilibrium counterparts. Studying the steady state and/or simple systems is thus one of the main interests. Hence in this work we have focused our attention on the driven lattice gas model (DLG model) consisting of interacting particles subject to an external field E. The dynamics of the system are given by hopping of particles to nearby empty sites with rates biased for jumps in the direction of E. Having used small two dimensional systems of DLG model, the stochastic properties at nonequilibrium steady state were analytically studied. To understand the non-equilibrium phenomena, we have applied the analytic approach via master equation to calculate probability function and analyze violation of detailed balance in term of the fluctuation-dissipation theorem. Monte Carlo simulations have been performed to validate the analytic results.

Performance Analysis of Quantum Cascaded Lasers

Improving the performance of the QCL through block diagram as well as mathematical models is the main scope of this paper. In order to enhance the performance of the underlined device, the mathematical model parameters are used in a reliable manner in such a way that the optimum behavior was achieved. These parameters play the central role in specifying the optical characteristics of the considered laser source. Moreover, it is important to have a large amount of radiated power, where increasing the amount of radiated power represents the main hopping process that can be predicted from the behavior of quantum laser devices. It was found that there is a good agreement between the calculated values from our mathematical model and those obtained with VisSim and experimental results. These demonstrate the strength of mplementation of both mathematical and block diagram models.

Extended Least Squares LS–SVM

Among neural models the Support Vector Machine (SVM) solutions are attracting increasing attention, mostly because they eliminate certain crucial questions involved by neural network construction. The main drawback of standard SVM is its high computational complexity, therefore recently a new technique, the Least Squares SVM (LS–SVM) has been introduced. In this paper we present an extended view of the Least Squares Support Vector Regression (LS–SVR), which enables us to develop new formulations and algorithms to this regression technique. Based on manipulating the linear equation set -which embodies all information about the regression in the learning process- some new methods are introduced to simplify the formulations, speed up the calculations and/or provide better results.

A Case Study on Appearance Based Feature Extraction Techniques and Their Susceptibility to Image Degradations for the Task of Face Recognition

Over the past decades, automatic face recognition has become a highly active research area, mainly due to the countless application possibilities in both the private as well as the public sector. Numerous algorithms have been proposed in the literature to cope with the problem of face recognition, nevertheless, a group of methods commonly referred to as appearance based have emerged as the dominant solution to the face recognition problem. Many comparative studies concerned with the performance of appearance based methods have already been presented in the literature, not rarely with inconclusive and often with contradictory results. No consent has been reached within the scientific community regarding the relative ranking of the efficiency of appearance based methods for the face recognition task, let alone regarding their susceptibility to appearance changes induced by various environmental factors. To tackle these open issues, this paper assess the performance of the three dominant appearance based methods: principal component analysis, linear discriminant analysis and independent component analysis, and compares them on equal footing (i.e., with the same preprocessing procedure, with optimized parameters for the best possible performance, etc.) in face verification experiments on the publicly available XM2VTS database. In addition to the comparative analysis on the XM2VTS database, ten degraded versions of the database are also employed in the experiments to evaluate the susceptibility of the appearance based methods on various image degradations which can occur in "real-life" operating conditions. Our experimental results suggest that linear discriminant analysis ensures the most consistent verification rates across the tested databases.

Total Quality Management: The Socio- Demographic and Operational-Financial Determinants for Users- Perception of the Services Quality

The aim of this paper is to know the sociodemographic and operational-financial determinants of the services quality perceived by users of the national health services. Through the use of an inquiry conducted by the Ministry of Health, comprehending 16.936 interviews in 2006, we intend to find out if there is any characteristic that determines the 2006 inquiry results. With the revision of the literature we also want to know if the operational-financial results have implications in hospitals users- perception on the quality of the received services. In order to achieve our main goals we will make use of the regression analysis to find out the possible dimensions that determine those results.

The Organizational Innovativeness of Public-Listed Housing Developers

This paper investigated the organizational innovativeness of public listed housing developers in Malaysia. We conceptualized organizational innovativeness as a multi-dimensional construct consisting of 5 dimensions: market innovativeness, product innovativeness, process innovativeness, behavior innovativeness and strategic innovativeness. We carried out questionnaire survey with all accessible public listed developers in Malaysia and received a 56 percent response. We found that the innovativeness of public listed housing developers is low. The study extends the knowledge on innovativeness theory by using a multi-dimensional contructs to conceptualize the innovativeness of public listed housing developers in Malaysia where all this while most studies focused on single dimensional construct of innovativeness. The paper ends by providing some explanations for the results.