Low Resolution Face Recognition Using Mixture of Experts

Human activity is a major concern in a wide variety of applications, such as video surveillance, human computer interface and face image database management. Detecting and recognizing faces is a crucial step in these applications. Furthermore, major advancements and initiatives in security applications in the past years have propelled face recognition technology into the spotlight. The performance of existing face recognition systems declines significantly if the resolution of the face image falls below a certain level. This is especially critical in surveillance imagery where often, due to many reasons, only low-resolution video of faces is available. If these low-resolution images are passed to a face recognition system, the performance is usually unacceptable. Hence, resolution plays a key role in face recognition systems. In this paper we introduce a new low resolution face recognition system based on mixture of expert neural networks. In order to produce the low resolution input images we down-sampled the 48 × 48 ORL images to 12 × 12 ones using the nearest neighbor interpolation method and after that applying the bicubic interpolation method yields enhanced images which is given to the Principal Component Analysis feature extractor system. Comparison with some of the most related methods indicates that the proposed novel model yields excellent recognition rate in low resolution face recognition that is the recognition rate of 100% for the training set and 96.5% for the test set.

Existence and Stability Analysis of Discrete-time Fuzzy BAM Neural Networks with Delays and Impulses

In this paper, the discrete-time fuzzy BAM neural network with delays and impulses is studied. Sufficient conditions are obtained for the existence and global stability of a unique equilibrium of this class of fuzzy BAM neural networks with Lipschitzian activation functions without assuming their boundedness, monotonicity or differentiability and subjected to impulsive state displacements at fixed instants of time. Some numerical examples are given to demonstrate the effectiveness of the obtained results.

Design of Stable IIR Digital Filters with Specified Group Delay Errors

The design problem of Infinite Impulse Response (IIR) digital filters is usually expressed as the minimization problem of the complex magnitude error that includes both the magnitude and phase information. However, the group delay of the filter obtained by solving such design problem may be far from the desired group delay. In this paper, we propose a design method of stable IIR digital filters with prespecified maximum group delay errors. In the proposed method, the approximation problems of the magnitude-phase and group delay are separately defined, and these two approximation problems are alternately solved using successive projections. As a result, the proposed method can design the IIR filters that satisfy the prespecified allowable errors for not only the complex magnitude but also the group delay by alternately executing the coefficient update for the magnitude-phase and the group delay approximation. The usefulness of the proposed method is verified through some examples.

Modified Fuzzy ARTMAP and Supervised Fuzzy ART: Comparative Study with Multispectral Classification

In this article a modification of the algorithm of the fuzzy ART network, aiming at returning it supervised is carried out. It consists of the search for the comparison, training and vigilance parameters giving the minimum quadratic distances between the output of the training base and those obtained by the network. The same process is applied for the determination of the parameters of the fuzzy ARTMAP giving the most powerful network. The modification consist in making learn the fuzzy ARTMAP a base of examples not only once as it is of use, but as many time as its architecture is in evolution or than the objective error is not reached . In this way, we don-t worry about the values to impose on the eight (08) parameters of the network. To evaluate each one of these three networks modified, a comparison of their performances is carried out. As application we carried out a classification of the image of Algiers-s bay taken by SPOT XS. We use as criterion of evaluation the training duration, the mean square error (MSE) in step control and the rate of good classification per class. The results of this study presented as curves, tables and images show that modified fuzzy ARTMAP presents the best compromise quality/computing time.

Modified Levenberg-Marquardt Method for Neural Networks Training

In this paper a modification on Levenberg-Marquardt algorithm for MLP neural network learning is proposed. The proposed algorithm has good convergence. This method reduces the amount of oscillation in learning procedure. An example is given to show usefulness of this method. Finally a simulation verifies the results of proposed method.

The Effects of Various Boundary Conditions on Thermal Buckling of Functionally Graded Beamwith Piezoelectric Layers Based on Third order Shear Deformation Theory

This article attempts to analyze functionally graded beam thermal buckling along with piezoelectric layers applying based on the third order shearing deformation theory considering various boundary conditions. The beam properties are assumed to vary continuously from the lower surface to the upper surface of the beam. The equilibrium equations are derived using the total potential energy equations, Euler equations, piezoelectric material constitutive equations and third order shear deformation theory assumptions. In order to fulfill such an aim, at first functionally graded beam with piezoelectric layers applying the third order shearing deformation theory along with clamped -clamped boundary conditions are thoroughly analyzed, and then following making sure of the correctness of all the equations, the very same beam is analyzed with piezoelectric layers through simply-simply and simply-clamped boundary conditions. In this article buckling critical temperature for functionally graded beam is derived in two different ways, without piezoelectric layer and with piezoelectric layer and the results are compared together. Finally, all the conclusions obtained will be compared and contrasted with the same samples in the same and distinguished conditions through tables and charts. It would be noteworthy that in this article, the software MAPLE has been applied in order to do the numeral calculations.

Learning to Order Terms: Supervised Interestingness Measures in Terminology Extraction

Term Extraction, a key data preparation step in Text Mining, extracts the terms, i.e. relevant collocation of words, attached to specific concepts (e.g. genetic-algorithms and decisiontrees are terms associated to the concept “Machine Learning" ). In this paper, the task of extracting interesting collocations is achieved through a supervised learning algorithm, exploiting a few collocations manually labelled as interesting/not interesting. From these examples, the ROGER algorithm learns a numerical function, inducing some ranking on the collocations. This ranking is optimized using genetic algorithms, maximizing the trade-off between the false positive and true positive rates (Area Under the ROC curve). This approach uses a particular representation for the word collocations, namely the vector of values corresponding to the standard statistical interestingness measures attached to this collocation. As this representation is general (over corpora and natural languages), generality tests were performed by experimenting the ranking function learned from an English corpus in Biology, onto a French corpus of Curriculum Vitae, and vice versa, showing a good robustness of the approaches compared to the state-of-the-art Support Vector Machine (SVM).

Numerical Solution of Riccati Differential Equations by Using Hybrid Functions and Tau Method

A numerical method for Riccati equation is presented in this work. The method is based on the replacement of unknown functions through a truncated series of hybrid of block-pulse functions and Chebyshev polynomials. The operational matrices of derivative and product of hybrid functions are presented. These matrices together with the tau method are then utilized to transform the differential equation into a system of algebraic equations. Corresponding numerical examples are presented to demonstrate the accuracy of the proposed method.

Suggestions for the Improvement of the Quality of Public Transportation Service in Campos,Brazil

In this paper the main objective is to analyze the quality of service of the bus companies operating in the city of Campos, located in the state of Rio de Janeiro, Brazil. This analysis, based on the opinion of the bus customers, will help to determine their degree of satisfaction with the service provided by the bus companies. The result of this assessment shows that the bus customers are displeased with the quality of service supplied by the bus companies. Therefore, it is necessary to identify alternative solutions to minimize the consequences of the main problems related to customers- dissatisfaction identified in our evaluation and to help the bus companies operating in Campos better fulfill their riders- needs.

Application of Adaptive Neuro-Fuzzy Inference System in the Prediction of Economic Crisis Periods in USA

In this paper discrete choice models, Logit and Probit are examined in order to predict the economic recession or expansion periods in USA. Additionally we propose an adaptive neuro-fuzzy inference system with triangular membership function. We examine the in-sample period 1947-2005 and we test the models in the out-of sample period 2006-2009. The forecasting results indicate that the Adaptive Neuro-fuzzy Inference System (ANFIS) model outperforms significant the Logit and Probit models in the out-of sample period. This indicates that neuro-fuzzy model provides a better and more reliable signal on whether or not a financial crisis will take place.

Performance Analysis of Chrominance Red and Chrominance Blue in JPEG

While compressing text files is useful, compressing still image files is almost a necessity. A typical image takes up much more storage than a typical text message and without compression images would be extremely clumsy to store and distribute. The amount of information required to store pictures on modern computers is quite large in relation to the amount of bandwidth commonly available to transmit them over the Internet and applications. Image compression addresses the problem of reducing the amount of data required to represent a digital image. Performance of any image compression method can be evaluated by measuring the root-mean-square-error & peak signal to noise ratio. The method of image compression that will be analyzed in this paper is based on the lossy JPEG image compression technique, the most popular compression technique for color images. JPEG compression is able to greatly reduce file size with minimal image degradation by throwing away the least “important" information. In JPEG, both color components are downsampled simultaneously, but in this paper we will compare the results when the compression is done by downsampling the single chroma part. In this paper we will demonstrate more compression ratio is achieved when the chrominance blue is downsampled as compared to downsampling the chrominance red in JPEG compression. But the peak signal to noise ratio is more when the chrominance red is downsampled as compared to downsampling the chrominance blue in JPEG compression. In particular we will use the hats.jpg as a demonstration of JPEG compression using low pass filter and demonstrate that the image is compressed with barely any visual differences with both methods.

Derivative Spectrophotometry Applied to the Determination of Triprolidine Hydrochloride and Pseudoephedrine Hydrochloride in Tablets and Dissolution Testing

A spectrophotometric method was developed for simultaneous quantification of pseudoephedrine hydrochloride (PSE) triprolidine hydrochloride (TRI) using second derivative method (zero-crossing technique). The second derivative amplitudes of PSE and TRI were measured at 271 and 321 nm, respectively. The calibration curves were linear in the range of 200 to 1,000 g/ml for PSE and 10 to 50 g/ml for TRI. The method was validated for specificity, accuracy, precision, limit of detection and limit of quantitation. The proposed method was applied to the assaying and dissolution of PSE and TRI in commercial tablets without any chemical separation. The results were compared with those obtained by the official USP31 method and statistical tests showed that there is no significant between the methods at 95% confidence level. The proposed method is simple, rapid and suitable for the routine quality control application. KeywordsTriprolidine, Pseudoephedrine, Derivative spectrophotometry, Dissolution testing.

A Robust LS-SVM Regression

In comparison to the original SVM, which involves a quadratic programming task; LS–SVM simplifies the required computation, but unfortunately the sparseness of standard SVM is lost. Another problem is that LS-SVM is only optimal if the training samples are corrupted by Gaussian noise. In Least Squares SVM (LS–SVM), the nonlinear solution is obtained, by first mapping the input vector to a high dimensional kernel space in a nonlinear fashion, where the solution is calculated from a linear equation set. In this paper a geometric view of the kernel space is introduced, which enables us to develop a new formulation to achieve a sparse and robust estimate.

Investigations of Free-to-Roll Motions and its Active Control under Pitch-up Maneuvers

Experiments have been carried out at sub-critical Reynolds number to investigate free-to-roll motions induced by forebody and/or wings complex flow on a 30° swept back nonslender wings-slender body-model for static and dynamic (pitch-up) cases. For the dynamic (pitch-up) case it has been observed that roll amplitude decreases and lag increases with increase in pitching speed. Decrease in roll amplitude with increase in pitch rate is attributed to low disturbing rolling moment due to weaker interaction between forebody and wing flow components. Asymmetric forebody vortices dominate and control the roll motion of the model in dynamic case when non-dimensional pitch rate ≥ 1x10-2. Effectiveness of the active control scheme utilizing rotating nose with artificial tip perturbation is observed to be low in the angle of attack region where the complex flow over the wings has contributions from both forebody and wings.

Gene Expression Data Classification Using Discriminatively Regularized Sparse Subspace Learning

Sparse representation which can represent high dimensional data effectively has been successfully used in computer vision and pattern recognition problems. However, it doesn-t consider the label information of data samples. To overcome this limitation, we develop a novel dimensionality reduction algorithm namely dscriminatively regularized sparse subspace learning(DR-SSL) in this paper. The proposed DR-SSL algorithm can not only make use of the sparse representation to model the data, but also can effective employ the label information to guide the procedure of dimensionality reduction. In addition,the presented algorithm can effectively deal with the out-of-sample problem.The experiments on gene-expression data sets show that the proposed algorithm is an effective tool for dimensionality reduction and gene-expression data classification.

Intellectual Capital and Competitive Advantage: An Analysis of the Biotechnology Industry

Intellectual capital measurement is a central aspect of knowledge management. The measurement and the evaluation of intangible assets play a key role in allowing an effective management of these assets as sources of competitiveness. For these reasons, managers and practitioners need conceptual and analytical tools taking into account the unique characteristics and economic significance of Intellectual Capital. Following this lead, we propose an efficiency and productivity analysis of Intellectual Capital, as a determinant factor of the company competitive advantage. The analysis is carried out by means of Data Envelopment Analysis (DEA) and Malmquist Productivity Index (MPI). These techniques identify Bests Practice companies that have accomplished competitive advantage implementing successful strategies of Intellectual Capital management, and offer to inefficient companies development paths by means of benchmarking. The proposed methodology is employed on the Biotechnology industry in the period 2007-2010.

Enhancements in Blended e-Learning Management System

A learning management system (commonly abbreviated as LMS) is a software application for the administration, documentation, tracking, and reporting of training programs, classroom and online events, e-learning programs, and training content (Ellis 2009). (Hall 2003) defines an LMS as \"software that automates the administration of training events. All Learning Management Systems manage the log-in of registered users, manage course catalogs, record data from learners, and provide reports to management\". Evidence of the worldwide spread of e-learning in recent years is easy to obtain. In April 2003, no fewer than 66,000 fully online courses and 1,200 complete online programs were listed on the TeleCampus portal from TeleEducation (Paulsen 2003). In the report \" The US market in the Self-paced eLearning Products and Services:2010-2015 Forecast and Analysis\" The number of student taken classes exclusively online will be nearly equal (1% less) to the number taken classes exclusively in physical campuses. Number of student taken online course will increase from 1.37 million in 2010 to 3.86 million in 2015 in USA. In another report by The Sloan Consortium three-quarters of institutions report that the economic downturn has increased demand for online courses and programs.

Biometric Authentication Using Fast Correlation of Near Infrared Hand Vein Patterns

This paper presents a hand vein authentication system using fast spatial correlation of hand vein patterns. In order to evaluate the system performance, a prototype was designed and a dataset of 50 persons of different ages above 16 and of different gender, each has 10 images per person was acquired at different intervals, 5 images for left hand and 5 images for right hand. In verification testing analysis, we used 3 images to represent the templates and 2 images for testing. Each of the 2 images is matched with the existing 3 templates. FAR of 0.02% and FRR of 3.00 % were reported at threshold 80. The system efficiency at this threshold was found to be 99.95%. The system can operate at a 97% genuine acceptance rate and 99.98 % genuine reject rate, at corresponding threshold of 80. The EER was reported as 0.25 % at threshold 77. We verified that no similarity exists between right and left hand vein patterns for the same person over the acquired dataset sample. Finally, this distinct 100 hand vein patterns dataset sample can be accessed by researchers and students upon request for testing other methods of hand veins matching.

Copper Contamination in the Sediments of Northern Kaohsiung Harbor, Taiwan

The distribution, enrichment, accumulation, and potential ecological risk of copper (Cu) in the surface sediments of northern Kaohsiung Harbor, Taiwan were investigated. Sediment samples from 12 locations of northern Kaohsiung Harbor were collected and characterized for Cu, aluminum, water content, organic matter, total nitrogen, total phosphorous, total grease and grain size. Results showed that the Cu concentrations varied from 6.9–244 mg/kg with an average of 109±66 mg/kg. The spatial distribution of Cu reveals that the Cu concentration is relatively high in the river mouth region, and gradually diminishes toward the harbor entrance region. This indicates that upstream industrial and municipal wastewater discharges along the river bank are major sources of Cu pollution. Results from the enrichment factor and geo-accumulation index analyses imply that the sediments collected from the river mouth can be characterized between moderate and moderately severe degree enrichment and between none to medium and moderate accumulation of Cu, respectively. However, results of potential ecological risk index indicate that the sediment has low ecological potential risk.

Minimizing Risk Costs through Optimal Responses in NPD Projects

In rapidly changing market environment, firms are investing a lot of time and resources into new product development (NPD) projects to make profit and to obtain competitive advantage. However, failure rate of NPD projects is becoming high due to various internal and external risks which hinder successful NPD projects. To reduce the failure rate, it is critical that risks have to be managed effectively and efficiently through good strategy, and treated by optimal responses to minimize risk cost. Four strategies are adopted to handle the risks in this study. The optimal responses are characterized by high reduction of risk costs with high efficiency. This study suggests a framework to decide the optimal responses considering the core risks, risk costs, response efficiency and response costs for successful NPD projects. Both binary particles warm optimization (BPSO) and multi-objective particle swarm optimization (MOPSO) methods are mainly used in the framework. Although several limitations exist in use for real industries, the frame work shows good strength for handling the risks with highly scientific ways through an example.