Face Localization and Recognition in Varied Expressions and Illumination

In this paper, we propose a robust scheme to work face alignment and recognition under various influences. For face representation, illumination influence and variable expressions are the important factors, especially the accuracy of facial localization and face recognition. In order to solve those of factors, we propose a robust approach to overcome these problems. This approach consists of two phases. One phase is preprocessed for face images by means of the proposed illumination normalization method. The location of facial features can fit more efficient and fast based on the proposed image blending. On the other hand, based on template matching, we further improve the active shape models (called as IASM) to locate the face shape more precise which can gain the recognized rate in the next phase. The other phase is to process feature extraction by using principal component analysis and face recognition by using support vector machine classifiers. The results show that this proposed method can obtain good facial localization and face recognition with varied illumination and local distortion.

An Efficient Watermarking Method for MP3 Audio Files

In this work, we present for the first time in our perception an efficient digital watermarking scheme for mpeg audio layer 3 files that operates directly in the compressed data domain, while manipulating the time and subband/channel domain. In addition, it does not need the original signal to detect the watermark. Our scheme was implemented taking special care for the efficient usage of the two limited resources of computer systems: time and space. It offers to the industrial user the capability of watermark embedding and detection in time immediately comparable to the real music time of the original audio file that depends on the mpeg compression, while the end user/audience does not face any artifacts or delays hearing the watermarked audio file. Furthermore, it overcomes the disadvantage of algorithms operating in the PCMData domain to be vulnerable to compression/recompression attacks, as it places the watermark in the scale factors domain and not in the digitized sound audio data. The strength of our scheme, that allows it to be used with success in both authentication and copyright protection, relies on the fact that it gives to the users the enhanced capability their ownership of the audio file not to be accomplished simply by detecting the bit pattern that comprises the watermark itself, but by showing that the legal owner knows a hard to compute property of the watermark.

Application of Multi-Dimensional Principal Component Analysis to Medical Data

Multi-dimensional principal component analysis (PCA) is the extension of the PCA, which is used widely as the dimensionality reduction technique in multivariate data analysis, to handle multi-dimensional data. To calculate the PCA the singular value decomposition (SVD) is commonly employed by the reason of its numerical stability. The multi-dimensional PCA can be calculated by using the higher-order SVD (HOSVD), which is proposed by Lathauwer et al., similarly with the case of ordinary PCA. In this paper, we apply the multi-dimensional PCA to the multi-dimensional medical data including the functional independence measure (FIM) score, and describe the results of experimental analysis.

Monitoring Patents Using the Statistical Process Control

The statistical process control (SPC) is one of the most powerful tools developed to assist ineffective control of quality, involves collecting, organizing and interpreting data during production. This article aims to show how the use of CEP industries can control and continuously improve product quality through monitoring of production that can detect deviations of parameters representing the process by reducing the amount of off-specification products and thus the costs of production. This study aimed to conduct a technological forecasting in order to characterize the research being done related to the CEP. The survey was conducted in the databases Spacenet, WIPO and the National Institute of Industrial Property (INPI). Among the largest are the United States depositors and deposits via PCT, the classification section that was presented in greater abundance to F.

A Comparative Study of Main Memory Databases and Disk-Resident Databases

Main Memory Database systems (MMDB) store their data in main physical memory and provide very high-speed access. Conventional database systems are optimized for the particular characteristics of disk storage mechanisms. Memory resident systems, on the other hand, use different optimizations to structure and organize data, as well as to make it reliable. This paper provides a brief overview on MMDBs and one of the memory resident systems named FastDB and compares the processing time of this system with a typical disc resident database based on the results of the implementation of TPC benchmarks environment on both.

64 bit Computer Architectures for Space Applications – A study

The more recent satellite projects/programs makes extensive usage of real – time embedded systems. 16 bit processors which meet the Mil-Std-1750 standard architecture have been used in on-board systems. Most of the Space Applications have been written in ADA. From a futuristic point of view, 32 bit/ 64 bit processors are needed in the area of spacecraft computing and therefore an effort is desirable in the study and survey of 64 bit architectures for space applications. This will also result in significant technology development in terms of VLSI and software tools for ADA (as the legacy code is in ADA). There are several basic requirements for a special processor for this purpose. They include Radiation Hardened (RadHard) devices, very low power dissipation, compatibility with existing operational systems, scalable architectures for higher computational needs, reliability, higher memory and I/O bandwidth, predictability, realtime operating system and manufacturability of such processors. Further on, these may include selection of FPGA devices, selection of EDA tool chains, design flow, partitioning of the design, pin count, performance evaluation, timing analysis etc. This project deals with a brief study of 32 and 64 bit processors readily available in the market and designing/ fabricating a 64 bit RISC processor named RISC MicroProcessor with added functionalities of an extended double precision floating point unit and a 32 bit signal processing unit acting as co-processors. In this paper, we emphasize the ease and importance of using Open Core (OpenSparc T1 Verilog RTL) and Open “Source" EDA tools such as Icarus to develop FPGA based prototypes quickly. Commercial tools such as Xilinx ISE for Synthesis are also used when appropriate.

Coreless Printed Circuit Board (PCB) Stepdown Transformers for DC-DC Converter Applications

In this paper, multilayered coreless printed circuit board (PCB) step-down power transformers for DC-DC converter applications have been designed, manufactured and evaluated. A set of two different circular spiral step-down transformers were fabricated in the four layered PCB. These transformers have been modelled with the assistance of high frequency equivalent circuit and characterized with both sinusoidal and square wave excitation. This paper provides the comparative results of these two different transformers in terms of their resistances, self, leakage, mutual inductances, coupling coefficient and also their energy efficiencies. The operating regions for optimal performance of these transformers for power transfer applications are determined. These transformers were tested for the output power levels of about 30 Watts within the input voltage range of 12-50 Vrms. The energy efficiency for these step down transformers is observed to be in the range of 90%-97% in MHz frequency region.

Nonlinear Model Predictive Control for Solid Oxide Fuel Cell System Based On Wiener Model

In this paper, we consider Wiener nonlinear model for solid oxide fuel cell (SOFC). The Wiener model of the SOFC consists of a linear dynamic block and a static output non-linearity followed by the block, in which linear part is approximated by state-space model and the nonlinear part is identified by a polynomial form. To control the SOFC system, we have to consider various view points such as operating conditions, another constraint conditions, change of load current and so on. A change of load current is the significant one of these for good performance of the SOFC system. In order to keep the constant stack terminal voltage by changing load current, the nonlinear model predictive control (MPC) is proposed in this paper. After primary control method is designed to guarantee the fuel utilization as a proper constant, a nonlinear model predictive control based on the Wiener model is developed to control the stack terminal voltage of the SOFC system. Simulation results verify the possibility of the proposed Wiener model and MPC method to control of SOFC system.

Aroma Composition and Polyphenol Content of Ciders Available in Latvian Market

Aroma forming volatiles are important components of fermented beverages. The aim of current research is to evaluate the volatile compounds and phenolic compounds of commercial ciders. Volatile aroma compounds and TPC of seven commercial ciders were determined. Extraction of aroma compounds was performed using solid phase microextraction (DVB/Car/PDMS fibre). Analysis of volatile aroma compounds was made using a Perkin Elmer Clarus 500 GC/MS. Total phenol content (TPC) was determined according to the Folin-Ciocalteu spectrophotometric method and results were expressed as gallic acid equivalents. The highest volatile compounds were in apple ciders with pear flavor. The highest TPC and lower content of volatile compounds were detected in French ciders.

Implementation of Parallel Interface for Microprocessor Trainer

In this paper, parallel interface for microprocessor trainer was implemented. A programmable parallel–port device such as the IC 8255A is initialized for simple input or output and for handshake input or output by choosing kinds of modes. The hardware connections and the programs can be used to interface microprocessor trainer and a personal computer by using IC 8255A. The assembly programs edited on PC-s editor can be downloaded to the trainer.

Unified Fusion Approach with Application to SLAM

In this paper, we propose the pre-processor based on the Evidence Supporting Measure of Similarity (ESMS) filter and also propose the unified fusion approach (UFA) based on the general fusion machine coupled with ESMS filter, which improve the correctness and precision of information fusion in any fields of application. Here we mainly apply the new approach to Simultaneous Localization And Mapping (SLAM) of Pioneer II mobile robots. A simulation experiment was performed, where an autonomous virtual mobile robot with sonar sensors evolves in a virtual world map with obstacles. By comparing the result of building map according to the general fusion machine (here DSmT-based fusing machine and PCR5-based conflict redistributor considereded) coupling with ESMS filter and without ESMS filter, it shows the benefit of the selection of the sources as a prerequisite for improvement of the information fusion, and also testifies the superiority of the UFA in dealing with SLAM.

An Enhanced Distributed System to improve theTime Complexity of Binary Indexed Trees

Distributed Computing Systems are usually considered the most suitable model for practical solutions of many parallel algorithms. In this paper an enhanced distributed system is presented to improve the time complexity of Binary Indexed Trees (BIT). The proposed system uses multi-uniform processors with identical architectures and a specially designed distributed memory system. The analysis of this system has shown that it has reduced the time complexity of the read query to O(Log(Log(N))), and the update query to constant complexity, while the naive solution has a time complexity of O(Log(N)) for both queries. The system was implemented and simulated using VHDL and Verilog Hardware Description Languages, with xilinx ISE 10.1, as the development environment and ModelSim 6.1c, similarly as the simulation tool. The simulation has shown that the overhead resulting by the wiring and communication between the system fragments could be fairly neglected, which makes it applicable to practically reach the maximum speed up offered by the proposed model.

View-Point Insensitive Human Pose Recognition using Neural Network and CUDA

Although lots of research work has been done for human pose recognition, the view-point of cameras is still critical problem of overall recognition system. In this paper, view-point insensitive human pose recognition is proposed. The aims of the proposed system are view-point insensitivity and real-time processing. Recognition system consists of feature extraction module, neural network and real-time feed forward calculation. First, histogram-based method is used to extract feature from silhouette image and it is suitable for represent the shape of human pose. To reduce the dimension of feature vector, Principle Component Analysis(PCA) is used. Second, real-time processing is implemented by using Compute Unified Device Architecture(CUDA) and this architecture improves the speed of feed-forward calculation of neural network. We demonstrate the effectiveness of our approach with experiments on real environment.

An Implementation of MacMahon's Partition Analysis in Ordering the Lower Bound of Processing Elements for the Algorithm of LU Decomposition

A lot of Scientific and Engineering problems require the solution of large systems of linear equations of the form bAx in an effective manner. LU-Decomposition offers good choices for solving this problem. Our approach is to find the lower bound of processing elements needed for this purpose. Here is used the so called Omega calculus, as a computational method for solving problems via their corresponding Diophantine relation. From the corresponding algorithm is formed a system of linear diophantine equalities using the domain of computation which is given by the set of lattice points inside the polyhedron. Then is run the Mathematica program DiophantineGF.m. This program calculates the generating function from which is possible to find the number of solutions to the system of Diophantine equalities, which in fact gives the lower bound for the number of processors needed for the corresponding algorithm. There is given a mathematical explanation of the problem as well. Keywordsgenerating function, lattice points in polyhedron, lower bound of processor elements, system of Diophantine equationsand : calculus.

Characterization of the O.ul-mS952 Intron:A Potential Molecular Marker to Distinguish Between Ophiostoma Ulmi and Ophiostoma Novo-Ulmi Subsp. Americana

The full length mitochondrial small subunit ribosomal (mt-rns) gene has been characterized for Ophiostoma novo-ulmi subspecies americana. The gene was also characterized for Ophiostoma ulmi and a group II intron was noted in the mt-rns gene of O. ulmi. The insertion in the mt-rns gene is at position S952 and it is a group IIB1 intron that encodes a double motif LAGLIDADG homing endonuclease from an open reading frame located within a loop of domain III. Secondary structure models for the mt-rns RNA of O. novo-ulmi subsp. americana and O. ulmi were generated to place the intron within the context of the ribosomal RNA. The in vivo splicing of the O.ul-mS952 group II intron was confirmed with reverse transcription-PCR. A survey of 182 strains of Dutch Elm Diseases causing agents showed that the mS952 intron was absent in what is considered to be the more aggressive species O. novo-ulmi but present in strains of the less aggressive O. ulmi. This observation suggests that the O.ul-mS952 intron can be used as a PCR-based molecular marker to discriminate between O. ulmi and O. novo-ulmi subsp. americana.

Palmprint Recognition by Wavelet Transform with Competitive Index and PCA

This manuscript presents, palmprint recognition by combining different texture extraction approaches with high accuracy. The Region of Interest (ROI) is decomposed into different frequencytime sub-bands by wavelet transform up-to two levels and only the approximate image of two levels is selected, which is known as Approximate Image ROI (AIROI). This AIROI has information of principal lines of the palm. The Competitive Index is used as the features of the palmprint, in which six Gabor filters of different orientations convolve with the palmprint image to extract the orientation information from the image. The winner-take-all strategy is used to select dominant orientation for each pixel, which is known as Competitive Index. Further, PCA is applied to select highly uncorrelated Competitive Index features, to reduce the dimensions of the feature vector, and to project the features on Eigen space. The similarity of two palmprints is measured by the Euclidean distance metrics. The algorithm is tested on Hong Kong PolyU palmprint database. Different AIROI of different wavelet filter families are also tested with the Competitive Index and PCA. AIROI of db7 wavelet filter achievs Equal Error Rate (EER) of 0.0152% and Genuine Acceptance Rate (GAR) of 99.67% on the palm database of Hong Kong PolyU.

Classification of Defects by the SVM Method and the Principal Component Analysis (PCA)

Analyses carried out on examples of detected defects echoes showed clearly that one can describe these detected forms according to a whole of characteristic parameters in order to be able to make discrimination between a planar defect and a volumic defect. This work answers to a problem of ultrasonics NDT like Identification of the defects. The problems as well as the objective of this realized work, are divided in three parts: Extractions of the parameters of wavelets from the ultrasonic echo of the detected defect - the second part is devoted to principal components analysis (PCA) for optimization of the attributes vector. And finally to establish the algorithm of classification (SVM, Support Vector Machine) which allows discrimination between a plane defect and a volumic defect. We have completed this work by a conclusion where we draw up a summary of the completed works, as well as the robustness of the various algorithms proposed in this study.

Durability of Mortar in Presence of Rice Husk Ash

The purpose of this paper is to investigate the durability of cement mortar in presence of Rice Husk Ash (RHA). The strength and durability of mortar with different replacement level (0%, 10%, 15%, 20%, 25% and 30%) of Ordinary Portland Cement (OPC) by RHA is investigated here. RHA was manufactured from an uncontrolled burning process. Test samples were prepared with river sand of FM 2.73. Samples were kept in controlled environment up to test time. The results show that addition of RHA was shown better results for 20% replacement level than OPC at 90 days. In durability test all samples passed for 20 cycles except 25% and 30% replacement level.

Investigation of 5,10,15,20-Tetrakis(3-,5--Di-Tert-Butylphenyl)Porphyrinatocopper(II) for Electronics Applications

In this work, an organic compound 5,10,15,20- Tetrakis(3,5-di-tertbutylphenyl)porphyrinatocopper(II) (TDTBPPCu) is studied as an active material for thin film electronic devices. To investigate the electrical properties of TDTBPPCu, junction of TDTBPPCu with heavily doped n-Si and Al is fabricated. TDTBPPCu film was sandwiched between Al and n-Si electrodes. Various electrical parameters of TDTBPPCu are determined. The current-voltage characteristics of the junction are nonlinear, asymmetric and show rectification behavior, which gives the clue of formation of depletion region. This behavior indicates the potential of TDTBPPCu for electronics applications. The current-voltage and capacitance-voltage techniques are used to find the different electronic parameters.

H-ARQ Techniques for Wireless Systems with Punctured Non-Binary LDPC as FEC Code

This paper presents the H-ARQ techniques comparison for OFDM systems with a new family of non-binary LDPC codes which has been developed within the EU FP7 DAVINCI project. The punctured NB-LDPC codes have been used in a simulated model of the transmission system. The link level performance has been evaluated in terms of spectral efficiency, codeword error rate and average number of retransmissions. The NB-LDPC codes can be easily and effective implemented with different methods of the retransmission needed if correct decoding of a codeword failed. Here the Optimal Symbol Selection method is proposed as a Chase Combining technique.