Speaker Independent Quranic Recognizer Basedon Maximum Likelihood Linear Regression

An automatic speech recognition system for the formal Arabic language is needed. The Quran is the most formal spoken book in Arabic, it is spoken all over the world. In this research, an automatic speech recognizer for Quranic based speakerindependent was developed and tested. The system was developed based on the tri-phone Hidden Markov Model and Maximum Likelihood Linear Regression (MLLR). The MLLR computes a set of transformations which reduces the mismatch between an initial model set and the adaptation data. It uses the regression class tree, as well as, estimates a set of linear transformations for the mean and variance parameters of a Gaussian mixture HMM system. The 30th Chapter of the Quran, with five of the most famous readers of the Quran, was used for the training and testing of the data. The chapter includes about 2000 distinct words. The advantages of using the Quranic verses as the database in this developed recognizer are the uniqueness of the words and the high level of orderliness between verses. The level of accuracy from the tested data ranged 68 to 85%.

Content Based Image Retrieval of Brain MR Images across Different Classes

Magnetic Resonance Imaging play a vital role in the decision-diagnosis process of brain MR images. For an accurate diagnosis of brain related problems, the experts mostly compares both T1 and T2 weighted images as the information presented in these two images are complementary. In this paper, rotational and translational invariant form of Local binary Pattern (LBP) with additional gray scale information is used to retrieve similar slices of T1 weighted images from T2 weighted images or vice versa. The incorporation of additional gray scale information on LBP can extract more local texture information. The accuracy of retrieval can be improved by extracting moment features of LBP and reweighting the features based on users feedback. Here retrieval is done in a single subject scenario where similar images of a particular subject at a particular level are retrieved, and multiple subjects scenario where relevant images at a particular level across the subjects are retrieved.

Extraction of Symbolic Rules from Artificial Neural Networks

Although backpropagation ANNs generally predict better than decision trees do for pattern classification problems, they are often regarded as black boxes, i.e., their predictions cannot be explained as those of decision trees. In many applications, it is desirable to extract knowledge from trained ANNs for the users to gain a better understanding of how the networks solve the problems. A new rule extraction algorithm, called rule extraction from artificial neural networks (REANN) is proposed and implemented to extract symbolic rules from ANNs. A standard three-layer feedforward ANN is the basis of the algorithm. A four-phase training algorithm is proposed for backpropagation learning. Explicitness of the extracted rules is supported by comparing them to the symbolic rules generated by other methods. Extracted rules are comparable with other methods in terms of number of rules, average number of conditions for a rule, and predictive accuracy. Extensive experimental studies on several benchmarks classification problems, such as breast cancer, iris, diabetes, and season classification problems, demonstrate the effectiveness of the proposed approach with good generalization ability.

Software Maintenance Severity Prediction for Object Oriented Systems

As the majority of faults are found in a few of its modules so there is a need to investigate the modules that are affected severely as compared to other modules and proper maintenance need to be done in time especially for the critical applications. As, Neural networks, which have been already applied in software engineering applications to build reliability growth models predict the gross change or reusability metrics. Neural networks are non-linear sophisticated modeling techniques that are able to model complex functions. Neural network techniques are used when exact nature of input and outputs is not known. A key feature is that they learn the relationship between input and output through training. In this present work, various Neural Network Based techniques are explored and comparative analysis is performed for the prediction of level of need of maintenance by predicting level severity of faults present in NASA-s public domain defect dataset. The comparison of different algorithms is made on the basis of Mean Absolute Error, Root Mean Square Error and Accuracy Values. It is concluded that Generalized Regression Networks is the best algorithm for classification of the software components into different level of severity of impact of the faults. The algorithm can be used to develop model that can be used for identifying modules that are heavily affected by the faults.

A New Approach for Controlling Overhead Traveling Crane Using Rough Controller

This paper presents the idea of a rough controller with application to control the overhead traveling crane system. The structure of such a controller is based on a suggested concept of a fuzzy logic controller. A measure of fuzziness in rough sets is introduced. A comparison between fuzzy logic controller and rough controller has been demonstrated. The results of a simulation comparing the performance of both controllers are shown. From these results we infer that the performance of the proposed rough controller is satisfactory.

Application of Smooth Ergodic Hidden Markov Model in Text to Speech Systems

In developing a text-to-speech system, it is well known that the accuracy of information extracted from a text is crucial to produce high quality synthesized speech. In this paper, a new scheme for converting text into its equivalent phonetic spelling is introduced and developed. This method is applicable to many applications in text to speech converting systems and has many advantages over other methods. The proposed method can also complement the other methods with a purpose of improving their performance. The proposed method is a probabilistic model and is based on Smooth Ergodic Hidden Markov Model. This model can be considered as an extension to HMM. The proposed method is applied to Persian language and its accuracy in converting text to speech phonetics is evaluated using simulations.

Research on the Influence of Emotional Labor Strategy used by Public Transportation Employee on Service Satisfaction

The aim of the research is to understand whether the accuracy of customer detection of employee emotional labor strategy would influence the overall service satisfaction. From path analysis, it was found that employee-s positive emotions positively influenced service quality. Service quality in turn influenced Customer detection of employee emotional deep action strategy and Customer detection of employee emotional surface action strategy. Lastly, Customer detection of employee emotional deep action strategy and Customer detection of employee emotional surface action strategy positively influenced service satisfaction. Based on the analysis results, suggestions are proposed to provide reference for human resource management and use in relative fields.

Neural Network Ensemble-based Solar Power Generation Short-Term Forecasting

This paper presents the applicability of artificial neural networks for 24 hour ahead solar power generation forecasting of a 20 kW photovoltaic system, the developed forecasting is suitable for a reliable Microgrid energy management. In total four neural networks were proposed, namely: multi-layred perceptron, radial basis function, recurrent and a neural network ensemble consisting in ensemble of bagged networks. Forecasting reliability of the proposed neural networks was carried out in terms forecasting error performance basing on statistical and graphical methods. The experimental results showed that all the proposed networks achieved an acceptable forecasting accuracy. In term of comparison the neural network ensemble gives the highest precision forecasting comparing to the conventional networks. In fact, each network of the ensemble over-fits to some extent and leads to a diversity which enhances the noise tolerance and the forecasting generalization performance comparing to the conventional networks.

A Kernel Classifier using Linearised Bregman Iteration

In this paper we introduce a novel kernel classifier based on a iterative shrinkage algorithm developed for compressive sensing. We have adopted Bregman iteration with soft and hard shrinkage functions and generalized hinge loss for solving l1 norm minimization problem for classification. Our experimental results with face recognition and digit classification using SVM as the benchmark have shown that our method has a close error rate compared to SVM but do not perform better than SVM. We have found that the soft shrinkage method give more accuracy and in some situations more sparseness than hard shrinkage methods.

The Differential Transform Method for Advection-Diffusion Problems

In this paper a class of numerical methods to solve linear and nonlinear PDEs and also systems of PDEs is developed. The Differential Transform method associated with the Method of Lines (MoL) is used. The theory for linear problems is extended to the nonlinear case, and a recurrence relation is established. This method can achieve an arbitrary high-order accuracy in time. A variable stepsize algorithm and some numerical results are also presented.

High Accuracy Eigensolutions in Elasticity for Boundary Integral Equations by Nyström Method

Elastic boundary eigensolution problems are converted into boundary integral equations by potential theory. The kernels of the boundary integral equations have both the logarithmic and Hilbert singularity simultaneously. We present the mechanical quadrature methods for solving eigensolutions of the boundary integral equations by dealing with two kinds of singularities at the same time. The methods possess high accuracy O(h3) and low computing complexity. The convergence and stability are proved based on Anselone-s collective compact theory. Bases on the asymptotic error expansion with odd powers, we can greatly improve the accuracy of the approximation, and also derive a posteriori error estimate which can be used for constructing self-adaptive algorithms. The efficiency of the algorithms are illustrated by numerical examples.

Using the Monte Carlo Simulation to Predict the Assembly Yield

Electronics Products that achieve high levels of integrated communications, computing and entertainment, multimedia features in small, stylish and robust new form factors are winning in the market place. Due to the high costs that an industry may undergo and how a high yield is directly proportional to high profits, IC (Integrated Circuit) manufacturers struggle to maximize yield, but today-s customers demand miniaturization, low costs, high performance and excellent reliability making the yield maximization a never ending research of an enhanced assembly process. With factors such as minimum tolerances, tighter parameter variations a systematic approach is needed in order to predict the assembly process. In order to evaluate the quality of upcoming circuits, yield models are used which not only predict manufacturing costs but also provide vital information in order to ease the process of correction when the yields fall below expectations. For an IC manufacturer to obtain higher assembly yields all factors such as boards, placement, components, the material from which the components are made of and processes must be taken into consideration. Effective placement yield depends heavily on machine accuracy and the vision of the system which needs the ability to recognize the features on the board and component to place the device accurately on the pads and bumps of the PCB. There are currently two methods for accurate positioning, using the edge of the package and using solder ball locations also called footprints. The only assumption that a yield model makes is that all boards and devices are completely functional. This paper will focus on the Monte Carlo method which consists in a class of computational algorithms (information processed algorithms) which depends on repeated random samplings in order to compute the results. This method utilized in order to recreate the simulation of placement and assembly processes within a production line.

Study of Integrated Vehicle Image System Including LDW, FCW, and AFS

The objective of this research is to develop an advanced driver assistance system characterized with the functions of lane departure warning (LDW), forward collision warning (FCW) and adaptive front-lighting system (AFS). The system is mainly configured a CCD/CMOS camera to acquire the images of roadway ahead in association with the analysis made by an image-processing unit concerning the lane ahead and the preceding vehicles. The input image captured by a camera is used to recognize the lane and the preceding vehicle positions by image detection and DROI (Dynamic Range of Interesting) algorithms. Therefore, the system is able to issue real-time auditory and visual outputs of warning when a driver is departing the lane or driving too close to approach the preceding vehicle unwittingly so that the danger could be prevented from occurring. During the nighttime, in addition to the foregoing warning functions, the system is able to control the bending light of headlamp to provide an immediate light illumination when making a turn at a curved lane and adjust the level automatically to reduce the lighting interference against the oncoming vehicles driving in the opposite direction by the curvature of lane and the vanishing point estimations. The experimental results show that the integrated vehicle image system is robust to most environments such as the lane detection and preceding vehicle detection average accuracy performances are both above 90 %.

Using ANSYS to Realize a Semi-Analytical Method for Predicting Temperature Profile in Injection/Production Well

Determination of wellbore problems during a production/injection process might be evaluated thorough temperature log analysis. Other applications of this kind of log analysis may also include evaluation of fluid distribution analysis along the wellbore and identification of anomalies encountered during production/injection process. While the accuracy of such prediction is paramount, the common method of determination of a wellbore temperature log includes use of steady-state energy balance equations, which hardly describe the real conditions as observed in typical oil and gas flowing wells during production operation; and thus increase level of uncertainties. In this study, a practical method has been proposed through development of a simplified semianalytical model to apply for predicting temperature profile along the wellbore. The developed model includes an overall heat transfer coefficient accounting all modes of heat transferring mechanism, which has been focused on the prediction of a temperature profile as a function of depth for the injection/production wells. The model has been validated with the results obtained from numerical simulation.

Measurement of Systemic Power Efficiency of Microwave Heating Application

Microwave heating process has been developed about sixty years while measurement system has also progressed. Because of irradiation of high frequency of microwave, researchers have been utilized many costly technical instrument measuring parameters to evaluate the performance of microwave heating system. Therefore, this paper is intended to present an easier and feasible efficiency measurement method. It can help inspecting efficiency of microwave heating system with good accuracy, while the method can also give reference to optimizing procedure for microwave heating system for various load material

An Information Theoretic Approach to Rescoring Peptides Produced by De Novo Peptide Sequencing

Tandem mass spectrometry (MS/MS) is the engine driving high-throughput protein identification. Protein mixtures possibly representing thousands of proteins from multiple species are treated with proteolytic enzymes, cutting the proteins into smaller peptides that are then analyzed generating MS/MS spectra. The task of determining the identity of the peptide from its spectrum is currently the weak point in the process. Current approaches to de novo sequencing are able to compute candidate peptides efficiently. The problem lies in the limitations of current scoring functions. In this paper we introduce the concept of proteome signature. By examining proteins and compiling proteome signatures (amino acid usage) it is possible to characterize likely combinations of amino acids and better distinguish between candidate peptides. Our results strongly support the hypothesis that a scoring function that considers amino acid usage patterns is better able to distinguish between candidate peptides. This in turn leads to higher accuracy in peptide prediction.

Identification of Reusable Software Modules in Function Oriented Software Systems using Neural Network Based Technique

The cost of developing the software from scratch can be saved by identifying and extracting the reusable components from already developed and existing software systems or legacy systems [6]. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. We have used metric based approach for characterizing a software module. In this present work, the metrics McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component are used as input attributes to the different types of Neural Network system and reusability of the software component is calculated. The results are recorded in terms of Accuracy, Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE).

Parametric Cost Estimating Relationships for Design Effort Estimation

The Canadian aerospace industry faces many challenges. One of them is the difficulty in estimating costs. In particular, the design effort required in a project impacts resource requirements and lead-time, and consequently the final cost. This paper presents the findings of a case study conducted for recognized global leader in the design and manufacturing of aircraft engines. The study models parametric cost estimation relationships to estimate the design effort of integrated blade-rotor low-pressure compressor fans. Several effort drivers are selected to model the relationship. Comparative analyses of three types of models are conducted. The model with the best accuracy and significance in design estimation is retained.

The Current Implementation Status of Manufacturing Control Systems for a Key Manufacturing Industry

Manufacturing, production and service industries within Libya have struggled with many problems during the past two decades due to many difficulties. These problems have created a negative impact on the productivity and utilization of many industries around the country. This paper studies the implementation levels of the manufacturing control systems known as Manufacturing Resource Planning (MRPII) being adapted within some Libyan industries. A survey methodology has been applied for this research, based on the survey analysis, the results pointed out that the system within these industries has a modest strategy towards most of the areas that are considered as being very crucial in implementing these systems successfully. The findings also show a variation within these implementation levels with a respect to the key-elements that related to MRPII, giving the highest levels in the emphasise on financial data accuracy. The paper has also identified limitations within the investigated manufacturing and managerial areas and has pointed to where senior managers should take immediate actions in order to achieve effective implementation of MRPII within their business area.

Improvement Approach on Rotor Time Constant Adaptation with Optimum Flux in IFOC for Induction Machines Drives

Induction machine models used for steady-state and transient analysis require machine parameters that are usually considered design parameters or data. The knowledge of induction machine parameters is very important for Indirect Field Oriented Control (IFOC). A mismatched set of parameters will degrade the response of speed and torque control. This paper presents an improvement approach on rotor time constant adaptation in IFOC for Induction Machines (IM). Our approach tends to improve the estimation accuracy of the fundamental model for flux estimation. Based on the reduced order of the IM model, the rotor fluxes and rotor time constant are estimated using only the stator currents and voltages. This reduced order model offers many advantages for real time identification parameters of the IM.