Detection and Classification of Faults on Parallel Transmission Lines Using Wavelet Transform and Neural Network

The protection of parallel transmission lines has been a challenging task due to mutual coupling between the adjacent circuits of the line. This paper presents a novel scheme for detection and classification of faults on parallel transmission lines. The proposed approach uses combination of wavelet transform and neural network, to solve the problem. While wavelet transform is a powerful mathematical tool which can be employed as a fast and very effective means of analyzing power system transient signals, artificial neural network has a ability to classify non-linear relationship between measured signals by identifying different patterns of the associated signals. The proposed algorithm consists of time-frequency analysis of fault generated transients using wavelet transform, followed by pattern recognition using artificial neural network to identify the type of the fault. MATLAB/Simulink is used to generate fault signals and verify the correctness of the algorithm. The adaptive discrimination scheme is tested by simulating different types of fault and varying fault resistance, fault location and fault inception time, on a given power system model. The simulation results show that the proposed scheme for fault diagnosis is able to classify all the faults on the parallel transmission line rapidly and correctly.

ANN Based Currency Recognition System using Compressed Gray Scale and Application for Sri Lankan Currency Notes - SLCRec

Automatic currency note recognition invariably depends on the currency note characteristics of a particular country and the extraction of features directly affects the recognition ability. Sri Lanka has not been involved in any kind of research or implementation of this kind. The proposed system “SLCRec" comes up with a solution focusing on minimizing false rejection of notes. Sri Lankan currency notes undergo severe changes in image quality in usage. Hence a special linear transformation function is adapted to wipe out noise patterns from backgrounds without affecting the notes- characteristic images and re-appear images of interest. The transformation maps the original gray scale range into a smaller range of 0 to 125. Applying Edge detection after the transformation provided better robustness for noise and fair representation of edges for new and old damaged notes. A three layer back propagation neural network is presented with the number of edges detected in row order of the notes and classification is accepted in four classes of interest which are 100, 500, 1000 and 2000 rupee notes. The experiments showed good classification results and proved that the proposed methodology has the capability of separating classes properly in varying image conditions.

Grocery Customer Behavior Analysis using RFID-based Shopping Paths Data

Knowing about the customer behavior in a grocery has been a long-standing issue in the retailing industry. The advent of RFID has made it easier to collect moving data for an individual shopper's behavior. Most of the previous studies used the traditional statistical clustering technique to find the major characteristics of customer behavior, especially shopping path. However, in using the clustering technique, due to various spatial constraints in the store, standard clustering methods are not feasible because moving data such as the shopping path should be adjusted in advance of the analysis, which is time-consuming and causes data distortion. To alleviate this problem, we propose a new approach to spatial pattern clustering based on the longest common subsequence. Experimental results using real data obtained from a grocery confirm the good performance of the proposed method in finding the hot spot, dead spot and major path patterns of customer movements.

Effect of Prandtl Number on Natural Convection Heat Transfer from a Heated Semi-Circular Cylinder

Natural convection heat transfer from a heated horizontal semi-circular cylinder (flat surface upward) has been investigated for the following ranges of conditions; Grashof number, and Prandtl number. The governing partial differential equations (continuity, Navier-Stokes and energy equations) have been solved numerically using a finite volume formulation. In addition, the role of the type of the thermal boundary condition imposed at cylinder surface, namely, constant wall temperature (CWT) and constant heat flux (CHF) are explored. Natural convection heat transfer from a heated horizontal semi-circular cylinder (flat surface upward) has been investigated for the following ranges of conditions; Grashof number, and Prandtl number, . The governing partial differential equations (continuity, Navier-Stokes and energy equations) have been solved numerically using a finite volume formulation. In addition, the role of the type of the thermal boundary condition imposed at cylinder surface, namely, constant wall temperature (CWT) and constant heat flux (CHF) are explored. The resulting flow and temperature fields are visualized in terms of the streamline and isotherm patterns in the proximity of the cylinder. The flow remains attached to the cylinder surface over the range of conditions spanned here except that for and ; at these conditions, a separated flow region is observed when the condition of the constant wall temperature is prescribed on the surface of the cylinder. The heat transfer characteristics are analyzed in terms of the local and average Nusselt numbers. The maximum value of the local Nusselt number always occurs at the corner points whereas it is found to be minimum at the rear stagnation point on the flat surface. Overall, the average Nusselt number increases with Grashof number and/ or Prandtl number in accordance with the scaling considerations. The numerical results are used to develop simple correlations as functions of Grashof and Prandtl number thereby enabling the interpolation of the present numerical results for the intermediate values of the Prandtl or Grashof numbers for both thermal boundary conditions.

Spatio-Temporal Patterns and Dynamics in Motion of Pathogenic Spirochetes: Implications toward Virulence and Treatment of Leptospirosis

We apply a particle tracking technique to track the motion of individual pathogenic Leptospira. We observe and capture images of motile Leptospira by means of CCD and darkfield microscope. Image processing, statistical theories and simulations are used for data analysis. Based on trajectory patterns, mean square displacement, and power spectral density characteristics, we found that the motion modes are most likely to be directed motion mode (70%) and the rest are either normal diffusion or unidentified mode. Our findings may support the fact that why leptospires are very well efficient toward targeting internal tissues as a result of increase in virulence factor.

A Study of Water Consumption in Two Malaysian Resorts

In the effort to reduce water consumption for resorts, more water conservation practices need to be implemented. Hence water audits need to be performed to obtain a baseline of water consumption, before planning water conservation practices. In this study, a water audit framework specifically for resorts was created, and the audit was performed on two resorts: Resort A in Langkawi, Malaysia; and Resort B in Miri, Malaysia. From the audit, the total daily water consumption for Resorts A and B were estimated to be 180m3 and 330 m3 respectively, while the actual water consumption (based on water meter readings) were 175 m3 and 325 m3. This suggests that the audit framework is reasonably accurate and may be used to account for most of the water consumption sources in a resort. The daily water consumption per guest is about 500 litres. The water consumption of both resorts is poorly rated compared with established benchmarks. Water conservation measures were suggested for both resorts.

Optimizing of Fuzzy C-Means Clustering Algorithm Using GA

Fuzzy C-means Clustering algorithm (FCM) is a method that is frequently used in pattern recognition. It has the advantage of giving good modeling results in many cases, although, it is not capable of specifying the number of clusters by itself. In FCM algorithm most researchers fix weighting exponent (m) to a conventional value of 2 which might not be the appropriate for all applications. Consequently, the main objective of this paper is to use the subtractive clustering algorithm to provide the optimal number of clusters needed by FCM algorithm by optimizing the parameters of the subtractive clustering algorithm by an iterative search approach and then to find an optimal weighting exponent (m) for the FCM algorithm. In order to get an optimal number of clusters, the iterative search approach is used to find the optimal single-output Sugenotype Fuzzy Inference System (FIS) model by optimizing the parameters of the subtractive clustering algorithm that give minimum least square error between the actual data and the Sugeno fuzzy model. Once the number of clusters is optimized, then two approaches are proposed to optimize the weighting exponent (m) in the FCM algorithm, namely, the iterative search approach and the genetic algorithms. The above mentioned approach is tested on the generated data from the original function and optimal fuzzy models are obtained with minimum error between the real data and the obtained fuzzy models.

A New Biologically Inspired Pattern Recognition Spproach for Face Recognition

This paper reports a new pattern recognition approach for face recognition. The biological model of light receptors - cones and rods in human eyes and the way they are associated with pattern vision in human vision forms the basis of this approach. The functional model is simulated using CWD and WPD. The paper also discusses the experiments performed for face recognition using the features extracted from images in the AT & T face database. Artificial Neural Network and k- Nearest Neighbour classifier algorithms are employed for the recognition purpose. A feature vector is formed for each of the face images in the database and recognition accuracies are computed and compared using the classifiers. Simulation results show that the proposed method outperforms traditional way of feature extraction methods prevailing for pattern recognition in terms of recognition accuracy for face images with pose and illumination variations.

Predictive Clustering Hybrid Regression(pCHR) Approach and Its Application to Sucrose-Based Biohydrogen Production

A predictive clustering hybrid regression (pCHR) approach was developed and evaluated using dataset from H2- producing sucrose-based bioreactor operated for 15 months. The aim was to model and predict the H2-production rate using information available about envirome and metabolome of the bioprocess. Selforganizing maps (SOM) and Sammon map were used to visualize the dataset and to identify main metabolic patterns and clusters in bioprocess data. Three metabolic clusters: acetate coupled with other metabolites, butyrate only, and transition phases were detected. The developed pCHR model combines principles of k-means clustering, kNN classification and regression techniques. The model performed well in modeling and predicting the H2-production rate with mean square error values of 0.0014 and 0.0032, respectively.

Eye Gesture Analysis with Head Movement for Advanced Driver Assistance Systems

Road traffic accidents are a major cause of death worldwide. In an attempt to reduce accidents, some research efforts have focused on creating Advanced Driver Assistance Systems (ADAS) able to detect vehicle, driver and environmental conditions and to use this information to identify cues for potential accidents. This paper presents continued work on a novel Non-intrusive Intelligent Driver Assistance and Safety System (Ni-DASS) for assessing driver point of regard within vehicles. It uses an on-board CCD camera to observe the driver-s face. A template matching approach is used to compare the driver-s eye-gaze pattern with a set of eye-gesture templates of the driver looking at different focal points within the vehicle. The windscreen is divided into cells and comparison of the driver-s eye-gaze pattern with templates of a driver-s eyes looking at each cell is used to determine the driver-s point of regard on the windscreen. Results indicate that the proposed technique could be useful in situations where low resolution estimates of driver point of regard are adequate. For instance, To allow ADAS systems to alert the driver if he/she has positively failed to observe a hazard.

Adaptive Digital Watermarking Integrating Fuzzy Inference HVS Perceptual Model

An adaptive Fuzzy Inference Perceptual model has been proposed for watermarking of digital images. The model depends on the human visual characteristics of image sub-regions in the frequency multi-resolution wavelet domain. In the proposed model, a multi-variable fuzzy based architecture has been designed to produce a perceptual membership degree for both candidate embedding sub-regions and strength watermark embedding factor. Different sizes of benchmark images with different sizes of watermarks have been applied on the model. Several experimental attacks have been applied such as JPEG compression, noises and rotation, to ensure the robustness of the scheme. In addition, the model has been compared with different watermarking schemes. The proposed model showed its robustness to attacks and at the same time achieved a high level of imperceptibility.

Face Recognition using a Kernelization of Graph Embedding

Linearization of graph embedding has been emerged as an effective dimensionality reduction technique in pattern recognition. However, it may not be optimal for nonlinearly distributed real world data, such as face, due to its linear nature. So, a kernelization of graph embedding is proposed as a dimensionality reduction technique in face recognition. In order to further boost the recognition capability of the proposed technique, the Fisher-s criterion is opted in the objective function for better data discrimination. The proposed technique is able to characterize the underlying intra-class structure as well as the inter-class separability. Experimental results on FRGC database validate the effectiveness of the proposed technique as a feature descriptor.

Automatic Building an Extensive Arabic FA Terms Dictionary

Field Association (FA) terms are a limited set of discriminating terms that give us the knowledge to identify document fields which are effective in document classification, similar file retrieval and passage retrieval. But the problem lies in the lack of an effective method to extract automatically relevant Arabic FA Terms to build a comprehensive dictionary. Moreover, all previous studies are based on FA terms in English and Japanese, and the extension of FA terms to other language such Arabic could be definitely strengthen further researches. This paper presents a new method to extract, Arabic FA Terms from domain-specific corpora using part-of-speech (POS) pattern rules and corpora comparison. Experimental evaluation is carried out for 14 different fields using 251 MB of domain-specific corpora obtained from Arabic Wikipedia dumps and Alhyah news selected average of 2,825 FA Terms (single and compound) per field. From the experimental results, recall and precision are 84% and 79% respectively. Therefore, this method selects higher number of relevant Arabic FA Terms at high precision and recall.

Some Separations in Covering Approximation Spaces

Adopting Zakowski-s upper approximation operator C and lower approximation operator C, this paper investigates granularity-wise separations in covering approximation spaces. Some characterizations of granularity-wise separations are obtained by means of Pawlak rough sets and some relations among granularitywise separations are established, which makes it possible to research covering approximation spaces by logical methods and mathematical methods in computer science. Results of this paper give further applications of Pawlak rough set theory in pattern recognition and artificial intelligence.

A Dynamic RGB Intensity Based Steganography Scheme

Steganography meaning covered writing. Steganography includes the concealment of information within computer files [1]. In other words, it is the Secret communication by hiding the existence of message. In this paper, we will refer to cover image, to indicate the images that do not yet contain a secret message, while we will refer to stego images, to indicate an image with an embedded secret message. Moreover, we will refer to the secret message as stego-message or hidden message. In this paper, we proposed a technique called RGB intensity based steganography model as RGB model is the technique used in this field to hide the data. The methods used here are based on the manipulation of the least significant bits of pixel values [3][4] or the rearrangement of colors to create least significant bit or parity bit patterns, which correspond to the message being hidden. The proposed technique attempts to overcome the problem of the sequential fashion and the use of stego-key to select the pixels.

An EOQ Model for Non-Instantaneous Deteriorating Items with Power Demand, Time Dependent Holding Cost, Partial Backlogging and Permissible Delay in Payments

In this paper, Economic Order Quantity (EOQ) based model for non-instantaneous Weibull distribution deteriorating items with power demand pattern is presented. In this model, the holding cost per unit of the item per unit time is assumed to be an increasing linear function of time spent in storage. Here the retailer is allowed a trade-credit offer by the supplier to buy more items. Also in this model, shortages are allowed and partially backlogged. The backlogging rate is dependent on the waiting time for the next replenishment. This model aids in minimizing the total inventory cost by finding the optimal time interval and finding the optimal order quantity. The optimal solution of the model is illustrated with the help of numerical examples. Finally sensitivity analysis and graphical representations are given to demonstrate the model.

A Study of Visual Attention in Diagnosing Cerebellar Tumours

Visual attention allows user to select the most relevant information to ongoing behaviour. This paper presents a study on; i) the performance of people measurements, ii) accurateness of people measurement of the peaks that correspond to chemical quantities from the Magnetic Resonance Spectroscopy (MRS) graphs and iii) affects of people measurements to the algorithm-based diagnosis. Participant-s eye-movement was recorded using eye-tracker tool (Eyelink II). This experiment involves three participants for examining 20 MRS graphs to estimate the peaks of chemical quantities which indicate the abnormalities associated with Cerebellar Tumours (CT). The status of each MRS is verified by using decision algorithm. Analysis involves determination of humans-s eye movement pattern in measuring the peak of spectrograms, scan path and determining the relationship of distributions of fixation durations with the accuracy of measurement. In particular, the eye-tracking data revealed which aspects of the spectrogram received more visual attention and in what order they were viewed. This preliminary investigation provides a proof of concept for use of the eye tracking technology as the basis for expanded CT diagnosis.

Standardization and Adaption Requirements in Production System Transplants

As German companies roll out their standardized production systems to offshore manufacturing plants, they face the challenge of implementing them in different cultural environments. Studies show that the local adaptation is one of the key factors for a successful implementation. Thus the question arises of where the line between standardization and adaptation can be drawn. To answer this question the influence of culture on production systems is analysed in this paper. The culturally contingent components of production systems are identified. Also the contingency factors are classified according to their impact on the necessary adaptation changes and implementation effort. Culturally specific decision making, coordination, communication and motivation patterns require one-time changes in organizational and process design. The attitude towards rules requires more intense coaching and controlling. Lastly a framework is developed to depict standardization and adaption needs when transplanting production systems into different cultural environments.

MTSSM - A Framework for Multi-Track Segmentation of Symbolic Music

Music segmentation is a key issue in music information retrieval (MIR) as it provides an insight into the internal structure of a composition. Structural information about a composition can improve several tasks related to MIR such as searching and browsing large music collections, visualizing musical structure, lyric alignment, and music summarization. The authors of this paper present the MTSSM framework, a twolayer framework for the multi-track segmentation of symbolic music. The strength of this framework lies in the combination of existing methods for local track segmentation and the application of global structure information spanning via multiple tracks. The first layer of the MTSSM uses various string matching techniques to detect the best candidate segmentations for each track of a multi-track composition independently. The second layer combines all single track results and determines the best segmentation for each track in respect to the global structure of the composition.

Model Discovery and Validation for the Qsar Problem using Association Rule Mining

There are several approaches in trying to solve the Quantitative 1Structure-Activity Relationship (QSAR) problem. These approaches are based either on statistical methods or on predictive data mining. Among the statistical methods, one should consider regression analysis, pattern recognition (such as cluster analysis, factor analysis and principal components analysis) or partial least squares. Predictive data mining techniques use either neural networks, or genetic programming, or neuro-fuzzy knowledge. These approaches have a low explanatory capability or non at all. This paper attempts to establish a new approach in solving QSAR problems using descriptive data mining. This way, the relationship between the chemical properties and the activity of a substance would be comprehensibly modeled.