Real-time Laser Monitoring based on Pipe Detective Operation

The pipe inspection operation is the difficult detective performance. Almost applications are mainly relies on a manual recognition of defective areas that have carried out detection by an engineer. Therefore, an automation process task becomes a necessary in order to avoid the cost incurred in such a manual process. An automated monitoring method to obtain a complete picture of the sewer condition is proposed in this work. The focus of the research is the automated identification and classification of discontinuities in the internal surface of the pipe. The methodology consists of several processing stages including image segmentation into the potential defect regions and geometrical characteristic features. Automatic recognition and classification of pipe defects are carried out by means of using an artificial neural network technique (ANN) based on Radial Basic Function (RBF). Experiments in a realistic environment have been conducted and results are presented.

Soccer Video Edition Using a Multimodal Annotation

In this paper, we present an approach for soccer video edition using a multimodal annotation. We propose to associate with each video sequence of a soccer match a textual document to be used for further exploitation like search, browsing and abstract edition. The textual document contains video meta data, match meta data, and match data. This document, generated automatically while the video is analyzed, segmented and classified, can be enriched semi automatically according to the user type and/or a specialized recommendation system.

Multivariate School Travel Demand Regression Based on Trip Attraction

Since primary school trips usually start from home, attention by many scholars have been focused on the home end for data gathering. Thereafter category analysis has often been relied upon when predicting school travel demands. In this paper, school end was relied on for data gathering and multivariate regression for future travel demand prediction. 9859 pupils were surveyed by way of questionnaires at 21 primary schools. The town was divided into 5 zones. The study was carried out in Skudai Town, Malaysia. Based on the hypothesis that the number of primary school trip ends are expected to be the same because school trips are fixed, the choice of trip end would have inconsequential effect on the outcome. The study compared empirical data for home and school trip end productions and attractions. Variance from both data results was insignificant, although some claims from home based family survey were found to be grossly exaggerated. Data from the school trip ends was relied on for travel demand prediction because of its completeness. Accessibility, trip attraction and trip production were then related to school trip rates under daylight and dry weather conditions. The paper concluded that, accessibility is an important parameter when predicting demand for future school trip rates.

Feature Reduction of Nearest Neighbor Classifiers using Genetic Algorithm

The design of a pattern classifier includes an attempt to select, among a set of possible features, a minimum subset of weakly correlated features that better discriminate the pattern classes. This is usually a difficult task in practice, normally requiring the application of heuristic knowledge about the specific problem domain. The selection and quality of the features representing each pattern have a considerable bearing on the success of subsequent pattern classification. Feature extraction is the process of deriving new features from the original features in order to reduce the cost of feature measurement, increase classifier efficiency, and allow higher classification accuracy. Many current feature extraction techniques involve linear transformations of the original pattern vectors to new vectors of lower dimensionality. While this is useful for data visualization and increasing classification efficiency, it does not necessarily reduce the number of features that must be measured since each new feature may be a linear combination of all of the features in the original pattern vector. In this paper a new approach is presented to feature extraction in which feature selection, feature extraction, and classifier training are performed simultaneously using a genetic algorithm. In this approach each feature value is first normalized by a linear equation, then scaled by the associated weight prior to training, testing, and classification. A knn classifier is used to evaluate each set of feature weights. The genetic algorithm optimizes a vector of feature weights, which are used to scale the individual features in the original pattern vectors in either a linear or a nonlinear fashion. By this approach, the number of features used in classifying can be finely reduced.

Statistics over Lyapunov Exponents for Feature Extraction: Electroencephalographic Changes Detection Case

A new approach based on the consideration that electroencephalogram (EEG) signals are chaotic signals was presented for automated diagnosis of electroencephalographic changes. This consideration was tested successfully using the nonlinear dynamics tools, like the computation of Lyapunov exponents. This paper presented the usage of statistics over the set of the Lyapunov exponents in order to reduce the dimensionality of the extracted feature vectors. Since classification is more accurate when the pattern is simplified through representation by important features, feature extraction and selection play an important role in classifying systems such as neural networks. Multilayer perceptron neural network (MLPNN) architectures were formulated and used as basis for detection of electroencephalographic changes. Three types of EEG signals (EEG signals recorded from healthy volunteers with eyes open, epilepsy patients in the epileptogenic zone during a seizure-free interval, and epilepsy patients during epileptic seizures) were classified. The selected Lyapunov exponents of the EEG signals were used as inputs of the MLPNN trained with Levenberg- Marquardt algorithm. The classification results confirmed that the proposed MLPNN has potential in detecting the electroencephalographic changes.

Exponential Stability Analysis for Switched Cellular Neural Networks with Time-varying Delays and Impulsive Effects

In this Letter, a class of impulsive switched cellular neural networks with time-varying delays is investigated. At the same time, parametric uncertainties assumed to be norm bounded are considered. By dividing the network state variables into subgroups according to the characters of the neural networks, some sufficient conditions guaranteeing exponential stability for all admissible parametric uncertainties are derived via constructing appropriate Lyapunov functional. One numerical example is provided to illustrate the validity of the main results obtained in this paper.

Automatic Fingerprint Classification Using Graph Theory

Using efficient classification methods is necessary for automatic fingerprint recognition system. This paper introduces a new structural approach to fingerprint classification by using the directional image of fingerprints to increase the number of subclasses. In this method, the directional image of fingerprints is segmented into regions consisting of pixels with the same direction. Afterwards the relational graph to the segmented image is constructed and according to it, the super graph including prominent information of this graph is formed. Ultimately we apply a matching technique to compare obtained graph with the model graphs in order to classify fingerprints by using cost function. Increasing the number of subclasses with acceptable accuracy in classification and faster processing in fingerprints recognition, makes this system superior.

Implementation of a New Neural Network Function Block to Programmable Logic Controllers Library Function

Programmable logic controllers are the main controllers in the today's industries; they are used for several applications in industrial control systems and there are lots of examples exist from the PLC applications in industries especially in big companies and plants such as refineries, power plants, petrochemical companies, steel companies, and food and production companies. In the PLCs there are some functions in the function library in software that can be used in PLC programs as basic program elements. The aim of this project are introducing and implementing a new function block of a neural network to the function library of PLC. This block can be applied for some control applications or nonlinear functions calculations after it has been trained for these applications. The implemented neural network is a Perceptron neural network with three layers, three input nodes and one output node. The block can be used in manual or automatic mode. In this paper the structure of the implemented function block, the parameters and the training method of the network are presented by considering the especial method of PLC programming and its complexities. Finally the application of the new block is compared with a classic simulated block and the results are presented.

Multi-Stakeholder Road Pricing Game: Solution Concepts

A road pricing game is a game where various stakeholders and/or regions with different (and usually conflicting) objectives compete for toll setting in a given transportation network to satisfy their individual objectives. We investigate some classical game theoretical solution concepts for the road pricing game. We establish results for the road pricing game so that stakeholders and/or regions playing such a game will beforehand know what is obtainable. This will save time and argument, and above all, get rid of the feelings of unfairness among the competing actors and road users. Among the classical solution concepts we investigate is Nash equilibrium. In particular, we show that no pure Nash equilibrium exists among the actors, and further illustrate that even “mixed Nash equilibrium" may not be achievable in the road pricing game. The paper also demonstrates the type of coalitions that are not only reachable, but also stable and profitable for the actors involved.

Cognitive Behaviour Therapy to Treat Social Anxiety Disorder: A Psychology Case

Rational Emotive Behaviour Therapy is the first cognitive behavior therapy which was introduced by Albert Ellis. This is a systematic and structured psychotherapy which is effective in treating various psychological problems. A patient, 25 years old male, experienced intense fear and situational panic attack to return to his faculty and to face his class-mates after a long absence (2 years). This social anxiety disorder was a major factor that impeded the progress of his study. He was treated with the use of behavioural technique such as relaxation breathing technique and cognitive techniques such as imagery, cognitive restructuring, rationalization technique and systematic desensitization. The patient reported positive improvement in the anxiety disorder, able to progress well in studies and lead a better quality of life as a student.

Performance Analysis of Wireless Ad-Hoc Network Based on EDCA IEEE802.11e

IEEE 802.11e is the enhanced version of the IEEE 802.11 MAC dedicated to provide Quality of Service of wireless network. It supports QoS by the service differentiation and prioritization mechanism. Data traffic receives different priority based on QoS requirements. Fundamentally, applications are divided into four Access Categories (AC). Each AC has its own buffer queue and behaves as an independent backoff entity. Every frame with a specific priority of data traffic is assigned to one of these access categories. IEEE 802.11e EDCA (Enhanced Distributed Channel Access) is designed to enhance the IEEE 802.11 DCF (Distributed Coordination Function) mechanisms by providing a distributed access method that can support service differentiation among different classes of traffic. Performance of IEEE 802.11e MAC layer with different ACs is evaluated to understand the actual benefits deriving from the MAC enhancements.

Banach Lattices with Weak Dunford-Pettis Property

We introduce and study the class of weak almost Dunford-Pettis operators. As an application, we characterize Banach lattices with the weak Dunford-Pettis property. Also, we establish some sufficient conditions for which each weak almost Dunford-Pettis operator is weak Dunford-Pettis. Finally, we derive some interesting results.

Characterization of Corn Cobs from Microwave and Potassium Hydroxide Pretreatment

The complexity of lignocellulosic biomass requires a pretreatment step to improve the yield of fermentable sugars. The efficient pretreatment of corn cobs using microwave and potassium hydroxide and enzymatic hydrolysis was investigated. The objective of this work was to characterize the optimal condition of pretreatment of corn cobs using microwave and potassium hydroxide enhance enzymatic hydrolysis. Corn cobs were submerged in different potassium hydroxide concentration at varies temperature and resident time. The pretreated corn cobs were hydrolyzed to produce the reducing sugar for analysis. The morphology and microstructure of samples were investigated by Thermal gravimetric analysis (TGA, scanning electron microscope (SEM), X-ray diffraction (XRD). The results showed that lignin and hemicellulose were removed by microwave/potassium hydroxide pretreatment. The crystallinity of the pretreated corn cobs was higher than the untreated. This method was compared with autoclave and conventional heating method. The results indicated that microwave-alkali treatment was an efficient way to improve the enzymatic hydrolysis rate by increasing its accessibility hydrolysis enzymes.

Bottom Up Text Mining through Hierarchical Document Representation

Most of the existing text mining approaches are proposed, keeping in mind, transaction databases model. Thus, the mined dataset is structured using just one concept: the “transaction", whereas the whole dataset is modeled using the “set" abstract type. In such cases, the structure of the whole dataset and the relationships among the transactions themselves are not modeled and consequently, not considered in the mining process. We believe that taking into account structure properties of hierarchically structured information (e.g. textual document, etc ...) in the mining process, can leads to best results. For this purpose, an hierarchical associations rule mining approach for textual documents is proposed in this paper and the classical set-oriented mining approach is reconsidered profits to a Direct Acyclic Graph (DAG) oriented approach. Natural languages processing techniques are used in order to obtain the DAG structure. Based on this graph model, an hierarchical bottom up algorithm is proposed. The main idea is that each node is mined with its parent node.

BIP-Based Alarm Declaration and Clearing in SONET Networks Employing Automatic Protection Switching

The paper examines the performance of bit-interleaved parity (BIP) methods in error rate monitoring, and in declaration and clearing of alarms in those transport networks that employ automatic protection switching (APS). The BIP-based error rate monitoring is attractive for its simplicity and ease of implementation. The BIP-based results are compared with exact results and are found to declare the alarms too late, and to clear the alarms too early. It is concluded that the standards development and systems implementation should take into account the fact of early clearing and late declaration of alarms. The window parameters defining the detection and clearing thresholds should be set so as to build sufficient hysteresis into the system to ensure that BIP-based implementations yield acceptable performance results.

Mining Frequent Patterns with Functional Programming

Frequent patterns are patterns such as sets of features or items that appear in data frequently. Finding such frequent patterns has become an important data mining task because it reveals associations, correlations, and many other interesting relationships hidden in a dataset. Most of the proposed frequent pattern mining algorithms have been implemented with imperative programming languages such as C, Cµ, Java. The imperative paradigm is significantly inefficient when itemset is large and the frequent pattern is long. We suggest a high-level declarative style of programming using a functional language. Our supposition is that the problem of frequent pattern discovery can be efficiently and concisely implemented via a functional paradigm since pattern matching is a fundamental feature supported by most functional languages. Our frequent pattern mining implementation using the Haskell language confirms our hypothesis about conciseness of the program. The performance studies on speed and memory usage support our intuition on efficiency of functional language.

School Homework and its Relationship with Student Academic Achievement in Malaysia

School homework has been synonymous with students- life in Chinese national type primary schools in Malaysia. Although many reports in the press claimed that students were burdened with too much of it, homework continues to be a common practice in national type schools that is believed to contribute to academic achievement. This study is conducted to identify the relationship between the burden of school homework and academic achievement among pupils in Chinese National Type Primary School in the state of Perak, Malaysia. A total of 284 students (142 from urban and 142 from rural) respectively were chosen as participants in this study. Variables of gender and location (urban/rural areas) has shown significant difference in student academic achievement. Female Chinese student from rural areas showed a higher mean score than males from urban area. Therefore, the Chinese language teachers should give appropriate and relevant homework to primary school students to achieve good academic performance.

Contemporary Housing Indicators in Poland on the Wroclaw Study Case

The paper presents the results of research on trends in shaping of multifamily buildings in Poland on the example of Wrocław, after Polish accession to the European Union. The study is conducted within the research project: “Trends in creating of multifamily housing development since 2004, on the Wrocław study case" supported by Polish Ministry of Science and Higher Education and will be completed in November 2011. The research involves multifamily buildings completed in the last decade, in term of fundamental urbanization factors such as: building-s coefficient area, useable area, green area (biologically active surface), intensity of building development, amount of dwellings, dwelling area, amount of parking places, numbers of floors, etc. The analysis of these indicators was conducted based on the date obtained in the study of approximately one hundred new housing units, completed in Wroclaw. The analysis attempts to formulate the main trends in creating of housing policy in Poland during the last 10 years in reference to local urban policy.

Drum-Buffer-Rope: The Technique to Plan and Control the Production Using Theory of Constraints

Theory of Constraints has been emerging as an important tool for optimization of manufacturing/service systems. Goldratt in his first book “ The Goal " gave the introduction on Theory of Constraints and its applications in a factory scenario. A large number of production managers around the globe read this book but only a few could implement it in their plants because the book did not explain the steps to implement TOC in the factory. To overcome these limitations, Goldratt wrote this book to explain TOC, DBR and the method to implement it. In this paper, an attempt has been made to summarize the salient features of TOC and DBR listed in the book and the correct approach to implement TOC in a factory setting. The simulator available along with the book was actually used by the authors and the claim of Goldratt regarding the use of DBR and Buffer management to ease the work of production managers was tested and was found to be correct.

Fractal Dimension: An Index to Quantify Parameters in Genetic Algorithms

Genetic Algorithms (GAs) are direct searching methods which require little information from design space. This characteristic beside robustness of these algorithms makes them to be very popular in recent decades. On the other hand, while this method is employed, there is no guarantee to achieve optimum results. This obliged designer to run such algorithms more than one time to achieve more reliable results. There are many attempts to modify the algorithms to make them more efficient. In this paper, by application of fractal dimension (particularly, Box Counting Method), the complexity of design space are established for determination of mutation and crossover probabilities (Pm and Pc). This methodology is followed by a numerical example for more clarification. It is concluded that this modification will improve efficiency of GAs and make them to bring about more reliable results especially for design space with higher fractal dimensions.