Comparison of S-transform and Wavelet Transform in Power Quality Analysis

In the power quality analysis non-stationary nature of voltage distortions require some precise and powerful analytical techniques. The time-frequency representation (TFR) provides a powerful method for identification of the non-stationary of the signals. This paper investigates a comparative study on two techniques for analysis and visualization of voltage distortions with time-varying amplitudes. The techniques include the Discrete Wavelet Transform (DWT), and the S-Transform. Several power quality problems are analyzed using both the discrete wavelet transform and S–transform, showing clearly the advantage of the S– transform in detecting, localizing, and classifying the power quality problems.

An Experimental Investigation on the Behavior of Pressure Tube under Symmetrical and Asymmetrical Heating Conditions in an Indian PHWR

Thermal behavior of fuel channel under loss of coolant accident (LOCA) is a major concern for nuclear reactor safety. LOCA along with failure of emergency cooling water system (ECC) may leads to mechanical deformations like sagging and ballooning. In order to understand the phenomenon an experiment has been carried out using 19 pin fuel element simulator. Main purpose of the experiment was to trace temperature profiles over the pressure tube, calandria tube and clad tubes of Indian Pressurized Heavy Water Reactor (IPHWR) under symmetrical and asymmetrical heat-up conditions. For simulating the fully voided scenario, symmetrical heating of pressure was carried out by injecting 13.2 KW (2 % of nominal power) to all the 19 pins and the temperatures of pressure tube, calandria tube and clad tubes were measured. During symmetrical heating the sagging of fuel channel was initiated at 460 °C and the highest temperature attained by PT was 650 °C . The decay heat from clad tubes was dissipated to moderator mainly by radiation and natural convection. The highest temperature of 680 °C was observed over the outer ring of clad tubes of fuel simulator. Again, to simulate partially voided condition, asymmetrical heating of pressure was carried out by supplying 8.0 kW power to upper 8 pins of fuel simulator and temperature profiles were measured. Along the circumference of pressure tube (PT) the highest temperature difference of 320 °C was observed, which highlights the magnitude of thermal stresses under partially voided conditions.

Kinematic Gait Analysis of Upper and Lower Limbs Joints in Hemiplegic Children

Children with hemiplgic cerebral palsy often walk with diminished reciprocal arm swing so the purpose of this study was to describe kinematic characteristics in children with hemiplegic cerebral palsy (CP) during the gait suphases, and find if there is a correlation between upper(shoulder and elbow) and lower(hip, knee, and ankle) limb joints either in involved or uninvolved.48 children with hemiplegic cerebral palsy (18boys, 30girls) with an average age of (5.1±0.87) years were selected randomly to evaluate joint angles during gait by 3D motion analysis system with 6 pro reflex cameras in a sagittal plane for both sides of the body. The results showed increased shoulder and elbow flexion, increased hip angular displacement, decreased knee and ankle arcs during gait cycle, also there is correlation between shoulder and elbow to hip, knee, and ankle joints during various subphases of gait.

Methods of Forming Informational Culture Students

Along with the basic features of students\' culture information, with its widely usage oriented on implementation of the new information technologies in educational process that determines the search for ways of pointing to the similarity of interdisciplinary connections content, aims and objectives of the study. In this regard, the article questions about students\' information culture, and also presented information about the aims and objectives of the information culture process among students. In the formation of a professional interest in relevant information, which is an opportunity to assist in informing the professional activities of the essence of effective use of interactive methods and innovative technologies in the learning process. The result of the experiment proves the effectiveness of the information culture process of students in training the system of higher education based on the credit technology. The main purpose of this paper is a comprehensive review of students\' information culture.

Texture Based Weed Detection Using Multi Resolution Combined Statistical and Spatial Frequency (MRCSF)

Texture classification is a trendy and a catchy technology in the field of texture analysis. Textures, the repeated patterns, have different frequency components along different orientations. Our work is based on Texture Classification and its applications. It finds its applications in various fields like Medical Image Classification, Computer Vision, Remote Sensing, Agricultural Field, and Textile Industry. Weed control has a major effect on agriculture. A large amount of herbicide has been used for controlling weeds in agriculture fields, lawns, golf courses, sport fields, etc. Random spraying of herbicides does not meet the exact requirement of the field. Certain areas in field have more weed patches than estimated. So, we need a visual system that can discriminate weeds from the field image which will reduce or even eliminate the amount of herbicide used. This would allow farmers to not use any herbicides or only apply them where they are needed. A machine vision precision automated weed control system could reduce the usage of chemicals in crop fields. In this paper, an intelligent system for automatic weeding strategy Multi Resolution Combined Statistical & spatial Frequency is used to discriminate the weeds from the crops and to classify them as narrow, little and broad weeds.

Simulation Study for Performance Comparison of Routing Protocols in Mobile Adhoc Network

Due to insufficient frequency band and tremendous growth of the mobile users, complex computation is needed for the use of resources. Long distance communication began with the introduction of telegraphs and simple coded pulses, which were used to transmit short messages. Since then numerous advances have rendered reliable transfer of information both easier and quicker. Wireless network refers to any type of computer network that is wireless, and is commonly associated with a telecommunications network whose interconnections between nodes is implemented without the use of wires. Wireless network can be broadly categorized in infrastructure network and infrastructure less network. Infrastructure network is one in which we have a base station to serve the mobile users and in the infrastructure less network is one in which no infrastructure is available to serve the mobile users this kind of networks are also known as mobile Adhoc networks. In this paper we have simulated the result for different scenarios with protocols like AODV and DSR; we simulated the result for throughput, delay and receiving traffic in the given scenario.

Dynamic-Stochastic Influence Diagrams: Integrating Time-Slices IDs and Discrete Event Systems Modeling

The Influence Diagrams (IDs) is a kind of Probabilistic Belief Networks for graphic modeling. The usage of IDs can improve the communication among field experts, modelers, and decision makers, by showing the issue frame discussed from a high-level point of view. This paper enhances the Time-Sliced Influence Diagrams (TSIDs, or called Dynamic IDs) based formalism from a Discrete Event Systems Modeling and Simulation (DES M&S) perspective, for Exploring Analysis (EA) modeling. The enhancements enable a modeler to specify times occurred of endogenous events dynamically with stochastic sampling as model running and to describe the inter- influences among them with variable nodes in a dynamic situation that the existing TSIDs fails to capture. The new class of model is named Dynamic-Stochastic Influence Diagrams (DSIDs). The paper includes a description of the modeling formalism and the hiberarchy simulators implementing its simulation algorithm, and shows a case study to illustrate its enhancements.

Application of Nano-Zero Valent Iron for Treating Metolachlor in Aqueous Solution

Water, soil and sediment contaminated with metolachlor poses a threat to the environment and human health. We determined the effectiveness of nano-zerovalent iron (NZVI) to dechlorinate metolachlor [2-chloro-n-(2-ethyl-6-methyl-phenyl)-n- (1-methoxypropan-2-yl)acetamide] in pH solution and the presence of aluminium salt. The optimum dosage of degradation of 100 mlL-1 metolachlor was 1% (w/v) NZVI. The degradation kinetic rate (kobs) was 0.218×10-3 min-1 and specific first-order rates (kSA) was 8.72×10-7 L m-2min-1. By treating aqueous solutions of metolachlor with NZVI, metolachlor destruction rate were increased as the pH decrease from 10 to 4. Lowering solution pH removes Fe (III) passivating layers from the NZVI and makes it free for reductive transformations. Destruction kinetic rates were 20.8×10-3 min-1 for pH4, 18.9×10-3 min-1 for pH7, 13.8×10-3 min-1 for pH10. In addition, destruction kinetic of metolachlor by NZVI was enhanced when aluminium sulfate was added. The destruction kinetic rate were 20.4×10-3 min-1 for 0.05% Al(SO4)3 and 60×10-3 min-1 for 0.1% Al(SO4)3.

On-line Testing of Software Components for Diagnosis of Embedded Systems

This paper studies the dependability of componentbased applications, especially embedded ones, from the diagnosis point of view. The principle of the diagnosis technique is to implement inter-component tests in order to detect and locate the faulty components without redundancy. The proposed approach for diagnosing faulty components consists of two main aspects. The first one concerns the execution of the inter-component tests which requires integrating test functionality within a component. This is the subject of this paper. The second one is the diagnosis process itself which consists of the analysis of inter-component test results to determine the fault-state of the whole system. Advantage of this diagnosis method when compared to classical redundancy faulttolerant techniques are application autonomy, cost-effectiveness and better usage of system resources. Such advantage is very important for many systems and especially for embedded ones.

A Fast Sensor Relocation Algorithm in Wireless Sensor Networks

Sensor relocation is to repair coverage holes caused by node failures. One way to repair coverage holes is to find redundant nodes to replace faulty nodes. Most researches took a long time to find redundant nodes since they randomly scattered redundant nodes around the sensing field. To record the precise position of sensor nodes, most researches assumed that GPS was installed in sensor nodes. However, high costs and power-consumptions of GPS are heavy burdens for sensor nodes. Thus, we propose a fast sensor relocation algorithm to arrange redundant nodes to form redundant walls without GPS. Redundant walls are constructed in the position where the average distance to each sensor node is the shortest. Redundant walls can guide sensor nodes to find redundant nodes in the minimum time. Simulation results show that our algorithm can find the proper redundant node in the minimum time and reduce the relocation time with low message complexity.

Analysis of Message Authentication in Turbo Coded Halftoned Images using Exit Charts

Considering payload, reliability, security and operational lifetime as major constraints in transmission of images we put forward in this paper a steganographic technique implemented at the physical layer. We suggest transmission of Halftoned images (payload constraint) in wireless sensor networks to reduce the amount of transmitted data. For low power and interference limited applications Turbo codes provide suitable reliability. Ensuring security is one of the highest priorities in many sensor networks. The Turbo Code structure apart from providing forward error correction can be utilized to provide for encryption. We first consider the Halftoned image and then the method of embedding a block of data (called secret) in this Halftoned image during the turbo encoding process is presented. The small modifications required at the turbo decoder end to extract the embedded data are presented next. The implementation complexity and the degradation of the BER (bit error rate) in the Turbo based stego system are analyzed. Using some of the entropy based crypt analytic techniques we show that the strength of our Turbo based stego system approaches that found in the OTPs (one time pad).

The Issues of Effectiveness of Advertisement Communication Process: A Case Study of Lithuania Consumers

The goal of this study was to disclose the core of the advertising research based on the psychological aspects by acquainting with the nature of advertising research and revealing the importance of psychological aspects of advertising during the advertising research. The growing interest in consumer response to advertisement served as an encouragement to make the analysis of psychological aspects of the advertising research, because the information obtained during the advertising research helps to answer the question how advertising really works. In the research analysis focuses on the nature of advertising research. The place of advertising research in advertisement planning process and the advertising research process are unfolded. Moreover, the importance of psychological aspects in the advertising research is being examined. The certain psychological aspects like the particularities of advertising communication process, psychological process that are active at advertising acceptance and awareness process as well as the advertising effects are analysed in more detail.

Chances and Challenges of Intelligent Technologies in the Production and Retail Sector

This paper provides an introduction into the evolution of information and communication technology and illustrates its usage in the work domain. The paper is sub-divided into two parts. The first part gives an overview over the different phases of information processing in the work domain. It starts by charting the past and present usage of computers in work environments and shows current technological trends, which are likely to influence future business applications. The second part starts by briefly describing, how the usage of computers changed business processes in the past, and presents first Ambient Intelligence applications based on identification and localization information, which are already used in the production and retail sector. Based on current systems and prototype applications, the paper gives an outlook of how Ambient Intelligence technologies could change business processes in the future.

The Tag Authentication Scheme using Self-Shrinking Generator on RFID System

Since communications between tag and reader in RFID system are by radio, anyone can access the tag and obtain its any information. And a tag always replies with the same ID so that it is hard to distinguish between a real and a fake tag. Thus, there are many security problems in today-s RFID System. Firstly, unauthorized reader can easily read the ID information of any Tag. Secondly, Adversary can easily cheat the legitimate reader using the collected Tag ID information, such as the any legitimate Tag. These security problems can be typically solved by encryption of messages transmitted between Tag and Reader and by authentication for Tag. In this paper, to solve these security problems on RFID system, we propose the Tag Authentication Scheme based on self shrinking generator (SSG). SSG Algorithm using in our scheme is proposed by W.Meier and O.Staffelbach in EUROCRYPT-94. This Algorithm is organized that only one LFSR and selection logic in order to generate random stream. Thus it is optimized to implement the hardware logic on devices with extremely limited resource, and the output generating from SSG at each time do role as random stream so that it is allow our to design the light-weight authentication scheme with security against some network attacks. Therefore, we propose the novel tag authentication scheme which use SSG to encrypt the Tag-ID transmitted from tag to reader and achieve authentication of tag.

Statistics over Lyapunov Exponents for Feature Extraction: Electroencephalographic Changes Detection Case

A new approach based on the consideration that electroencephalogram (EEG) signals are chaotic signals was presented for automated diagnosis of electroencephalographic changes. This consideration was tested successfully using the nonlinear dynamics tools, like the computation of Lyapunov exponents. This paper presented the usage of statistics over the set of the Lyapunov exponents in order to reduce the dimensionality of the extracted feature vectors. Since classification is more accurate when the pattern is simplified through representation by important features, feature extraction and selection play an important role in classifying systems such as neural networks. Multilayer perceptron neural network (MLPNN) architectures were formulated and used as basis for detection of electroencephalographic changes. Three types of EEG signals (EEG signals recorded from healthy volunteers with eyes open, epilepsy patients in the epileptogenic zone during a seizure-free interval, and epilepsy patients during epileptic seizures) were classified. The selected Lyapunov exponents of the EEG signals were used as inputs of the MLPNN trained with Levenberg- Marquardt algorithm. The classification results confirmed that the proposed MLPNN has potential in detecting the electroencephalographic changes.

Adaptive Hierarchical Key Structure Generation for Key Management in Wireless Sensor Networks using A*

Wireless Sensor networks have a wide spectrum of civil and military applications that call for secure communication such as the terrorist tracking, target surveillance in hostile environments. For the secure communication in these application areas, we propose a method for generating a hierarchical key structure for the efficient group key management. In this paper, we apply A* algorithm in generating a hierarchical key structure by considering the history data of the ratio of addition and eviction of sensor nodes in a location where sensor nodes are deployed. Thus generated key tree structure provides an efficient way of managing the group key in terms of energy consumption when addition and eviction event occurs. A* algorithm tries to minimize the number of messages needed for group key management by the history data. The experimentation with the tree shows efficiency of the proposed method.

Labeling Method in Steganography

In this paper a way of hiding text message (Steganography) in the gray image has been presented. In this method tried to find binary value of each character of text message and then in the next stage, tried to find dark places of gray image (black) by converting the original image to binary image for labeling each object of image by considering on 8 connectivity. Then these images have been converted to RGB image in order to find dark places. Because in this way each sequence of gray color turns into RGB color and dark level of grey image is found by this way if the Gary image is very light the histogram must be changed manually to find just dark places. In the final stage each 8 pixels of dark places has been considered as a byte and binary value of each character has been put in low bit of each byte that was created manually by dark places pixels for increasing security of the main way of steganography (LSB).

Effect of Superplasticizer and NaOH Molarity on Workability, Compressive Strength and Microstructure Properties of Self-Compacting Geopolymer Concrete

The research investigates the effects of super plasticizer and molarity of sodium hydroxide alkaline solution on the workability, microstructure and compressive strength of self compacting geopolymer concrete (SCGC). SCGC is an improved way of concreting execution that does not require compaction and is made by complete elimination of ordinary Portland cement content. The parameters studied were superplasticizer (SP) dosage and molarity of NaOH solution. SCGC were synthesized from low calcium fly ash, activated by combinations of sodium hydroxide and sodium silicate solutions, and by incorporation of superplasticizer for self compactability. The workability properties such as filling ability, passing ability and resistance to segregation were assessed using slump flow, T-50, V-funnel, L-Box and J-ring test methods. It was found that the essential workability requirements for self compactability according to EFNARC were satisfied. Results showed that the workability and compressive strength improved with the increase in superplasticizer dosage. An increase in strength and a decrease in workability of these concrete samples were observed with the increase in molarity of NaOH solution from 8M to 14M. Improvement of interfacial transition zone (ITZ) and micro structure with the increase of SP and increase of concentration from 8M to 12M were also identified.

Minimizing of Target Localization Error using Multi-robot System and Particle Filters

In recent years a number of applications with multirobot systems (MRS) is growing in various areas. But their design is in practice often difficult and algorithms are proposed for the theoretical background and do not consider errors and noise in real conditions, so they are not usable in real environment. These errors are visible also in task of target localization enough, when robots try to find and estimate the position of the target by the sensors. Localization of target is possible also with one robot but as it was examined target finding and localization with group of mobile robots can estimate the target position more accurately and faster. The accuracy of target position estimation is made by cooperation of MRS and particle filtering. Advantage of usage the MRS with particle filtering was tested on task of fixed target localization by group of mobile robots.

Mining Frequent Patterns with Functional Programming

Frequent patterns are patterns such as sets of features or items that appear in data frequently. Finding such frequent patterns has become an important data mining task because it reveals associations, correlations, and many other interesting relationships hidden in a dataset. Most of the proposed frequent pattern mining algorithms have been implemented with imperative programming languages such as C, Cµ, Java. The imperative paradigm is significantly inefficient when itemset is large and the frequent pattern is long. We suggest a high-level declarative style of programming using a functional language. Our supposition is that the problem of frequent pattern discovery can be efficiently and concisely implemented via a functional paradigm since pattern matching is a fundamental feature supported by most functional languages. Our frequent pattern mining implementation using the Haskell language confirms our hypothesis about conciseness of the program. The performance studies on speed and memory usage support our intuition on efficiency of functional language.