Application of Neural Networks in Power Systems; A Review

The electric power industry is currently undergoing an unprecedented reform. One of the most exciting and potentially profitable recent developments is increasing usage of artificial intelligence techniques. The intention of this paper is to give an overview of using neural network (NN) techniques in power systems. According to the growth rate of NNs application in some power system subjects, this paper introduce a brief overview in fault diagnosis, security assessment, load forecasting, economic dispatch and harmonic analyzing. Advantages and disadvantages of using NNs in above mentioned subjects and the main challenges in these fields have been explained, too.

Signed Approach for Mining Web Content Outliers

The emergence of the Internet has brewed the revolution of information storage and retrieval. As most of the data in the web is unstructured, and contains a mix of text, video, audio etc, there is a need to mine information to cater to the specific needs of the users without loss of important hidden information. Thus developing user friendly and automated tools for providing relevant information quickly becomes a major challenge in web mining research. Most of the existing web mining algorithms have concentrated on finding frequent patterns while neglecting the less frequent ones that are likely to contain outlying data such as noise, irrelevant and redundant data. This paper mainly focuses on Signed approach and full word matching on the organized domain dictionary for mining web content outliers. This Signed approach gives the relevant web documents as well as outlying web documents. As the dictionary is organized based on the number of characters in a word, searching and retrieval of documents takes less time and less space.

Two Scenarios for Ultra-Light Overhead Conveyor System in Logistics Applications

Overhead conveyor systems are in use in many installations around the world, meeting the widest range of applications possible. Overhead conveyor systems are particularly preferred in automotive industry but also at post offices. Overhead conveyor systems must always be integrated with a logistical process by finding the best way for a cheaper material flow in order to guarantee precise and fast workflows. With their help, any transport can take place without wasting ground and space, without excessive company capacity, lost or damaged products, erroneous delivery, endless travels and without wasting time. Ultra-light overhead conveyor systems are rope-based conveying systems with individually driven vehicles. The vehicles can move automatically on the rope and this can be realized by energy and signals. Crossings are realized by switches. Ultra-light overhead conveyor systems provide optimal material flow, which produces profit and saves time. This article introduces two new ultra-light overhead conveyor designs in logistics and explains their components. According to the explanation of the components, scenarios are created by means of their technical characteristics. The scenarios are visualized with the help of CAD software. After that, assumptions are made for application area. According to these assumptions scenarios are visualized. These scenarios help logistics companies achieve lower development costs as well as quicker market maturity.

Ultra Fast Solid State Ground Fault Isolator

Personnel protection devices are cardinal in safety hazard applications. They are widely used in home, office and in industry environments to reduce the risk of lethal shock to human being and equipment safety. This paper briefly reviews various personnel protection devices also describes the basic working principle of conventional ground fault circuit interrupter (GFCI) or ground fault isolator (GFI), its disadvantages and ways to overcome the disadvantages with solid-state relay (SSR) based GFI with ultrafast response up on fault implemented in printed circuit board. This solid state GFI comprises discrete MOSFET based alternating current (AC) switches, linear optical amplifier, photovoltaic isolator and sense resistor. In conventional GFI, current transformer is employed as a sensing element to detect the difference in current flow between live and neutral conductor. If there is no fault in equipment powered through GFI, due to insulation failure of internal wires and windings of motors, both live and neutral currents will be equal in magnitude and opposite in phase.

Biosorption of Heavy Metals Contaminating the Wonderfonteinspruit Catchment Area using Desmodesmus sp.

A vast array of biological materials, especially algae have received increasing attention for heavy metal removal. Algae have been proven to be cheaper, more effective for the removal of metallic elements in aqueous solutions. A fresh water algal strain was isolated from Zoo Lake, Johannesburg, South Africa and identified as Desmodesmus sp. This paper investigates the efficacy of Desmodesmus sp.in removing heavy metals contaminating the Wonderfonteinspruit Catchment Area (WCA) water bodies. The biosorption data fitted the pseudo-second order and Langmuir isotherm models. The Langmuir maximum uptakes gave the sequence: Mn2+>Ni2+>Fe2+. The best results for kinetic study was obtained in concentration 120 ppm for Fe3+ and Mn2+, whilst for Ni2+ was at 20 ppm, which is about the same concentrations found in contaminated water in the WCA (Fe3+115 ppm, Mn2+ 121 ppm and Ni2+ 26.5 ppm).

Goal Based Episodic Processing in Implicit Learning

Research has suggested that implicit learning tasks may rely on episodic processing to generate above chance performance on the standard classification tasks. The current research examines the invariant features task (McGeorge and Burton, 1990) and argues that such episodic processing is indeed important. The results of the experiment suggest that both rejection and similarity strategies are used by participants in this task to simultaneously reject unfamiliar items and to accept (falsely) familiar items. Primarily these decisions are based on the presence of low or high frequency goal based features of the stimuli presented in the incidental learning phase. It is proposed that a goal based analysis of the incidental learning task provides a simple step in understanding which features of the episodic processing are most important for explaining the match between incidental, implicit learning and test performance.

Simulation of the Airflow Characteristic inside a Hard Disk Drive by Applying a Computational Fluid Dynamics Software

Now-a-days, numbers of simulation software are being used all over the world to solve Computational Fluid Dynamics (CFD) related problems. In this present study, a commercial CFD simulation software namely STAR-CCM+ is applied to analyze the airflow characteristics inside a 2.5" hard disk drive. Each step of the software is described adequately to obtain the output and the data are verified with the theories to justify the robustness of the simulation outcome. This study gives an insight about the accuracy level of the CFD simulation software to compute CFD related problems although it largely depends upon the computer speed. Also this study will open avenues for further research.

Comparison of Full Graph Methods of Switched Circuits Solution

As there are also graph methods of circuit analysis in addition to algebraic methods, it is, in theory, clearly possible to carry out an analysis of a whole switched circuit in two-phase switching exclusively by the graph method as well. This article deals with two methods of full-graph solving of switched circuits: by transformation graphs and by two-graphs. It deals with the circuit switched capacitors and the switched current, too. All methods are presented in an equally detailed steps to be able to compare.

Systematic Analysis of Dynamic Association of Health Outcomes with Computer Usage for Office Staff

This paper systematically investigates the timedependent health outcomes for office staff during computer work using the developed mathematical model. The model describes timedependent health outcomes in multiple body regions associated with computer usage. The association is explicitly presented with a doseresponse relationship which is parametrized by body region parameters. Using the developed model we perform extensive investigations of the health outcomes statically and dynamically. We compare the risk body regions and provide various severity rankings of the discomfort rate changes with respect to computer-related workload dynamically for the study population. Application of the developed model reveals a wide range of findings. Such broad spectrum of investigations in a single report literature is lacking. Based upon the model analysis, it is discovered that the highest average severity level of the discomfort exists in neck, shoulder, eyes, shoulder joint/upper arm, upper back, low back and head etc. The biggest weekly changes of discomfort rates are in eyes, neck, head, shoulder, shoulder joint/upper arm and upper back etc. The fastest discomfort rate is found in neck, followed by shoulder, eyes, head, shoulder joint/upper arm and upper back etc. Most of our findings are consistent with the literature, which demonstrates that the developed model and results are applicable and valuable and can be utilized to assess correlation between the amount of computer-related workload and health risk.

Automatic 2D/2D Registration using Multiresolution Pyramid based Mutual Information in Image Guided Radiation Therapy

Medical image registration is the key technology in image guided radiation therapy (IGRT) systems. On the basis of the previous work on our IGRT prototype with a biorthogonal x-ray imaging system, we described a method focused on the 2D/2D rigid-body registration using multiresolution pyramid based mutual information in this paper. Three key steps were involved in the method : firstly, four 2D images were obtained including two x-ray projection images and two digital reconstructed radiographies(DRRs ) as the input for the registration ; Secondly, each pair of the corresponding x-ray image and DRR image were matched using multiresolution pyramid based mutual information under the ITK registration framework ; Thirdly, we got the final couch offset through a coordinate transformation by calculating the translations acquired from the two pairs of the images. A simulation example of a parotid gland tumor case and a clinical example of an anthropomorphic head phantom were employed in the verification tests. In addition, the influence of different CT slice thickness were tested. The simulation results showed that the positioning errors were 0.068±0.070, 0.072±0.098, 0.154±0.176mm along three axes which were lateral, longitudinal and vertical. The clinical test indicated that the positioning errors of the planned isocenter were 0.066, 0.07, 2.06mm on average with a CT slice thickness of 2.5mm. It can be concluded that our method with its verified accuracy and robustness can be effectively used in IGRT systems for patient setup.

Robust Face Recognition Using Eigen Faces and Karhunen-Loeve Algorithm

The current research paper is an implementation of Eigen Faces and Karhunen-Loeve Algorithm for face recognition. The designed program works in a manner where a unique identification number is given to each face under trial. These faces are kept in a database from where any particular face can be matched and found out of the available test faces. The Karhunen –Loeve Algorithm has been implemented to find out the appropriate right face (with same features) with respect to given input image as test data image having unique identification number. The procedure involves usage of Eigen faces for the recognition of faces.

Matching Pursuit based Removal of Cardiac Pulse-Related Artifacts in EEG/fMRI

Cardiac pulse-related artifacts in the EEG recorded simultaneously with fMRI are complex and highly variable. Their effective removal is an unsolved problem. Our aim is to develop an adaptive removal algorithm based on the matching pursuit (MP) technique and to compare it to established methods using a visual evoked potential (VEP). We recorded the VEP inside the static magnetic field of an MR scanner (with artifacts) as well as in an electrically shielded room (artifact free). The MP-based artifact removal outperformed average artifact subtraction (AAS) and optimal basis set removal (OBS) in terms of restoring the EEG field map topography of the VEP. Subsequently, a dipole model was fitted to the VEP under each condition using a realistic boundary element head model. The source location of the VEP recorded inside the MR scanner was closest to that of the artifact free VEP after cleaning with the MP-based algorithm as well as with AAS. While none of the tested algorithms offered complete removal, MP showed promising results due to its ability to adapt to variations of latency, frequency and amplitude of individual artifact occurrences while still utilizing a common template.

Predicting Oil Content of Fresh Palm Fruit Using Transmission-Mode Ultrasonic Technique

In this paper, an ultrasonic technique is proposed to predict oil content in a fresh palm fruit. This is accomplished by measuring the attenuation based on ultrasonic transmission mode. Several palm fruit samples with known oil content by Soxhlet extraction (ISO9001:2008) were tested with our ultrasonic measurement. Amplitude attenuation data results for all palm samples were collected. The Feedforward Neural Networks (FNNs) are applied to predict the oil content for the samples. The Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) of the FNN model for predicting oil content percentage are 7.6186 and 5.2287 with the correlation coefficient (R) of 0.9193.

Performance Evaluation of a Limited Round-Robin System

Performance of a limited Round-Robin (RR) rule is studied in order to clarify the characteristics of a realistic sharing model of a processor. Under the limited RR rule, the processor allocates to each request a fixed amount of time, called a quantum, in a fixed order. The sum of the requests being allocated these quanta is kept below a fixed value. Arriving requests that cannot be allocated quanta because of such a restriction are queued or rejected. Practical performance measures, such as the relationship between the mean sojourn time, the mean number of requests, or the loss probability and the quantum size are evaluated via simulation. In the evaluation, the requested service time of an arriving request is converted into a quantum number. One of these quanta is included in an RR cycle, which means a series of quanta allocated to each request in a fixed order. The service time of the arriving request can be evaluated using the number of RR cycles required to complete the service, the number of requests receiving service, and the quantum size. Then an increase or decrease in the number of quanta that are necessary before service is completed is reevaluated at the arrival or departure of other requests. Tracking these events and calculations enables us to analyze the performance of our limited RR rule. In particular, we obtain the most suitable quantum size, which minimizes the mean sojourn time, for the case in which the switching time for each quantum is considered.

The Effect of Ambient Occlusion Shading on Perception of Sign Language Animations

The goal of the study reported in the paper was to determine whether Ambient Occlusion Shading (AOS) has a significant effect on users' perception of American Sign Language (ASL) finger spelling animations. Seventy-one (71) subjects participated in the study; all subjects were fluent in ASL. The participants were asked to watch forty (40) sign language animation clips representing twenty (20) finger spelled words. Twenty (20) clips did not show ambient occlusion, whereas the other twenty (20) were rendered using ambient occlusion shading. After viewing each animation, subjects were asked to type the word being finger-spelled and rate its legibility. Findings show that the presence of AOS had a significant effect on the subjects perception of the signed words. Subjects were able to recognize the animated words rendered with AOS with higher level of accuracy, and the legibility ratings of the animations showing AOS were consistently higher across subjects.

Traditionally Sustainability Analyses of Hydraulic-Architectural Bridge Construction in Iran

Bridge is an architectural symbol in Iran as Badgir (wind catcher); fire temples and arch are vaults are such. Therefore, from the very old ages, construction of bridges in Iran has mixed with architecture, social customs, alms and charity and holiness. Since long ago, from Mad, Achaemenid, Parthian and Sassanid times which construction of bridges got an inseparable relation with social dependency and architecture, based on those dependency bridges and dams got holy names; as Dokhtar castle and Dokhtar bridges were constructed. This method continued even after Islam and whenever Iranians got free from political fights and the immunity of roads were established the bridge construction did also prospered. In ancient times bridge construction passes through it growing and completion process and in Sassanid time in some way it reached to the peak of art and glory; as after Islam especially during 4th. century (Arab calendar) it put behind a period of glory and in Safavid time it reached to an exceptional glory and magnificence by constructing glorious bridges on Zayandeh Roud River in Isfahan. Having a combined style and changeability into bridge barrier, some of these bridges develop into magnificent constructions. The sustainable structures, mentioned above, are constructed for various reasons as follows: connecting two sides of a river, storing water, controlling floods, using water energy to operate water windmills, making lanes of streams for farms- use, and building recreational places for people, etc. These studies carried in bridges reveals the fact that in construction and designing mentioned above, lots of technological factors have been taken into consideration such as exceeding floods in the rives, hydraulic and hydrology of the rivers and bridges, geology, foundation, structure, construction material, and adopting appropriate executing methods, all of which are being analyzed in this article.

Advanced Geolocation of IP Addresses

Tracing and locating the geographical location of users (Geolocation) is used extensively in todays Internet. Whenever we, e.g., request a page from google we are - unless there was a specific configuration made - automatically forwarded to the page with the relevant language and amongst others, dependent on our location identified, specific commercials are presented. Especially within the area of Network Security, Geolocation has a significant impact. Because of the way the Internet works, attacks can be executed from almost everywhere. Therefore, for an attribution, knowledge of the origination of an attack - and thus Geolocation - is mandatory in order to be able to trace back an attacker. In addition, Geolocation can also be used very successfully to increase the security of a network during operation (i.e. before an intrusion actually has taken place). Similar to greylisting in emails, Geolocation allows to (i) correlate attacks detected with new connections and (ii) as a consequence to classify traffic a priori as more suspicious (thus particularly allowing to inspect this traffic in more detail). Although numerous techniques for Geolocation are existing, each strategy is subject to certain restrictions. Following the ideas of Endo et al., this publication tries to overcome these shortcomings with a combined solution of different methods to allow improved and optimized Geolocation. Thus, we present our architecture for improved Geolocation, by designing a new algorithm, which combines several Geolocation techniques to increase the accuracy.

A Family Cars- Life Cycle Cost (LCC)-Oriented Hybrid Modelling Approach Combining ANN and CBR

Design for cost (DFC) is a method that reduces life cycle cost (LCC) from the angle of designers. Multiple domain features mapping (MDFM) methodology was given in DFC. Using MDFM, we can use design features to estimate the LCC. From the angle of DFC, the design features of family cars were obtained, such as all dimensions, engine power and emission volume. At the conceptual design stage, cars- LCC were estimated using back propagation (BP) artificial neural networks (ANN) method and case-based reasoning (CBR). Hamming space was used to measure the similarity among cases in CBR method. Levenberg-Marquardt (LM) algorithm and genetic algorithm (GA) were used in ANN. The differences of LCC estimation model between CBR and artificial neural networks (ANN) were provided. ANN and CBR separately each method has its shortcomings. By combining ANN and CBR improved results accuracy was obtained. Firstly, using ANN selected some design features that affect LCC. Then using LCC estimation results of ANN could raise the accuracy of LCC estimation in CBR method. Thirdly, using ANN estimate LCC errors and correct errors in CBR-s estimation results if the accuracy is not enough accurate. Finally, economically family cars and sport utility vehicle (SUV) was given as LCC estimation cases using this hybrid approach combining ANN and CBR.

Removal of Basic Blue 3 from Aqueous Solution by Adsorption Onto Durio Ziberthinus

Durian husk (DH), a fruit waste, was studied for its ability to remove Basic blue 3 (BB3) from aqueous solutions. Batch kinetic studies were carried out to study the sorption characteristics under various experimental conditions. The optimum pH for the dye removal occurred in the pH range of 3-10. Sorption was found to be concentration and agitation dependent. The kinetics of dye sorption fitted a pseudo-second order rate expression. Both Langmuir and Freundlich models appeared to provide reasonable fittings for the sorption data of BB3 on durian husk. Maximum sorption capacity calculated from the Langmuir model is 49.50 mg g-1.

Low Power and Less Area Architecture for Integer Motion Estimation

Full search block matching algorithm is widely used for hardware implementation of motion estimators in video compression algorithms. In this paper we are proposing a new architecture, which consists of a 2D parallel processing unit and a 1D unit both working in parallel. The proposed architecture reduces both data access power and computational power which are the main causes of power consumption in integer motion estimation. It also completes the operations with nearly the same number of clock cycles as compared to a 2D systolic array architecture. In this work sum of absolute difference (SAD)-the most repeated operation in block matching, is calculated in two steps. The first step is to calculate the SAD for alternate rows by a 2D parallel unit. If the SAD calculated by the parallel unit is less than the stored minimum SAD, the SAD of the remaining rows is calculated by the 1D unit. Early termination, which stops avoidable computations has been achieved with the help of alternate rows method proposed in this paper and by finding a low initial SAD value based on motion vector prediction. Data reuse has been applied to the reference blocks in the same search area which significantly reduced the memory access.