Optimal Embedded Generation Allocation in Distribution System Employing Real Coded Genetic Algorithm Method

This paper proposes a new methodology for the optimal allocation and sizing of Embedded Generation (EG) employing Real Coded Genetic Algorithm (RCGA) to minimize the total power losses and to improve voltage profiles in the radial distribution networks. RCGA is a method that uses continuous floating numbers as representation which is different from conventional binary numbers. The RCGA is used as solution tool, which can determine the optimal location and size of EG in radial system simultaneously. This method is developed in MATLAB. The effect of EG units- installation and their sizing to the distribution networks are demonstrated using 24 bus system.

Performance Analysis of Heat Pipe Using Copper Nanofluid with Aqueous Solution of n-Butanol

This study presents the improvement of thermal performance of heat pipe using copper nanofluid with aqueous solution of n-Butanol. The nanofluids kept in the suspension of conventional fluids have the potential of superior heat transfer capability than the conventional fluids due to their improved thermal conductivity. In this work, the copper nanofluid which has a 40 nm size with a concentration of 100 mg/lit is kept in the suspension of the de-ionized (DI) water and an aqueous solution of n-Butanol and these fluids are used as a working medium in the heat pipe. The study discusses about the effect of heat pipe inclination, type of working fluid and heat input on the thermal efficiency and thermal resistance. The experimental results are evaluated in terms of its performance metrics and are compared with that of DI water.

Ribeirinhos: A Sustainability Assessment of Housing Typologies in the Amazon Region

The 20th century has brought much development to the practice of Architecture worldwide, and technology has bridged inhabitation limits in many regions of the world with high levels of comfort and conveniences, most times at high costs to the environment. Throughout the globe, the tropical countries are being urbanized at an unprecedented rate and housing has become a major issue worldwide, in light of increased demand and lack of appropriate infra-structure and planning. Buildings and urban spaces designed in tropical cities have mainly adopted external concepts that in most cases do not fit the needs of the inhabitants living in such harsh climatic environment, and when they do, do so at high financial, environmental and cultural costs. Traditional architectural practices can provide valuable understanding on how self-reliance and autonomy of construction can be reinforced in rural-urban tropical environments. From traditional housing knowledge, it is possible to derive lessons for the development of new construction materials that are affordable, environmentally friendly, culturally acceptable and accesible to all.Specifically to the urban context, such solutions are of outmost importance, given the needs to a more democratic society, where access to housing is considered high in the agenda for development. Traditional or rural constructions are also ongoing through extensive changes eventhough they have mostly adopted climate-responsive building practices relying on local resources (with minimum embodied energy) and energy (for comfort and quality of life). It is important to note that many of these buildings can actually be called zero-energy, and hold potential answers to enable transition from high energy, high cost, low comfort urban habitations to zero/low energy habitations with high quality urban livelihood. Increasing access to modern urban lifestyels have also an effect on the aspirations from people in terms of performance, comfort and convenience in terms of their housing and the way it is produced and used. These aspirations are resulting in transitions from localresource dependent habitations- to non-local resource based highenergy urban style habitations. And such transitions are resulting in the habitations becoming increasingly unsuited to the local climatic conditions with increasing discomfort, ill-health, and increased CO2 emissions and local environmental disruption. This research studies one specific transition group in the context of 'water communities' in tropical-equatorial regions: Ribeirinhos housing typology (Amazonas, Brazil). The paper presents the results of a qualitative sustainability assessment of the housing typologies under transition, found at the Ribeirinhos communities.

An Efficient Graph Query Algorithm Based on Important Vertices and Decision Features

Graph has become increasingly important in modeling complicated structures and schemaless data such as proteins, chemical compounds, and XML documents. Given a graph query, it is desirable to retrieve graphs quickly from a large database via graph-based indices. Different from the existing methods, our approach, called VFM (Vertex to Frequent Feature Mapping), makes use of vertices and decision features as the basic indexing feature. VFM constructs two mappings between vertices and frequent features to answer graph queries. The VFM approach not only provides an elegant solution to the graph indexing problem, but also demonstrates how database indexing and query processing can benefit from data mining, especially frequent pattern mining. The results show that the proposed method not only avoids the enumeration method of getting subgraphs of query graph, but also effectively reduces the subgraph isomorphism tests between the query graph and graphs in candidate answer set in verification stage.

Plastic Flow through Taper Dies: A Threedimensional Analysis

The plastic flow of metal in the extrusion process is an important factor in controlling the mechanical properties of the extruded products. It is, however, difficult to predict the metal flow in three dimensional extrusions of sections due to the involvement of re-entrant corners. The present study is to find an upper bound solution for the extrusion of triangular sectioned through taper dies from round sectioned billet. A discontinuous kinematically admissible velocity field (KAVF) is proposed. From the proposed KAVF, the upper bound solution on non-dimensional extrusion pressure is determined with respect to the chosen process parameters. The theoretical results are compared with experimental results to check the validity of the proposed velocity field. An extrusion setup is designed and fabricated for the said purpose, and all extrusions are carried out using circular billets. Experiments are carried out with commercially available lead at room temperature.

Removal of Volatile Organic Compounds from Contaminated Surfactant Solution using Co-Curren Vacuum Stripping

There has been a growing interest in utilizing surfactants in remediation processes to separate the hydrophobic volatile organic compounds (HVOCs) from aqueous solution. One attractive process is cloud point extraction (CPE), which utilizes nonionic surfactants as a separating agent. Since the surfactant cost is a key determination of the economic viability of the process, it is important that the surfactants are recycled and reused. This work aims to study the performance of the co-current vacuum stripping using a packed column for HVOCs removal from contaminated surfactant solution. Six types HVOCs are selected as contaminants. The studied surfactant is the branched secondary alcohol ethoxylates (AEs), Tergitol TMN-6 (C14H30O2). The volatility and the solubility of HVOCs in surfactant system are determined in terms of an apparent Henry’s law constant and a solubilization constant, respectively. Moreover, the HVOCs removal efficiency of vacuum stripping column is assessed in terms of percentage of HVOCs removal and the overall liquid phase volumetric mass transfer coefficient. The apparent Henry’s law constant of benzenz , toluene, and ethyl benzene were 7.00×10-5, 5.38×10-5, 3.35× 10-5 respectively. The solubilization constant of benzene, toluene, and ethyl benzene were 1.71, 2.68, 7.54 respectively. The HVOCs removal for all solute were around 90 percent.

Augmented Reality Interaction System in 3D Environment

It is important to give input information without other device in AR system. One solution is using hand for augmented reality application. Many researchers have proposed different solutions for hand interface in augmented reality. Analyze Histogram and connecting factor is can be example for that. Various Direction searching is one of robust way to recognition hand but it takes too much calculating time. And background should be distinguished with skin color. This paper proposes a hand tracking method to control the 3D object in augmented reality using depth device and skin color. Also in this work discussed relationship between several markers, which is based on relationship between camera and marker. One marker used for displaying virtual object and three markers for detecting hand gesture and manipulating the virtual object.

Enhancement of Stereo Video Pairs Using SDNs To Aid In 3D Reconstruction

This paper presents the results of enhancing images from a left and right stereo pair in order to increase the resolution of a 3D representation of a scene generated from that same pair. A new neural network structure known as a Self Delaying Dynamic Network (SDN) has been used to perform the enhancement. The advantage of SDNs over existing techniques such as bicubic interpolation is their ability to cope with motion and noise effects. SDNs are used to generate two high resolution images, one based on frames taken from the left view of the subject, and one based on the frames from the right. This new high resolution stereo pair is then processed by a disparity map generator. The disparity map generated is compared to two other disparity maps generated from the same scene. The first is a map generated from an original high resolution stereo pair and the second is a map generated using a stereo pair which has been enhanced using bicubic interpolation. The maps generated using the SDN enhanced pairs match more closely the target maps. The addition of extra noise into the input images is less problematic for the SDN system which is still able to out perform bicubic interpolation.

Scale Time Offset Robust Modulation (STORM) in a Code Division Multiaccess Environment

Scale Time Offset Robust Modulation (STORM) [1]– [3] is a high bandwidth waveform design that adds time-scale to embedded reference modulations using only time-delay [4]. In an environment where each user has a specific delay and scale, identification of the user with the highest signal power and that user-s phase is facilitated by the STORM processor. Both of these parameters are required in an efficient multiuser detection algorithm. In this paper, the STORM modulation approach is evaluated with a direct sequence spread quadrature phase shift keying (DS-QPSK) system. A misconception of the STORM time scale modulation is that a fine temporal resolution is required at the receiver. STORM will be applied to a QPSK code division multiaccess (CDMA) system by modifying the spreading codes. Specifically, the in-phase code will use a typical spreading code, and the quadrature code will use a time-delayed and time-scaled version of the in-phase code. Subsequently, the same temporal resolution in the receiver is required before and after the application of STORM. In this paper, the bit error performance of STORM in a synchronous CDMA system is evaluated and compared to theory, and the bit error performance of STORM incorporated in a single user WCDMA downlink is presented to demonstrate the applicability of STORM in a modern communication system.

Sulphur-Mediated Precipitation of Pt/Fe/Co/CrIons in Liquid-Liquid and Gas-Liquid Chloride Systems

The proof of concept experiments were conducted to determine the feasibility of using small amounts of Dissolved Sulphur (DS) from the gaseous phase to precipitate platinum ions in chloride media. Two sets of precipitation experiments were performed in which the source of sulphur atoms was either a thiosulphate solution (Na2S2O3) or a sulphur dioxide gas (SO2). In liquid-liquid (L-L) system, complete precipitation of Pt was achieved at small dosages of Na2S2O3 (0.01 – 1.0 M) in a time interval of 3-5 minutes. On the basis of this result, gas absorption tests were carried out mainly to achieve sulphur solubility equivalent to 0.018 M. The idea that huge amounts of precious metals could be recovered selectively from their dilute solutions by utilizing the waste SO2 streams at low pressure seemed attractive from the economic and environmental point of views. Therefore, mass transfer characteristics of SO2 gas associated with reactive absorption across the gas-liquid (G-L) interface were evaluated under different conditions of pressure (0.5 – 2 bar), solution temperature ranges from 20 – 50 oC and acid strength (1 – 4 M, HCl). This paper concludes with information about selective precipitation of Pt in the presence of cations (Fe2+, Co2+, and Cr3+) in a CSTR and recommendation to scale up laboratory data to industrial pilot scale operations.

Adomian Decomposition Method Associated with Boole-s Integration Rule for Goursat Problem

The Goursat partial differential equation arises in linear and non linear partial differential equations with mixed derivatives. This equation is a second order hyperbolic partial differential equation which occurs in various fields of study such as in engineering, physics, and applied mathematics. There are many approaches that have been suggested to approximate the solution of the Goursat partial differential equation. However, all of the suggested methods traditionally focused on numerical differentiation approaches including forward and central differences in deriving the scheme. An innovation has been done in deriving the Goursat partial differential equation scheme which involves numerical integration techniques. In this paper we have developed a new scheme to solve the Goursat partial differential equation based on the Adomian decomposition (ADM) and associated with Boole-s integration rule to approximate the integration terms. The new scheme can easily be applied to many linear and non linear Goursat partial differential equations and is capable to reduce the size of computational work. The accuracy of the results reveals the advantage of this new scheme over existing numerical method.

Generational PipeLined Genetic Algorithm (PLGA)using Stochastic Selection

In this paper, a pipelined version of genetic algorithm, called PLGA, and a corresponding hardware platform are described. The basic operations of conventional GA (CGA) are made pipelined using an appropriate selection scheme. The selection operator, used here, is stochastic in nature and is called SA-selection. This helps maintaining the basic generational nature of the proposed pipelined GA (PLGA). A number of benchmark problems are used to compare the performances of conventional roulette-wheel selection and the SA-selection. These include unimodal and multimodal functions with dimensionality varying from very small to very large. It is seen that the SA-selection scheme is giving comparable performances with respect to the classical roulette-wheel selection scheme, for all the instances, when quality of solutions and rate of convergence are considered. The speedups obtained by PLGA for different benchmarks are found to be significant. It is shown that a complete hardware pipeline can be developed using the proposed scheme, if parallel evaluation of the fitness expression is possible. In this connection a low-cost but very fast hardware evaluation unit is described. Results of simulation experiments show that in a pipelined hardware environment, PLGA will be much faster than CGA. In terms of efficiency, PLGA is found to outperform parallel GA (PGA) also.

Solar Photo-Fenton Induced Degradation of Combined Chlorpyrifos, Cypermethrin and Chlorothalonil Pesticides in Aqueous Solution

The feasibility of employing solar radiation for enhanced Fenton process in degradation of combined chlorpyrifos, cypermethrin and chlorothalonil pesticides was examined. The effect of various operating conditions of the process on biodegradability improvement and mineralization of the pesticides were also evaluated. The optimum operating conditions for treatment of aqueous solution containing 100, 50 and 250 mg L-1 chlorpyrifos cypermethrin and chlorothalonil, respectively were observed to be H2O2/COD molar ratio 2, H2O2/Fe2+ molar ratio 25 and pH 3. Under the optimum operating conditions, complete degradation of the pesticides occurred in 1 min. Biodegradability (BOD5/COD) increased from zero to 0.36 in 60 min, and COD and TOC removal were 74.19 and 58.32%, respectively in 60 min. Due to mineralization of organic carbon, decrease in ammonia-nitrogen from 22 to 4.3 mg L-1 and increase in nitrate from 0.7 to 18.1 mg L-1 in 60 min were recorded. The study indicated that solar photo-Fenton process can be used for pretreatment of chlorpyrifos, cypermethrin and chlorothalonil pesticides in aqueous solution for further biological treatment.

Alignment of Emission Gamma Ray Sources with Nai(Ti) Scintillation Detectors by Two Laser Beams to Pre-Operation using Alternating Minimization Technique

Accurate timing alignment and stability is important to maximize the true counts and minimize the random counts in positron emission tomography So signals output from detectors must be centering with the two isotopes to pre-operation and fed signals into four units of pulse-processing units, each unit can accept up to eight inputs. The dual source computed tomography consist two units on the left for 15 detector signals of Cs-137 isotope and two units on the right are for 15 detectors signals of Co-60 isotope. The gamma spectrum consisting of either single or multiple photo peaks. This allows for the use of energy discrimination electronic hardware associated with the data acquisition system to acquire photon counts data with a specific energy, even if poor energy resolution detectors are used. This also helps to avoid counting of the Compton scatter counts especially if a single discrete gamma photo peak is emitted by the source as in the case of Cs-137. In this study the polyenergetic version of the alternating minimization algorithm is applied to the dual energy gamma computed tomography problem.

Osmotic Dehydration of Beetroot in Salt Solution: Optimization of Parameters through Statistical Experimental Design

Response surface methodology was used for quantitative investigation of water and solids transfer during osmotic dehydration of beetroot in aqueous solution of salt. Effects of temperature (25 – 45oC), processing time (30–150 min), salt concentration (5–25%, w/w) and solution to sample ratio (5:1 – 25:1) on osmotic dehydration of beetroot were estimated. Quadratic regression equations describing the effects of these factors on the water loss and solids gain were developed. It was found that effects of temperature and salt concentrations were more significant on the water loss than the effects of processing time and solution to sample ratio. As for solids gain processing time and salt concentration were the most significant factors. The osmotic dehydration process was optimized for water loss, solute gain, and weight reduction. The optimum conditions were found to be: temperature – 35oC, processing time – 90 min, salt concentration – 14.31% and solution to sample ratio 8.5:1. At these optimum values, water loss, solid gain and weight reduction were found to be 30.86 (g/100 g initial sample), 9.43 (g/100 g initial sample) and 21.43 (g/100 g initial sample) respectively.

A New Voting Approach to Texture Defect Detection Based on Multiresolutional Decomposition

Wavelets have provided the researchers with significant positive results, by entering the texture defect detection domain. The weak point of wavelets is that they are one-dimensional by nature so they are not efficient enough to describe and analyze two-dimensional functions. In this paper we present a new method to detect the defect of texture images by using curvelet transform. Simulation results of the proposed method on a set of standard texture images confirm its correctness. Comparing the obtained results indicates the ability of curvelet transform in describing discontinuity in two-dimensional functions compared to wavelet transform

Extracting Tongue Shape Dynamics from Magnetic Resonance Image Sequences

An important problem in speech research is the automatic extraction of information about the shape and dimensions of the vocal tract during real-time speech production. We have previously developed Southampton dynamic magnetic resonance imaging (SDMRI) as an approach to the solution of this problem.However, the SDMRI images are very noisy so that shape extraction is a major challenge. In this paper, we address the problem of tongue shape extraction, which poses difficulties because this is a highly deforming non-parametric shape. We show that combining active shape models with the dynamic Hough transform allows the tongue shape to be reliably tracked in the image sequence.

Using Emotional Learning in Rescue Simulation Environment

RoboCup Rescue simulation as a large-scale Multi agent system (MAS) is one of the challenging environments for keeping coordination between agents to achieve the objectives despite sensing and communication limitations. The dynamicity of the environment and intensive dependency between actions of different kinds of agents make the problem more complex. This point encouraged us to use learning-based methods to adapt our decision making to different situations. Our approach is utilizing reinforcement leaning. Using learning in rescue simulation is one of the current ways which has been the subject of several researches in recent years. In this paper we present an innovative learning method implemented for Police Force (PF) Agent. This method can cope with the main difficulties that exist in other learning approaches. Different methods used in the literature have been examined. Their drawbacks and possible improvements have led us to the method proposed in this paper which is fast and accurate. The Brain Emotional Learning Based Intelligent Controller (BELBIC) is our solution for learning in this environment. BELBIC is a physiologically motivated approach based on a computational model of amygdale and limbic system. The paper presents the results obtained by the proposed approach, showing the power of BELBIC as a decision making tool in complex and dynamic situation.

A Fuzzy Multi-objective Model for a Machine Selection Problem in a Flexible Manufacturing System

This research presents a fuzzy multi-objective model for a machine selection problem in a flexible manufacturing system of a tire company. Two main objectives are minimization of an average machine error and minimization of the total setup time. Conventionally, the working team uses trial and error in selecting a pressing machine for each task due to the complexity and constraints of the problem. So, both objectives may not satisfy. Moreover, trial and error takes a lot of time to get the final decision. Therefore, in this research preemptive fuzzy goal programming model is developed for solving this multi-objective problem. The proposed model can obtain the appropriate results that the Decision Making (DM) is satisfied for both objectives. Besides, alternative choice can be easily generated by varying the satisfaction level. Additionally, decision time can be reduced by using the model, which includes all constraints of the system to generate the solutions. A numerical example is also illustrated to show the effectiveness of the proposed model.

Performance Boundaries for Interactive Finite Element Applications

This paper presents work characterizing finite element performance boundaries within which live, interactive finite element modeling is feasible on current and emerging systems. These results are based on wide-ranging tests performed using a prototype finite element program implemented specifically for this study, thereby enabling the unified investigation of numerous direct and iterative solver strategies and implementations in a variety of modeling contexts. The results are intended to be useful for researchers interested in interactive analysis by providing baseline performance estimates, to give guidance in matching solution strategies to problem domains, and to spur further work addressing the challenge of extending the present boundaries.