Chelate Enhanced Modified Fenton Treatment for Polycyclic Aromatic Hydrocarbons Contaminated Soils

This work focuses on the remediation of polycyclic aromatic hydrocarbons (PAHs)-contaminated soil via Fenton treatment coupled with novel chelating agent (CA). The feasibility of chelated modified Fenton (MF) treatment to promote PAH oxidation in artificially contaminated soils was investigated in laboratory scale batch experiments at natural pH. The effects of adding inorganic and organic CA are discussed. Experiments using different iron catalyst to CA ratios were conducted, resulting in hydrogen peroxide: soil: iron: CA weight ratios that varied from 0.049: 1: 0.072: 0.008 to 0.049: 1: 0.072: 0.067. The results revealed that (1) inorganic CA could provide much higher PAH removal efficiency and (2) most of the proposed CAs were more efficient than commonly utilised CAs even at mild ratio. This work highlights the potential of novel chelating agents in maintaining a suitable environment throughout the Fenton treatment, particularly in soils with high buffer capacity.

Generational PipeLined Genetic Algorithm (PLGA)using Stochastic Selection

In this paper, a pipelined version of genetic algorithm, called PLGA, and a corresponding hardware platform are described. The basic operations of conventional GA (CGA) are made pipelined using an appropriate selection scheme. The selection operator, used here, is stochastic in nature and is called SA-selection. This helps maintaining the basic generational nature of the proposed pipelined GA (PLGA). A number of benchmark problems are used to compare the performances of conventional roulette-wheel selection and the SA-selection. These include unimodal and multimodal functions with dimensionality varying from very small to very large. It is seen that the SA-selection scheme is giving comparable performances with respect to the classical roulette-wheel selection scheme, for all the instances, when quality of solutions and rate of convergence are considered. The speedups obtained by PLGA for different benchmarks are found to be significant. It is shown that a complete hardware pipeline can be developed using the proposed scheme, if parallel evaluation of the fitness expression is possible. In this connection a low-cost but very fast hardware evaluation unit is described. Results of simulation experiments show that in a pipelined hardware environment, PLGA will be much faster than CGA. In terms of efficiency, PLGA is found to outperform parallel GA (PGA) also.

Face Detection using Gabor Wavelets and Neural Networks

This paper proposes new hybrid approaches for face recognition. Gabor wavelets representation of face images is an effective approach for both facial action recognition and face identification. Perform dimensionality reduction and linear discriminate analysis on the down sampled Gabor wavelet faces can increase the discriminate ability. Nearest feature space is extended to various similarity measures. In our experiments, proposed Gabor wavelet faces combined with extended neural net feature space classifier shows very good performance, which can achieve 93 % maximum correct recognition rate on ORL data set without any preprocessing step.

On the Reduction of Side Effects in Tomography

As the Computed Tomography(CT) requires normally hundreds of projections to reconstruct the image, patients are exposed to more X-ray energy, which may cause side effects such as cancer. Even when the variability of the particles in the object is very less, Computed Tomography requires many projections for good quality reconstruction. In this paper, less variability of the particles in an object has been exploited to obtain good quality reconstruction. Though the reconstructed image and the original image have same projections, in general, they need not be the same. In addition to projections, if a priori information about the image is known, it is possible to obtain good quality reconstructed image. In this paper, it has been shown by experimental results why conventional algorithms fail to reconstruct from a few projections, and an efficient polynomial time algorithm has been given to reconstruct a bi-level image from its projections along row and column, and a known sub image of unknown image with smoothness constraints by reducing the reconstruction problem to integral max flow problem. This paper also discusses the necessary and sufficient conditions for uniqueness and extension of 2D-bi-level image reconstruction to 3D-bi-level image reconstruction.

A Study on the Location and Range of Obstacle Region in Robot's Point Placement Task based on the Vision Control Algorithm

This paper is concerned with the application of the vision control algorithm for robot's point placement task in discontinuous trajectory caused by obstacle. The presented vision control algorithm consists of four models, which are the robot kinematic model, vision system model, parameters estimation model, and robot joint angle estimation model.When the robot moves toward a target along discontinuous trajectory, several types of obstacles appear in two obstacle regions. Then, this study is to investigate how these changes will affect the presented vision control algorithm.Thus, the practicality of the vision control algorithm is demonstrated experimentally by performing the robot's point placement task in discontinuous trajectory by obstacle.

The Development of Flying Type Moving Robot Using Image Processing

Wheel-running type moving robot has the restriction on the moving range caused by obstacles or stairs. Solving this weakness, we studied the development of moving robot using airship. Our airship robot moves by recognizing arrow marks on the path. To have the airship robot recognize arrow marks, we used edge-based template matching. To control propeller units, we used PID and PD controller. The results of experiments demonstrated that the airship robot can move along the marks and can go up and down the stairs. It is shown the possibility that airship robot can become a robot which can move at wide range facilities.

Identifying Attack Code through an Ontology-Based Multiagent Tool: FROID

This paper describes the design and results of FROID, an outbound intrusion detection system built with agent technology and supported by an attacker-centric ontology. The prototype features a misuse-based detection mechanism that identifies remote attack tools in execution. Misuse signatures composed of attributes selected through entropy analysis of outgoing traffic streams and process runtime data are derived from execution variants of attack programs. The core of the architecture is a mesh of self-contained detection cells organized non-hierarchically that group agents in a functional fashion. The experiments show performance gains when the ontology is enabled as well as an increase in accuracy achieved when correlation cells combine detection evidence received from independent detection cells.

An Algorithm for Secure Visible Logo Embedding and Removing in Compression Domain

Digital watermarking is the process of embedding information into a digital signal which can be used in DRM (digital rights managements) system. The visible watermark (often called logo) can indicate the owner of the copyright which can often be seen in the TV program and protects the copyright in an active way. However, most of the schemes do not consider the visible watermark removing process. To solve this problem, a visible watermarking scheme with embedding and removing process is proposed under the control of a secure template. The template generates different version of watermarks which can be seen visually the same for different users. Users with the right key can completely remove the watermark and recover the original image while the unauthorized user is prevented to remove the watermark. Experiment results show that our watermarking algorithm obtains a good visual quality and is hard to be removed by the illegally users. Additionally, the authorized users can completely remove the visible watermark and recover the original image with a good quality.

Model Predictive Fuzzy Control of Air-ratio for Automotive Engines

Automotive engine air-ratio plays an important role of emissions and fuel consumption reduction while maintains satisfactory engine power among all of the engine control variables. In order to effectively control the air-ratio, this paper presents a model predictive fuzzy control algorithm based on online least-squares support vector machines prediction model and fuzzy logic optimizer. The proposed control algorithm was also implemented on a real car for testing and the results are highly satisfactory. Experimental results show that the proposed control algorithm can regulate the engine air-ratio to the stoichiometric value, 1.0, under external disturbance with less than 5% tolerance.

Estimation of Buffer Size of Internet Gateway Server via G/M/1 Queuing Model

How to efficiently assign system resource to route the Client demand by Gateway servers is a tricky predicament. In this paper, we tender an enhanced proposal for autonomous recital of Gateway servers under highly vibrant traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.

Performance Trade-Off of File System between Overwriting and Dynamic Relocation on a Solid State Drive

Most file systems overwrite modified file data and metadata in their original locations, while the Log-structured File System (LFS) dynamically relocates them to other locations. We design and implement the Evergreen file system that can select between overwriting or relocation for each block of a file or metadata. Therefore, the Evergreen file system can achieve superior write performance by sequentializing write requests (similar to LFS-style relocation) when space utilization is low and overwriting when utilization is high. Another challenging issue is identifying performance benefits of LFS-style relocation over overwriting on a newly introduced SSD (Solid State Drive) which has only Flash-memory chips and control circuits without mechanical parts. Our experimental results measured on a SSD show that relocation outperforms overwriting when space utilization is below 80% and vice versa.

CNC Wire-Cut Parameter Optimized Determination of the Stair Shape Workpiece

The objective of this research is parameters optimized of the stair shape workpiece which is cut by CNC Wire-Cut EDM (WEDW). The experiment material is SKD-11 steel of stair-shaped with variable height workpiece 10, 20, 30 and 40 mm. with the same 10 mm. thickness are cut by Sodick's CNC Wire-Cut EDM model AD325L. The experiments are designed by 3k full factorial experimental design at 3 level 2 factors and 9 experiments with 2 replicate. The selected two factor are servo voltage (SV) and servo feed rate (SF) and the response is cutting thickness error. The experiment is divided in two experiments. The first experiment determines the significant effective factor at confidential interval 95%. The SV factor is the significant effective factor from first result. In order to result smallest cutting thickness error of workpieces is 17 micron with the SV value is 46 volt. Also show that the lower SV value, the smaller different thickness error of workpiece. Then the second experiment is done to reduce different cutting thickness error of workpiece as small as possible by lower SV. The second experiment result show the significant effective factor at confidential interval 95% is the SV factor and the smallest cutting thickness error of workpieces reduce to 11 micron with the experiment SV value is 36 volt.

Hydrodynamic Modeling of a Surface Water Treatment Pilot Plant

A mathematical model for the hydrodynamics of a surface water treatment pilot plant was developed and validated by the determination of the residence time distribution (RTD) for the main equipments of the unit. The well known models of ideal/real mixing, ideal displacement (plug flow) and (one-dimensional axial) dispersion model were combined in order to identify the structure that gives the best fitting of the experimental data for each equipment of the pilot plant. RTD experimental results have shown that pilot plant hydrodynamics can be quite well approximated by a combination of simple mathematical models, structure which is suitable for engineering applications. Validated hydrodynamic models will be further used in the evaluation and selection of the most suitable coagulation-flocculation reagents, optimum operating conditions (injection point, reaction times, etc.), in order to improve the quality of the drinking water.

A Study on Fuzzy Adaptive Control of Enteral Feeding Pump

Recent medical studies have investigated the importance of enteral feeding and the use of feeding pumps for recovering patients unable to feed themselves or gain nourishment and nutrients by natural means. The most of enteral feeding system uses a peristaltic tube pump. A peristaltic pump is a form of positive displacement pump in which a flexible tube is progressively squeezed externally to allow the resulting enclosed pillow of fluid to progress along it. The squeezing of the tube requires a precise and robust controller of the geared motor to overcome parametric uncertainty of the pumping system which generates due to a wide variation of friction and slip between tube and roller. So, this paper proposes fuzzy adaptive controller for the robust control of the peristaltic tube pump. This new adaptive controller uses a fuzzy multi-layered architecture which has several independent fuzzy controllers in parallel, each with different robust stability area. Out of several independent fuzzy controllers, the most suited one is selected by a system identifier which observes variations in the controlled system parameter. This paper proposes a design procedure which can be carried out mathematically and systematically from the model of a controlled system. Finally, the good control performance, accurate dose rate and robust system stability, of the developed feeding pump is confirmed through experimental and clinic testing.

Analysis of Social Network Using Clever Ant Colony Metaphor

A social network is a set of people or organization or other social entities connected by some form of relationships. Analysis of social network broadly elaborates visual and mathematical representation of that relationship. Web can also be considered as a social network. This paper presents an innovative approach to analyze a social network using a variant of existing ant colony optimization algorithm called as Clever Ant Colony Metaphor. Experiments are performed and interesting findings and observations have been inferred based on the proposed model.

Using Simulation for Prediction of Units Movements in Case of Communication Failure

Command and Control (C2) system and its interfacethe Common Operational Picture (COP) are main means that supports commander in its decision making process. COP contains information about friendly and enemy unit positions. The friendly position is gathered via tactical network. In the case of tactical network failure the information about units are not available. The tactical simulator can be used as a tool that is capable to predict movements of units in respect of terrain features. Article deals with an experiment that was based on Czech C2 system that is in the case of connectivity lost fed by VR Forces simulator. Article analyzes maximum time interval in which the position created by simulator is still usable and truthful for commander in real time.

Experimental Results about the Dynamics of the Generalized Belief Propagation Used on LDPC Codes

In the context of channel coding, the Generalized Belief Propagation (GBP) is an iterative algorithm used to recover the transmission bits sent through a noisy channel. To ensure a reliable transmission, we apply a map on the bits, that is called a code. This code induces artificial correlations between the bits to send, and it can be modeled by a graph whose nodes are the bits and the edges are the correlations. This graph, called Tanner graph, is used for most of the decoding algorithms like Belief Propagation or Gallager-B. The GBP is based on a non unic transformation of the Tanner graph into a so called region-graph. A clear advantage of the GBP over the other algorithms is the freedom in the construction of this graph. In this article, we explain a particular construction for specific graph topologies that involves relevant performance of the GBP. Moreover, we investigate the behavior of the GBP considered as a dynamic system in order to understand the way it evolves in terms of the time and in terms of the noise power of the channel. To this end we make use of classical measures and we introduce a new measure called the hyperspheres method that enables to know the size of the attractors.

Osmotic Dehydration of Beetroot in Salt Solution: Optimization of Parameters through Statistical Experimental Design

Response surface methodology was used for quantitative investigation of water and solids transfer during osmotic dehydration of beetroot in aqueous solution of salt. Effects of temperature (25 – 45oC), processing time (30–150 min), salt concentration (5–25%, w/w) and solution to sample ratio (5:1 – 25:1) on osmotic dehydration of beetroot were estimated. Quadratic regression equations describing the effects of these factors on the water loss and solids gain were developed. It was found that effects of temperature and salt concentrations were more significant on the water loss than the effects of processing time and solution to sample ratio. As for solids gain processing time and salt concentration were the most significant factors. The osmotic dehydration process was optimized for water loss, solute gain, and weight reduction. The optimum conditions were found to be: temperature – 35oC, processing time – 90 min, salt concentration – 14.31% and solution to sample ratio 8.5:1. At these optimum values, water loss, solid gain and weight reduction were found to be 30.86 (g/100 g initial sample), 9.43 (g/100 g initial sample) and 21.43 (g/100 g initial sample) respectively.

Comprehensive Study on the Linear Hydrodynamic Analysis of a Truss Spar in Random Waves

Truss spars are used for oil exploitation in deep and ultra-deep water if storage crude oil is not needed. The linear hydrodynamic analysis of truss spar in random sea wave load is necessary for determining the behaviour of truss spar. This understanding is not only important for design of the mooring lines, but also for optimising the truss spar design. In this paper linear hydrodynamic analysis of truss spar is carried out in frequency domain. The hydrodynamic forces are calculated using the modified Morison equation and diffraction theory. Added mass and drag coefficients of truss section computed by transmission matrix and normal acceleration and velocity component acting on each element and for hull section computed by strip theory. The stiffness properties of the truss spar can be separated into two components; hydrostatic stiffness and mooring line stiffness. Then, platform response amplitudes obtained by solved the equation of motion. This equation is non-linear due to viscous damping term therefore linearised by iteration method [1]. Finally computed RAOs and significant response amplitude and results are compared with experimental data.

Experimental Parallel Architecture for Rendering 3D Model into MPEG-4 Format

This paper will present the initial findings of a research into distributed computer rendering. The goal of the research is to create a distributed computer system capable of rendering a 3D model into an MPEG-4 stream. This paper outlines the initial design, software architecture and hardware setup for the system. Distributed computing means designing and implementing programs that run on two or more interconnected computing systems. Distributed computing is often used to speed up the rendering of graphical imaging. Distributed computing systems are used to generate images for movies, games and simulations. A topic of interest is the application of distributed computing to the MPEG-4 standard. During the course of the research, a distributed system will be created that can render a 3D model into an MPEG-4 stream. It is expected that applying distributed computing principals will speed up rendering, thus improving the usefulness and efficiency of the MPEG-4 standard