Development and Optimization of Automated Dry-Wafer Separation

In a state-of-the-art industrial production line of photovoltaic products the handling and automation processes are of particular importance and implication. While processing a fully functional crystalline solar cell an as-cut photovoltaic wafer is subject to numerous repeated handling steps. With respect to stronger requirements in productivity and decreasing rejections due to defects the mechanical stress on the thin wafers has to be reduced to a minimum as the fragility increases by decreasing wafer thicknesses. In relation to the increasing wafer fragility, researches at the Fraunhofer Institutes IPA and CSP showed a negative correlation between multiple handling processes and the wafer integrity. Recent work therefore focused on the analysis and optimization of the dry wafer stack separation process with compressed air. The achievement of a wafer sensitive process capability and a high production throughput rate is the basic motivation in this research.

Ordinary Differential Equations with Inverted Functions

Equations with differentials relating to the inverse of an unknown function rather than to the unknown function itself are solved exactly for some special cases and numerically for the general case. Invertibility combined with differentiability over connected domains forces solutions always to be monotone. Numerical function inversion is key to all solution algorithms which either are of a forward type or a fixed point type considering whole approximate solution functions in each iteration. The given considerations are restricted to ordinary differential equations with inverted functions (ODEIs) of first order. Forward type computations, if applicable, admit consistency of order one and, under an additional accuracy condition, convergence of order one.

Web Application Security, Attacks and Mitigation

Today’s technology is heavily dependent on web applications. Web applications are being accepted by users at a very rapid pace. These have made our work efficient. These include webmail, online retail sale, online gaming, wikis, departure and arrival of trains and flights and list is very long. These are developed in different languages like PHP, Python, C#, ASP.NET and many more by using scripts such as HTML and JavaScript. Attackers develop tools and techniques to exploit web applications and legitimate websites. This has led to rise of web application security; which can be broadly classified into Declarative Security and Program Security. The most common attacks on the applications are by SQL Injection and XSS which give access to unauthorized users who totally damage or destroy the system. This paper presents a detailed literature description and analysis on Web Application Security, examples of attacks and steps to mitigate the vulnerabilities.

Matrix Based Synthesis of EXOR dominated Combinational Logic for Low Power

This paper discusses a new, systematic approach to the synthesis of a NP-hard class of non-regenerative Boolean networks, described by FON[FOFF]={mi}[{Mi}], where for every mj[Mj]∈{mi}[{Mi}], there exists another mk[Mk]∈{mi}[{Mi}], such that their Hamming distance HD(mj, mk)=HD(Mj, Mk)=O(n), (where 'n' represents the number of distinct primary inputs). The method automatically ensures exact minimization for certain important selfdual functions with 2n-1 points in its one-set. The elements meant for grouping are determined from a newly proposed weighted incidence matrix. Then the binary value corresponding to the candidate pair is correlated with the proposed binary value matrix to enable direct synthesis. We recommend algebraic factorization operations as a post processing step to enable reduction in literal count. The algorithm can be implemented in any high level language and achieves best cost optimization for the problem dealt with, irrespective of the number of inputs. For other cases, the method is iterated to subsequently reduce it to a problem of O(n-1), O(n-2),.... and then solved. In addition, it leads to optimal results for problems exhibiting higher degree of adjacency, with a different interpretation of the heuristic, and the results are comparable with other methods. In terms of literal cost, at the technology independent stage, the circuits synthesized using our algorithm enabled net savings over AOI (AND-OR-Invert) logic, AND-EXOR logic (EXOR Sum-of- Products or ESOP forms) and AND-OR-EXOR logic by 45.57%, 41.78% and 41.78% respectively for the various problems. Circuit level simulations were performed for a wide variety of case studies at 3.3V and 2.5V supply to validate the performance of the proposed method and the quality of the resulting synthesized circuits at two different voltage corners. Power estimation was carried out for a 0.35micron TSMC CMOS process technology. In comparison with AOI logic, the proposed method enabled mean savings in power by 42.46%. With respect to AND-EXOR logic, the proposed method yielded power savings to the tune of 31.88%, while in comparison with AND-OR-EXOR level networks; average power savings of 33.23% was obtained.

A Black-Box Approach in Modeling Valve Stiction

Several valve stiction models have been proposed in the literature to help understand and study the behavior of sticky valves. In this paper, an alternative black-box modeling approach based on Neural Network (NN) is presented. It is shown that with proper network type and optimum model structures, the performance of the developed NN stiction model is comparable to other established method. The resulting NN model is also tested for its robustness against the uncertainty in the stiction parameter values. Predictive mode operation also shows excellent performance of the proposed model for multi-steps ahead prediction.

Three Steps of One-way Nested Grid for Energy Balance Equations by Wave Model

The three steps of the standard one-way nested grid for a regional scale of the third generation WAve Model Cycle 4 (WAMC4) is scrutinized. The model application is enabled to solve the energy balance equation on a coarse resolution grid in order to produce boundary conditions for a smaller area by the nested grid technique. In the present study, the model takes a full advantage of the fine resolution of wind fields in space and time produced by the available U.S. Navy Global Atmospheric Prediction System (NOGAPS) model with 1 degree resolution. The nested grid application of the model is developed in order to gradually increase the resolution from the open ocean towards the South China Sea (SCS) and the Gulf of Thailand (GoT) respectively. The model results were compared with buoy observations at Ko Chang, Rayong and Huahin locations which were obtained from the Seawatch project. In addition, the results were also compared with Satun based weather station which was provided from Department of Meteorology, Thailand. The data collected from this station presented the significant wave height (Hs) reached 12.85 m. The results indicated that the tendency of the Hs from the model in the spherical coordinate propagation with deep water condition in the fine grid domain agreed well with the Hs from the observations.

A Middleware Transparent Framework for Applying MDA to SOA

Although Model Driven Architecture has taken successful steps toward model-based software development, this approach still faces complex situations and ambiguous questions while applying to real world software systems. One of these questions - which has taken the most interest and focus - is how model transforms between different abstraction levels, MDA proposes. In this paper, we propose an approach based on Story Driven Modeling and Aspect Oriented Programming to ease these transformations. Service Oriented Architecture is taken as the target model to test the proposed mechanism in a functional system. Service Oriented Architecture and Model Driven Architecture [1] are both considered as the frontiers of their own domain in the software world. Following components - which was the greatest step after object oriented - SOA is introduced, focusing on more integrated and automated software solutions. On the other hand - and from the designers' point of view - MDA is just initiating another evolution. MDA is considered as the next big step after UML in designing domain.

Stochastic Simulation of Reaction-Diffusion Systems

Reactiondiffusion systems are mathematical models that describe how the concentration of one or more substances distributed in space changes under the influence of local chemical reactions in which the substances are converted into each other, and diffusion which causes the substances to spread out in space. The classical representation of a reaction-diffusion system is given by semi-linear parabolic partial differential equations, whose general form is ÔêétX(x, t) = DΔX(x, t), where X(x, t) is the state vector, D is the matrix of the diffusion coefficients and Δ is the Laplace operator. If the solute move in an homogeneous system in thermal equilibrium, the diffusion coefficients are constants that do not depend on the local concentration of solvent and of solutes and on local temperature of the medium. In this paper a new stochastic reaction-diffusion model in which the diffusion coefficients are function of the local concentration, viscosity and frictional forces of solvent and solute is presented. Such a model provides a more realistic description of the molecular kinetics in non-homogenoeus and highly structured media as the intra- and inter-cellular spaces. The movement of a molecule A from a region i to a region j of the space is described as a first order reaction Ai k- → Aj , where the rate constant k depends on the diffusion coefficient. Representing the diffusional motion as a chemical reaction allows to assimilate a reaction-diffusion system to a pure reaction system and to simulate it with Gillespie-inspired stochastic simulation algorithms. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the specific speed of reaction and diffusion events. Redi is the software tool, developed to implement the model of reaction-diffusion kinetics and dynamics. It is a free software, that can be downloaded from http://www.cosbi.eu. To demonstrate the validity of the new reaction-diffusion model, the simulation results of the chaperone-assisted protein folding in cytoplasm obtained with Redi are reported. This case study is redrawing the attention of the scientific community due to current interests on protein aggregation as a potential cause for neurodegenerative diseases.

Ultra-Light Overhead Conveyor Systems for Logistics Applications

Overhead conveyor systems satisfy by their simple construction, wide application range and their full compatibility with other manufacturing systems, which are designed according to international standards. Ultra-light overhead conveyor systems are rope-based conveying systems with individually driven vehicles. The vehicles can move automatically on the rope and this can be realized by energy and signals. Crossings are realized by switches. Overhead conveyor systems are particularly used in the automotive industry but also at post offices. Overhead conveyor systems always must be integrated with a logistical process by finding the best way for a cheaper material flow and in order to guarantee precise and fast workflows. With their help, any transport can take place without wasting ground and space, without excessive company capacity, lost or damaged products, erroneous delivery, endless travels and without wasting time. Ultra-light overhead conveyor systems provide optimal material flow, which produces profit and saves time. This article illustrates the advantages of the structure of the ultra-light overhead conveyor systems in logistics applications and explains the steps of their system design. After an illustration of the steps, currently available systems on the market will be shown by means of their technical characteristics. Due to their simple construction, demands to an ultra-light overhead conveyor system will be illustrated.

Some Issues on Integrating Telepresence Technology into Industrial Robotic Assembly

Since the 1940s, many promising telepresence research results have been obtained. However, telepresence technology still has not reached industrial usage. As human intelligence is necessary for successful execution of most manual assembly tasks, the ability of the human is hindered in some cases, such as the assembly of heavy parts of small/medium lots or prototypes. In such a case of manual assembly, the help of industrial robots is mandatory. The telepresence technology can be considered as a solution for performing assembly tasks, where the human intelligence and haptic sense are needed to identify and minimize the errors during an assembly process and a robot is needed to carry heavy parts. In this paper, preliminary steps to integrate the telepresence technology into industrial robot systems are introduced. The system described here combines both, the human haptic sense and the industrial robot capability to perform a manual assembly task remotely using a force feedback joystick. Mapping between the joystick-s Degrees of Freedom (DOF) and the robot-s ones are introduced. Simulation and experimental results are shown and future work is discussed.

Real Time Speed Estimation of Vehicles

this paper gives a novel approach towards real-time speed estimation of multiple traffic vehicles using fuzzy logic and image processing techniques with proper arrangement of camera parameters. The described algorithm consists of several important steps. First, the background is estimated by computing median over time window of specific frames. Second, the foreground is extracted using fuzzy similarity approach (FSA) between estimated background pixels and the current frame pixels containing foreground and background. Third, the traffic lanes are divided into two parts for both direction vehicles for parallel processing. Finally, the speeds of vehicles are estimated by Maximum a Posterior Probability (MAP) estimator. True ground speed is determined by utilizing infrared sensors for three different vehicles and the results are compared to the proposed algorithm with an accuracy of ± 0.74 kmph.

Bioethanol - A Viable Answer to India-s Surging Energy Needs

India is currently the second most populous nation in the world with over 1.2 billion people, growing annually at the rate of 1.5%. It is experiencing a surge in energy demands, expected to grow more than three to four times in 25 years. Most of the energy requirements are currently satisfied by the import of fossil fuels – coal, petroleum-based products and natural gas. Biofuels can satisfy these energy needs in an environmentally benign and cost effective manner while reducing dependence on import of fossil fuels, thus providing National Energy Security. Among various forms of bioenergy, bioethanol is one of the major options for India because of availability of feed stock crops. This paper presents an overview on bioethanol production and technology, steps taken by the Indian government to facilitate and bring about optimal development and utilization of indigenous biomass feedstocks for production of this biofuel.

Enhancing Multi-Frame Images Using Self-Delaying Dynamic Networks

This paper presents the use of a newly created network structure known as a Self-Delaying Dynamic Network (SDN) to create a high resolution image from a set of time stepped input frames. These SDNs are non-recurrent temporal neural networks which can process time sampled data. SDNs can store input data for a lifecycle and feature dynamic logic based connections between layers. Several low resolution images and one high resolution image of a scene were presented to the SDN during training by a Genetic Algorithm. The SDN was trained to process the input frames in order to recreate the high resolution image. The trained SDN was then used to enhance a number of unseen noisy image sets. The quality of high resolution images produced by the SDN is compared to that of high resolution images generated using Bi-Cubic interpolation. The SDN produced images are superior in several ways to the images produced using Bi-Cubic interpolation.

Effect Comparison of Speckle Noise Reduction Filters on 2D-Echocardigraphic Images

Echocardiography imaging is one of the most common diagnostic tests that are widely used for assessing the abnormalities of the regional heart ventricle function. The main goal of the image enhancement task in 2D-echocardiography (2DE) is to solve two major anatomical structure problems; speckle noise and low quality. Therefore, speckle noise reduction is one of the important steps that used as a pre-processing to reduce the distortion effects in 2DE image segmentation. In this paper, we present the common filters that based on some form of low-pass spatial smoothing filters such as Mean, Gaussian, and Median. The Laplacian filter was used as a high-pass sharpening filter. A comparative analysis was presented to test the effectiveness of these filters after being applied to original 2DE images of 4-chamber and 2-chamber views. Three statistical quantity measures: root mean square error (RMSE), peak signal-to-ratio (PSNR) and signal-tonoise ratio (SNR) are used to evaluate the filter performance quantitatively on the output enhanced image.

Adaptive Gaussian Mixture Model for Skin Color Segmentation

Skin color based tracking techniques often assume a static skin color model obtained either from an offline set of library images or the first few frames of a video stream. These models can show a weak performance in presence of changing lighting or imaging conditions. We propose an adaptive skin color model based on the Gaussian mixture model to handle the changing conditions. Initial estimation of the number and weights of skin color clusters are obtained using a modified form of the general Expectation maximization algorithm, The model adapts to changes in imaging conditions and refines the model parameters dynamically using spatial and temporal constraints. Experimental results show that the method can be used in effectively tracking of hand and face regions.

A Supervisory Scheme for Step-Wise Safe Switching Controllers

A supervisory scheme is proposed that implements Stepwise Safe Switching Logic. The functionality of the supervisory scheme is organized in the following eight functional units: Step- Wise Safe Switching unit, Common controllers design unit, Experimentation unit, Simulation unit, Identification unit, Trajectory cruise unit, Operating points unit and Expert system unit. The supervisory scheme orchestrates both the off-line preparative actions, as well as the on-line actions that implement the Stepwise Safe Switching Logic. The proposed scheme is a generic tool, that may be easily applied for a variety of industrial control processes and may be implemented as an automation software system, with the use of a high level programming environment, like Matlab.

Finding Sparse Features in Face Detection Using Genetic Algorithms

Although Face detection is not a recent activity in the field of image processing, it is still an open area for research. The greatest step in this field is the work reported by Viola and its recent analogous is Huang et al. Both of them use similar features and also similar training process. The former is just for detecting upright faces, but the latter can detect multi-view faces in still grayscale images using new features called 'sparse feature'. Finding these features is very time consuming and inefficient by proposed methods. Here, we propose a new approach for finding sparse features using a genetic algorithm system. This method requires less computational cost and gets more effective features in learning process for face detection that causes more accuracy.

Modeling of Pulping of Sugar Maple Using Advanced Neural Network Learning

This paper reports work done to improve the modeling of complex processes when only small experimental data sets are available. Neural networks are used to capture the nonlinear underlying phenomena contained in the data set and to partly eliminate the burden of having to specify completely the structure of the model. Two different types of neural networks were used for the application of Pulping of Sugar Maple problem. A three layer feed forward neural networks, using the Preconditioned Conjugate Gradient (PCG) methods were used in this investigation. Preconditioning is a method to improve convergence by lowering the condition number and increasing the eigenvalues clustering. The idea is to solve the modified problem where M is a positive-definite preconditioner that is closely related to A. We mainly focused on Preconditioned Conjugate Gradient- based training methods which originated from optimization theory, namely Preconditioned Conjugate Gradient with Fletcher-Reeves Update (PCGF), Preconditioned Conjugate Gradient with Polak-Ribiere Update (PCGP) and Preconditioned Conjugate Gradient with Powell-Beale Restarts (PCGB). The behavior of the PCG methods in the simulations proved to be robust against phenomenon such as oscillations due to large step size.

Transient Stress Analysis on Medium Modules Spur Gear by Using Mode Super Position Technique

Natural frequencies and dynamic response of a spur gear sector are investigated using a two dimensional finite element model that offers significant advantages for dynamic gear analyses. The gear teeth are analyzed for different operating speeds. A primary feature of this modeling is determination of mesh forces using a detailed contact analysis for each time step as the gears roll through the mesh. ANSYS software has been used on the proposed model to find the natural frequencies by Block Lanczos technique and displacements and dynamic stresses by transient mode super position method. The effect of rotational speed of the gear on the dynamic response of gear tooth has been studied and design limits have been discussed.

High-Speed Train Planning in France, Lessons from Mediterranean TGV-Line

To fight against the economic crisis, French Government, like many others in Europe, has decided to give a boost to high-speed line projects. This paper explores the implementation and decision-making process in TGV projects, their evolutions, especially since the Mediterranean TGV-line. This project was probably the most controversial, but paradoxically represents today a huge success for all the actors involved. What kind of lessons we can learn from this experience? How to evaluate the impact of this project on TGV-line planning? How can we characterize this implementation and decision-making process regards to the sustainability challenges? The construction of Mediterranean TGV-line was the occasion to make several innovations: to introduce more dialog into the decisionmaking process, to take into account the environment, to introduce a new project management and technological innovations. That-s why this project appears today as an example in terms of integration of sustainable development. In this paper we examine the different kinds of innovations developed in this project, by using concepts from sociology of innovation to understand how these solutions emerged in a controversial situation. Then we analyze the lessons which were drawn from this decision-making process (in the immediacy and a posteriori) and the way in which procedures evolved: creation of new tools and devices (public consultation, project management...). Finally we try to highlight the impact of this evolution on TGV projects governance. In particular, new methods of implementation and financing involve a reconfiguration of the system of actors. The aim of this paper is to define the impact of this reconfiguration on negotiations between stakeholders.