Significance of Splitting Method in Non-linear Grid system for the Solution of Navier-Stokes Equation

Solution to unsteady Navier-Stokes equation by Splitting method in physical orthogonal algebraic curvilinear coordinate system, also termed 'Non-linear grid system' is presented. The linear terms in Navier-Stokes equation are solved by Crank- Nicholson method while the non-linear term is solved by the second order Adams-Bashforth method. This work is meant to bring together the advantage of Splitting method as pressure-velocity solver of higher efficiency with the advantage of consuming Non-linear grid system which produce more accurate results in relatively equal number of grid points as compared to Cartesian grid. The validation of Splitting method as a solution of Navier-Stokes equation in Nonlinear grid system is done by comparison with the benchmark results for lid driven cavity flow by Ghia and some case studies including Backward Facing Step Flow Problem.

Numerical Simulation of the Turbulent Flow over a Three-Dimensional Flat Roof

The flow field over a flat roof model building has been numerically investigated in order to determine threedimensional CFD guidelines for the calculation of the turbulent flow over a structure immersed in an atmospheric boundary layer. To this purpose, a complete validation campaign has been performed through a systematic comparison of numerical simulations with wind tunnel experimental data. Wind tunnel measurements and numerical predictions have been compared for five different vertical positions, respectively from the upstream leading edge to the downstream bottom edge of the analyzed model. Flow field characteristics in the neighborhood of the building model have been numerically investigated, allowing a quantification of the capabilities of the CFD code to predict the flow separation and the extension of the recirculation regions. The proposed calculations have allowed the development of a preliminary procedure to be used as guidance in selecting the appropriate grid configuration and corresponding turbulence model for the prediction of the flow field over a three-dimensional roof architecture dominated by flow separation.

Design Process and Real-Time Validation of an Innovative Autonomous Mid-Air Flight and Landing System

This paper describes the design process and the realtime validation of an innovative autonomous mid-air flight and landing system developed by the Italian Aerospace Research Center in the framework of the Italian national funded project TECVOL (Technologies for the Autonomous Flight). In the paper it is provided an insight of the whole development process of the system under study. In particular, the project framework is illustrated at first, then the functional context and the adopted design and testing approach are described, and finally the on-ground validation test rig on purpose designed is addressed in details. Furthermore, the hardwarein- the-loop validation of the autonomous mid-air flight and landing system by means of the real-time test rig is described and discussed.

Dynamic Clustering using Particle Swarm Optimization with Application in Unsupervised Image Classification

A new dynamic clustering approach (DCPSO), based on Particle Swarm Optimization, is proposed. This approach is applied to unsupervised image classification. The proposed approach automatically determines the "optimum" number of clusters and simultaneously clusters the data set with minimal user interference. The algorithm starts by partitioning the data set into a relatively large number of clusters to reduce the effects of initial conditions. Using binary particle swarm optimization the "best" number of clusters is selected. The centers of the chosen clusters is then refined via the Kmeans clustering algorithm. The experiments conducted show that the proposed approach generally found the "optimum" number of clusters on the tested images.

Passive Flow Control in Twin Air-Intakes

Aircraft propulsion systems often use Y-shaped subsonic diffusing ducts as twin air-intakes to supply the ambient air into the engine compressor for thrust generation. Due to space constraint, the diffusers need to be curved, which causes severe flow non-uniformity at the engine face. The present study attempt to control flow in a mild-curved Y-duct diffuser using trapezoidalshaped vortex generators (VG) attached on either both the sidewalls or top and bottom walls of the diffuser at the inflexion plane. A commercial computational fluid dynamics (CFD) code is modified and is used to simulate the effects of SVG in flow of a Y-duct diffuser. A few experiments are conducted for CFD code validation, while the rest are done computationally. The best combination of Yduct diffuser is found with VG-2 arranged in co-rotating sequence and attached to both the sidewalls, which ensures highest static pressure recovery, lowest total pressure loss, minimum flow distortion and less flow separation in Y-duct diffuser. The decrease in VG height while attached to top and bottom walls further improves axial flow uniformity at the diffuser outlet by a great margin as compared to the bare duct.

Experimental Modal Analysis and Model Validation of Antenna Structures

Numerical design optimization is a powerful tool that can be used by engineers during any stage of the design process. There are many different applications for structural optimization. A specific application that will be discussed in the following paper is experimental data matching. Data obtained through tests on a physical structure will be matched with data from a numerical model of that same structure. The data of interest will be the dynamic characteristics of an antenna structure focusing on the mode shapes and modal frequencies. The structure used was a scaled and simplified model of the Karoo Array Telescope-7 (KAT-7) antenna structure. This kind of data matching is a complex and difficult task. This paper discusses how optimization can assist an engineer during the process of correlating a finite element model with vibration test data.

Evaluating the Tool Wear Rate in Ultrasonic Machining of Titanium using Design of Experiments Approach

Ultrasonic machining (USM) is a non-traditional machining process being widely used for commercial machining of brittle and fragile materials such as glass, ceramics and semiconductor materials. However, USM could be a viable alternative for machining a tough material such as titanium; and this aspect needs to be explored through experimental research. This investigation is focused on exploring the use of ultrasonic machining for commercial machining of pure titanium (ASTM Grade-I) and evaluation of tool wear rate (TWR) under controlled experimental conditions. The optimal settings of parameters are determined through experiments planned, conducted and analyzed using Taguchi method. In all, the paper focuses on parametric optimization of ultrasonic machining of pure titanium metal with TWR as response, and validation of the optimized value of TWR by conducting confirmatory experiments.

A Traffic Simulation Package Based on Travel Demand

In this paper we propose a new traffic simulation package, TDMSim, which supports both macroscopic and microscopic simulation on free-flowing and regulated traffic systems. Both simulators are based on travel demands, which specify the numbers of vehicles departing from origins to arrive at different destinations. The microscopic simulator implements the carfollowing model given the pre-defined routes of the vehicles but also supports the rerouting of vehicles. We also propose a macroscopic simulator which is built in integration with the microscopic simulator to allow the simulation to be scaled for larger networks without sacrificing the precision achievable through the microscopic simulator. The macroscopic simulator also enables the reuse of previous simulation results when simulating traffic on the same networks at later time. Validations have been conducted to show the correctness of both simulators.

Integrating Decision Tree and Spatial Cluster Analysis for Landslide Susceptibility Zonation

Landslide susceptibility map delineates the potential zones for landslide occurrence. Previous works have applied multivariate methods and neural networks for mapping landslide susceptibility. This study proposed a new approach to integrate decision tree model and spatial cluster statistic for assessing landslide susceptibility spatially. A total of 2057 landslide cells were digitized for developing the landslide decision tree model. The relationships of landslides and instability factors were explicitly represented by using tree graphs in the model. The local Getis-Ord statistics were used to cluster cells with high landslide probability. The analytic result from the local Getis-Ord statistics was classed to create a map of landslide susceptibility zones. The map was validated using new landslide data with 482 cells. Results of validation show an accuracy rate of 86.1% in predicting new landslide occurrence. This indicates that the proposed approach is useful for improving landslide susceptibility mapping.

Numerical Analysis and Experimental Validation of a Downhole Stress/Strain Measurement Tool

Real-time measurement of applied forces, like tension, compression, torsion, and bending moment, identifies the transferred energies being applied to the bottomhole assembly (BHA). These forces are highly detrimental to measurement/logging-while-drilling tools and downhole equipment. Real-time measurement of the dynamic downhole behavior, including weight, torque, bending on bit, and vibration, establishes a real-time feedback loop between the downhole drilling system and drilling team at the surface. This paper describes the numerical analysis of the strain data acquired by the measurement tool at different locations on the strain pockets. The strain values obtained by FEA for various loading conditions (tension, compression, torque, and bending moment) are compared against experimental results obtained from an identical experimental setup. Numerical analyses results agree with experimental data within 8% and, therefore, substantiate and validate the FEA model. This FEA model can be used to analyze the combined loading conditions that reflect the actual drilling environment.

Optimal Allocation of FACTS Devices for ATC Enhancement Using Bees Algorithm

In this paper, a novel method using Bees Algorithm is proposed to determine the optimal allocation of FACTS devices for maximizing the Available Transfer Capability (ATC) of power transactions between source and sink areas in the deregulated power system. The algorithm simultaneously searches the FACTS location, FACTS parameters and FACTS types. Two types of FACTS are simulated in this study namely Thyristor Controlled Series Compensator (TCSC) and Static Var Compensator (SVC). A Repeated Power Flow with FACTS devices including ATC is used to evaluate the feasible ATC value within real and reactive power generation limits, line thermal limits, voltage limits and FACTS operation limits. An IEEE30 bus system is used to demonstrate the effectiveness of the algorithm as an optimization tool to enhance ATC. A Genetic Algorithm technique is used for validation purposes. The results clearly indicate that the introduction of FACTS devices in a right combination of location and parameters could enhance ATC and Bees Algorithm can be efficiently used for this kind of nonlinear integer optimization.

Suggestion of Ultrasonic System for Diagnosis of Functional Gastrointestinal Disorders: Finite Difference Analysis, Development and Clinical Trials

The disaster from functional gastrointestinal disorders has detrimental impact on the quality of life of the effected population and imposes a tremendous social and economic burden. There are, however, rare diagnostic methods for the functional gastrointestinal disorders. Our research group identified recently that the gastrointestinal tract well in the patients with the functional gastrointestinal disorders becomes more rigid than healthy people when palpating the abdominal regions overlaying the gastrointestinal tract. Objective of current study is, therefore, identify feasibility of a diagnostic system for the functional gastrointestinal disorders based on ultrasound technique, which can quantify the characteristics above. Two-dimensional finite difference (FD) models (one normal and two rigid model) were developed to analyze the reflective characteristic (displacement) on each soft-tissue layer responded after application of ultrasound signals. The FD analysis was then based on elastic ultrasound theory. Validation of the model was performed via comparison of the characteristic of the ultrasonic responses predicted by FD analysis with that determined from the actual specimens for the normal and rigid conditions. Based on the results from FD analysis, ultrasound system for diagnosis of the functional gastrointestinal disorders was developed and clinically tested via application of it to 40 human subjects with/without functional gastrointestinal disorders who were assigned to Normal and Patient Groups. The FD models were favorably validated. The results from FD analysis showed that the maximum displacement amplitude in the rigid models (0.12 and 0.16) at the interface between the fat and muscle layers was explicitly less than that in the normal model (0.29). The results from actual specimens showed that the maximum amplitude of the ultrasonic reflective signal in the rigid models (0.2±0.1Vp-p) at the interface between the fat and muscle layers was explicitly higher than that in the normal model (0.1±0.2 Vp-p). Clinical tests using our customized ultrasound system showed that the maximum amplitudes of the ultrasonic reflective signals near to the gastrointestinal tract well for the patient group (2.6±0.3 Vp-p) were generally higher than those in normal group (0.1±0.2 Vp-p). Here, maximum reflective signals was appeared at 20mm depth approximately from abdominal skin for all human subjects, corresponding to the location of the boundary layer close to gastrointestinal tract well. These findings suggest that our customized ultrasound system using the ultrasonic reflective signal may be helpful to the diagnosis of the functional gastrointestinal disorders.

A Comparison of Some Thresholding Selection Methods for Wavelet Regression

In wavelet regression, choosing threshold value is a crucial issue. A too large value cuts too many coefficients resulting in over smoothing. Conversely, a too small threshold value allows many coefficients to be included in reconstruction, giving a wiggly estimate which result in under smoothing. However, the proper choice of threshold can be considered as a careful balance of these principles. This paper gives a very brief introduction to some thresholding selection methods. These methods include: Universal, Sure, Ebays, Two fold cross validation and level dependent cross validation. A simulation study on a variety of sample sizes, test functions, signal-to-noise ratios is conducted to compare their numerical performances using three different noise structures. For Gaussian noise, EBayes outperforms in all cases for all used functions while Two fold cross validation provides the best results in the case of long tail noise. For large values of signal-to-noise ratios, level dependent cross validation works well under correlated noises case. As expected, increasing both sample size and level of signal to noise ratio, increases estimation efficiency.

Computing Entropy for Ortholog Detection

Biological sequences from different species are called or-thologs if they evolved from a sequence of a common ancestor species and they have the same biological function. Approximations of Kolmogorov complexity or entropy of biological sequences are already well known to be useful in extracting similarity information between such sequences -in the interest, for example, of ortholog detection. As is well known, the exact Kolmogorov complexity is not algorithmically computable. In prac-tice one can approximate it by computable compression methods. How-ever, such compression methods do not provide a good approximation to Kolmogorov complexity for short sequences. Herein is suggested a new ap-proach to overcome the problem that compression approximations may notwork well on short sequences. This approach is inspired by new, conditional computations of Kolmogorov entropy. A main contribution of the empir-ical work described shows the new set of entropy-based machine learning attributes provides good separation between positive (ortholog) and nega-tive (non-ortholog) data - better than with good, previously known alter-natives (which do not employ some means to handle short sequences well).Also empirically compared are the new entropy based attribute set and a number of other, more standard similarity attributes sets commonly used in genomic analysis. The various similarity attributes are evaluated by cross validation, through boosted decision tree induction C5.0, and by Receiver Operating Characteristic (ROC) analysis. The results point to the conclu-sion: the new, entropy based attribute set by itself is not the one giving the best prediction; however, it is the best attribute set for use in improving the other, standard attribute sets when conjoined with them.

Complexity of Mathematical Expressions in Adaptive Multimodal Multimedia System Ensuring Access to Mathematics for Visually Impaired Users

Our adaptive multimodal system aims at correctly presenting a mathematical expression to visually impaired users. Given an interaction context (i.e. combination of user, environment and system resources) as well as the complexity of the expression itself and the user-s preferences, the suitability scores of different presentation format are calculated. Unlike the current state-of-the art solutions, our approach takes into account the user-s situation and not imposes a solution that is not suitable to his context and capacity. In this wok, we present our methodology for calculating the mathematical expression complexity and the results of our experiment. Finally, this paper discusses the concepts and principles applied on our system as well as their validation through cases studies. This work is our original contribution to an ongoing research to make informatics more accessible to handicapped users.

Dynamic Modeling and Simulation of Threephase Small Power Induction Motor

This paper is proposed the dynamic simulation of small power induction motor based on Mathematical modeling. The dynamic simulation is one of the key steps in the validation of the design process of the motor drive systems and it is needed for eliminating inadvertent design mistakes and the resulting error in the prototype construction and testing. This paper demonstrates the simulation of steady-state performance of induction motor by MATLAB Program Three phase 3 hp induction motor is modeled and simulated with SIMULINK model.

Model Based Monitoring Using Integrated Data Validation, Simulation and Parameter Estimation

Efficient and safe plant operation can only be achieved if the operators are able to monitor all key process parameters. Instrumentation is used to measure many process variables, like temperatures, pressures, flow rates, compositions or other product properties. Therefore Performance monitoring is a suitable tool for operators. In this paper, we integrate rigorous simulation model, data reconciliation and parameter estimation to monitor process equipments and determine key performance indicator (KPI) of them. The applied method here has been implemented in two case studies.

Signature Recognition and Verification using Hybrid Features and Clustered Artificial Neural Network(ANN)s

Signature represents an individual characteristic of a person which can be used for his / her validation. For such application proper modeling is essential. Here we propose an offline signature recognition and verification scheme which is based on extraction of several features including one hybrid set from the input signature and compare them with the already trained forms. Feature points are classified using statistical parameters like mean and variance. The scanned signature is normalized in slant using a very simple algorithm with an intention to make the system robust which is found to be very helpful. The slant correction is further aided by the use of an Artificial Neural Network (ANN). The suggested scheme discriminates between originals and forged signatures from simple and random forgeries. The primary objective is to reduce the two crucial parameters-False Acceptance Rate (FAR) and False Rejection Rate (FRR) with lesser training time with an intension to make the system dynamic using a cluster of ANNs forming a multiple classifier system.

Experimental Validation of Treatment Planning for Multiple Radiotherapy Fields by EDR2 Film Dosimeter

To investigate the applicability of the EDR-2 film for clinical radiation dosimetry, percentage depth-doses, profiles and distributions in open and dynamically wedged fields were measured using film and compared with data from a Treatment Planning system.The validity of the EDR2 film to measure dose in a plane parallel to the beam was tested by irradiating 10 cm×10 cm and 4 cm×4 cm fields from a Siemens, primus linac with a 6MV beam and a source-to-surface distance of 100 cm. The film was placed Horizontally between solid water phantom blocks and marked with pin holes at a depth of 10 cm from the incident beam surface. The film measurement results, in absolute dose, were compared with ion chamber measurements using a Welhoffer scanning water tank system and Treatment Planning system. Our results indicate a maximum underestimate of calculated dose of 8 % with Treatment Planning system.

Adaptation of State/Transition-Based Methods for Embedded System Testing

In this paper test generation methods and appropriate fault models for testing and analysis of embedded systems described as (extended) finite state machines ((E)FSMs) are presented. Compared to simple FSMs, EFSMs specify not only the control flow but also the data flow. Thus, we define a two-level fault model to cover both aspects. The goal of this paper is to reuse well-known FSM-based test generation methods for automation of embedded system testing. These methods have been widely used in testing and validation of protocols and communicating systems. In particular, (E)FSMs-based specification and testing is more advantageous because (E)FSMs support the formal semantic of already standardised formal description techniques (FDTs) despite of their popularity in the design of hardware and software systems.