Semi-Lagrangian Method for Advection Equation on GPU in Unstructured R3 Mesh for Fluid Dynamics Application

Numerical integration of initial boundary problem for advection equation in 3 ℜ is considered. The method used is  conditionally stable semi-Lagrangian advection scheme with high order interpolation on unstructured mesh. In order to increase time step integration the BFECC method with limiter TVD correction is used. The method is adopted on parallel graphic processor unit environment using NVIDIA CUDA and applied in Navier-Stokes solver. It is shown that the calculation on NVIDIA GeForce 8800  GPU is 184 times faster than on one processor AMDX2 4800+ CPU. The method is extended to the incompressible fluid dynamics solver. Flow over a Cylinder for 3D case is compared to the experimental data.

Renewal of The Swedish Million Dwelling Program, the Public Housing Company and the Local Community, Hindrances and Mutual Aid

Public housing is a vital factor in community development. Successful city, housing and eco system regeneration design is essential in providing positive community development. This concerns work places, nice dwellings, providing premises for child care, care of the elderly, providing qualitative premises for different kinds of commercial service, providing a nice built environment and housing areas and not the least activating tenants. The public housing companies give value to society by stimulating people, renovating socially and economically sustainable as well as being partners to local business and authorities. By their activities the housing companies contribute to sustainable local and regional growth and the identity and reputation of cities. A Social, Economic and Ecological Reputation Effect (SEERE) model for actions to promote housing and community reputation is presented. The model emphasizes regenerative actions to restore natural eco systems as part of housing renewal strategies and to strengthen municipality reputation.

Genetic-Based Multi Resolution Noisy Color Image Segmentation

Segmentation of a color image composed of different kinds of regions can be a hard problem, namely to compute for an exact texture fields. The decision of the optimum number of segmentation areas in an image when it contains similar and/or un stationary texture fields. A novel neighborhood-based segmentation approach is proposed. A genetic algorithm is used in the proposed segment-pass optimization process. In this pass, an energy function, which is defined based on Markov Random Fields, is minimized. In this paper we use an adaptive threshold estimation method for image thresholding in the wavelet domain based on the generalized Gaussian distribution (GGD) modeling of sub band coefficients. This method called Normal Shrink is computationally more efficient and adaptive because the parameters required for estimating the threshold depend on sub band data energy that used in the pre-stage of segmentation. A quad tree is employed to implement the multi resolution framework, which enables the use of different strategies at different resolution levels, and hence, the computation can be accelerated. The experimental results using the proposed segmentation approach are very encouraging.

Target and Kaizen Costing

increased competition and increased costs of designing made it important for the firms to identify the right products and the right methods for manufacturing the products. Firms should focus on customers and identify customer demands directly to design the right products. Several management methods and techniques that are currently available improve one or more functions or processes in an industry and do not take the complete product life cycle into consideration. On the other hand target costing is a method / philosophy that takes financial, manufacturing and customer aspects into consideration during designing phase and helps firms in making product design decisions to increase the profit / value of the company. It uses various techniques to identify customer demands, to decrease costs of manufacturing and finally to achieve strategic goals. Target Costing forms an integral part of total product design / redesign based on strategic plans.

Performance Analysis of a WiMax/Wi-Fi System Whilst Streaming a Video Conference Application

WiMAX and Wi-Fi are considered as the promising broadband access solutions for wireless MAN’s and LANs, respectively. In the recent works WiMAX is considered suitable as a backhaul service to connect multiple dispersed Wi-Fi ‘hotspots’. Hence a new integrated WiMAX/Wi-Fi architecture has been proposed in literatures. In this paper the performance of an integrated WiMAX/Wi-Fi network has been investigated by streaming a video conference application. The difference in performance between the two protocols is compared with respect to video conferencing. The Heterogeneous network was simulated in the OPNET simulator.

Control-flow Complexity Measurement of Processes and Weyuker's Properties

Process measurement is the task of empirically and objectively assigning numbers to the properties of business processes in such a way as to describe them. Desirable attributes to study and measure include complexity, cost, maintainability, and reliability. In our work we will focus on investigating process complexity. We define process complexity as the degree to which a business process is difficult to analyze, understand or explain. One way to analyze a process- complexity is to use a process control-flow complexity measure. In this paper, an attempt has been made to evaluate the control-flow complexity measure in terms of Weyuker-s properties. Weyuker-s properties must be satisfied by any complexity measure to qualify as a good and comprehensive one.

Auto Tuning PID Controller based on Improved Genetic Algorithm for Reverse Osmosis Plant

An optimal control of Reverse Osmosis (RO) plant is studied in this paper utilizing the auto tuning concept in conjunction with PID controller. A control scheme composing an auto tuning stochastic technique based on an improved Genetic Algorithm (GA) is proposed. For better evaluation of the process in GA, objective function defined newly in sense of root mean square error has been used. Also in order to achieve better performance of GA, more pureness and longer period of random number generation in operation are sought. The main improvement is made by replacing the uniform distribution random number generator in conventional GA technique to newly designed hybrid random generator composed of Cauchy distribution and linear congruential generator, which provides independent and different random numbers at each individual steps in Genetic operation. The performance of newly proposed GA tuned controller is compared with those of conventional ones via simulation.

Binary Phase-Only Filter Watermarking with Quantized Embedding

The binary phase-only filter digital watermarking embeds the phase information of the discrete Fourier transform of the image into the corresponding magnitudes for better image authentication. The paper proposed an approach of how to implement watermark embedding by quantizing the magnitude, with discussing how to regulate the quantization steps based on the frequencies of the magnitude coefficients of the embedded watermark, and how to embed the watermark at low frequency quantization. The theoretical analysis and simulation results show that algorithm flexibility, security, watermark imperceptibility and detection performance of the binary phase-only filter digital watermarking can be effectively improved with quantization based watermark embedding, and the robustness against JPEG compression will also be increased to some extent.

Dynamic Modeling of Underplateform Damper used in Turbomachinery

The present work deals with the structural analysis of turbine blades and modeling of turbine blades. A common failure mode for turbine machines is high cycle of fatigue of compressor and turbine blades due to high dynamic stresses caused by blade vibration and resonance within the operation range of the machinery. In this work, proper damping system will be analyzed to reduce the vibrating blade. The main focus of the work is the modeling of under platform damper to evaluate the dynamic analysis of turbine-blade vibrations. The system is analyzed using Bond graph technique. Bond graph is one of the most convenient ways to represent a system from the physical aspect in foreground. It has advantage of putting together multi-energy domains of a system in a single representation in a unified manner. The bond graph model of dry friction damper is simulated on SYMBOLS-shakti® software. In this work, the blades are modeled as Timoshenko beam. Blade Vibrations under different working conditions are being analyzed numerically.

A Systems Modeling Approach to Support Environmentally Sustainable Business Development in Manufacturing SMEs

Small and Medium Sized Enterprises (SMEs) play an important role in many economies. In New Zealand, for example, 97% of all manufacturing companies employ less than 100 staff, and generate the predominant part of this industry sector-s economic output. Manufacturing SMEs as a group also have a significant impact on the environment. This situation is similar in many developed economies, including the European Union. Sustainable economic development therefore needs to strongly consider the role of manufacturing SMEs, who generally find it challenging to move towards more environmentally friendly business practices. This paper presents a systems thinking approach to modelling and understanding the factors which have an influence on the successful uptake of environmental practices in small and medium sized manufacturing companies. It presents a number of causal loop diagrams which have been developed based on primary action research, and a thorough understanding of the literature in this area. The systems thinking model provides the basis for further development of a strategic framework for the successful uptake of environmental innovation in manufacturing SMEs.

Optical Limiting Characteristics of Core-Shell Nanoparticles

TiO2 nanoparticles were synthesized by hydrothermal method at 180°C from TiOSO4 aqueous solution with1m/l concentration. The obtained products were coated with silica by means of a seeded polymerization technique for a coating time of 1440 minutes to obtain well defined TiO2@SiO2 core-shell structure. The uncoated and coated nanoparticles were characterized by using X-Ray diffraction technique (XRD), Fourier Transform Infrared Spectroscopy (FT-IR) to study their physico-chemical properties. Evidence from XRD and FTIR results show that SiO2 is homogenously coated on the surface of titania particles. FTIR spectra show that there exists an interaction between TiO2 and SiO2 and results in the formation of Ti-O-Si chemical bonds at the interface of TiO2 particles and SiO2 coating layer. The non linear optical limiting properties of TiO2 and TiO2@SiO2 nanoparticles dispersed in ethylene glycol were studied at 532nm using 5ns Nd:YAG laser pulses. Three-photon absorption is responsible for optical limiting characteristics in these nanoparticles and it is seen that the optical nonlinearity is enhanced in core-shell structures when compared with single counterparts. This effective three-photon type absorption at this wavelength, is of potential application in fabricating optical limiting devices.

A Comparative Study into Observer based Fault Detection and Diagnosis in DC Motors: Part-I

A model based fault detection and diagnosis technique for DC motor is proposed in this paper. Fault detection using Kalman filter and its different variants are compared. Only incipient faults are considered for the study. The Kalman Filter iterations and all the related computations required for fault detection and fault confirmation are presented. A second order linear state space model of DC motor is used for this work. A comparative assessment of the estimates computed from four different observers and their relative performance is evaluated.

Optimal Algorithm for Constructing the Delaunay Triangulation in Ed

In this paper we propose a new approach to constructing the Delaunay Triangulation and the optimum algorithm for the case of multidimensional spaces (d ≥ 2). Analysing the modern state, it is possible to draw a conclusion, that the ideas for the existing effective algorithms developed for the case of d ≥ 2 are not simple to generalize on a multidimensional case, without the loss of efficiency. We offer for the solving this problem an effective algorithm that satisfies all the given requirements. But theoretical complexity of the problem it is impossible to improve as the Worst - Case Optimality for algorithms of solving such a problem is proved.

Semantic Spatial Objects Data Structure for Spatial Access Method

Modern spatial database management systems require a unique Spatial Access Method (SAM) in order solve complex spatial quires efficiently. In this case the spatial data structure takes a prominent place in the SAM. Inadequate data structure leads forming poor algorithmic choices and forging deficient understandings of algorithm behavior on the spatial database. A key step in developing a better semantic spatial object data structure is to quantify the performance effects of semantic and outlier detections that are not reflected in the previous tree structures (R-Tree and its variants). This paper explores a novel SSRO-Tree on SAM to the Topo-Semantic approach. The paper shows how to identify and handle the semantic spatial objects with outlier objects during page overflow/underflow, using gain/loss metrics. We introduce a new SSRO-Tree algorithm which facilitates the achievement of better performance in practice over algorithms that are superior in the R*-Tree and RO-Tree by considering selection queries.

Weak Measurement Theory for Discrete Scales

With the increasing spread of computers and the internet among culturally, linguistically and geographically diverse communities, issues of internationalization and localization and becoming increasingly important. For some of the issues such as different scales for length and temperature, there is a well-developed measurement theory. For others such as date formats no such theory will be possible. This paper fills a gap by developing a measurement theory for a class of scales previously overlooked, based on discrete and interval-valued scales such as spanner and shoe sizes. The paper gives a theoretical foundation for a class of data representation problems.

Molecular Docking on Recomposed versus Crystallographic Structures of Zn-Dependent Enzymes and their Natural Inhibitors

Matrix metalloproteinases (MMP) are a class of structural and functional related enzymes involved in altering the natural elements of the extracellular matrix. Most of the MMP structures are cristalographycally determined and published in WorldWide ProteinDataBank, isolated, in full structure or bound to natural or synthetic inhibitors. This study proposes an algorithm to replace missing crystallographic structures in PDB database. We have compared the results of a chosen docking algorithm with a known crystallographic structure in order to validate enzyme sites reconstruction there where crystallographic data are missing.

Density Estimation using Generalized Linear Model and a Linear Combination of Gaussians

In this paper we present a novel approach for density estimation. The proposed approach is based on using the logistic regression model to get initial density estimation for the given empirical density. The empirical data does not exactly follow the logistic regression model, so, there will be a deviation between the empirical density and the density estimated using logistic regression model. This deviation may be positive and/or negative. In this paper we use a linear combination of Gaussian (LCG) with positive and negative components as a model for this deviation. Also, we will use the expectation maximization (EM) algorithm to estimate the parameters of LCG. Experiments on real images demonstrate the accuracy of our approach.

A Comparison Study of the Removal of Selected Pharmaceuticals in Waters by Chemical Oxidation Treatments

The degradation of selected pharmaceuticals in some water matrices was studied by using several chemical treatments. The pharmaceuticals selected were the beta-blocker metoprolol, the nonsteroidal anti-inflammatory naproxen, the antibiotic amoxicillin, and the analgesic phenacetin; and their degradations were conducted by using UV radiation alone, ozone, Fenton-s reagent, Fenton-like system, photo-Fenton system, and combinations of UV radiation and ozone with H2O2, TiO2, Fe(II), and Fe(III). The water matrices, in addition to ultra-pure water, were a reservoir water, a groundwater, and two secondary effluents from two municipal WWTP. The results reveal that the presence of any second oxidant enhanced the oxidation rates, with the systems UV/TiO2 and O3/TiO2 providing the highest degradation rates. It is also observed in most of the investigated oxidation systems that the degradation rate followed the sequence: amoxicillin > naproxen > metoprolol > phenacetin. Lower rates were obtained with the pharmaceuticals dissolved in natural waters and secondary effluents due to the organic matter present which consume some amounts of the oxidant agents.

Principal Component Analysis-Ranking as a Variable Selection Method for the Simultaneous Spectrophotometric Determination of Phenol, Resorcinol and Catechol in Real Samples

Simultaneous determination of multicomponents of phenol, resorcinol and catechol with a chemometric technique a PCranking artificial neural network (PCranking-ANN) algorithm is reported in this study. Based on the data correlation coefficient method, 3 representative PCs are selected from the scores of original UV spectral data (35 PCs) as the original input patterns for ANN to build a neural network model. The results obtained by iterating 8000 .The RMSEP for phenol, resorcinol and catechol with PCranking- ANN were 0.6680, 0.0766 and 0.1033, respectively. Calibration matrices were 0.50-21.0, 0.50-15.1 and 0.50-20.0 μg ml-1 for phenol, resorcinol and catechol, respectively. The proposed method was successfully applied for the determination of phenol, resorcinol and catechol in synthetic and water samples.