New Nonlinear Filtering Strategies for Eliminating Short and Long Tailed Noise in Images with Edge Preservation Properties

Midpoint filter is quite effective in recovering the images confounded by the short-tailed (uniform) noise. It, however, performs poorly in the presence of additive long-tailed (impulse) noise and it does not preserve the edge structures of the image signals. Median smoother discards outliers (impulses) effectively, but it fails to provide adequate smoothing for images corrupted with nonimpulse noise. In this paper, two nonlinear techniques for image filtering, namely, New Filter I and New Filter II are proposed based on a nonlinear high-pass filter algorithm. New Filter I is constructed using a midpoint filter, a highpass filter and a combiner. It suppresses uniform noise quite well. New Filter II is configured using an alpha trimmed midpoint filter, a median smoother of window size 3x3, the high pass filter and the combiner. It is robust against impulse noise and attenuates uniform noise satisfactorily. Both the filters are shown to exhibit good response at the image boundaries (edges). The proposed filters are evaluated for their performance on a test image and the results obtained are included.

Low-Cost and Highly Accurate Motion Models for Three-Dimensional Local Landmark-based Autonomous Navigation

Recently, the Spherical Motion Models (SMM-s) have been introduced [1]. These new models have been developed for 3D local landmark-base Autonomous Navigation (AN). This paper is revealing new arguments and experimental results to support the SMM-s characteristics. The accuracy and the robustness in performing a specific task are the main concerns of the new investigations. To analyze their performances of the SMM-s, the most powerful tools of estimation theory, the extended Kalman filter (EKF) and unscented Kalman filter (UKF), which give the best estimations in noisy environments, have been employed. The Monte Carlo validation implementations used to test the stability and robustness of the models have been employed as well.

New Laguerre-s Type Method for Solving of a Polynomial Equations Systems

In this paper we present a substantiation of a new Laguerre-s type iterative method for solving of a nonlinear polynomial equations systems with real coefficients. The problems of its implementation, including relating to the structural choice of initial approximations, were considered. Test examples demonstrate the effectiveness of the method at the solving of many practical problems solving.

The Influence of Biofuels on the Permeability of Sand-Bentonite Liners

Liners are made to protect the groundwater table from the infiltration of leachate which normally carries different kinds of toxic materials from landfills. Although these liners are engineered to last for long period of time; unfortunately these liners fail; therefore, toxic materials pass to groundwater. This paper focuses on the changes of the hydraulic conductivity of a sand-bentonite liner due to the infiltration of biofuel and ethanol fuel. Series of laboratory tests were conducted in 20-cm-high PVC columns. Several compositions of sand-bentonite liners were tested: 95% sand: 5% bentonite; 90% sand: 10% bentonite; and 100% sand (passed mesh #40). The columns were subjected to extreme pressures of 40 kPa, and 100 kPa to evaluate the transport of alternative fuels (biofuel and ethanol fuel). For comparative studies, similar tests were carried out using water. Results showed that hydraulic conductivity increased due to the infiltration of alternative fuels through the liners. Accordingly, the increase in the hydraulic conductivity showed significant dependency on the type of liner mixture and the characteristics of the liquid. The hydraulic conductivity of a liner (subjected to biofuel infiltration) consisting of 5% bentonite: 95% sand under pressure of 40 kPa and 100 kPa had increased by one fold. In addition, the hydraulic conductivity of a liner consisting of 10% bentonite: 90% sand under pressure of 40 kPa and 100 kPa and infiltrated by biofuel had increased by three folds. On the other hand, the results obtained by water infiltration under 40 kPa showed lower hydraulic conductivities of 1.50×10-5 and 1.37×10-9 cm/s for 5% bentonite: 95% sand, and 10% bentonite: 90% sand, respectively. Similarly, under 100 kPa, the hydraulic conductivities were 2.30×10-5 and 1.90×10-9 cm/s for 5% bentonite: 95% sand, and 10% bentonite: 90% sand, respectively.

Towards a Sustainable Regeneration: The Case Study of the San Mateo Neighborhood, in Jerez de la Frontera (Andalusia)

Based on different experiences in the historic centers of Spain, we propose an global strategy for the regeneration of the pre-tertiary fabrics and its application to the specific case of San Mateo neighborhood, in Jerez de la Frontera (Andalusia), through a diagnosis that focus particularly on the punishments the last-decade economic situation (building boom and crisis) and shows the tragic transition from economic center to an imminent disappearance with an image similar to the ruins of war, due to the loss of their traditional roles. From it we will learn their historically-tested mechanisms of environment adaptation, which distill the vernacular architecture essence and that we will apply to our strategy of action based on a dotacional-and-free-space rhizome which rediscovers its hidden character. The architectural fact will be crystallized in one of the example-pieces proposed: The Artistic Revitalization Center.

A Study of the Lighting Control System for a Daylit Office

Increasing user comfort and reducing operation costs have always been primary objectives of lighting control strategies in a building. This paper proposes an architecture of the lighting control system for a daylit office. The system consists of the lighting controller, A/D & D/A converter, dimmable LED lights, and the lighting management software. Verification tests are conducted using the proposed system specialized for the interior lighting of a open-plan office. The results showed the proposed architecture of the lighting system would improve the overall system reliability, lower the system cost, and provide ease of installation and maintenance.

Validation and Application of a New Optimized RP-HPLC-Fluorescent Detection Method for Norfloxacin

A new reverse phase-high performance liquid chromatography (RP-HPLC) method with fluorescent detector (FLD) was developed and optimized for Norfloxacin determination in human plasma. Mobile phase specifications, extraction method and excitation and emission wavelengths were varied for optimization. HPLC system contained a reverse phase C18 (5 μm, 4.6 mm×150 mm) column with FLD operated at excitation 330 nm and emission 440 nm. The optimized mobile phase consisted of 14% acetonitrile in buffer solution. The aqueous phase was prepared by mixing 2g of citric acid, 2g sodium acetate and 1 ml of triethylamine in 1 L of Milli-Q water was run at a flow rate of 1.2 mL/min. The standard curve was linear for the range tested (0.156–20 μg/mL) and the coefficient of determination was 0.9978. Aceclofenac sodium was used as internal standard. A detection limit of 0.078 μg/mL was achieved. Run time was set at 10 minutes because retention time of norfloxacin was 0.99 min. which shows the rapidness of this method of analysis. The present assay showed good accuracy, precision and sensitivity for Norfloxacin determination in human plasma with a new internal standard and can be applied pharmacokinetic evaluation of Norfloxacin tablets after oral administration in human.

Cold Flow Investigation of Primary Zone Characteristics in Combustor Utilizing Axial Air Swirler

This paper presents a cold flow simulation study of a small gas turbine combustor performed using laboratory scale test rig. The main objective of this investigation is to obtain physical insight of the main vortex, responsible for the efficient mixing of fuel and air. Such models are necessary for predictions and optimization of real gas turbine combustors. Air swirler can control the combustor performance by assisting in the fuel-air mixing process and by producing recirculation region which can act as flame holders and influences residence time. Thus, proper selection of a swirler is needed to enhance combustor performance and to reduce NOx emissions. Three different axial air swirlers were used based on their vane angles i.e., 30°, 45°, and 60°. Three-dimensional, viscous, turbulent, isothermal flow characteristics of the combustor model operating at room temperature were simulated via Reynolds- Averaged Navier-Stokes (RANS) code. The model geometry has been created using solid model, and the meshing has been done using GAMBIT preprocessing package. Finally, the solution and analysis were carried out in a FLUENT solver. This serves to demonstrate the capability of the code for design and analysis of real combustor. The effects of swirlers and mass flow rate were examined. Details of the complex flow structure such as vortices and recirculation zones were obtained by the simulation model. The computational model predicts a major recirculation zone in the central region immediately downstream of the fuel nozzle and a second recirculation zone in the upstream corner of the combustion chamber. It is also shown that swirler angles changes have significant effects on the combustor flowfield as well as pressure losses.

Automata Theory Approach for Solving Frequent Pattern Discovery Problems

The various types of frequent pattern discovery problem, namely, the frequent itemset, sequence and graph mining problems are solved in different ways which are, however, in certain aspects similar. The main approach of discovering such patterns can be classified into two main classes, namely, in the class of the levelwise methods and in that of the database projection-based methods. The level-wise algorithms use in general clever indexing structures for discovering the patterns. In this paper a new approach is proposed for discovering frequent sequences and tree-like patterns efficiently that is based on the level-wise issue. Because the level-wise algorithms spend a lot of time for the subpattern testing problem, the new approach introduces the idea of using automaton theory to solve this problem.

Particle Swarm Optimization with Reduction for Global Optimization Problems

This paper presents an algorithm of particle swarm optimization with reduction for global optimization problems. Particle swarm optimization is an algorithm which refers to the collective motion such as birds or fishes, and a multi-point search algorithm which finds a best solution using multiple particles. Particle swarm optimization is so flexible that it can adapt to a number of optimization problems. When an objective function has a lot of local minimums complicatedly, the particle may fall into a local minimum. For avoiding the local minimum, a number of particles are initially prepared and their positions are updated by particle swarm optimization. Particles sequentially reduce to reach a predetermined number of them grounded in evaluation value and particle swarm optimization continues until the termination condition is met. In order to show the effectiveness of the proposed algorithm, we examine the minimum by using test functions compared to existing algorithms. Furthermore the influence of best value on the initial number of particles for our algorithm is discussed.

Development of Molecular Imprinted Polymers (MIPs) for the Selective Removal of Carbamazepine from Aqueous Solution

The occurrence and removal of trace organic contaminants in the aquatic environment has become a focus of environmental concern. For the selective removal of carbamazepine from loaded waters molecularly imprinted polymers (MIPs) were synthesized with carbamazepine as template. Parameters varied were the type of monomer, crosslinker, and porogen, the ratio of starting materials, and the synthesis temperature. Best results were obtained with a template to crosslinker ratio of 1:20, toluene as porogen, and methacrylic acid (MAA) as monomer. MIPs were then capable to recover carbamazepine by 93% from a 10-5 M landfill leachate solution containing also caffeine and salicylic acid. By comparison, carbamazepine recoveries of 75% were achieved using a nonimprinted polymer (NIP) synthesized under the same conditions, but without template. In landfill leachate containing solutions carbamazepine was adsorbed by 93-96% compared with an uptake of 73% by activated carbon. The best solvent for desorption was acetonitrile, with which the amount of solvent necessary and dilution with water was tested. Selected MIPs were tested for their reusability and showed good results for at least five cycles. Adsorption isotherms were prepared with carbamazepine solutions in the concentration range of 0.01 M to 5*10-6 M. The heterogeneity index showed a more homogenous binding site distribution.

Pharmaceutical Microencapsulation Technology for Development of Controlled Release Drug Delivery systems

This article demonstrated development of controlled release system of an NSAID drug, Diclofenac sodium employing different ratios of Ethyl cellulose. Diclofenac sodium and ethyl cellulose in different proportions were processed by microencapsulation based on phase separation technique to formulate microcapsules. The prepared microcapsules were then compressed into tablets to obtain controlled release oral formulations. In-vitro evaluation was performed by dissolution test of each preparation was conducted in 900 ml of phosphate buffer solution of pH 7.2 maintained at 37 ± 0.5 °C and stirred at 50 rpm. At predetermined time intervals (0, 0.5, 1.0, 1.5, 2, 3, 4, 6, 8, 10, 12, 16, 20 and 24 hrs). The drug concentration in the collected samples was determined by UV spectrophotometer at 276 nm. The physical characteristics of diclofenac sodium microcapsules were according to accepted range. These were off-white, free flowing and spherical in shape. The release profile of diclofenac sodium from microcapsules was found to be directly proportional to the proportion of ethylcellulose and coat thickness. The in-vitro release pattern showed that with ratio of 1:1 and 1:2 (drug: polymer), the percentage release of drug at first hour was 16.91 and 11.52 %, respectively as compared to 1:3 which is only 6.87 % with in this time. The release mechanism followed higuchi model for its release pattern. Tablet Formulation (F2) of present study was found comparable in release profile the marketed brand Phlogin-SR, microcapsules showed an extended release beyond 24 h. Further, a good correlation was found between drug release and proportion of ethylcellulose in the microcapsules. Microencapsulation based on coacervation found as good technique to control release of diclofenac sodium for making the controlled release formulations.

Data Structures and Algorithms of Intelligent Web-Based System for Modular Design

In recent years, new product development became more and more competitive and globalized, and the designing phase is critical for the product success. The concept of modularity can provide the necessary foundation for organizations to design products that can respond rapidly to market needs. The paper describes data structures and algorithms of intelligent Web-based system for modular design taking into account modules compatibility relationship and given design requirements. The system intelligence is realized by developed algorithms for choice of modules reflecting all system restrictions and requirements. The proposed data structure and algorithms are illustrated by case study of personal computer configuration. The applicability of the proposed approach is tested through a prototype of Web-based system.

Determination of Non Uniform Sinusoidal Microstrip Leaky-Wave Antenna Radiating Performances in Millimeter Band

Here we have considered non uniform microstrip leaky-wave antenna implemented on a dielectric waveguide by a sinusoidal profile of periodic metallic grating. The non distribution of the attenuation constant α along propagation axis, optimize the radiating characteristics and performances of such antennas. The method developped here is based on an integral method where the formalism of the admittance operator is combined to a BKW approximation. First, the effect of the modeling in the modal analysis of complex waves is studied in detail. Then, the BKW model is used for the dispersion analysis of the antenna of interest. According to antenna theory, a forced continuity of the leaky-wave magnitude at discontinuities of the non uniform structure is established. To test the validity of our dispersion analysis, computed radiation patterns are presented and compared in the millimeter band.

Higher Plants Ability to Assimilate Explosives

The ability of agricultural and decorative plants to absorb and detoxify TNT and RDX has been studied. All tested 8 plants, grown hydroponically, were able to absorb these explosives from water solutions: Alfalfa > Soybean > Chickpea> Chikling vetch >Ryegrass > Mung bean> China bean > Maize. Differently from TNT, RDX did not exhibit negative influence on seed germination and plant growth. Moreover, some plants, exposed to RDX containing solution were increased in their biomass by 20%. Study of the fate of absorbed [1-14ðí]-TNT revealed the label distribution in low and high-molecular mass compounds, both in roots and above ground parts of plants, prevailing in the later. Content of 14ðí in lowmolecular compounds in plant roots are much higher than in above ground parts. On the contrary, high-molecular compounds are more intensively labeled in aboveground parts of soybean. Most part (up to 70%) of metabolites of TNT, formed either by enzymatic reduction or oxidation, is found in high molecular insoluble conjugates. Activation of enzymes, responsible for reduction, oxidation and conjugation of TNT, such as nitroreductase, peroxidase, phenoloxidase and glutathione S-transferase has been demonstrated. Among these enzymes, only nitroreductase was shown to be induced in alfalfa, exposed to RDX. The increase in malate dehydrogenase activities in plants, exposed to both explosives, indicates intensification of Tricarboxylic Acid Cycle, that generates reduced equivalents of NAD(P)H, necessary for functioning of the nitroreductase. The hypothetic scheme of TNT metabolism in plants is proposed.

Modeling and Identification of Hammerstein System by using Triangular Basis Functions

This paper deals with modeling and parameter identification of nonlinear systems described by Hammerstein model having Piecewise nonlinear characteristics such as Dead-zone nonlinearity characteristic. The simultaneous use of both an easy decomposition technique and the triangular basis functions leads to a particular form of Hammerstein model. The approximation by using Triangular basis functions for the description of the static nonlinear block conducts to a linear regressor model, so that least squares techniques can be used for the parameter estimation. Singular Values Decomposition (SVD) technique has been applied to separate the coupled parameters. The proposed approach has been efficiently tested on academic examples of simulation.

An Implementation of Stipple Operations

Stipples are desired for pattern fillings and transparency effects. In contrast, some graphics standards, including OpenGL ES 1.1 and 2.0, omitted this feature. We represent details of providing line stipples and polygon stipples, through combining texture mapping and alpha blending functions. We start from the OpenGL-specified stipple-related API functions. The details of mathematical transformations are explained to get the correct texture coordinates. Then, the overall algorithm is represented, and its implementation results are followed. We accomplished both of line and polygon stipples, and verified its result with conformance test routines.

Comparative Study of Transformed and Concealed Data in Experimental Designs and Analyses

This paper presents the comparative study of coded data methods for finding the benefit of concealing the natural data which is the mercantile secret. Influential parameters of the number of replicates (rep), treatment effects (τ) and standard deviation (σ) against the efficiency of each transformation method are investigated. The experimental data are generated via computer simulations under the specified condition of the process with the completely randomized design (CRD). Three ways of data transformation consist of Box-Cox, arcsine and logit methods. The difference values of F statistic between coded data and natural data (Fc-Fn) and hypothesis testing results were determined. The experimental results indicate that the Box-Cox results are significantly different from natural data in cases of smaller levels of replicates and seem to be improper when the parameter of minus lambda has been assigned. On the other hand, arcsine and logit transformations are more robust and obviously, provide more precise numerical results. In addition, the alternate ways to select the lambda in the power transformation are also offered to achieve much more appropriate outcomes.

An Algorithm for Detecting Seam Cracks in Steel Plates

In this study, we developed an algorithm for detecting seam cracks in a steel plate. Seam cracks are generated in the edge region of a steel plate. We used the Gabor filter and an adaptive double threshold method to detect them. To reduce the number of pseudo defects, features based on the shape of seam cracks were used. To evaluate the performance of the proposed algorithm, we tested 989 images with seam cracks and 9470 defect-free images. Experimental results show that the proposed algorithm is suitable for detecting seam cracks. However, it should be improved to increase the true positive rate.

A New Brazilian Friction-Resistant Low Alloy High Strength Steel – A Life Testing Approach

In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.