Cold Flow Investigation of Primary Zone Characteristics in Combustor Utilizing Axial Air Swirler

This paper presents a cold flow simulation study of a small gas turbine combustor performed using laboratory scale test rig. The main objective of this investigation is to obtain physical insight of the main vortex, responsible for the efficient mixing of fuel and air. Such models are necessary for predictions and optimization of real gas turbine combustors. Air swirler can control the combustor performance by assisting in the fuel-air mixing process and by producing recirculation region which can act as flame holders and influences residence time. Thus, proper selection of a swirler is needed to enhance combustor performance and to reduce NOx emissions. Three different axial air swirlers were used based on their vane angles i.e., 30°, 45°, and 60°. Three-dimensional, viscous, turbulent, isothermal flow characteristics of the combustor model operating at room temperature were simulated via Reynolds- Averaged Navier-Stokes (RANS) code. The model geometry has been created using solid model, and the meshing has been done using GAMBIT preprocessing package. Finally, the solution and analysis were carried out in a FLUENT solver. This serves to demonstrate the capability of the code for design and analysis of real combustor. The effects of swirlers and mass flow rate were examined. Details of the complex flow structure such as vortices and recirculation zones were obtained by the simulation model. The computational model predicts a major recirculation zone in the central region immediately downstream of the fuel nozzle and a second recirculation zone in the upstream corner of the combustion chamber. It is also shown that swirler angles changes have significant effects on the combustor flowfield as well as pressure losses.

A Tabu Search Heuristic for Scratch-Pad Memory Management

Reducing energy consumption of embedded systems requires careful memory management. It has been shown that Scratch- Pad Memories (SPMs) are low size, low cost, efficient (i.e. energy saving) data structures directly managed at the software level. In this paper, the focus is on heuristic methods for SPMs management. A method is efficient if the number of accesses to SPM is as large as possible and if all available space (i.e. bits) is used. A Tabu Search (TS) approach for memory management is proposed which is, to the best of our knowledge, a new original alternative to the best known existing heuristic (BEH). In fact, experimentations performed on benchmarks show that the Tabu Search method is as efficient as BEH (in terms of energy consumption) but BEH requires a sorting which can be computationally expensive for a large amount of data. TS is easy to implement and since no sorting is necessary, unlike BEH, the corresponding sorting time is saved. In addition to that, in a dynamic perspective where the maximum capacity of the SPM is not known in advance, the TS heuristic will perform better than BEH.

Automata Theory Approach for Solving Frequent Pattern Discovery Problems

The various types of frequent pattern discovery problem, namely, the frequent itemset, sequence and graph mining problems are solved in different ways which are, however, in certain aspects similar. The main approach of discovering such patterns can be classified into two main classes, namely, in the class of the levelwise methods and in that of the database projection-based methods. The level-wise algorithms use in general clever indexing structures for discovering the patterns. In this paper a new approach is proposed for discovering frequent sequences and tree-like patterns efficiently that is based on the level-wise issue. Because the level-wise algorithms spend a lot of time for the subpattern testing problem, the new approach introduces the idea of using automaton theory to solve this problem.

Influence of Laminated Textile Structures on Mechanical Performance of NF-Epoxy Composites

Textile structures are engineered and fabricated to meet worldwide structural applications. Nevertheless, research varying textile structure on natural fibre as composite reinforcement was found to be very limited. Most of the research is focusing on short fibre and random discontinuous orientation of the reinforcement structure. Realizing that natural fibre (NF) composite had been widely developed to be used as synthetic fibre composite replacement, this research attempted to examine the influence of woven and cross-ply laminated structure towards its mechanical performances. Laminated natural fibre composites were developed using hand lay-up and vacuum bagging technique. Impact and flexural strength were investigated as a function of fibre type (coir and kenaf) and reinforcement structure (imbalanced plain woven, 0°/90° cross-ply and +45°/-45° cross-ply). Multi-level full factorial design of experiment (DOE) and analysis of variance (ANOVA) was employed to impart data as to how fibre type and reinforcement structure parameters affect the mechanical properties of the composites. This systematic experimentation has led to determination of significant factors that predominant influences the impact and flexural properties of the textile composites. It was proven that both fibre type and reinforcement structure demonstrated significant difference results. Overall results indicated that coir composite and woven structure exhibited better impact and flexural strength. Yet, cross-ply composite structure demonstrated better fracture resistance.

Objects Extraction by Cooperating Optical Flow, Edge Detection and Region Growing Procedures

The image segmentation method described in this paper has been developed as a pre-processing stage to be used in methodologies and tools for video/image indexing and retrieval by content. This method solves the problem of whole objects extraction from background and it produces images of single complete objects from videos or photos. The extracted images are used for calculating the object visual features necessary for both indexing and retrieval processes. The segmentation algorithm is based on the cooperation among an optical flow evaluation method, edge detection and region growing procedures. The optical flow estimator belongs to the class of differential methods. It permits to detect motions ranging from a fraction of a pixel to a few pixels per frame, achieving good results in presence of noise without the need of a filtering pre-processing stage and includes a specialised model for moving object detection. The first task of the presented method exploits the cues from motion analysis for moving areas detection. Objects and background are then refined using respectively edge detection and seeded region growing procedures. All the tasks are iteratively performed until objects and background are completely resolved. The method has been applied to a variety of indoor and outdoor scenes where objects of different type and shape are represented on variously textured background.

Development of Molecular Imprinted Polymers (MIPs) for the Selective Removal of Carbamazepine from Aqueous Solution

The occurrence and removal of trace organic contaminants in the aquatic environment has become a focus of environmental concern. For the selective removal of carbamazepine from loaded waters molecularly imprinted polymers (MIPs) were synthesized with carbamazepine as template. Parameters varied were the type of monomer, crosslinker, and porogen, the ratio of starting materials, and the synthesis temperature. Best results were obtained with a template to crosslinker ratio of 1:20, toluene as porogen, and methacrylic acid (MAA) as monomer. MIPs were then capable to recover carbamazepine by 93% from a 10-5 M landfill leachate solution containing also caffeine and salicylic acid. By comparison, carbamazepine recoveries of 75% were achieved using a nonimprinted polymer (NIP) synthesized under the same conditions, but without template. In landfill leachate containing solutions carbamazepine was adsorbed by 93-96% compared with an uptake of 73% by activated carbon. The best solvent for desorption was acetonitrile, with which the amount of solvent necessary and dilution with water was tested. Selected MIPs were tested for their reusability and showed good results for at least five cycles. Adsorption isotherms were prepared with carbamazepine solutions in the concentration range of 0.01 M to 5*10-6 M. The heterogeneity index showed a more homogenous binding site distribution.

Retrieving Extended High Dynamic Range from Digital Negative Image - An Experiment on Architectural Photo Imaging

The paper explores the development of an optimization of method and apparatus for retrieving extended high dynamic range from digital negative image. Architectural photo imaging can benefit from high dynamic range imaging (HDRI) technique for preserving and presenting sufficient luminance in the shadow and highlight clipping image areas. The HDRI technique that requires multiple exposure images as the source of HDRI rendering may not be effective in terms of time efficiency during the acquisition process and post-processing stage, considering it has numerous potential imaging variables and technical limitations during the multiple exposure process. This paper explores an experimental method and apparatus that aims to expand the dynamic range from digital negative image in HDRI environment. The method and apparatus explored is based on a single source of RAW image acquisition for the use of HDRI post-processing. It will cater the optimization in order to avoid and minimize the conventional HDRI photographic errors caused by different physical conditions during the photographing process and the misalignment of multiple exposed image sequences. The study observes the characteristics and capabilities of RAW image format as digital negative used for the retrieval of extended high dynamic range process in HDRI environment.

Comparation Treatment Method for Industrial Tempeh Waste by Constructed Wetland and Activated Sludge

Ever since industrial revolution began, our ecosystem has changed. And indeed, the negatives outweigh the positives. Industrial waste usually released into all kinds of body of water, such as river or sea. Tempeh waste is one example of waste that carries many hazardous and unwanted substances that will affect the surrounding environment. Tempeh is a popular fermented food in Asia which is rich in nutrients and active substances. Tempeh liquid waste- in particular- can cause an air pollution, and if penetrates through the soil, it will contaminates ground-water, making it unavailable for the water to be consumed. Moreover, bacteria will thrive within the polluted water, which often responsible for causing many kinds of diseases. The treatment used for this chemical waste is biological treatment such as constructed wetland and activated sludge. These kinds of treatment are able to reduce both physical and chemical parameters altogether such as temperature, TSS, pH, BOD, COD, NH3-N, NO3-N, and PO4-P. These treatments are implemented before the waste is released into the water. The result is a comparation between constructed wetland and activated sludge, along with determining which method is better suited to reduce the physical and chemical subtances of the waste.

Comparative Finite Element Simulation of Nonlinear Vibrations and Sensor Output Voltage of Smart Piezolaminated Structures

Two geometrically nonlinear plate theories, based either on first- or third-order transverse shear deformation theory are used for finite element modeling and simulation of the transient response of smart structures incorporating piezoelectric layers. In particular the time histories of nonlinear vibrations and sensor voltage output of a thin beam with a piezoelectric patch bonded to the surface due to an applied step force are studied.

An Efficient Algorithm for Computing all Program Forward Static Slices

Program slicing is the task of finding all statements in a program that directly or indirectly influence the value of a variable occurrence. The set of statements that can affect the value of a variable at some point in a program is called a program backward slice. In several software engineering applications, such as program debugging and measuring program cohesion and parallelism, several slices are computed at different program points. The existing algorithms for computing program slices are introduced to compute a slice at a program point. In these algorithms, the program, or the model that represents the program, is traversed completely or partially once. To compute more than one slice, the same algorithm is applied for every point of interest in the program. Thus, the same program, or program representation, is traversed several times. In this paper, an algorithm is introduced to compute all forward static slices of a computer program by traversing the program representation graph once. Therefore, the introduced algorithm is useful for software engineering applications that require computing program slices at different points of a program. The program representation graph used in this paper is called Program Dependence Graph (PDG).

Modeling Directional Thermal Radiance Anisotropy for Urban Canopy

one of the significant factors for improving the accuracy of Land Surface Temperature (LST) retrieval is the correct understanding of the directional anisotropy for thermal radiance. In this paper, the multiple scattering effect between heterogeneous non-isothermal surfaces is described rigorously according to the concept of configuration factor, based on which a directional thermal radiance model is built, and the directional radiant character for urban canopy is analyzed. The model is applied to a simple urban canopy with row structure to simulate the change of Directional Brightness Temperature (DBT). The results show that the DBT is aggrandized because of the multiple scattering effects, whereas the change range of DBT is smoothed. The temperature difference, spatial distribution, emissivity of the components can all lead to the change of DBT. The “hot spot" phenomenon occurs when the proportion of high temperature component in the vision field came to a head. On the other hand, the “cool spot" phenomena occur when low temperature proportion came to the head. The “spot" effect disappears only when the proportion of every component keeps invariability. The model built in this paper can be used for the study of directional effect on emissivity, the LST retrieval over urban areas and the adjacency effect of thermal remote sensing pixels.

Creative Thinking Skill Approach Through Problem-Based Learning: Pedagogy and Practice in the Engineering Classroom

Problem-based learning (PBL) is one of the student centered approaches and has been considered by a number of higher educational institutions in many parts of the world as a method of delivery. This paper presents a creative thinking approach for implementing Problem-based Learning in Mechanics of Structure within a Malaysian Polytechnics environment. In the learning process, students learn how to analyze the problem given among the students and sharing classroom knowledge into practice. Further, through this course-s emphasis on problem-based learning, students acquire creative thinking skills and professional skills as they tackle complex, interdisciplinary and real-situation problems. Once the creative ideas are generated, there are useful additional techniques for tender ideas that will grow into a productive concept or solution. The combination of creative skills and technical abilities will enable the students to be ready to “hit-the-ground-running" and produce in industry when they graduate.

Virtual Reality Models used on the Visualization of Construction Activities in Civil Engineering Education

Three-dimensional geometric models have been used to present architectural and engineering works, showing their final configuration. When the clarification of a detail or the constitution of a construction step in needed, these models are not appropriate. They do not allow the observation of the construction progress of a building. Models that could present dynamically changes of the building geometry are a good support to the elaboration of projects. Techniques of geometric modeling and virtual reality were used to obtain models that could visually simulate the construction activity. The applications explain the construction work of a cavity wall and a bridge. These models allow the visualization of the physical progression of the work following a planned construction sequence, the observation of details of the form of every component of the works and support the study of the type and method of operation of the equipment applied in the construction. These models presented distinct advantage as educational aids in first-degree courses in Civil Engineering. The use of Virtual Reality techniques in the development of educational applications brings new perspectives to the teaching of subjects related to the field of civil construction.

Determination of Non Uniform Sinusoidal Microstrip Leaky-Wave Antenna Radiating Performances in Millimeter Band

Here we have considered non uniform microstrip leaky-wave antenna implemented on a dielectric waveguide by a sinusoidal profile of periodic metallic grating. The non distribution of the attenuation constant α along propagation axis, optimize the radiating characteristics and performances of such antennas. The method developped here is based on an integral method where the formalism of the admittance operator is combined to a BKW approximation. First, the effect of the modeling in the modal analysis of complex waves is studied in detail. Then, the BKW model is used for the dispersion analysis of the antenna of interest. According to antenna theory, a forced continuity of the leaky-wave magnitude at discontinuities of the non uniform structure is established. To test the validity of our dispersion analysis, computed radiation patterns are presented and compared in the millimeter band.

Modeling and Identification of Hammerstein System by using Triangular Basis Functions

This paper deals with modeling and parameter identification of nonlinear systems described by Hammerstein model having Piecewise nonlinear characteristics such as Dead-zone nonlinearity characteristic. The simultaneous use of both an easy decomposition technique and the triangular basis functions leads to a particular form of Hammerstein model. The approximation by using Triangular basis functions for the description of the static nonlinear block conducts to a linear regressor model, so that least squares techniques can be used for the parameter estimation. Singular Values Decomposition (SVD) technique has been applied to separate the coupled parameters. The proposed approach has been efficiently tested on academic examples of simulation.

E-government Adoption in Romania

The Romanian government has been making significant attempts to make its services and information available on the Internet. According to the UN e-government survey conducted in 2008, Romania comes under mid range countries by utilization of egovernment (percent of utilization 41%). Romania-s national portal www.e-guvernare.ro aims at progressively making all services and information accessible through the portal. However, the success of these efforts depends, to a great extent, on how well the targeted users for such services, citizens in general, make use of them. For this reason, the purpose of the presented study was to identify what factors could affect the citizens' adoption of e-government services. The study is an extension of the Technology Acceptance Model. The proposed model was validated using data collected from 481 citizens. The results provided substantial support for all proposed hypotheses and showed the significance of the extended constructs.

An Algorithm for Detecting Seam Cracks in Steel Plates

In this study, we developed an algorithm for detecting seam cracks in a steel plate. Seam cracks are generated in the edge region of a steel plate. We used the Gabor filter and an adaptive double threshold method to detect them. To reduce the number of pseudo defects, features based on the shape of seam cracks were used. To evaluate the performance of the proposed algorithm, we tested 989 images with seam cracks and 9470 defect-free images. Experimental results show that the proposed algorithm is suitable for detecting seam cracks. However, it should be improved to increase the true positive rate.

A New Brazilian Friction-Resistant Low Alloy High Strength Steel – A Life Testing Approach

In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.

Towards an Enhanced Stochastic Simulation Model for Risk Analysis in Highway Construction

Over the years, there is a growing trend towards quality-based specifications in highway construction. In many Quality Control/Quality Assurance (QC/QA) specifications, the contractor is primarily responsible for quality control of the process, whereas the highway agency is responsible for testing the acceptance of the product. A cooperative investigation was conducted in Illinois over several years to develop a prototype End-Result Specification (ERS) for asphalt pavement construction. The final characteristics of the product are stipulated in the ERS and the contractor is given considerable freedom in achieving those characteristics. The risk for the contractor or agency depends on how the acceptance limits and processes are specified. Stochastic simulation models are very useful in estimating and analyzing payment risk in ERS systems and these form an integral part of the Illinois-s prototype ERS system. This paper describes the development of an innovative methodology to estimate the variability components in in-situ density, air voids and asphalt content data from ERS projects. The information gained from this would be crucial in simulating these ERS projects for estimation and analysis of payment risks associated with asphalt pavement construction. However, these methods require at least two parties to conduct tests on all the split samples obtained according to the sampling scheme prescribed in present ERS implemented in Illinois.

Optimization Using Simulation of the Vehicle Routing Problem

A key element of many distribution systems is the routing and scheduling of vehicles servicing a set of customers. A wide variety of exact and approximate algorithms have been proposed for solving the vehicle routing problems (VRP). Exact algorithms can only solve relatively small problems of VRP, which is classified as NP-Hard. Several approximate algorithms have proven successful in finding a feasible solution not necessarily optimum. Although different parts of the problem are stochastic in nature; yet, limited work relevant to the application of discrete event system simulation has addressed the problem. Presented here is optimization using simulation of VRP; where, a simplified problem has been developed in the ExtendSimTM simulation environment; where, ExtendSimTM evolutionary optimizer is used to minimize the total transportation cost of the problem. Results obtained from the model are very satisfactory. Further complexities of the problem are proposed for consideration in the future.