Study of Damage in Beams with Different Boundary Conditions

–In this paper the damage in clamped-free, clampedclamped and free-free beam are analyzed considering samples without and with structural modifications. The damage location is investigated by the use of the bispectrum and wavelet analysis. The mathematical models are obtained using 2D elasticity theory and the Finite Element Method (FEM). The numerical and experimental data are approximated using the Particle Swarm Optimizer (PSO) method and this way is possible to adjust the localization and the severity of the damage. The experimental data are obtained through accelerometers placed along the sample. The system is excited using impact hammer.

A Modification of Wireless and Internet Technologies for Logistics- Analysis

This research is designed for helping a WAPbased mobile phone-s user in order to analyze of logistics in the traffic area by applying and designing the accessible processes from mobile user to server databases. The research-s design comprises Mysql 4.1.8-nt database system for being the server which there are three sub-databases, traffic light – times of intersections in periods of the day, distances on the road of area-blocks where are divided from the main sample-area and speeds of sample vehicles (motorcycle, personal car and truck) in periods of the day. For interconnections between the server and user, PHP is used to calculate distances and travelling times from the beginning point to destination, meanwhile XHTML applied for receiving, sending and displaying data from PHP to user-s mobile. In this research, the main sample-area is focused at the Huakwang-Ratchada-s area, Bangkok, Thailand where usually the congested point and 6.25 km2 surrounding area which are split into 25 blocks, 0.25 km2 for each. For simulating the results, the designed server-database and all communicating models of this research have been uploaded to www.utccengineering.com/m4tg and used the mobile phone which supports WAP 2.0 XHTML/HTML multimode browser for observing values and displayed pictures. According to simulated results, user can check the route-s pictures from the requiring point to destination along with analyzed consuming times when sample vehicles travel in various periods of the day.

Bridging the Mental Gap between Convolution Approach and Compartmental Modeling in Functional Imaging: Typical Embedding of an Open Two-Compartment Model into the Systems Theory Approach of Indicator Dilution Theory

Functional imaging procedures for the non-invasive assessment of tissue microcirculation are highly requested, but require a mathematical approach describing the trans- and intercapillary passage of tracer particles. Up to now, two theoretical, for the moment different concepts have been established for tracer kinetic modeling of contrast agent transport in tissues: pharmacokinetic compartment models, which are usually written as coupled differential equations, and the indicator dilution theory, which can be generalized in accordance with the theory of lineartime- invariant (LTI) systems by using a convolution approach. Based on mathematical considerations, it can be shown that also in the case of an open two-compartment model well-known from functional imaging, the concentration-time course in tissue is given by a convolution, which allows a separation of the arterial input function from a system function being the impulse response function, summarizing the available information on tissue microcirculation. Due to this reason, it is possible to integrate the open two-compartment model into the system-theoretic concept of indicator dilution theory (IDT) and thus results known from IDT remain valid for the compartment approach. According to the long number of applications of compartmental analysis, even for a more general context similar solutions of the so-called forward problem can already be found in the extensively available appropriate literature of the seventies and early eighties. Nevertheless, to this day, within the field of biomedical imaging – not from the mathematical point of view – there seems to be a trench between both approaches, which the author would like to get over by exemplary analysis of the well-known model.

Does Leisure Time Use Contribute to a Wage Increase of the Thai People?

This paper develops models to analyze the relationship between leisure time and wage change. Using Thailand-s Time Use Survey and Labor Force Survey data, the estimation of wage changes in response to leisure time change indicates that media receiving, personal care and social participation and volunteer activities are the ones that significantly raise hourly wages. Thus, the finding suggests the stimulation in time use for media access to enhance knowledge and productivity, personal care for attractiveness and healthiness in order to raise productivity, and social activities to develop connections for possible future opportunities including wage increase. These activities should be promoted for productive leisure time and for welfare improvement.

Defect Cause Modeling with Decision Tree and Regression Analysis

The main aim of this study is to identify the most influential variables that cause defects on the items produced by a casting company located in Turkey. To this end, one of the items produced by the company with high defective percentage rates is selected. Two approaches-the regression analysis and decision treesare used to model the relationship between process parameters and defect types. Although logistic regression models failed, decision tree model gives meaningful results. Based on these results, it can be claimed that the decision tree approach is a promising technique for determining the most important process variables.

Comparison of Imputation Techniques for Efficient Prediction of Software Fault Proneness in Classes

Missing data is a persistent problem in almost all areas of empirical research. The missing data must be treated very carefully, as data plays a fundamental role in every analysis. Improper treatment can distort the analysis or generate biased results. In this paper, we compare and contrast various imputation techniques on missing data sets and make an empirical evaluation of these methods so as to construct quality software models. Our empirical study is based on NASA-s two public dataset. KC4 and KC1. The actual data sets of 125 cases and 2107 cases respectively, without any missing values were considered. The data set is used to create Missing at Random (MAR) data Listwise Deletion(LD), Mean Substitution(MS), Interpolation, Regression with an error term and Expectation-Maximization (EM) approaches were used to compare the effects of the various techniques.

Application of Artificial Neural Network to Classification Surface Water Quality

Water quality is a subject of ongoing concern. Deterioration of water quality has initiated serious management efforts in many countries. This study endeavors to automatically classify water quality. The water quality classes are evaluated using 6 factor indices. These factors are pH value (pH), Dissolved Oxygen (DO), Biochemical Oxygen Demand (BOD), Nitrate Nitrogen (NO3N), Ammonia Nitrogen (NH3N) and Total Coliform (TColiform). The methodology involves applying data mining techniques using multilayer perceptron (MLP) neural network models. The data consisted of 11 sites of canals in Dusit district in Bangkok, Thailand. The data is obtained from the Department of Drainage and Sewerage Bangkok Metropolitan Administration during 2007-2011. The results of multilayer perceptron neural network exhibit a high accuracy multilayer perception rate at 96.52% in classifying the water quality of Dusit district canal in Bangkok Subsequently, this encouraging result could be applied with plan and management source of water quality.

Exploring the Combinatorics of Motif Alignments Foraccurately Computing E-values from P-values

In biological and biomedical research motif finding tools are important in locating regulatory elements in DNA sequences. There are many such motif finding tools available, which often yield position weight matrices and significance indicators. These indicators, p-values and E-values, describe the likelihood that a motif alignment is generated by the background process, and the expected number of occurrences of the motif in the data set, respectively. The various tools often estimate these indicators differently, making them not directly comparable. One approach for comparing motifs from different tools, is computing the E-value as the product of the p-value and the number of possible alignments in the data set. In this paper we explore the combinatorics of the motif alignment models OOPS, ZOOPS, and ANR, and propose a generic algorithm for computing the number of possible combinations accurately. We also show that using the wrong alignment model can give E-values that significantly diverge from their true values.

Design Optimization of Cutting Parameters when Turning Inconel 718 with Cermet Inserts

Inconel 718, a nickel based super-alloy is an extensively used alloy, accounting for about 50% by weight of materials used in an aerospace engine, mainly in the gas turbine compartment. This is owing to their outstanding strength and oxidation resistance at elevated temperatures in excess of 5500 C. Machining is a requisite operation in the aircraft industries for the manufacture of the components especially for gas turbines. This paper is concerned with optimization of the surface roughness when turning Inconel 718 with cermet inserts. Optimization of turning operation is very useful to reduce cost and time for machining. The approach is based on Response Surface Method (RSM). In this work, second-order quadratic models are developed for surface roughness, considering the cutting speed, feed rate and depth of cut as the cutting parameters, using central composite design. The developed models are used to determine the optimum machining parameters. These optimized machining parameters are validated experimentally, and it is observed that the response values are in reasonable agreement with the predicted values.

About Analysis and Modelling of the Open Message Switching System

The modern queueing theory is one of the powerful tools for a quantitative and qualitative analysis of communication systems, computer networks, transportation systems, and many other technical systems. The paper is designated to the analysis of queueing systems, arising in the networks theory and communications theory (called open queueing network). The authors of this research in the sphere of queueing theory present the theorem about the law of the iterated logarithm (LIL) for the queue length of a customers in open queueing network and its application to the mathematical model of the open message switching system.

Some Factors Affecting the Compressive Behaviour of Structural Masonry at Small Scales

This paper presents part of a research into the small scale modelling of masonry. Small scale testing of masonry has been carried out by many authors, but few have attempted a systematic determination of the parameters that affect masonry at a small scale. The effect of increasing mortar strength and different sand gradings under compression were investigated. The results show masonry strength at small scale is influenced by increasing mortar strength and different sand gradings.

Geostatistical Analysis and Mapping of Groundlevel Ozone in a Medium Sized Urban Area

Ground-level tropospheric ozone is one of the air pollutants of most concern. It is mainly produced by photochemical processes involving nitrogen oxides and volatile organic compounds in the lower parts of the atmosphere. Ozone levels become particularly high in regions close to high ozone precursor emissions and during summer, when stagnant meteorological conditions with high insolation and high temperatures are common. In this work, some results of a study about urban ozone distribution patterns in the city of Badajoz, which is the largest and most industrialized city in Extremadura region (southwest Spain) are shown. Fourteen sampling campaigns, at least one per month, were carried out to measure ambient air ozone concentrations, during periods that were selected according to favourable conditions to ozone production, using an automatic portable analyzer. Later, to evaluate the ozone distribution at the city, the measured ozone data were analyzed using geostatistical techniques. Thus, first, during the exploratory analysis of data, it was revealed that they were distributed normally, which is a desirable property for the subsequent stages of the geostatistical study. Secondly, during the structural analysis of data, theoretical spherical models provided the best fit for all monthly experimental variograms. The parameters of these variograms (sill, range and nugget) revealed that the maximum distance of spatial dependence is between 302-790 m and the variable, air ozone concentration, is not evenly distributed in reduced distances. Finally, predictive ozone maps were derived for all points of the experimental study area, by use of geostatistical algorithms (kriging). High prediction accuracy was obtained in all cases as cross-validation showed. Useful information for hazard assessment was also provided when probability maps, based on kriging interpolation and kriging standard deviation, were produced.

Further the Effectiveness of Software Testability Measure

Software testability is proposed to address the problem of increasing cost of test and the quality of software. Testability measure provides a quantified way to denote the testability of software. Since 1990s, many testability measure models are proposed to address the problem. By discussing the contradiction between domain testability and domain range ratio (DRR), a new testability measure, semantic fault distance, is proposed. Its validity is discussed.

Finite Element Analysis of Full Ceramic Crowns with and without Zirconia Framework

Simulation of occlusal function during laboratory material-s testing becomes essential in predicting long-term performance before clinical usage. The aim of the study was to assess the influence of chamfer preparation depth on failure risk of heat pressed ceramic crowns with and without zirconia framework by means of finite element analysis. 3D models of maxillary central incisor, prepared for full ceramic crowns with different depths of the chamfer margin (between 0.8 and 1.2 mm) and 6-degree tapered walls together with the overlying crowns were generated using literature data (Fig. 1, 2). The crowns were designed with and without a zirconia framework with a thickness of 0.4 mm. For all preparations and crowns, stresses in the pressed ceramic crown, zirconia framework, pressed ceramic veneer, and dentin were evaluated separately. The highest stresses were registered in the dentin. The depth of the preparations had no significant influence on the stress values of the teeth and pressed ceramics for the studied cases, only for the zirconia framework. The zirconia framework decreases the stress values in the veneer.

A Formulation of the Latent Class Vector Model for Pairwise Data

In this research, a latent class vector model for pairwise data is formulated. As compared to the basic vector model, this model yields consistent estimates of the parameters since the number of parameters to be estimated does not increase with the number of subjects. The result of the analysis reveals that the model was stable and could classify each subject to the latent classes representing the typical scales used by these subjects.

Study of Aero-thermal Effects with Heat Radiation in Optical Side Window

In hypersonic environments, the aerothermal effect makes it difficult for the optical side windows of optical guided missiles to withstand high heat. This produces cracking or breaking, resulting in an inability to function. This study used computational fluid mechanics to investigate the external cooling jet conditions of optical side windows. The turbulent models k-ε and k-ω were simulated. To be in better accord with actual aerothermal environments, a thermal radiation model was added to examine suitable amounts of external coolants and the optical window problems of aero-thermodynamics. The simulation results indicate that when there are no external cooling jets, because airflow on the optical window and the tail groove produce vortices, the temperatures in these two locations reach a peak of approximately 1600 K. When the external cooling jets worked at 0.15 kg/s, the surface temperature of the optical windows dropped to approximately 280 K. When adding thermal radiation conditions, because heat flux dissipation was faster, the surface temperature of the optical windows fell from 280 K to approximately 260 K. The difference in influence of the different turbulence models k-ε and k-ω on optical window surface temperature was not significant.

An Investigation into Turbine Blade Tip Leakage Flows at High Speeds

The effect of the blade tip geometry of a high pressure gas turbine is studied experimentally and computationally for high speed leakage flows. For this purpose two simplified models are constructed, one models a flat tip of the blade and the second models a cavity tip of the blade. Experimental results are obtained from a transonic wind tunnel to show the static pressure distribution along the tip wall and provide flow visualization. RANS computations were carried to provide further insight into the mean flow behavior and to calculate the discharge coefficient which is a measure of the flow leaking over the tip. It is shown that in both geometries of tip the flow separates over the tip to form a separation bubble. The bubble is higher for the cavity tip while a complete shock wave system of oblique waves ending with a normal wave can be seen for the flat tip. The discharge coefficient for the flat tip shows less dependence on the pressure ratio over the blade tip than the cavity tip. However, the discharge coefficient for the cavity tip is lower than that of the flat tip, showing a better ability to reduce the leakage flow and thus increase the turbine efficiency.

A Study of Classification Models to Predict Drill-Bit Breakage Using Degradation Signals

Cutting tools are widely used in manufacturing processes and drilling is the most commonly used machining process. Although drill-bits used in drilling may not be expensive, their breakage can cause damage to expensive work piece being drilled and at the same time has major impact on productivity. Predicting drill-bit breakage, therefore, is important in reducing cost and improving productivity. This study uses twenty features extracted from two degradation signals viz., thrust force and torque. The methodology used involves developing and comparing decision tree, random forest, and multinomial logistic regression models for classifying and predicting drill-bit breakage using degradation signals.

Dimensional Modeling of HIV Data Using Open Source

Selecting the data modeling technique for an information system is determined by the objective of the resultant data model. Dimensional modeling is the preferred modeling technique for data destined for data warehouses and data mining, presenting data models that ease analysis and queries which are in contrast with entity relationship modeling. The establishment of data warehouses as components of information system landscapes in many organizations has subsequently led to the development of dimensional modeling. This has been significantly more developed and reported for the commercial database management systems as compared to the open sources thereby making it less affordable for those in resource constrained settings. This paper presents dimensional modeling of HIV patient information using open source modeling tools. It aims to take advantage of the fact that the most affected regions by the HIV virus are also heavily resource constrained (sub-Saharan Africa) whereas having large quantities of HIV data. Two HIV data source systems were studied to identify appropriate dimensions and facts these were then modeled using two open source dimensional modeling tools. Use of open source would reduce the software costs for dimensional modeling and in turn make data warehousing and data mining more feasible even for those in resource constrained settings but with data available.