On the Comparison of Several Goodness of Fit tests under Simple Random Sampling and Ranked Set Sampling

Many works have been carried out to compare the efficiency of several goodness of fit procedures for identifying whether or not a particular distribution could adequately explain a data set. In this paper a study is conducted to investigate the power of several goodness of fit tests such as Kolmogorov Smirnov (KS), Anderson-Darling(AD), Cramer- von- Mises (CV) and a proposed modification of Kolmogorov-Smirnov goodness of fit test which incorporates a variance stabilizing transformation (FKS). The performances of these selected tests are studied under simple random sampling (SRS) and Ranked Set Sampling (RSS). This study shows that, in general, the Anderson-Darling (AD) test performs better than other GOF tests. However, there are some cases where the proposed test can perform as equally good as the AD test.

Tracing Quality Cost in a Luggage Manufacturing Industry

Quality costs are the costs associated with preventing, finding, and correcting defective work. Since the main language of corporate management is money, quality-related costs act as means of communication between the staff of quality engineering departments and the company managers. The objective of quality engineering is to minimize the total quality cost across the life of product. Quality costs provide a benchmark against which improvement can be measured over time. It provides a rupee-based report on quality improvement efforts. It is an effective tool to identify, prioritize and select quality improvement projects. After reviewing through the literature it was noticed that a simplified methodology for data collection of quality cost in a manufacturing industry was required. The quantified standard methodology is proposed for collecting data of various elements of quality cost categories for manufacturing industry. Also in the light of research carried out so far, it is felt necessary to standardise cost elements in each of the prevention, appraisal, internal failure and external failure costs. . Here an attempt is made to standardise the various cost elements applicable to manufacturing industry and data is collected by using the proposed quantified methodology. This paper discusses the case study carried in luggage manufacturing industry.

Bridging the Mental Gap between Convolution Approach and Compartmental Modeling in Functional Imaging: Typical Embedding of an Open Two-Compartment Model into the Systems Theory Approach of Indicator Dilution Theory

Functional imaging procedures for the non-invasive assessment of tissue microcirculation are highly requested, but require a mathematical approach describing the trans- and intercapillary passage of tracer particles. Up to now, two theoretical, for the moment different concepts have been established for tracer kinetic modeling of contrast agent transport in tissues: pharmacokinetic compartment models, which are usually written as coupled differential equations, and the indicator dilution theory, which can be generalized in accordance with the theory of lineartime- invariant (LTI) systems by using a convolution approach. Based on mathematical considerations, it can be shown that also in the case of an open two-compartment model well-known from functional imaging, the concentration-time course in tissue is given by a convolution, which allows a separation of the arterial input function from a system function being the impulse response function, summarizing the available information on tissue microcirculation. Due to this reason, it is possible to integrate the open two-compartment model into the system-theoretic concept of indicator dilution theory (IDT) and thus results known from IDT remain valid for the compartment approach. According to the long number of applications of compartmental analysis, even for a more general context similar solutions of the so-called forward problem can already be found in the extensively available appropriate literature of the seventies and early eighties. Nevertheless, to this day, within the field of biomedical imaging – not from the mathematical point of view – there seems to be a trench between both approaches, which the author would like to get over by exemplary analysis of the well-known model.

Phase Behavior of CO2 and CH4 Hydrate in Porous Media

Hydrate phase equilibria for the binary CO2+water and CH4+water mixtures in silica gel pore of nominal diameters 6, 30, and 100 nm were measured and compared with the calculated results based on van der Waals and Platteeuw model. At a specific temperature, three-phase hydrate-water-vapor (HLV) equilibrium curves for pore hydrates were shifted to the higher-pressure condition depending on pore sizes when compared with those of bulk hydrates. Notably, hydrate phase equilibria for the case of 100 nominal nm pore size were nearly identical with those of bulk hydrates. The activities of water in porous silica gels were modified to account for capillary effect, and the calculation results were generally in good agreement with the experimental data. The structural characteristics of gas hydrates in silica gel pores were investigated through NMR spectroscopy.

Sloshing Control in Tilting Phases of the Pouring Process

We propose a control design scheme that aims to prevent undesirable liquid outpouring and suppress sloshing during the forward and backward tilting phases of the pouring process, for the case of liquid containers carried by manipulators. The proposed scheme combines a partial inverse dynamics controller with a PID controller, tuned with the use of a “metaheuristic" search algorithm. The “metaheuristic" search algorithm tunes the PID controller based on simulation results of the plant-s linearization around the operating point corresponding to the critical tilting angle, where outpouring initiates. Liquid motion is modeled using the well-known pendulumtype model. However, the proposed controller does not require measurements of the liquid-s motion within the tank.

Finding Authoritative Researchers on Academic Web Sites

In this paper, we present a methodology for finding authoritative researchers by analyzing academic Web sites. We show a case study in which we concentrate on a set of Czech computer science departments- Web sites. We analyze the relations between them via hyperlinks and find the most important ones using several common ranking algorithms. We then examine the contents of the research papers present on these sites and determine the most authoritative Czech authors.

Microarrays Denoising via Smoothing of Coefficients in Wavelet Domain

We describe a novel method for removing noise (in wavelet domain) of unknown variance from microarrays. The method is based on a smoothing of the coefficients of the highest subbands. Specifically, we decompose the noisy microarray into wavelet subbands, apply smoothing within each highest subband, and reconstruct a microarray from the modified wavelet coefficients. This process is applied a single time, and exclusively to the first level of decomposition, i.e., in most of the cases, it is not necessary a multirresoltuion analysis. Denoising results compare favorably to the most of methods in use at the moment.

Enhancement of Methane Productivity of Anaerobic Reactors of Wastewater Treatment Plants

This paper describes technological possibilities to enhance methane productionin the anaerobic stabilization of wastewater treatment plant excess sludge. This objective can be achieved by the addition of waste residues: crude glycerol from biodiesel production and residues from fishery. The addition ofglycerol in an amount by weight of 2 – 5% causes enhancement of methane production of about 250 – 400%. At the same time the percentage increase of total solids concentration in the outgoing sludge is ten or more times less. The containment of methane in biogas is higher in case of admixed substrate.

Prevalence of Epstein-Barr Virus Latent Membrane Protein-1 in Jordanian Patients with Hodgkin's Lymphoma and Non- Hodgkin's Lymphoma

The aim of this study was to estimate the frequency of EBV infection in Hodgkin's lymphoma (HL) and non-Hodgkin's lymphoma (NHL) occurring in Jordanian patients. A total of 55 patients with lymphoma were examined in this study. Of 55 patients, 30 and 25 were diagnosed as HL and NHL, respectively. The four HL subtypes were observed with the majority of the cases exhibited the mixed cellularity (MC) subtype followed by the nodular sclerosis (NS). The high grade was found to be the commonest subtype of NHL in our sample, followed by the low grade. The presence of EBV virus was detected by immunostating for expression of latent membrane protein-1 (LMP-1). The frequency of LMP-1 expression occurred more frequent in patients with HL (60.0%) than in patients with NHL (32.0%). The frequency of LMP-1 expression was also higher in patients with MC subtype (61.11%) than those patients with NS (28.57%). No age or gender difference in occurrence of EBV infection was observed among patient with HL. By contrast, the prevalence of EBV infection in NHL patients aged below 50 was lower (16.66%) than in NHL patients aged 50 or above (46.15%). In addition, EBV infection was more frequent in females with NHL (38.46%) than in male with NHL (25%). In NHL cases, the frequency of EBV infection in intermediate grade (60.0%) was high when compared with frequency of low (25%) or high grades (25%). In conclusion, analysis of LMP-1 expression indicates an important role for this viral oncogene in the pathogenesis of EBV-associated malignant lymphomas. These data also support the previous findings that people with EBV may develop lymphoma and that efforts to maintain low lymphoma should be considered for people with EBV infection.

Global Product Development Ways in Modern Thai Economy – Case Studies, Good Practices and Ways to Implement in Thailand

Advances in technology (e.g. the internet, telecommunication) and political changes (fewer trade barriers and an enlarged European Union, ASEAN, NAFTA and other organizations) have led to develop international competition and expand into new markets. Companies in Thailand, Asia and around the globe are increasingly being pressured on price and for faster time to enter the market. At the same time, new markets are appearing and many companies are looking for changes and shifts in their domestic markets. These factors have enabled the rapid growth for companies and globalizing many different business activities during the product development process from research and development (R&D) to production. This research will show and clarify methods how to develop global product. Also, it will show how important is a global product impact into Thai Economy development.

A Dynamic Filter for Removal DC - Offset In Current and Voltage Waveforms

In power systems, protective relays must filter their inputs to remove undesirable quantities and retain signal quantities of interest. This job must be performed accurate and fast. A new method for filtering the undesirable components such as DC and harmonic components associated with the fundamental system signals. The method is s based on a dynamic filtering algorithm. The filtering algorithm has many advantages over some other classical methods. It can be used as dynamic on-line filter without the need of parameters readjusting as in the case of classic filters. The proposed filter is tested using different signals. Effects of number of samples and sampling window size are discussed. Results obtained are presented and discussed to show the algorithm capabilities.

An Adverse Model for Price Discrimination in the Case of Monopoly

We consider a Principal-Agent model with the Principal being a seller who does not know perfectly how much the buyer (the Agent) is willing to pay for the good. The buyer-s preferences are hence his private information. The model corresponds to the nonlinear pricing problem of Maskin and Riley. We assume there are three types of Agents. The model is solved using “informational rents" as variables. In the last section we present the main characteristics of the optimal contracts in asymmetric information and some possible extensions of the model.

Multimodal Reasoning in a Knowledge Engineering Framework for Product Support

Problem solving has traditionally been one of the principal research areas for artificial intelligence. Yet, although artificial intelligence reasoning techniques have been employed in several product support systems, the benefit of integrating product support, knowledge engineering, and problem solving, is still unclear. This paper studies the synergy of these areas and proposes a knowledge engineering framework that integrates product support systems and artificial intelligence techniques. The framework includes four spaces; the data, problem, hypothesis, and solution ones. The data space incorporates the knowledge needed for structured reasoning to take place, the problem space contains representations of problems, and the hypothesis space utilizes a multimodal reasoning approach to produce appropriate solutions in the form of virtual documents. The solution space is used as the gateway between the system and the user. The proposed framework enables the development of product support systems in terms of smaller, more manageable steps while the combination of different reasoning techniques provides a way to overcome the lack of documentation resources.

Comparison of Imputation Techniques for Efficient Prediction of Software Fault Proneness in Classes

Missing data is a persistent problem in almost all areas of empirical research. The missing data must be treated very carefully, as data plays a fundamental role in every analysis. Improper treatment can distort the analysis or generate biased results. In this paper, we compare and contrast various imputation techniques on missing data sets and make an empirical evaluation of these methods so as to construct quality software models. Our empirical study is based on NASA-s two public dataset. KC4 and KC1. The actual data sets of 125 cases and 2107 cases respectively, without any missing values were considered. The data set is used to create Missing at Random (MAR) data Listwise Deletion(LD), Mean Substitution(MS), Interpolation, Regression with an error term and Expectation-Maximization (EM) approaches were used to compare the effects of the various techniques.

New Hybrid Method to Correct for Wind Tunnel Wall- and Support Interference On-line

Because support interference corrections are not properly understood, engineers mostly rely on expensive dummy measurements or CFD calculations. This paper presents a method based on uncorrected wind tunnel measurements and fast calculation techniques (it is a hybrid method) to calculate wall interference, support interference and residual interference (when e.g. a support member closely approaches the wind tunnel walls) for any type of wind tunnel and support configuration. The method provides with a simple formula for the calculation of the interference gradient. This gradient is based on the uncorrected measurements and a successive calculation of the slopes of the interference-free aerodynamic coefficients. For the latter purpose a new vortex-lattice routine is developed that corrects the slopes for viscous effects. A test case of a measurement on a wing proves the value of this hybrid method as trends and orders of magnitudes of the interference are correctly determined.

Correlating Site-Specific Meteorological Data and Power Availability for Small-Scale, Multi-Source Renewable Energy Systems

The paper presents a modelling methodology for small scale multi-source renewable energy systems. Using historical site-specific weather data, the relationships of cost, availability and energy form are visualised as a function of the sizing of photovoltaic arrays, wind turbines, and battery capacity. The specific dependency of each site on its own particular weather patterns show that unique solutions exist for each site. It is shown that in certain cases the capital component cost can be halved if the desired theoretical demand availability is reduced from 100% to 99%.

GEP Considering Purchase Prices, Profits of IPPs and Reliability Criteria Using Hybrid GA and PSO

In this paper, optimal generation expansion planning (GEP) is investigated considering purchase prices, profits of independent power producers (IPPs) and reliability criteria using a new method based on hybrid coded Genetic Algorithm (GA) and Particle Swarm Optimization (PSO). In this approach, optimal purchase price of each IPP is obtained by HCGA and reliability criteria are calculated by PSO technique. It should be noted that reliability criteria and the rate of carbon dioxide (CO2) emission have been considered as constraints of the GEP problem. Finally, the proposed method has been tested on the case study system. The results evaluation show that the proposed method can simply obtain optimal purchase prices of IPPs and is a fast method for calculation of reliability criteria in expansion planning. Also, considering the optimal purchase prices and profits of IPPs in generation expansion planning are caused that the expansion costs are decreased and the problem is solved more exactly.

Modeling, Simulation and Monitoring of Nuclear Reactor Using Directed Graph and Bond Graph

The main objective developed in this paper is to find a graphic technique for modeling, simulation and diagnosis of the industrial systems. This importance is much apparent when it is about a complex system such as the nuclear reactor with pressurized water of several form with various several non-linearity and time scales. In this case the analytical approach is heavy and does not give a fast idea on the evolution of the system. The tool Bond Graph enabled us to transform the analytical model into graphic model and the software of simulation SYMBOLS 2000 specific to the Bond Graphs made it possible to validate and have the results given by the technical specifications. We introduce the analysis of the problem involved in the faults localization and identification in the complex industrial processes. We propose a method of fault detection applied to the diagnosis and to determine the gravity of a detected fault. We show the possibilities of application of the new diagnosis approaches to the complex system control. The industrial systems became increasingly complex with the faults diagnosis procedures in the physical systems prove to become very complex as soon as the systems considered are not elementary any more. Indeed, in front of this complexity, we chose to make recourse to Fault Detection and Isolation method (FDI) by the analysis of the problem of its control and to conceive a reliable system of diagnosis making it possible to apprehend the complex dynamic systems spatially distributed applied to the standard pressurized water nuclear reactor.

Optimization by Ant Colony Hybryde for the Bin-Packing Problem

The problem of bin-packing in two dimensions (2BP) consists in placing a given set of rectangular items in a minimum number of rectangular and identical containers, called bins. This article treats the case of objects with a free orientation of 90Ôùª. We propose an approach of resolution combining optimization by colony of ants (ACO) and the heuristic method IMA to resolve this NP-Hard problem.

Performance Optimization of Data Mining Application Using Radial Basis Function Classifier

Text data mining is a process of exploratory data analysis. Classification maps data into predefined groups or classes. It is often referred to as supervised learning because the classes are determined before examining the data. This paper describes proposed radial basis function Classifier that performs comparative crossvalidation for existing radial basis function Classifier. The feasibility and the benefits of the proposed approach are demonstrated by means of data mining problem: direct Marketing. Direct marketing has become an important application field of data mining. Comparative Cross-validation involves estimation of accuracy by either stratified k-fold cross-validation or equivalent repeated random subsampling. While the proposed method may have high bias; its performance (accuracy estimation in our case) may be poor due to high variance. Thus the accuracy with proposed radial basis function Classifier was less than with the existing radial basis function Classifier. However there is smaller the improvement in runtime and larger improvement in precision and recall. In the proposed method Classification accuracy and prediction accuracy are determined where the prediction accuracy is comparatively high.