Six Sigma Solutions and its Benefit-Cost Ratio for Quality Improvement

This is an application research presenting the improvement of production quality using the six sigma solutions and the analyses of benefit-cost ratio. The case of interest is the production of tile-concrete. Such production has faced with the problem of high nonconforming products from an inappropriate surface coating and had low process capability based on the strength property of tile. Surface coating and tile strength are the most critical to quality of this product. The improvements followed five stages of six sigma solutions. After the improvement, the production yield was improved to 80% as target required and the defective products from coating process was remarkably reduced from 29.40% to 4.09%. The process capability based on the strength quality was increased from 0.87 to 1.08 as customer oriented. The improvement was able to save the materials loss for 3.24 millions baht or 0.11 million dollars. The benefits from the improvement were analyzed from (1) the reduction of the numbers of non conforming tile using its factory price for surface coating improvement and (2) the materials saved from the increment of process capability. The benefit-cost ratio of overall improvement was high as 7.03. It was non valuable investment in define, measure, analyses and the initial of improve stages after that it kept increasing. This was due to there were no benefits in define, measure, and analyze stages of six sigma since these three stages mainly determine the cause of problem and its effects rather than improve the process. The benefit-cost ratio starts existing in the improve stage and go on. Within each stage, the individual benefitcost ratio was much higher than the accumulative one as there was an accumulation of cost since the first stage of six sigma. The consideration of the benefit-cost ratio during the improvement project helps make decisions for cost saving of similar activities during the improvement and for new project. In conclusion, the determination of benefit-cost ratio behavior through out six sigma implementation period provides the useful data for managing quality improvement for the optimal effectiveness. This is the additional outcome from the regular proceeding of six sigma.

Design of High Torque Elbow Joint for Above Elbow Prosthesis

Above Elbow Prosthesis is one of the most commonly amputated or missing limbs. The research is done for modelling techniques of upper limb prosthesis and design of high torque, light weight and compact in size elbow actuator. The purposed actuator consists of a DC motor, planetary gear set and a harmonic drive. The calculations show that the actuator is good enough to be used in real life powered prosthetic upper limb or rehabilitation exoskeleton.

Electrical Resistivity of Subsurface: Field and Laboratory Assessment

The objective of this paper is to study the electrical resistivity complexity between field and laboratory measurement, in order to improve the effectiveness of data interpretation for geophysical ground resistivity survey. The geological outcrop in Penang, Malaysia with an obvious layering contact was chosen as the study site. Two dimensional geoelectrical resistivity imaging were used in this study to maps the resistivity distribution of subsurface, whereas few subsurface sample were obtained for laboratory advance. In this study, resistivity of samples in original conditions is measured in laboratory by using time domain low-voltage technique, particularly for granite core sample and soil resistivity measuring set for soil sample. The experimentation results from both schemes are studied, analyzed, calibrated and verified, including basis and correlation, degree of tolerance and characteristics of substance. Consequently, the significant different between both schemes is explained comprehensively within this paper.

Analytical Proposal to Damage Assessment of Buried Continuous Pipelines during External Blast Loading

In this paper, transversal vibration of buried pipelines during loading induced by underground explosions is analyzed. The pipeline is modeled as an infinite beam on an elastic foundation, so that soil-structure interaction is considered by means of transverse linear springs along the pipeline. The pipeline behavior is assumed to be ideal elasto-plastic which an ultimate strain value limits the plastic behavior. The blast loading is considered as a point load, considering the affected length at some point of the pipeline, in which the magnitude decreases exponentially with time. A closed-form solution for the quasi-static problem is carried out for both elastic and elasticperfect plastic behaviors of pipe materials. At the end, a comparative study on steel and polyethylene pipes with different sizes buried in various soil conditions, affected by a predefined underground explosion is conducted, in which effect of each parameter is discussed.

Concurrent Approach to Data Parallel Model using Java

Parallel programming models exist as an abstraction of hardware and memory architectures. There are several parallel programming models in commonly use; they are shared memory model, thread model, message passing model, data parallel model, hybrid model, Flynn-s models, embarrassingly parallel computations model, pipelined computations model. These models are not specific to a particular type of machine or memory architecture. This paper expresses the model program for concurrent approach to data parallel model through java programming.

Study of Mechanical Properties for the Aluminum Bronze Matrix Composites of Hot Pressing

The aluminum bronze matrix alumina composites using hot press and resin infiltration were investigated to study their porosities, hardness, bending strengths, and microstructures. The experiment results show that the hardness of the sintered composites with the decrease of porosity increases. The composites without and with resin infiltration have about HRF 42-61 of about 34-40% of porosity and about HRF 62-83 of about 30-36% of porosity, respectively. Besides, the alumina composites contain a more amount of iron and nickel powders would cause a lower bending strength due to forming some weaker bonding among the iron, nickel, copper, aluminum under this hot pressing of shorter time.

A New Design of Mobile Thermoelectric Power Generation System

This paper presents a compact thermoelectric power generator system based on temperature difference across the element. The system can transfer the burning heat energy to electric energy directly. The proposed system has a thermoelectric generator and a power control box. In the generator, there are 4 thermoelectric modules (TEMs), each of which uses 2 thermoelectric chips (TEs) and 2 cold sinks, 1 thermal absorber, and 1 thermal conduction flat board. In the power control box, there are 1 storing energy device, 1 converter, and 1 inverter. The total net generating power is about 11W. This system uses commercial portable gas stoves or burns timber or the coal as the heat source, which is easily obtained. It adopts solid-state thermoelectric chips as heat inverter parts. The system has the advantages of being light-weight, quite, and mobile, requiring no maintenance, and havng easily-supplied heat source. The system can be used a as long as burning is allowed. This system works well for highly-mobilized outdoors situations by providing a power for illumination, entertainment equipment or the wireless equipment at refuge. Under heavy storms such as typhoon, when the solar panels become ineffective and the wind-powered machines malfunction, the thermoelectric power generator can continue providing the vital power.

Comparative Analysis of the Public Funding for Greek Universities: An Ordinal DEA/MCDM Approach

This study performs a comparative analysis of the 21 Greek Universities in terms of their public funding, awarded for covering their operating expenditure. First it introduces a DEA/MCDM model that allocates the fund into four expenditure factors in the most favorable way for each university. Then, it presents a common, consensual assessment model to reallocate the amounts, remaining in the same level of total public budget. From the analysis it derives that a number of universities cannot justify the public funding in terms of their size and operational workload. For them, the sufficient reduction of their public funding amount is estimated as a future target. Due to the lack of precise data for a number of expenditure criteria, the analysis is based on a mixed crisp-ordinal data set.

Meta-Classification using SVM Classifiers for Text Documents

Text categorization is the problem of classifying text documents into a set of predefined classes. In this paper, we investigated three approaches to build a meta-classifier in order to increase the classification accuracy. The basic idea is to learn a metaclassifier to optimally select the best component classifier for each data point. The experimental results show that combining classifiers can significantly improve the accuracy of classification and that our meta-classification strategy gives better results than each individual classifier. For 7083 Reuters text documents we obtained a classification accuracies up to 92.04%.

An Overview of Islanding Detection Methods in Photovoltaic Systems

The issue of unintentional islanding in PV grid interconnection still remains as a challenge in grid-connected photovoltaic (PV) systems. This paper discusses the overview of popularly used anti-islanding detection methods, practically applied in PV grid-connected systems. Anti-islanding methods generally can be classified into four major groups, which include passive methods, active methods, hybrid methods and communication base methods. Active methods have been the preferred detection technique over the years due to very small non-detected zone (NDZ) in small scale distribution generation. Passive method is comparatively simpler than active method in terms of circuitry and operations. However, it suffers from large NDZ that significantly reduces its performance. Communication base methods inherit the advantages of active and passive methods with reduced drawbacks. Hybrid method which evolved from the combination of both active and passive methods has been proven to achieve accurate anti-islanding detection by many researchers. For each of the studied anti-islanding methods, the operation analysis is described while the advantages and disadvantages are compared and discussed. It is difficult to pinpoint a generic method for a specific application, because most of the methods discussed are governed by the nature of application and system dependent elements. This study concludes that the setup and operation cost is the vital factor for anti-islanding method selection in order to achieve minimal compromising between cost and system quality.

Combining Color and Layout Features for the Identification of Low-resolution Documents

This paper proposes a method, combining color and layout features, for identifying documents captured from lowresolution handheld devices. On one hand, the document image color density surface is estimated and represented with an equivalent ellipse and on the other hand, the document shallow layout structure is computed and hierarchically represented. The combined color and layout features are arranged in a symbolic file, which is unique for each document and is called the document-s visual signature. Our identification method first uses the color information in the signatures in order to focus the search space on documents having a similar color distribution, and finally selects the document having the most similar layout structure in the remaining search space. Finally, our experiment considers slide documents, which are often captured using handheld devices.

Data Envelopment Analysis with Partially Perfect Objects

This paper presents a simplified version of Data Envelopment Analysis (DEA) - a conventional approach to evaluating the performance and ranking of competitive objects characterized by two groups of factors acting in opposite directions: inputs and outputs. DEA with a Perfect Object (DEA PO) augments the group of actual objects with a virtual Perfect Object - the one having greatest outputs and smallest inputs. It allows for obtaining an explicit analytical solution and making a step to an absolute efficiency. This paper develops this approach further and introduces a DEA model with Partially Perfect Objects. DEA PPO consecutively eliminates the smallest relative inputs or greatest relative outputs, and applies DEA PO to the reduced collections of indicators. The partial efficiency scores are combined to get the weighted efficiency score. The computational scheme remains simple, like that of DEA PO, but the advantage of the DEA PPO is taking into account all of the inputs and outputs for each actual object. Firm evaluation is considered as an example.

A Novel Arabic Text Steganography Method Using Letter Points and Extensions

This paper presents a new steganography approach suitable for Arabic texts. It can be classified under steganography feature coding methods. The approach hides secret information bits within the letters benefiting from their inherited points. To note the specific letters holding secret bits, the scheme considers the two features, the existence of the points in the letters and the redundant Arabic extension character. We use the pointed letters with extension to hold the secret bit 'one' and the un-pointed letters with extension to hold 'zero'. This steganography technique is found attractive to other languages having similar texts to Arabic such as Persian and Urdu.

Dynamic Bus Binding for Low Power Using Multiple Binding Tables

A conventional binding method for low power in a high-level synthesis mainly focuses on finding an optimal binding for an assumed input data, and obtains only one binding table. In this paper, we show that a binding method which uses multiple binding tables gets better solution compared with the conventional methods which use a single binding table, and propose a dynamic bus binding scheme for low power using multiple binding tables. The proposed method finds multiple binding tables for the proper partitions of an input data, and switches binding tables dynamically to produce the minimum total switching activity. Experimental result shows that the proposed method obtains a binding solution having 12.6-28.9% smaller total switching activity compared with the conventional methods.

Modeling of Pulsatile Blood Flow in a Weak Magnetic Field

Blood pulse is an important human physiological signal commonly used for the understanding of the individual physical health. Current methods of non-invasive blood pulse sensing require direct contact or access to the human skin. As such, the performances of these devices tend to vary with time and are subjective to human body fluids (e.g. blood, perspiration and skin-oil) and environmental contaminants (e.g. mud, water, etc). This paper proposes a simulation model for the novel method of non-invasive acquisition of blood pulse using the disturbance created by blood flowing through a localized magnetic field. The simulation model geometry represents a blood vessel, a permanent magnet, a magnetic sensor, surrounding tissues and air in 2-dimensional. In this model, the velocity and pressure fields in the blood stream are described based on Navier-Stroke equations and the walls of the blood vessel are assumed to have no-slip condition. The blood assumes a parabolic profile considering a laminar flow for blood in major artery near the skin. And the inlet velocity follows a sinusoidal equation. This will allow the computational software to compute the interactions between the magnetic vector potential generated by the permanent magnet and the magnetic nanoparticles in the blood. These interactions are simulated based on Maxwell equations at the location where the magnetic sensor is placed. The simulated magnetic field at the sensor location is found to assume similar sinusoidal waveform characteristics as the inlet velocity of the blood. The amplitude of the simulated waveforms at the sensor location are compared with physical measurements on human subjects and found to be highly correlated.

Predicting Extrusion Process Parameters Using Neural Networks

The objective of this paper is to estimate realistic principal extrusion process parameters by means of artificial neural network. Conventionally, finite element analysis is used to derive process parameters. However, the finite element analysis of the extrusion model does not consider the manufacturing process constraints in its modeling. Therefore, the process parameters obtained through such an analysis remains highly theoretical. Alternatively, process development in industrial extrusion is to a great extent based on trial and error and often involves full-size experiments, which are both expensive and time-consuming. The artificial neural network-based estimation of the extrusion process parameters prior to plant execution helps to make the actual extrusion operation more efficient because more realistic parameters may be obtained. And so, it bridges the gap between simulation and real manufacturing execution system. In this work, a suitable neural network is designed which is trained using an appropriate learning algorithm. The network so trained is used to predict the manufacturing process parameters.

Polyurethane Nanofibers Obtained By Electrospinning Process

Electrospinning is a broadly used technology to obtain polymeric nanofibers ranging from several micrometers down to several hundred nanometers for a wide range of applications. It offers unique capabilities to produce nanofibers with controllable porous structure. With smaller pores and higher surface area than regular fibers, electrospun fibers have been successfully applied in various fields, such as, nanocatalysis, tissue engineering scaffolds, protective clothing, filtration, biomedical, pharmaceutical, optical electronics, healthcare, biotechnology, defense and security, and environmental engineering. In this study, polyurethane nanofibers were obtained under different electrospinning parameters. Fiber morphology and diameter distribution were investigated in order to understand them as a function of process parameters.

A Fiber Optic Interferometric Sensor for Dynamic Measurement

An optical fiber Fabry-Perot interferometer (FFPI) is proposed and demonstrated for dynamic measurements in a mechanical vibrating target. A polishing metal with a low reflectance value adhered to a mechanical vibrator was excited via a function generator at various excitation frequencies. Output interference fringes were generated by modulating the reference and sensing signal at the output arm. A fringe-counting technique was used for interpreting the displacement information on the dedicated computer. The fiber interferometer has been found the capability of the displacement measurements of 1.28 μm – 96.01 μm. A commercial displacement sensor was employed as a reference sensor for investigating the measurement errors from the fiber sensor. A maximum percentage measurement error of approximately 1.59 % was obtained.

A Proposed Technique for Software Development Risks Identification by using FTA Model

Software Development Risks Identification (SDRI), using Fault Tree Analysis (FTA), is a proposed technique to identify not only the risk factors but also the causes of the appearance of the risk factors in software development life cycle. The method is based on analyzing the probable causes of software development failures before they become problems and adversely affect a project. It uses Fault tree analysis (FTA) to determine the probability of a particular system level failures that are defined by A Taxonomy for Sources of Software Development Risk to deduce failure analysis in which an undesired state of a system by using Boolean logic to combine a series of lower-level events. The major purpose of this paper is to use the probabilistic calculations of Fault Tree Analysis approach to determine all possible causes that lead to software development risk occurrence

Computational Fluid Dynamics Expert System using Artificial Neural Networks

The design of a modern aircraft is based on three pillars: theoretical results, experimental test and computational simulations. As a results of this, Computational Fluid Dynamic (CFD) solvers are widely used in the aeronautical field. These solvers require the correct selection of many parameters in order to obtain successful results. Besides, the computational time spent in the simulation depends on the proper choice of these parameters. In this paper we create an expert system capable of making an accurate prediction of the number of iterations and time required for the convergence of a computational fluid dynamic (CFD) solver. Artificial neural network (ANN) has been used to design the expert system. It is shown that the developed expert system is capable of making an accurate prediction the number of iterations and time required for the convergence of a CFD solver.