Measurement of Steady Streaming from an Oscillating Bubble Using Particle Image Velocimetry

Steady streaming flow fields induced by a 500 mm bubble oscillating at 12 kHz were measured using microscopic particle image velocimetry (PIV). The accuracy of velocity measurement using a micro PIV system was checked by comparing the measured velocity fields with the theoretical velocity profiles in fully developed laminar flow. The steady streaming flow velocities were measured in the sagittal plane of the bubble attached on the wall. Measured velocity fields showed upward jet flow with two symmetric counter-rotating vortices, and the maximum streaming velocity was about 12 mm/s, which was within the velocity ranges measured by other researchers. The measured streamlines were compared with the analytical solution, and they also showed a reasonable agreement.

2-D Ablated Plasma Production Process for Pulsed Ion Beam-Solid Target Interaction

This paper presents a 2-D hydrodynamic model of the ablated plasma when irradiating a 50 μm Al solid target with a single pulsed ion beam. The Lagrange method is used to solve the moving fluid for the ablated plasma production and formation mechanism. In the calculations, a 10-ns-single-pulsed of ion beam with a total energy density of 120 J/cm2, is used. The results show that the ablated plasma was formed after 2 ns of ion beam irradiation and it started to expand right after 4-6 ns. In addition, the 2-D model give a better understanding of pulsed ion beam-solid target ablated plasma production and expansion process clearer.

Learning Styles of University Students in Bangkok: The Characteristics and the Relevant Instructional Context

The purposes of this study are 1) to identify learning styles of university students in Bangkok, and 2) to study the frequency of the relevant instructional context of the identified learning styles. Learning Styles employed in this study are those of Honey and Mumford, which include 1) Reflectors, 2) Theorists, 3) Pragmatists, and 4) Activists. The population comprises 1383 students and 5 lecturers. Research tools are 2 questionnaires – one used for identifying students- learning styles, and the other used for identifying the frequency of the relevant instructional context of the identified learning styles. The research findings reveal that 32.30 percent - are Activists, while 28.10 percent are Theorists, 20.10 are Reflectors, and 19.50 are Pragmatists. In terms of the relevant instructional context of the identified 4 learning styles, it is found that the frequency level of the instructional context is totally in high level. Moreover, 2 lists of the context being conducted most frequently are 'Lead'in activity to review background knowledge,- and 'Information retrieval report.' And these two activities serve the learning styles of theorists and activists. It is, therefore, suggested that more instructional context supporting the activists, the majority of the population, learning best by doing, as well as emotional learning situation should be added.

Correction of Frequent English Writing Errors by Using Coded Indirect Corrective Feedback and Error Treatment

The purposes of this study are 1) to study the frequent English writing errors of students registering the course: Reading and Writing English for Academic Purposes II, and 2) to find out the results of writing error correction by using coded indirect corrective feedback and writing error treatments. Samples include 28 2nd year English Major students, Faculty of Education, Suan Sunandha Rajabhat University. Tool for experimental study includes the lesson plan of the course; Reading and Writing English for Academic Purposes II, and tool for data collection includes 4 writing tests of short texts. The research findings disclose that frequent English writing errors found in this course comprise 7 types of grammatical errors, namely Fragment sentence, Subject-verb agreement, Wrong form of verb tense, Singular or plural noun endings, Run-ons sentence, Wrong form of verb pattern and Lack of parallel structure. Moreover, it is found that the results of writing error correction by using coded indirect corrective feedback and error treatment reveal the overall reduction of the frequent English writing errors and the increase of students’ achievement in the writing of short texts with the significance at .05.

Pilot-scale Study of Horizontal Anaerobic Digester for Biogas Production using Food Waste

A horizontal anaerobic digester was developed and tested in pilot scale for Korean food waste with high water contents (>80%). The hydrogen sulfide in the biogas was removed by a biological desulfurization equipment integrated in the horizontal digester. A mixer of the horizontal digester was designed to easily remove the sediment in the bottom and scum layers on surface in the digester. Experimental result for 120 days of operation of the pilot plant showed a high removal efficiency of 81.2% for organic substance and high stability during the whole operation period were acquired. Also food waste was treated at high organic loading rates over 4 kg•VS/m3∙day and a methane gas production rate of 0.62 m3/kg•VSremoved was accomplished. The biological desulfurization equipment inside the horizontal digester was proven to be an economic and effective method to reduce the biogas desulfurization cost by removing hydrogen sulfide more than 90% without external desulfurization equipments.

A Framework for Product Development Process including HW and SW Components

This paper proposes a framework for product development including hardware and software components. It provides separation of hardware dependent software, modifications of current product development process, and integration of software modules with existing product configuration models and assembly product structures. In order to decide the dependent software, the framework considers product configuration modules and engineering changes of associated software and hardware components. In order to support efficient integration of the two different hardware and software development, a modified product development process is proposed. The process integrates the dependent software development into product development through the interchanges of specific product information. By using existing product data models in Product Data Management (PDM), the framework represents software as modules for product configurations and software parts for product structure. The framework is applied to development of a robot system in order to show its effectiveness.

Object Recognition on Horse Riding Simulator System

In recent years, IT convergence technology has been developed to get creative solution by combining robotics or sports science technology. Object detection and recognition have mainly applied to sports science field that has processed by recognizing face and by tracking human body. But object detection and recognition using vision sensor is challenge task in real world because of illumination. In this paper, object detection and recognition using vision sensor applied to sports simulator has been introduced. Face recognition has been processed to identify user and to update automatically a person athletic recording. Human body has tracked to offer a most accurate way of riding horse simulator. Combined image processing has been processed to reduce illumination adverse affect because illumination has caused low performance in detection and recognition in real world application filed. Face has recognized using standard face graph and human body has tracked using pose model, which has composed of feature nodes generated diverse face and pose images. Face recognition using Gabor wavelet and pose recognition using pose graph is robust to real application. We have simulated using ETRI database, which has constructed on horse riding simulator.

Production of As Isotopes in the Interaction of natGe with 14-30 MeV Protons

Cross sections of As radionuclides in the interaction of natGe with 14-30 MeV protons have been deduced by off-line y-ray spectroscopy to find optimal reaction channels leading to radiotracers for positron emission tomography. The experimental results were compared with the previous results and those estimated by the compound nucleus reaction model.

Automatic Choice of Topics for Seminars by Clustering Students According to Their Profile

The new framework the Higher Education is immersed in involves a complete change in the way lecturers must teach and students must learn. Whereas the lecturer was the main character in traditional education, the essential goal now is to increase the students' participation in the process. Thus, one of the main tasks of lecturers in this new context is to design activities of different nature in order to encourage such participation. Seminars are one of the activities included in this environment. They are active sessions that enable going in depth into specific topics as support of other activities. They are characterized by some features such as favoring interaction between students and lecturers or improving their communication skills. Hence, planning and organizing strategic seminars is indeed a great challenge for lecturers with the aim of acquiring knowledge and abilities. This paper proposes a method using Artificial Intelligence techniques to obtain student profiles from their marks and preferences. The goal of building such profiles is twofold. First, it facilitates the task of splitting the students into different groups, each group with similar preferences and learning difficulties. Second, it makes it easy to select adequate topics to be a candidate for the seminars. The results obtained can be either a guarantee of what the lecturers could observe during the development of the course or a clue to reconsider new methodological strategies in certain topics.

An Algorithm of Finite Capacity Material Requirement Planning System for Multi-stage Assembly Flow Shop

This paper aims to develop an algorithm of finite capacity material requirement planning (FCMRP) system for a multistage assembly flow shop. The developed FCMRP system has two main stages. The first stage is to allocate operations to the first and second priority work centers and also determine the sequence of the operations on each work center. The second stage is to determine the optimal start time of each operation by using a linear programming model. Real data from a factory is used to analyze and evaluate the effectiveness of the proposed FCMRP system and also to guarantee a practical solution to the user. There are five performance measures, namely, the total tardiness, the number of tardy orders, the total earliness, the number of early orders, and the average flow-time. The proposed FCMRP system offers an adjustable solution which is a compromised solution among the conflicting performance measures. The user can adjust the weight of each performance measure to obtain the desired performance. The result shows that the combination of FCMRP NP3 and EDD outperforms other combinations in term of overall performance index. The calculation time for the proposed FCMRP system is about 10 minutes which is practical for the planners of the factory.

Measuring Process Component Design on Achieving Managerial Goals

Process-oriented software development is a new software development paradigm in which software design is modeled by a business process which is in turn translated into a process execution language for execution. The building blocks of this paradigm are software units that are composed together to work according to the flow of the business process. This new paradigm still exhibits the characteristic of the applications built with the traditional software component technology. This paper discusses an approach to apply a traditional technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses, and these process components can be reused to design the business processes of other application domains. The decomposition considers five managerial goals, namely cost effectiveness, ease of assembly, customization, reusability, and maintainability. The paper presents how to design or decompose process components from a business process model and measure some technical features of the design that would affect the managerial goals. A comparison between the measurement values from different designs can tell which process component design is more appropriate for the managerial goals that have been set. The proposed approach can be applied in Web Services environment which accommodates process-oriented software development.

FleGSens – Secure Area Monitoring Using Wireless Sensor Networks

In the project FleGSens, a wireless sensor network (WSN) for the surveillance of critical areas and properties is currently developed which incorporates mechanisms to ensure information security. The intended prototype consists of 200 sensor nodes for monitoring a 500m long land strip. The system is focused on ensuring integrity and authenticity of generated alarms and availability in the presence of an attacker who may even compromise a limited number of sensor nodes. In this paper, two of the main protocols developed in the project are presented, a tracking protocol to provide secure detection of trespasses within the monitored area and a protocol for secure detection of node failures. Simulation results of networks containing 200 and 2000 nodes as well as the results of the first prototype comprising a network of 16 nodes are presented. The focus of the simulations and prototype are functional testing of the protocols and particularly demonstrating the impact and cost of several attacks.

Public R and D Risk and Risk Management Policy

R&D risk management has been suggested as one of the management approaches for accomplishing the goals of public R&D investment. The investment in basic science and core technology development is the essential roles of government for securing the social base needed for continuous economic growth. And, it is also an important role of the science and technology policy sectors to generate a positive environment in which the outcomes of public R&D can be diffused in a stable fashion by controlling the uncertainties and risk factors in advance that may arise during the application of such achievements to society and industry. Various policies have already been implemented to manage uncertainties and variables that may have negative impact on accomplishing public R& investment goals. But we may derive new policy measures for complementing the existing policies and for exploring progress direction by analyzing them in a policy package from the viewpoint of R&D risk management.

Biosignal Measurement System Based On Ultra-Wide Band Human Body Communication

A wrist-band type biosignal measurement system and its data transfer through human body communication (HBC) were investigated. An HBC method based on pulses of ultra-wide band instead of using frequency or amplitude modulations was studied and implemented since the system became very compact and it was more suited for personal or mobile health monitoring. Our system measured photo-plethysmogram (PPG) and measured PPG signals were transmitted through a finger to a monitoring PC system. The device was compact and low-power consuming. HBC communication has very strongsecurity measures since it does not use wireless network.Furthermore, biosignal monitoring system becomes handy because it does not need to have wire connections.

A Knowledge Engineering Workshop: Application for Choise Car

This paper proposes a declarative language for knowledge representation (Ibn Rochd), and its environment of exploitation (DeGSE). This DeGSE system was designed and developed to facilitate Ibn Rochd writing applications. The system was tested on several knowledge bases by ascending complexity, culminating in a system for recognition of a plant or a tree, and advisors to purchase a car, for pedagogical and academic guidance, or for bank savings and credit. Finally, the limits of the language and research perspectives are stated.

Robust Face Recognition using AAM and Gabor Features

In this paper, we propose a face recognition algorithm using AAM and Gabor features. Gabor feature vectors which are well known to be robust with respect to small variations of shape, scaling, rotation, distortion, illumination and poses in images are popularly employed for feature vectors for many object detection and recognition algorithms. EBGM, which is prominent among face recognition algorithms employing Gabor feature vectors, requires localization of facial feature points where Gabor feature vectors are extracted. However, localization method employed in EBGM is based on Gabor jet similarity and is sensitive to initial values. Wrong localization of facial feature points affects face recognition rate. AAM is known to be successfully applied to localization of facial feature points. In this paper, we devise a facial feature point localization method which first roughly estimate facial feature points using AAM and refine facial feature points using Gabor jet similarity-based facial feature localization method with initial points set by the rough facial feature points obtained from AAM, and propose a face recognition algorithm using the devised localization method for facial feature localization and Gabor feature vectors. It is observed through experiments that such a cascaded localization method based on both AAM and Gabor jet similarity is more robust than the localization method based on only Gabor jet similarity. Also, it is shown that the proposed face recognition algorithm using this devised localization method and Gabor feature vectors performs better than the conventional face recognition algorithm using Gabor jet similarity-based localization method and Gabor feature vectors like EBGM.

The Performance of Predictive Classification Using Empirical Bayes

This research is aimed to compare the percentages of correct classification of Empirical Bayes method (EB) to Classical method when data are constructed as near normal, short-tailed and long-tailed symmetric, short-tailed and long-tailed asymmetric. The study is performed using conjugate prior, normal distribution with known mean and unknown variance. The estimated hyper-parameters obtained from EB method are replaced in the posterior predictive probability and used to predict new observations. Data are generated, consisting of training set and test set with the sample sizes 100, 200 and 500 for the binary classification. The results showed that EB method exhibited an improved performance over Classical method in all situations under study.

A Quantitative Approach to Strategic Design of Component-Based Business Process Models

A new paradigm for software design and development models software by its business process, translates the model into a process execution language, and has it run by a supporting execution engine. This process-oriented paradigm promotes modeling of software by less technical users or business analysts as well as rapid development. Since business process models may be shared by different organizations and sometimes even by different business domains, it is interesting to apply a technique used in traditional software component technology to design reusable business processes. This paper discusses an approach to apply a technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses with an aim that the process components can be reusable in different process-based software models. The approach is quantitative because the quality of process component design is measured from technical features of the process components. The approach is also strategic because the measured quality is determined against business-oriented component management goals. A software tool has been developed to measure how good a process component design is, according to the required managerial goals and comparing to other designs. We also discuss how we benefit from reusable process components.

VISUAL JESS: AN Expandable Visual Generator of Oriented Object Expert systems

The utility of expert system generators has been widely recognized in many applications. Several generators based on concept of the paradigm object, have been recently proposed. The generator of oriented object expert system (GSEOO) offers languages that are often complex and difficult to use. We propose in this paper an extension of the expert system generator, JESS, which permits a friendly use of this expert system. The new tool, called VISUAL JESS, bring two main improvements to JESS. The first improvement concerns the easiness of its utilization while giving back transparency to the syntax and semantic aspects of the JESS programming language. The second improvement permits an easy access and modification of the JESS knowledge basis. The implementation of VISUAL JESS is made so that it is extensible and portable.

Computational Investigation of Air-Gas Venturi Mixer for Powered Bi-Fuel Diesel Engine

In a bi-fuel diesel engine, the carburetor plays a vital role in switching from fuel gas to petrol mode operation and viceversa. The carburetor is the most important part of the fuel system of a diesel engine. All diesel engines carry variable venturi mixer carburetors. The basic operation of the carburetor mainly depends on the restriction barrel called the venturi. When air flows through the venturi, its speed increases and its pressure decreases. The main challenge focuses on designing a mixing device which mixes the supplied gas is the incoming air at an optimum ratio. In order to surmount the identified problems, the way fuel gas and air flow in the mixer have to be analyzed. In this case, the Computational Fluid Dynamics or CFD approach is applied in design of the prototype mixer. The present work is aimed at further understanding of the air and fuel flow structure by performing CFD studies using a software code. In this study for mixing air and gas in the condition that has been mentioned in continuance, some mixers have been designed. Then using of computational fluid dynamics, the optimum mixer has been selected. The results indicated that mixer with 12 holes can produce a homogenous mixture than those of 8-holes and 6-holes mixer. Also the result showed that if inlet convergency was smoother than outlet divergency, the mixture get more homogenous, the reason of that is in increasing turbulence in outlet divergency.