Application of Formal Methods for Designing a Separation Kernel for Embedded Systems

A separation-kernel-based operating system (OS) has been designed for use in secure embedded systems by applying formal methods to the design of the separation-kernel part. The separation kernel is a small OS kernel that provides an abstract distributed environment on a single CPU. The design of the separation kernel was verified using two formal methods, the B method and the Spin model checker. A newly designed semi-formal method, the extended state transition method, was also applied. An OS comprising the separation-kernel part and additional OS services on top of the separation kernel was prototyped on the Intel IA-32 architecture. Developing and testing of a prototype embedded application, a point-of-sale application, on the prototype OS demonstrated that the proposed architecture and the use of formal methods to design its kernel part are effective for achieving a secure embedded system having a high-assurance separation kernel.

Foot Anthropometry of Primary School Children in the South of Thailand

The objective of the research was to study of foot anthropometry of children aged 7-12 years in the South of Thailand Thirty-three dimensions were measured on 305 male and 295 female subjects with 3 age ranges (7-12 years old). The instrumentation consists of four types of anthropometer, digital vernier caliper, digital height gauge and measuring tape. The mean values and standard deviations of average age, height, and weight of the male subjects were 9.52(±1.70) years, 137.80(±11.55) cm, and 37.57(±11.65) kg. Female average age, height, and weight subjects were 9.53(±1.70) years, 137.88(±11.55) cm, and 34.90(±11.57) kg respectively. The comparison of the 33 comparison measured anthropometric. Between male and female subjects were sexual differences in size on women in almost all areas of significance (p

Three Dimensional Analysis of Pollution Dispersion in Street Canyon

Three dimensional simulations are carried out to estimate the effect of wind direction, wind speed and geometry on the flow and dispersion of vehicular pollutant in a street canyon. The pollutant sources are motor vehicles passing between the two buildings. Suitable emission factors for petrol and diesel vehicles at varying vehicle speed are used for the estimation of the rate of emission from the streets. The dispersion of automobile pollutant released from the street is simulated by introducing vehicular emission source term as a fixed-flux boundary condition at the ground level over the road. The emission source term is suitably calculated by adopting emission factors from literature for varying conditions of street traffic. It is observed that increase in wind angle disturbs the symmetric pattern of pollution distribution along the street length. The concentration increases in the far end of the street as compared to the near end.

RFID-ready Master Data Management for Reverse Logistics

Sharing consistent and correct master data among disparate applications in a reverse-logistics chain has long been recognized as an intricate problem. Although a master data management (MDM) system can surely assume that responsibility, applications that need to co-operate with it must comply with proprietary query interfaces provided by the specific MDM system. In this paper, we present a RFID-ready MDM system which makes master data readily available for any participating applications in a reverse-logistics chain. We propose a RFID-wrapper as a part of our MDM. It acts as a gateway between any data retrieval request and query interfaces that process it. With the RFID-wrapper, any participating applications in a reverse-logistics chain can easily retrieve master data in a way that is analogous to retrieval of any other RFID-based logistics transactional data.

CFD Analysis on Aerodynamic Design Optimization of Wind Turbine Rotor Blades

Wind energy has been shown to be one of the most viable sources of renewable energy. With current technology, the low cost of wind energy is competitive with more conventional sources of energy such as coal. Most blades available for commercial grade wind turbines incorporate a straight span-wise profile and airfoil shaped cross sections. These blades are found to be very efficient at lower wind speeds in comparison to the potential energy that can be extracted. However as the oncoming wind speed increases the efficiency of the blades decreases as they approach a stall point. This paper explores the possibility of increasing the efficiency of the blades at higher wind speeds while maintaining efficiency at the lower wind speeds. The design intends to maintain efficiency at lower wind speeds by selecting the appropriate orientation and size of the airfoil cross sections based on a low oncoming wind speed and given constant rotation rate. The blades will be made more efficient at higher wind speeds by implementing a swept blade profile. Performance was investigated using the computational fluid dynamics (CFD).

UD Covariance Factorization for Unscented Kalman Filter using Sequential Measurements Update

Extended Kalman Filter (EKF) is probably the most widely used estimation algorithm for nonlinear systems. However, not only it has difficulties arising from linearization but also many times it becomes numerically unstable because of computer round off errors that occur in the process of its implementation. To overcome linearization limitations, the unscented transformation (UT) was developed as a method to propagate mean and covariance information through nonlinear transformations. Kalman filter that uses UT for calculation of the first two statistical moments is called Unscented Kalman Filter (UKF). Square-root form of UKF (SRUKF) developed by Rudolph van der Merwe and Eric Wan to achieve numerical stability and guarantee positive semi-definiteness of the Kalman filter covariances. This paper develops another implementation of SR-UKF for sequential update measurement equation, and also derives a new UD covariance factorization filter for the implementation of UKF. This filter is equivalent to UKF but is computationally more efficient.

Li4SiO4 Prepared by Sol-gel Method as Potential Host for LISICON Structured Solid Electrolytes

In this study, Li4SiO4 powder was successfully synthesized via sol gel method followed by drying at 150oC. Lithium oxide, Li2O and silicon oxide, SiO2 were used as the starting materials with citric acid as the chelating agent. The obtained powder was then sintered at various temperatures. Crystallographic phase analysis, morphology and ionic conductivity were investigated systematically employing X-ray diffraction, Fourier Transform Infrared, Scanning Electron Microscopy and AC impedance spectroscopy. XRD result showed the formation of pure monoclinic Li4SiO4 crystal structure with lattice parameters a = 5.140 Å, b = 6.094 Å, c = 5.293 Å, β = 90o in the sample sintered at 750oC. This observation was confirmed by FTIR analysis. The bulk conductivity of this sample at room temperature was 3.35 × 10-6 S cm-1 and the highest bulk conductivity of 1.16 × 10-4 S cm-1 was obtained at 100°C. The results indicated that, the Li4SiO4 compound has potential to be used as host for LISICON structured solid electrolyte for low temperature application.

Management of Multimedia Contents for Distributed e-Learning System

We have developed a distributed asynchronous Web based training system. In order to improve the scalability and robustness of this system, all contents and functions are realized on mobile agents. These agents are distributed to computers, and they can use a Peer to Peer network that modified Content-Addressable Network. In the proposed system, only text data can be included in a exercise. To make our proposed system more useful, the mechanism that it not only adapts to multimedia data but also it doesn-t influence the user-s learning even if the size of exercise becomes large is necessary.

Determining Factors for ISO14001 EMS Implementation among SMEs in Malaysia: A Resource Based View

This research aimed to find out the determining factors for ISO 14001 EMS implementation among SMEs in Malaysia from the Resource based view. A cross-sectional approach using survey was conducted. A research model been proposed which comprises of ISO 14001 EMS implementation as the criterion variable while physical capital resources (i.e. environmental performance tracking and organizational infrastructures), human capital resources (i.e. top management commitment and support, training and education, employee empowerment and teamwork) and organizational capital resources (i.e. recognition and reward, organizational culture and organizational communication) as the explanatory variables. The research findings show that only environmental performance tracking, top management commitment and support and organizational culture are found to be positively and significantly associated with ISO 14001 EMS implementation. It is expected that this research will shed new knowledge and provide a base for future studies about the role played by firm-s internal resources.

Parallel Algorithm for Numerical Solution of Three-Dimensional Poisson Equation

In this paper developed and realized absolutely new algorithm for solving three-dimensional Poisson equation. This equation used in research of turbulent mixing, computational fluid dynamics, atmospheric front, and ocean flows and so on. Moreover in the view of rising productivity of difficult calculation there was applied the most up-to-date and the most effective parallel programming technology - MPI in combination with OpenMP direction, that allows to realize problems with very large data content. Resulted products can be used in solving of important applications and fundamental problems in mathematics and physics.

Performance Evaluation of Routing Protocols For High Density Ad Hoc Networks based on Qos by GlomoSim Simulator

Ad hoc networks are characterized by multihop wireless connectivity, frequently changing network topology and the need for efficient dynamic routing protocols. We compare the performance of three routing protocols for mobile ad hoc networks: Dynamic Source Routing (DSR) , Ad Hoc On-Demand Distance Vector Routing (AODV), location-aided routing(LAR1).The performance differentials are analyzed using varying network load, mobility, and network size. We simulate protocols with GLOMOSIM simulator. Based on the observations, we make recommendations about when the performance of either protocol can be best.

Protein-Protein Interaction Detection Based on Substring Sensitivity Measure

Detecting protein-protein interactions is a central problem in computational biology and aberrant such interactions may have implicated in a number of neurological disorders. As a result, the prediction of protein-protein interactions has recently received considerable attention from biologist around the globe. Computational tools that are capable of effectively identifying protein-protein interactions are much needed. In this paper, we propose a method to detect protein-protein interaction based on substring similarity measure. Two protein sequences may interact by the mean of the similarities of the substrings they contain. When applied on the currently available protein-protein interaction data for the yeast Saccharomyces cerevisiae, the proposed method delivered reasonable improvement over the existing ones.

Improving Co-integration Trading Rule Profitability with Forecasts from an Artificial Neural Network

Co-integration models the long-term, equilibrium relationship of two or more related financial variables. Even if cointegration is found, in the short run, there may be deviations from the long run equilibrium relationship. The aim of this work is to forecast these deviations using neural networks and create a trading strategy based on them. A case study is used: co-integration residuals from Australian Bank Bill futures are forecast and traded using various exogenous input variables combined with neural networks. The choice of the optimal exogenous input variables chosen for each neural network, undertaken in previous work [1], is validated by comparing the forecasts and corresponding profitability of each, using a trading strategy.

Approaches and Schemes for Storing DTD-Independent XML Data in Relational Databases

The volume of XML data exchange is explosively increasing, and the need for efficient mechanisms of XML data management is vital. Many XML storage models have been proposed for storing XML DTD-independent documents in relational database systems. Benchmarking is the best way to highlight pros and cons of different approaches. In this study, we use a common benchmarking scheme, known as XMark to compare the most cited and newly proposed DTD-independent methods in terms of logical reads, physical I/O, CPU time and duration. We show the effect of Label Path, extracting values and storing in another table and type of join needed for each method's query answering.

Condition Monitoring in the Management of Maintenance in a Large Scale Precision CNC Machining Manufacturing Facility

The manufacture of large-scale precision aerospace components using CNC requires a highly effective maintenance strategy to ensure that the required accuracy can be achieved over many hours of production. This paper reviews a strategy for a maintenance management system based on Failure Mode Avoidance, which uses advanced techniques and technologies to underpin a predictive maintenance strategy. It is shown how condition monitoring (CM) is important to predict potential failures in high precision machining facilities and achieve intelligent and integrated maintenance management. There are two distinct ways in which CM can be applied. One is to monitor key process parameters and observe trends which may indicate a gradual deterioration of accuracy in the product. The other is the use of CM techniques to monitor high status machine parameters enables trends to be observed which can be corrected before machine failure and downtime occurs. It is concluded that the key to developing a flexible and intelligent maintenance framework in any precision manufacturing operation is the ability to evaluate reliably and routinely machine tool condition using condition monitoring techniques within a framework of Failure Mode Avoidance.

Optimization of Process Parameters of Pressure Die Casting using Taguchi Methodology

The present work analyses different parameters of pressure die casting to minimize the casting defects. Pressure diecasting is usually applied for casting of aluminium alloys. Good surface finish with required tolerances and dimensional accuracy can be achieved by optimization of controllable process parameters such as solidification time, molten temperature, filling time, injection pressure and plunger velocity. Moreover, by selection of optimum process parameters the pressure die casting defects such as porosity, insufficient spread of molten material, flash etc. are also minimized. Therefore, a pressure die casting component, carburetor housing of aluminium alloy (Al2Si2O5) has been considered. The effects of selected process parameters on casting defects and subsequent setting of parameters with the levels have been accomplished by Taguchi-s parameter design approach. The experiments have been performed as per the combination of levels of different process parameters suggested by L18 orthogonal array. Analyses of variance have been performed for mean and signal-to-noise ratio to estimate the percent contribution of different process parameters. Confidence interval has also been estimated for 95% consistency level and three conformational experiments have been performed to validate the optimum level of different parameters. Overall 2.352% reduction in defects has been observed with the help of suggested optimum process parameters.

MONARC: A Case Study on Simulation Analysis for LHC Activities

The scale, complexity and worldwide geographical spread of the LHC computing and data analysis problems are unprecedented in scientific research. The complexity of processing and accessing this data is increased substantially by the size and global span of the major experiments, combined with the limited wide area network bandwidth available. We present the latest generation of the MONARC (MOdels of Networked Analysis at Regional Centers) simulation framework, as a design and modeling tool for large scale distributed systems applied to HEP experiments. We present simulation experiments designed to evaluate the capabilities of the current real-world distributed infrastructure to support existing physics analysis processes and the means by which the experiments bands together to meet the technical challenges posed by the storage, access and computing requirements of LHC data analysis within the CMS experiment.

Personal Digital Assistants for Fieldwork Training in College Campus

Education supported by mobile computers has been widely done for some time. Teachers have attempted to use mobile computers and to find concrete subjects for student-s fieldwork training in college education. The purpose of this research is to develop software for Personal Digital Assistant (PDA) to conduct fieldwork in our campus, and to report a fieldwork class using PDAs in the curriculum of the Department of Regional Environment Studies.

A Quantitative Tool for Analyze Process Design

Some quality control tools use non metric subjective information coming from experts, who qualify the intensity of relations existing inside processes, but without quantifying them. In this paper we have developed a quality control analytic tool, measuring the impact or strength of the relationship between process operations and product characteristics. The tool includes two models: a qualitative model, allowing relationships description and analysis; and a formal quantitative model, by means of which relationship quantification is achieved. In the first one, concepts from the Graphs Theory were applied to identify those process elements which can be sources of variation, that is, those quality characteristics or operations that have some sort of prelacy over the others and that should become control items. Also the most dependent elements can be identified, that is those elements receiving the effects of elements identified as variation sources. If controls are focused in those dependent elements, efficiency of control is compromised by the fact that we are controlling effects, not causes. The second model applied adapts the multivariate statistical technique of Covariance Structural Analysis. This approach allowed us to quantify the relationships. The computer package LISREL was used to obtain statistics and to validate the model.

An Autonomous Collaborative Forecasting System Implementation – The First Step towards Successful CPFR System

In the past decade, artificial neural networks (ANNs) have been regarded as an instrument for problem-solving and decision-making; indeed, they have already done with a substantial efficiency and effectiveness improvement in industries and businesses. In this paper, the Back-Propagation neural Networks (BPNs) will be modulated to demonstrate the performance of the collaborative forecasting (CF) function of a Collaborative Planning, Forecasting and Replenishment (CPFR®) system. CPFR functions the balance between the sufficient product supply and the necessary customer demand in a Supply and Demand Chain (SDC). Several classical standard BPN will be grouped, collaborated and exploited for the easy implementation of the proposed modular ANN framework based on the topology of a SDC. Each individual BPN is applied as a modular tool to perform the task of forecasting SKUs (Stock-Keeping Units) levels that are managed and supervised at a POS (point of sale), a wholesaler, and a manufacturer in an SDC. The proposed modular BPN-based CF system will be exemplified and experimentally verified using lots of datasets of the simulated SDC. The experimental results showed that a complex CF problem can be divided into a group of simpler sub-problems based on the single independent trading partners distributed over SDC, and its SKU forecasting accuracy was satisfied when the system forecasted values compared to the original simulated SDC data. The primary task of implementing an autonomous CF involves the study of supervised ANN learning methodology which aims at making “knowledgeable" decision for the best SKU sales plan and stocks management.