The Application of Adaptive Tabu Search Algorithm and Averaging Model to the Optimal Controller Design of Buck Converters

The paper presents the applications of artificial intelligence technique called adaptive tabu search to design the controller of a buck converter. The averaging model derived from the DQ and generalized state-space averaging methods is applied to simulate the system during a searching process. The simulations using such averaging model require the faster computational time compared with that of the full topology model from the software packages. The reported model is suitable for the work in the paper in which the repeating calculation is needed for searching the best solution. The results will show that the proposed design technique can provide the better output waveforms compared with those designed from the classical method.

SATA: A Web Based Scheduling Support System

Developing a university course schedule is difficult. This is due to the limitations in the resources available. The process is made even harder with different faculties or departments having different ways of stating their schedule requirements. The person in charge of taking the schedule requirements and turning them into a proper course schedule is not only burden with the task of allocating the appropriate classes and time to lecturers and students, they also need to understand the schedule requirements. Therefore a scheduling support system named SATA is developed to assist ICRESS in the course scheduling process. SATA has been put to use for several semesters and the results have been encouraging. It won a bronze medal in the 2008 Invention, Innovation and Design competition (IID-08) and has been submitted to be patented in October 2008

Thai Perception on Litecoin Value

This research analyzes factors affecting the success of Litecoin Value within Thailand and develops a guideline for selfreliance for effective business implementation. Samples in this study included 119 people through surveys. The results revealed four main factors affecting the success as follows: 1) Future Career training should be pursued in applied Litecoin development. 2) Didn't grasp the concept of a digital currency or see the benefit of a digital currency. 3) There is a great need to educate the next generation of learners on the benefits of Litecoin within the community. 4) A great majority didn't know what Litecoin was. The guideline for self-reliance planning consisted of 4 aspects: 1) Development planning: by arranging meet up groups to conduct further education on Litecoin and share solutions on adoption into every day usage. Local communities need to develop awareness of the usefulness of Litecoin and share the value of Litecoin among friends and family. 2) Computer Science and Business Management staff should develop skills to expand on the benefits of Litecoin within their departments. 3) Further research should be pursued on how Litecoin Value can improve business and tourism within Thailand. 4) Local communities should focus on developing Litecoin awareness by encouraging street vendors to accept Litecoin as another form of payment for services rendered.

Procurement for Management Services in Delivery of Public Construction Projects in Poland

Construction projects can be implemented under various contractual and organizational systems. They can be divided into two groups: systems without the managing company where the Client manages the process, and systems with the managing company, where management is entrusted to an external company. In the public sector of the Polish market there are two ways of delivery of construction projects with the participation of the manager: one is to assign operations to another party, the so called Project Supervisor, whilst the other results from the application of FIDIC conditions of contract, which entail appointment of the Engineer. The decision is to be made by the Client and depends on various factors. On the public procurement market in Poland the selection of construction project manager boils down to awarding the contract for such a service. The selection can be done by one of eight public procurement procedures identified by the procurement law. The paper provides the analysis of 96 contracts for services awarded in 2011, which employed construction management. The study aimed to investigate the methods and criteria for selecting managers, applied in practice by the Polish public Clients.

Evaluate Aging Effect of SBS Modified Bitumen

One of the important factors of cracks on the asphalt pavements is bitumen aging that associated with the loss of volatile components and oxidation of asphalt binder. This paper is about effect of Styrene-Butadiene-Styrene (SBS) polymer on asphalt aging In order to decrease asphalt aging effects. For this purpose samples of base bitumen and SBS modified bitumen aged according to the rolling thin film oven test (RTFOT) and pressure aging vessel (PAV), respectively. Properties of each sample were evaluated using Fourier Transform Infrared (FTIR) spectroscopy, n-heptane precipitation, viscosity test, softening point test and penetration test. FT-IR Analysis, showed lower oxidation of SBS modified bitumen than base bitumen, after aging.

Asymptotic Analysis of Instant Messaging Service with Relay Nodes

In this paper, we provide complete end-to-end delay analyses including the relay nodes for instant messages. Message Session Relay Protocol (MSRP) is used to provide congestion control for large messages in the Instant Messaging (IM) service. Large messages are broken into several chunks. These chunks may traverse through a maximum number of two relay nodes before reaching destination according to the IETF specification of the MSRP relay extensions. We discuss the current solutions of sending large instant messages and introduce a proposal to reduce message flows in the IM service. We consider virtual traffic parameter i.e., the relay nodes are stateless non-blocking for scalability purpose. This type of relay node is also assumed to have input rate at constant bit rate. We provide a new scheduling policy that schedules chunks according to their previous node?s delivery time stamp tags. Validation and analysis is shown for such scheduling policy. The performance analysis with the model introduced in this paper is simple and straight forward, which lead to reduced message flows in the IM service.

Grid Coordination with Marketmaker Agents

Market based models are frequently used in the resource allocation on the computational grid. However, as the size of the grid grows, it becomes difficult for the customer to negotiate directly with all the providers. Middle agents are introduced to mediate between the providers and customers and facilitate the resource allocation process. The most frequently deployed middle agents are the matchmakers and the brokers. The matchmaking agent finds possible candidate providers who can satisfy the requirements of the consumers, after which the customer directly negotiates with the candidates. The broker agents are mediating the negotiation with the providers in real time. In this paper we present a new type of middle agent, the marketmaker. Its operation is based on two parallel operations - through the investment process the marketmaker is acquiring resources and resource reservations in large quantities, while through the resale process it sells them to the customers. The operation of the marketmaker is based on the fact that through its global view of the grid it can perform a more efficient resource allocation than the one possible in one-to-one negotiations between the customers and providers. We present the operation and algorithms governing the operation of the marketmaker agent, contrasting it with the matchmaker and broker agents. Through a series of simulations in the task oriented domain we compare the operation of the three agents types. We find that the use of marketmaker agent leads to a better performance in the allocation of large tasks and a significant reduction of the messaging overhead.

Semi-automatic Background Detection in Microscopic Images

The last years have seen an increasing use of image analysis techniques in the field of biomedical imaging, in particular in microscopic imaging. The basic step for most of the image analysis techniques relies on a background image free of objects of interest, whether they are cells or histological samples, to perform further analysis, such as segmentation or mosaicing. Commonly, this image consists of an empty field acquired in advance. However, many times achieving an empty field could not be feasible. Or else, this could be different from the background region of the sample really being studied, because of the interaction with the organic matter. At last, it could be expensive, for instance in case of live cell analyses. We propose a non parametric and general purpose approach where the background is built automatically stemming from a sequence of images containing even objects of interest. The amount of area, in each image, free of objects just affects the overall speed to obtain the background. Experiments with different kinds of microscopic images prove the effectiveness of our approach.

Knowledge Management Model for Managing Knowledge among Related Organizations

Transferring information developed by other peoples is an ordinary event that happens during daily conversations, for example when employees sea each other in the organization, or when they are having lunch together, or attending a meeting, they use to talk about their experience, and discuss about their current projects, and talk about their successes over some specific problems. Despite the potential value of leveraging organizational memory and expertise by using OMS and ER, still small organizations haven-t been able to capitalize on its promised value. Each organization has its internal knowledge management system, in some of organizations the system face the lack of expert people to save their experience in the repository and in another hand on some other organizations there are lots of expert people but the organization doesn-t have the maximum use of their knowledge.

The Roles of Community Based Telecenters in Bridging the Digital Divide in Rural Malaysia

Malaysia is aggressive in promoting the usage of ICT to its mass population through the support by the government policies and programs targeting the general population. However, with the uneven distribution of the basic telecommunication infrastructure between the urban and rural area, cost for being “interconnected" that is considered high among the poorer rural population and the lack of local contents that suit the rural population needs or lifestyles, it is still a challenge for Malaysia to achieve its Vision 2020 Agenda moving the nation towards an information society by the year 2020. Among the existing programs that have been carried out by the government to encourage the usage of ICT by the rural population is “Kedaikom", a community telecenter with the general aim is to engage the community to get exposed and to use the ICT, encouraging the diffusion of the ICT technology to the rural population. The research investigated by using a questionnaire survey of how Kedaikom, as a community telecenter could play a role in encouraging the rural or underserved community to use the ICT. The result from the survey has proven that the community telecenter could bridge the digital divide between the underserved rural population and the well-accessed urban population in Malaysia. More of the rural population, especially from the younger generation and those with higher educational background are using the community telecenter to be connected to the ICT.

Adaptive Weighted Averaging Filter Using the Appropriate Number of Consecutive Frames

In this paper, we propose a novel adaptive spatiotemporal filter that utilizes image sequences in order to remove noise. The consecutive frames include: current, previous and next noisy frames. The filter proposed in this paper is based upon the weighted averaging pixels intensity and noise variance in image sequences. It utilizes the Appropriate Number of Consecutive Frames (ANCF) based on the noisy pixels intensity among the frames. The number of consecutive frames is adaptively calculated for each region in image and its value may change from one region to another region depending on the pixels intensity within the region. The weights are determined by a well-defined mathematical criterion, which is adaptive to the feature of spatiotemporal pixels of the consecutive frames. It is experimentally shown that the proposed filter can preserve image structures and edges under motion while suppressing noise, and thus can be effectively used in image sequences filtering. In addition, the AWA filter using ANCF is particularly well suited for filtering sequences that contain segments with abruptly changing scene content due to, for example, rapid zooming and changes in the view of the camera.

Towards Self-ware via Swarm-Array Computing

The work reported in this paper proposes Swarm-Array computing, a novel technique inspired by swarm robotics, and built on the foundations of autonomic and parallel computing. The approach aims to apply autonomic computing constructs to parallel computing systems and in effect achieve the self-ware objectives that describe self-managing systems. The constitution of swarm-array computing comprising four constituents, namely the computing system, the problem/task, the swarm and the landscape is considered. Approaches that bind these constituents together are proposed. Space applications employing FPGAs are identified as a potential area for applying swarm-array computing for building reliable systems. The feasibility of a proposed approach is validated on the SeSAm multi-agent simulator and landscapes are generated using the MATLAB toolkit.

Body Mass Index for Australian Athletes Participating in Rugby Union, Soccer and Touch Football at the World Masters Games

Whilst there is growing evidence that activity across the lifespan is beneficial for improved health, there are also many changes involved with the aging process and subsequently the potential for reduced indices of health. Data gathered on a subsample of 535 football code athletes, aged 31-72 yrs ( = 47.4, s = ±7.1), competing at the Sydney World Masters Games (2009) demonstrated a significantly (p < 0.001), reduced classification of obesity using Body Mass Index (BMI) when compared to the general Australian population. This evidence of improved classification in one index of health (BMI < 30) for master athletes (when compared to the general population) implies there are either improved levels of this index of health due to adherence to sport or possibly the reduced BMI is advantageous and contributes to this cohort adhering (or being attracted) to masters sport. Demonstration of this proportionately under-investigated World Masters Games population having improved health over the general population is of particular interest.

Retrospective Synthetic Focusing with Correlation Weighting for Very High Frame Rate Ultrasound

The need of high frame-rate imaging has been triggered by the new applications of ultrasound imaging to transient elastography and real-time 3D ultrasound. Using plane wave excitation (PWE) is one of the methods to achieve very high frame-rate imaging since an image can be formed with a single insonification. However, due to the lack of transmit focusing, the image quality with PWE is lower compared with those using conventional focused transmission. To solve this problem, we propose a filter-retrieved transmit focusing (FRF) technique combined with cross-correlation weighting (FRF+CC weighting) for high frame-rate imaging with PWE. A restrospective focusing filter is designed to simultaneously minimize the predefined sidelobe energy associated with single PWE and the filter energy related to the signal-to-noise-ratio (SNR). This filter attempts to maintain the mainlobe signals and to reduce the sidelobe ones, which gives similar mainlobe signals and different sidelobes between the original PWE and the FRF baseband data. Normalized cross-correlation coefficient at zero lag is calculated to quantify the degree of similarity at each imaging point and used as a weighting matrix to the FRF baseband data to further suppress sidelobes, thus improving the filter-retrieved focusing quality.

Effects of Feeding Glycerol to Lactating Dairy Cows on Milk Production and Composition

A study was conducted to determine the effect of feeding glycerol on dairy cows performance. Twenty four Holstein Friesian crossbred (>87.5% Holstein Friesian) lactating dairy cows in early lactation; averaging 13+2.4 kg of milk, 64+45 days in milk, 55+16 months old and 325+26 kg live weight, were stratified for milk yield, days in milk, age, stage of lactation and body weight, and then randomly allocated to three treatment groups. All cows were fed approximate 8 kg of concentrate together with ad libitum corn silage and freely access to clean water. Nil or 150 and 300g of glycerol were supplemented to the cows according to treatment groups. All cows consumed similar concentrate, corn silage and total DM and NELP. There were no significant differences in DM intake, CP intake, NELP intake, milk and milk composition yields. All cows had similar fat, protein, lactose, solid not fat and total solid percentage. All cows gain similar live weight. The present study indicated that, supplementation of glycerol did not enhance milk yield, milk composition and live weight change.

Probabilistic Electrical Power Generation Modeling Using Decimal to Binary Conversion

Generation system reliability assessment is an important task which can be performed using deterministic or probabilistic techniques. The probabilistic approaches have significant advantages over the deterministic methods. However, more complicated modeling is required by the probabilistic approaches. Power generation model is a basic requirement for this assessment. One form of the generation models is the well known capacity outage probability table (COPT). Different analytical techniques have been used to construct the COPT. These approaches require considerable mathematical modeling of the generating units. The unit-s models are combined to build the COPT which will add more burdens on the process of creating the COPT. Decimal to Binary Conversion (DBC) technique is widely and commonly applied in electronic systems and computing This paper proposes a novel utilization of the DBC to create the COPT without engaging in analytical modeling or time consuming simulations. The simple binary representation , “0 " and “1 " is used to model the states o f generating units. The proposed technique is proven to be an effective approach to build the generation model.

Source Optimisation of Laser-Plasma Bremmstrahlung for Applications in Engineering Imaging

High Power Lasers produce an intense burst of Bremmstrahlung radiation which has potential applications in broadband x-ray radiography. Since the radiation produced is through the interaction of accelerated electrons with the remaining laser target, these bursts are extremely short – in the region of a few ps. As a result, the laser-produced x-rays are capable of imaging complex dynamic objects with zero motion blur.

A Basic Study on Ubiquitous Overloaded Vehicles Regulation System

Load managing method on road became necessary since overloaded vehicles occur damage on road facilities and existing systems for preventing this damage still show many problems.Accordingly, efficient managing system for preventing overloaded vehicles could be organized by using the road itself as a scale by applying genetic algorithm to analyze the load and the drive information of vehicles.Therefore, this paper organized Ubiquitous sensor network system for development of intelligent overload vehicle regulation system, also in this study, to use the behavior of road, the transformation was measured by installing underground box type indoor model and indoor experiment was held using genetic algorithm. And we examined wireless possibility of overloaded vehicle regulation system through experiment of the transmission and reception distance.If this system will apply to road and bridge, might be effective for economy and convenience through establishment of U-IT system..

Deriving Causal Explanation from Qualitative Model Reasoning

This paper discusses a qualitative simulator QRiOM that uses Qualitative Reasoning (QR) technique, and a process-based ontology to model, simulate and explain the behaviour of selected organic reactions. Learning organic reactions requires the application of domain knowledge at intuitive level, which is difficult to be programmed using traditional approach. The main objective of QRiOM is to help learners gain a better understanding of the fundamental organic reaction concepts, and to improve their conceptual comprehension on the subject by analyzing the multiple forms of explanation generated by the software. This paper focuses on the generation of explanation based on causal theories to explicate various phenomena in the chemistry subject. QRiOM has been tested with three classes problems related to organic chemistry, with encouraging results. This paper also presents the results of preliminary evaluation of QRiOM that reveal its explanation capability and usefulness.

Novel Use of a Quality Assurance Tool for Integrating Technology to HSE

The product development process (PDP) in the Technology group plays a very important role in the launch of any product. While a manufacturing process encourages the use of certain measures to reduce health, safety and environmental (HSE) risks on the shop floor, the PDP concentrates on the use of Geometric Dimensioning and Tolerancing (GD&T) to develop a flawless design. Furthermore, PDP distributes and coordinates activities between different departments such as marketing, purchasing, and manufacturing. However, it is seldom realized that PDP makes a significant contribution to developing a product that reduces HSE risks by encouraging the Technology group to use effective GD&T. The GD&T is a precise communication tool that uses a set of symbols, rules, and definitions to mathematically define parts to be manufactured. It is a quality assurance method widely used in the oil and gas sector. Traditionally it is used to ensure the interchangeability of a part without affecting its form, fit, and function. Parts that do not meet these requirements are rejected during quality audits. This paper discusses how the Technology group integrates this quality assurance tool into the PDP and how the tool plays a major role in helping the HSE department in its goal towards eliminating HSE incidents. The PDP involves a thorough risk assessment and establishes a method to address those risks during the design stage. An illustration shows how GD&T helped reduce safety risks by ergonomically improving assembling operations. A brief discussion explains how tolerances provided on a part help prevent finger injury. This tool has equipped Technology to produce fixtures, which are used daily in operations as well as manufacturing. By applying GD&T to create good fits, HSE risks are mitigated for operating personnel. Both customers and service providers benefit from reduced safety risks.