Functionality of Negotiation Agent on Value-based Design Decision

This paper presents functionality of negotiation agent on value-based design decision. The functionality is based on the characteristics of the system and goal specification. A Prometheus Design Tool model was used for developing the system. Group functionality will be the attribute for negotiation agents, which comprises a coordinator agent and decision- maker agent. The results of the testing of the system to a building system selection on valuebased decision environment are also presented.

Simulating Climate Change (Temperature and Soil Moisture) in a Mixed-Deciduous Forest, Ontario, Canada

To simulate expected climate change, we implemented a two-factor (temperature and soil moisture) field design in a forest in Ontario, Canada. To manipulate moisture input, we erected rain-exclusion structures. Under each structure, plots were watered with one of three treatments and thermally controlled with three heat treatments to simulate changes in air temperature and rainfall based on the climate model (GCM) predictions for the study area. Environmental conditions (including untreated controls) were monitored tracking air temperature, soil temperature, soil moisture, and photosynthetically active radiation. We measured rainfall and relative humidity at the site outside the rain-exclusion structures. Analyses of environmental conditions demonstrates that the temperature manipulation was most effective at maintaining target temperature during the early part of the growing season, but it was more difficult to keep the warmest treatment at 5º C above ambient by late summer. Target moisture regimes were generally achieved however incoming solar radiation was slightly attenuated by the structures.

Experimental Results about the Dynamics of the Generalized Belief Propagation Used on LDPC Codes

In the context of channel coding, the Generalized Belief Propagation (GBP) is an iterative algorithm used to recover the transmission bits sent through a noisy channel. To ensure a reliable transmission, we apply a map on the bits, that is called a code. This code induces artificial correlations between the bits to send, and it can be modeled by a graph whose nodes are the bits and the edges are the correlations. This graph, called Tanner graph, is used for most of the decoding algorithms like Belief Propagation or Gallager-B. The GBP is based on a non unic transformation of the Tanner graph into a so called region-graph. A clear advantage of the GBP over the other algorithms is the freedom in the construction of this graph. In this article, we explain a particular construction for specific graph topologies that involves relevant performance of the GBP. Moreover, we investigate the behavior of the GBP considered as a dynamic system in order to understand the way it evolves in terms of the time and in terms of the noise power of the channel. To this end we make use of classical measures and we introduce a new measure called the hyperspheres method that enables to know the size of the attractors.

Osmotic Dehydration of Beetroot in Salt Solution: Optimization of Parameters through Statistical Experimental Design

Response surface methodology was used for quantitative investigation of water and solids transfer during osmotic dehydration of beetroot in aqueous solution of salt. Effects of temperature (25 – 45oC), processing time (30–150 min), salt concentration (5–25%, w/w) and solution to sample ratio (5:1 – 25:1) on osmotic dehydration of beetroot were estimated. Quadratic regression equations describing the effects of these factors on the water loss and solids gain were developed. It was found that effects of temperature and salt concentrations were more significant on the water loss than the effects of processing time and solution to sample ratio. As for solids gain processing time and salt concentration were the most significant factors. The osmotic dehydration process was optimized for water loss, solute gain, and weight reduction. The optimum conditions were found to be: temperature – 35oC, processing time – 90 min, salt concentration – 14.31% and solution to sample ratio 8.5:1. At these optimum values, water loss, solid gain and weight reduction were found to be 30.86 (g/100 g initial sample), 9.43 (g/100 g initial sample) and 21.43 (g/100 g initial sample) respectively.

Experimental Parallel Architecture for Rendering 3D Model into MPEG-4 Format

This paper will present the initial findings of a research into distributed computer rendering. The goal of the research is to create a distributed computer system capable of rendering a 3D model into an MPEG-4 stream. This paper outlines the initial design, software architecture and hardware setup for the system. Distributed computing means designing and implementing programs that run on two or more interconnected computing systems. Distributed computing is often used to speed up the rendering of graphical imaging. Distributed computing systems are used to generate images for movies, games and simulations. A topic of interest is the application of distributed computing to the MPEG-4 standard. During the course of the research, a distributed system will be created that can render a 3D model into an MPEG-4 stream. It is expected that applying distributed computing principals will speed up rendering, thus improving the usefulness and efficiency of the MPEG-4 standard

Automatic Inspection of Percussion Caps by Means of Combined 2D and 3D Machine Vision Techniques

The exhaustive quality control is becoming more and more important when commercializing competitive products in the world's globalized market. Taken this affirmation as an undeniable truth, it becomes critical in certain sector markets that need to offer the highest restrictions in quality terms. One of these examples is the percussion cap mass production, a critical element assembled in firearm ammunition. These elements, built in great quantities at a very high speed, must achieve a minimum tolerance deviation in their fabrication, due to their vital importance in firing the piece of ammunition where they are built in. This paper outlines a machine vision development for the 100% inspection of percussion caps obtaining data from 2D and 3D simultaneous images. The acquisition speed and precision of these images from a metallic reflective piece as a percussion cap, the accuracy of the measures taken from these images and the multiple fabrication errors detected make the main findings of this work.

A Numerical Study of Single-phase Forced Convective Heat Transfer in Tube in Tube Heat Exchangers

Three dimensional simulations in tube in tube heat exchangers are investigated numerically in this study. In these simulations forced convective heat transfer and laminar flow of single-phase water are considered. In order to measure heat transfer parameters in these heat exchangers, FLUENT CFD Solver is used in this numerical method. For the purpose of creating geometry and exert boundary and initial conditions in the present model, finite volume method in Computational Fluid Dynamics is used in this study. In the present study, at each Z-location, variation of local temperatures, heat flux and Nusselt number at the whole tube is investigated in detail. Thereafter, averaged computational Nusselt number in this model is calculated. In addition, conceivable pressure drops have been obtained at each Z-location in this model. Then, pressure drop values in the present model are explored. Finally, all the numerical results for this kind of heat exchanger will be discussed precisely.

Analyzing Multi-Labeled Data Based on the Roll of a Concept against a Semantic Range

Classifying data hierarchically is an efficient approach to analyze data. Data is usually classified into multiple categories, or annotated with a set of labels. To analyze multi-labeled data, such data must be specified by giving a set of labels as a semantic range. There are some certain purposes to analyze data. This paper shows which multi-labeled data should be the target to be analyzed for those purposes, and discusses the role of a label against a set of labels by investigating the change when a label is added to the set of labels. These discussions give the methods for the advanced analysis of multi-labeled data, which are based on the role of a label against a semantic range.

The Effect of Breaststroke Swimming Exercise to Increase the Value of Peak Expiratory Flow

The purpose of this study is to investigate the influence of breaststroke swimming exercise to improving the peak expiratory flow. Methode: This study used 17 students of men aged 19-21 years, APE values measured before and after the study. Style swimming workout done in accordance with a program that has been made. Result: Value of peak expiratory flow in male students obtained on average before exercise (530 ± 15 811) liters / min and after doing the exercises (540.59 ± 17 092) liters / minute. Paired ttest showed t = -6.446 and p = 0.000, which means there are differences in peak expiratory flow values before and after exercise swimming breaststroke. Conclusion: The conclusion is the breaststroke swimming exercise can be improving of peak expiratory flow.

A Comparison of First and Second Order Training Algorithms for Artificial Neural Networks

Minimization methods for training feed-forward networks with Backpropagation are compared. Feedforward network training is a special case of functional minimization, where no explicit model of the data is assumed. Therefore due to the high dimensionality of the data, linearization of the training problem through use of orthogonal basis functions is not desirable. The focus is functional minimization on any basis. A number of methods based on local gradient and Hessian matrices are discussed. Modifications of many methods of first and second order training methods are considered. Using share rates data, experimentally it is proved that Conjugate gradient and Quasi Newton?s methods outperformed the Gradient Descent methods. In case of the Levenberg-Marquardt algorithm is of special interest in financial forecasting.

Auto Tuning of PID Controller for MIMO Processes

One of the most basic functions of control engineers is tuning of controllers. There are always several process loops in the plant necessitate of tuning. The auto tuned Proportional Integral Derivative (PID) Controllers are designed for applications where large load changes are expected or the need for extreme accuracy and fast response time exists. The algorithm presented in this paper is used for the tuning PID controller to obtain its parameters with a minimum computing complexity. It requires continuous analysis of variation in few parameters, and let the program to do the plant test and calculate the controller parameters to adjust and optimize the variables for the best performance. The algorithm developed needs less time as compared to a normal step response test for continuous tuning of the PID through gain scheduling.

Flow Discharge Determination in Straight Compound Channels Using ANNs

Although many researchers have studied the flow hydraulics in compound channels, there are still many complicated problems in determination of their flow rating curves. Many different methods have been presented for these channels but extending them for all types of compound channels with different geometrical and hydraulic conditions is certainly difficult. In this study, by aid of nearly 400 laboratory and field data sets of geometry and flow rating curves from 30 different straight compound sections and using artificial neural networks (ANNs), flow discharge in compound channels was estimated. 13 dimensionless input variables including relative depth, relative roughness, relative width, aspect ratio, bed slope, main channel side slopes, flood plains side slopes and berm inclination and one output variable (flow discharge), have been used in ANNs. Comparison of ANNs model and traditional method (divided channel method-DCM) shows high accuracy of ANNs model results. The results of Sensitivity analysis showed that the relative depth with 47.6 percent contribution, is the most effective input parameter for flow discharge prediction. Relative width and relative roughness have 19.3 and 12.2 percent of importance, respectively. On the other hand, shape parameter, main channel and flood plains side slopes with 2.1, 3.8 and 3.8 percent of contribution, have the least importance.

Computer Models of the Vestibular Head Tilt Response, and Their Relationship to EVestG and Meniere's Disease

This paper attempts to explain response components of Electrovestibulography (EVestG) using a computer simulation of a three-canal model of the vestibular system. EVestG is a potentially new diagnostic method for Meniere's disease. EVestG is a variant of Electrocochleography (ECOG), which has been used as a standard method for diagnosing Meniere's disease - it can be used to measure the SP/AP ratio, where an SP/AP ratio greater than 0.4-0.5 is indicative of Meniere-s Disease. In EVestG, an applied head tilt replaces the acoustic stimulus of ECOG. The EVestG output is also an SP/AP type plot, where SP is the summing potential, and AP is the action potential amplitude. AP is thought of as being proportional to the size of a population of afferents in an excitatory neural firing state. A simulation of the fluid volume displacement in the vestibular labyrinth in response to various types of head tilts (ipsilateral, backwards and horizontal rotation) was performed, and a simple neural model based on these simulations developed. The simple neural model shows that the change in firing rate of the utricle is much larger in magnitude than the change in firing rates of all three semi-circular canals following a head tilt (except in a horizontal rotation). The data suggests that the change in utricular firing rate is a minimum 2-3 orders of magnitude larger than changes in firing rates of the canals during ipsilateral/backward tilts. Based on these results, the neural response recorded by the electrode in our EVestG recordings is expected to be dominated by the utricle in ipsilateral/backward tilts (It is important to note that the effect of the saccule and efferent signals were not taken into account in this model). If the utricle response dominates the EVestG recordings as the modeling results suggest, then EVestG has the potential to diagnose utricular hair cell damage due to a viral infection (which has been cited as one possible cause of Meniere's Disease).

Innovation Knowledge and Capability, Work Efficiency of Accountants and the Success of SME Registered in Nakorn Pathom Province

The objectives of this research were to compare the success of SME registered in Nakorn Pathom Province divided in personal data also to study the relations between the innovation knowledge and capability and the success of SME registered in Nakorn Pathom Province and to study the relations between the work efficiency and the success of SME registered in Nakorn Pathom Province. A questionnaire was utilized as a tool to collect data. Statistics utilized in this research included frequency, percentage, mean, standard deviation, and multiple regression analysis. Data were analyzed by using Statistical Package for the Social Sciences.The findings revealed that the majority of respondents were male with the age between 25-34 years old, hold undergraduate degree, married and stay together. The average income of respondents was between 10,001-20,000 baht. It also found that in terms of innovation knowledge and capability, there were two variables had an influence on the amount of innovation knowledge and capability, innovation evaluation which were physical characteristic and innovation process.

Wavelet Based Identification of Second Order Linear System

In this paper, a wavelet based method is proposed to identify the constant coefficients of a second order linear system and is compared with the least squares method. The proposed method shows improved accuracy of parameter estimation as compared to the least squares method. Additionally, it has the advantage of smaller data requirement and storage requirement as compared to the least squares method.

Scheduling a Project to Minimize Costs of Material Requirements

Traditionally, project scheduling and material planning have been treated independently. In this research, a mixed integer programming model is presented to integrate project scheduling and materials ordering problems. The goal is to minimize the total material holding and ordering costs. In addition, an efficient metaheuristic algorithm is proposed to solve the model. The proposed algorithm is computationally tested, the results are analyzed, and conclusions are given.

Mathematical modeling of Bi-Substrate Enzymatic Reactions with Ping-Pong Mechanism in the Presence of Competitive Inhibitors

The mathematical modeling of different biological processes is usually used to predict or assess behavior of systems in which these processes take place. This study deals with mathematical and computer modeling of bi-substrate enzymatic reactions with ping-pong mechanism, which play an important role in different biochemical pathways. Besides that, three models of competitive inhibition were designed using different software packages. The main objective of this study is to represent the results from in silico investigation of bi-substrate enzymatic reactions with ordered pingpong mechanism in the presence of competitive inhibitors, as well as to describe in details the inhibition effects. The simulation of the models with certain kinetic parameters allowed investigating the behavior of reactions as well as determined some interesting aspects concerning influence of different cases of competitive inhibition. Simultaneous presence of two inhibitors, competitive to the S1 and S2 substrates have been studied. Moreover, we have found the pattern of simultaneous influence of both inhibitors.

A Nano-Scaled SRAM Guard Band Design with Gaussian Mixtures Model of Complex Long Tail RTN Distributions

This paper proposes, for the first time, how the challenges facing the guard-band designs including the margin assist-circuits scheme for the screening-test in the coming process generations should be addressed. The increased screening error impacts are discussed based on the proposed statistical analysis models. It has been shown that the yield-loss caused by the misjudgment on the screening test would become 5-orders of magnitude larger than that for the conventional one when the amplitude of random telegraph noise (RTN) caused variations approaches to that of random dopant fluctuation. Three fitting methods to approximate the RTN caused complex Gamma mixtures distributions by the simple Gaussian mixtures model (GMM) are proposed and compared. It has been verified that the proposed methods can reduce the error of the fail-bit predictions by 4-orders of magnitude.

Performance Boundaries for Interactive Finite Element Applications

This paper presents work characterizing finite element performance boundaries within which live, interactive finite element modeling is feasible on current and emerging systems. These results are based on wide-ranging tests performed using a prototype finite element program implemented specifically for this study, thereby enabling the unified investigation of numerous direct and iterative solver strategies and implementations in a variety of modeling contexts. The results are intended to be useful for researchers interested in interactive analysis by providing baseline performance estimates, to give guidance in matching solution strategies to problem domains, and to spur further work addressing the challenge of extending the present boundaries.

The Impact of Parent Involvement in Preschool Disabled Children

The purpose of this study was to investigate the relationship between parent involvement and preschool disabled children’s development. Parents of 3 year old disabled children (N=440) and 5 year old disabled children (N=937) participating in the Special Needs Education Longitudinal Study were interviewed or answered the web design questionnaire about their actions in parenting their disabled children. These children’s developments were also evaluated by their teachers. Data were analyzed using Structural Equation Modeling. Results were showed by tables and figures. Based on the results, the researcher made some suggestions for future studies.