EGCL: An Extended G-Code Language with Flow Control, Functions and Mnemonic Variables

In the context of computer numerical control (CNC) and computer aided manufacturing (CAM), the capabilities of programming languages such as symbolic and intuitive programming, program portability and geometrical portfolio have special importance. They allow to save time and to avoid errors during part programming and permit code re-usage. Our updated literature review indicates that the current state of art presents voids in parametric programming, program portability and programming flexibility. In response to this situation, this article presents a compiler implementation for EGCL (Extended G-code Language), a new, enriched CNC programming language which allows the use of descriptive variable names, geometrical functions and flow-control statements (if-then-else, while). Our compiler produces low-level generic, elementary ISO-compliant Gcode, thus allowing for flexibility in the choice of the executing CNC machine and in portability. Our results show that readable variable names and flow control statements allow a simplified and intuitive part programming and permit re-usage of the programs. Future work includes allowing the programmer to define own functions in terms of EGCL, in contrast to the current status of having them as library built-in functions.

Analysis of Textual Data Based On Multiple 2-Class Classification Models

This paper proposes a new method for analyzing textual data. The method deals with items of textual data, where each item is described based on various viewpoints. The method acquires 2- class classification models of the viewpoints by applying an inductive learning method to items with multiple viewpoints. The method infers whether the viewpoints are assigned to the new items or not by using the models. The method extracts expressions from the new items classified into the viewpoints and extracts characteristic expressions corresponding to the viewpoints by comparing the frequency of expressions among the viewpoints. This paper also applies the method to questionnaire data given by guests at a hotel and verifies its effect through numerical experiments.

An Examination of the Factors Influencing Software Development Effort

Effective evaluation of software development effort is an important aspect of successful project management. Based on a large database with 4106 projects ever developed, this study statistically examines the factors that influence development effort. The factors found to be significant for effort are project size, average number of developers that worked on the project, type of development, development language, development platform, and the use of rapid application development. Among these factors, project size is the most critical cost driver. Unsurprisingly, this study found that the use of CASE tools does not necessarily reduce development effort, which adds support to the claim that the use of tools is subtle. As many of the current estimation models are rarely or unsuccessfully used, this study proposes a parsimonious parametric model for the prediction of effort which is both simple and more accurate than previous models.

An Agent Oriented Approach to Operational Profile Management

Software reliability, defined as the probability of a software system or application functioning without failure or errors over a defined period of time, has been an important area of research for over three decades. Several research efforts aimed at developing models to improve reliability are currently underway. One of the most popular approaches to software reliability adopted by some of these research efforts involves the use of operational profiles to predict how software applications will be used. Operational profiles are a quantification of usage patterns for a software application. The research presented in this paper investigates an innovative multiagent framework for automatic creation and management of operational profiles for generic distributed systems after their release into the market. The architecture of the proposed Operational Profile MAS (Multi-Agent System) is presented along with detailed descriptions of the various models arrived at following the analysis and design phases of the proposed system. The operational profile in this paper is extended to comprise seven different profiles. Further, the criticality of operations is defined using a new composed metrics in order to organize the testing process as well as to decrease the time and cost involved in this process. A prototype implementation of the proposed MAS is included as proof-of-concept and the framework is considered as a step towards making distributed systems intelligent and self-managing.

Surrogate based Evolutionary Algorithm for Design Optimization

Optimization is often a critical issue for most system design problems. Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, finding optimal solution to complex high dimensional, multimodal problems often require highly computationally expensive function evaluations and hence are practically prohibitive. The Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model presented in our earlier work [14] reduced computation time by controlled use of meta-models to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the meta-model are generated from a single uniform model. Situations like model formation involving variable input dimensions and noisy data certainly can not be covered by this assumption. In this paper we present an enhanced version of DAFHEA that incorporates a multiple-model based learning approach for the SVM approximator. DAFHEA-II (the enhanced version of the DAFHEA framework) also overcomes the high computational expense involved with additional clustering requirements of the original DAFHEA framework. The proposed framework has been tested on several benchmark functions and the empirical results illustrate the advantages of the proposed technique.

Parametric Analysis in the Electronic Sensor Frequency Adjustment Process

The use of electronic sensors in the electronics industry has become increasingly popular over the past few years, and it has become a high competition product. The frequency adjustment process is regarded as one of the most important process in the electronic sensor manufacturing process. Due to inaccuracies in the frequency adjustment process, up to 80% waste can be caused due to rework processes; therefore, this study aims to provide a preliminary understanding of the role of parameters used in the frequency adjustment process, and also make suggestions in order to further improve performance. Four parameters are considered in this study: air pressure, dispensing time, vacuum force, and the distance between the needle tip and the product. A full factorial design for experiment 2k was considered to determine those parameters that significantly affect the accuracy of the frequency adjustment process, where a deviation in the frequency after adjustment and the target frequency is expected to be 0 kHz. The experiment was conducted on two levels, using two replications and with five center-points added. In total, 37 experiments were carried out. The results reveal that air pressure and dispensing time significantly affect the frequency adjustment process. The mathematical relationship between these two parameters was formulated, and the optimal parameters for air pressure and dispensing time were found to be 0.45 MPa and 458 ms, respectively. The optimal parameters were examined by carrying out a confirmation experiment in which an average deviation of 0.082 kHz was achieved.

Ruin Probabilities with Dependent Rates of Interest and Autoregressive Moving Average Structures

This paper studies ruin probabilities in two discrete-time risk models with premiums, claims and rates of interest modelled by three autoregressive moving average processes. Generalized Lundberg inequalities for ruin probabilities are derived by using recursive technique. A numerical example is given to illustrate the applications of these probability inequalities.

An Evaluation Framework for Participation: The VAAs Case Study

The number of electronic participation (eParticipation) projects introduced by different governments and international organisations is considerably high and increasing. In order to have an overview of the development of these projects, various evaluation frameworks have been proposed. In this paper, a five-level participation model, which takes into account the advantages of the Social Web or Web 2.0, together with a quantitative approach for the evaluation of eParticipation projects is presented. Each participation level is evaluated independently, taking into account three main components: Web evolution, media richness, and communication channels. This paper presents the evaluation of a number of existing Voting Advice Applications (VAAs). The results provide an overview of the main features implemented by each project, their strengths and weaknesses, and the participation levels reached.

Dosimetric Comparison of aSi1000 EPID and ImatriXX 2-D Array System for Volumetric Modulated Arc and Intensity Modulated Radiotherapy Patient Specific Quality Assurance

Prior to the use of detectors, characteristics comparison study was performed and baseline established. In patient specific QA, the portal dosimetry mean values of area gamma, average gamma and maximum gamma were 1.02, 0.31 and 1.31 with standard deviation of 0.33, 0.03 and 0.14 for IMRT and the corresponding values were 1.58, 0.48 and 1.73 with standard deviation of 0.31, 0.06 and 0.66 for VMAT. With ImatriXX 2-D array system, on an average 99.35% of the pixels passed the criteria of 3%-3 mm gamma with standard deviation of 0.24 for dynamic IMRT. For VMAT, the average value was 98.16% with a standard deviation of 0.86. The results showed that both the systems can be used in patient specific QA measurements for IMRT and VMAT. The values obtained with the portal dosimetry system were found to be relatively more consistent compared to those obtained with ImatriXX 2-D array system.

A New Design of Temperature-Controlled Chamber for OLED Panels

This paper presents an inexpensive and effective temperature-controlled chamber for temperature environment tests of Organic Light Emitting Diode (OLED) panels. The proposed chamber is a compact warmer and cooler with an exact temperature control system. In the temperature-controlled space of the chamber, thermoelectric modules (TEMs) are utilized to cool or to heat OLED panels, novel fixtures are designed to flexibly clamp the OLED panels of different size, and special connectors for wiring between the OLED panels and the test instrument are supplied. The proposed chamber has the following features. (1) The TEMs are solid semi-conductive devices, so they operate without noise and without pollution. (2) The volume of the temperature-controlled space of the chamber about 160mm*160mm*120mm, so the chamber are compact and easy to move. (3) The range of the controlled temperatures is from -10 oC to +80 oC, and the precision is ?0.5 oC. (4) The test instrument can conveniently and easily measure the OLED panels via the novel fixtures and special connectors. In addition to a constant temperature being maintained in the chamber, a temperature shock experiments can run for a long time. Therefore, the chamber will be convenient and useful for temperature environment tests of OLED panels.

Experimental and Numerical Investigation of the Dispersion of Microparticles Emitted by Machining Operation

As a part of the development of a numerical method of close capture exhausts systems for machining devices, a test rig recreating a situation similar to a grinding operation, but in a perfectly controlled environment, is used. The properties of the obtained spray of solid particles are initially characterized using particle tracking velocimetry (PTV), in order to obtain input and validation parameters for numerical simulations. The dispersion of a tracer gas (SF6) emitted simultaneously with the particle jet is then studied experimentally, as the dispersion of such a gas is representative of that of finer particles, whose aerodynamic response time is negligible. Finally, complete modeling of the test rig is achieved to allow comparison with experimental results and thus to progress towards validation of the models used to describe a twophase flow generated by machining operation.

Coreless Printed Circuit Board (PCB) Stepdown Transformers for DC-DC Converter Applications

In this paper, multilayered coreless printed circuit board (PCB) step-down power transformers for DC-DC converter applications have been designed, manufactured and evaluated. A set of two different circular spiral step-down transformers were fabricated in the four layered PCB. These transformers have been modelled with the assistance of high frequency equivalent circuit and characterized with both sinusoidal and square wave excitation. This paper provides the comparative results of these two different transformers in terms of their resistances, self, leakage, mutual inductances, coupling coefficient and also their energy efficiencies. The operating regions for optimal performance of these transformers for power transfer applications are determined. These transformers were tested for the output power levels of about 30 Watts within the input voltage range of 12-50 Vrms. The energy efficiency for these step down transformers is observed to be in the range of 90%-97% in MHz frequency region.

Evaluation of Groundwater Trend of Arsanjan Plain

Groundwater resources in Arsanjan plain provide water for agriculture, industry, and human consumption. Continued agricultural development in this area needs to additional groundwater resources for, particularly during of drought periods, and effects on the quantity and quality of ground water available. The purpose of this study is to evaluate water level changes in the aquifer of Arsanjan plain in the Fars province in order to determine the areas of greatest depletion and the causes of depletion. In this plain, farmers and other users are pumping groundwater faster than its natural replenishment rate, causing a continuous drop in groundwater tables and depletion of this resource. In this research variation of groundwater level, their effects and ways to help control groundwater levels in aquifer of the Arsanjan plains were evaluated .Excessive exploitation of groundwater in this aquifer caused the groundwater levels fall too fast or to unacceptable levels. The average drawdown of the groundwater level in this plain were 19.66 meters during 1996 to 2003.

An Improved Preprocessing for Biosonar Target Classification

An improved processing description to be employed in biosonar signal processing in a cochlea model is proposed and examined. It is compared to conventional models using a modified discrimination analysis and both are tested. Their performances are evaluated with echo data captured from natural targets (trees).Results indicate that the phase characteristics of low-pass filters employed in the echo processing have a significant effect on class separability for this data.

A Purpose Based Usage Access Control Model

As privacy becomes a major concern for consumers and enterprises, many research have been focused on the privacy protecting technology in recent years. In this paper, we present a comprehensive approach for usage access control based on the notion purpose. In our model, purpose information associated with a given data element specifies the intended use of the subjects and objects in the usage access control model. A key feature of our model is that it allows when an access is required, the access purpose is checked against the intended purposes for the data item. We propose an approach to represent purpose information to support access control based on purpose information. Our proposed solution relies on usage access control (UAC) models as well as the components which based on the notions of the purpose information used in subjects and objects. Finally, comparisons with related works are analyzed.

Optimization and GIS-Based Intelligent Decision Support System for Urban Transportation Systems Analysis

Optimization plays an important role in most real world applications that support decision makers to take the right decision regarding the strategic directions and operations of the system they manage. Solutions for traffic management and traffic congestion problems are considered major problems that most decision making authorities for cities around the world are looking for. This review paper gives a full description of the traffic problem as part of the transportation planning process and present a view as a framework of urban transportation system analysis where the core of the system is a transportation network equilibrium model that is based on optimization techniques and that can also be used for evaluating an alternative solution or a combination of alternative solutions for the traffic congestion. Different transportation network equilibrium models are reviewed from the sequential approach to the multiclass combining trip generation, trip distribution, modal split, trip assignment and departure time model. A GIS-Based intelligent decision support system framework for urban transportation system analysis is suggested for implementation where the selection of optimized alternative solutions, single or packages, will be based on an intelligent agent rather than human being which would lead to reduction in time, cost and the elimination of the difficulty, by human being, for finding the best solution to the traffic congestion problem.

3-D Reconstruction of Objects Using Digital Fringe Projection: Survey and Experimental Study

Three-dimensional reconstruction of small objects has been one of the most challenging problems over the last decade. Computer graphics researchers and photography professionals have been working on improving 3D reconstruction algorithms to fit the high demands of various real life applications. Medical sciences, animation industry, virtual reality, pattern recognition, tourism industry, and reverse engineering are common fields where 3D reconstruction of objects plays a vital role. Both lack of accuracy and high computational cost are the major challenges facing successful 3D reconstruction. Fringe projection has emerged as a promising 3D reconstruction direction that combines low computational cost to both high precision and high resolution. It employs digital projection, structured light systems and phase analysis on fringed pictures. Research studies have shown that the system has acceptable performance, and moreover it is insensitive to ambient light. This paper presents an overview of fringe projection approaches. It also presents an experimental study and implementation of a simple fringe projection system. We tested our system using two objects with different materials and levels of details. Experimental results have shown that, while our system is simple, it produces acceptable results.

Comparison of Alternative Models to Predict Lean Meat Percentage of Lamb Carcasses

The objective of this study was to develop and compare alternative prediction equations of lean meat proportion (LMP) of lamb carcasses. Forty (40) male lambs, 22 of Churra Galega Bragançana Portuguese local breed and 18 of Suffolk breed were used. Lambs were slaughtered, and carcasses weighed approximately 30 min later in order to obtain hot carcass weight (HCW). After cooling at 4º C for 24-h a set of seventeen carcass measurements was recorded. The left side of carcasses was dissected into muscle, subcutaneous fat, inter-muscular fat, bone, and remainder (major blood vessels, ligaments, tendons, and thick connective tissue sheets associated with muscles), and the LMP was evaluated as the dissected muscle percentage. Prediction equations of LMP were developed, and fitting quality was evaluated through the coefficient of determination of estimation (R2 e) and standard error of estimate (SEE). Models validation was performed by k-fold crossvalidation and the coefficient of determination of prediction (R2 p) and standard error of prediction (SEP) were computed. The BT2 measurement was the best single predictor and accounted for 37.8% of the LMP variation with a SEP of 2.30%. The prediction of LMP of lamb carcasses can be based simple models, using as predictors the HCW and one fat thickness measurement.

Multiresolution Approach to Subpixel Registration by Linear Approximation of PSF

Linear approximation of point spread function (PSF) is a new method for determining subpixel translations between images. The problem with the actual algorithm is the inability of determining translations larger than 1 pixel. In this paper a multiresolution technique is proposed to deal with the problem. Its performance is evaluated by comparison with two other well known registration method. In the proposed technique the images are downsampled in order to have a wider view. Progressively decreasing the downsampling rate up to the initial resolution and using linear approximation technique at each step, the algorithm is able to determine translations of several pixels in subpixel levels.

Mental Illness Stigma and Causal Beliefs: Among Potential Mental Health Professionals

Mental health professionals views about mental illness is an important issue which has not received enough attention. The negative stigma associated with mental illness can have many negative consequences. Unfortunately, health professionals working with the mentally ill can also exhibit stigma. It has been suggested that causal explanations or beliefs around the causes of mental illness may influence stigma. This study aims to gain a greater insight into stigma through examining stigma among potential mental health professionals. Firstly, results found that potential mental health professionals had relatively low social distance t(205) = -3.62, p