Virtual Reality for Mutual Understanding in Landscape Planning

This paper argues that fostering mutual understanding in landscape planning is as much about the planners educating stakeholder groups as the stakeholders educating the planners. In other words it is an epistemological agreement as to the meaning and nature of place, especially where an effort is made to go beyond the quantitative aspects, which can be achieved by the phenomenological experience of the Virtual Reality (VR) environment. This education needs to be a bi-directional process in which distance can be both temporal as well as spatial separation of participants, that there needs to be a common framework of understanding in which neither 'side' is disadvantaged during the process of information exchange and it follows that a medium such as VR offers an effective way of overcoming some of the shortcomings of traditional media by taking advantage of continuing technological advances in Information, Technology and Communications (ITC). In this paper we make particular reference to this as an extension to Geographical Information Systems (GIS). VR as a two-way communication tool offers considerable potential particularly in the area of Public Participation GIS (PPGIS). Information rich virtual environments that can operate over broadband networks are now possible and thus allow for the representation of large amounts of qualitative and quantitative information 'side-by-side'. Therefore, with broadband access becoming standard for households and enterprises alike, distributed virtual reality environments have great potential to contribute to enabling stakeholder participation and mutual learning within the planning context.

Optimal All-to-All Personalized Communication in All-Port Tori

All-to-all personalized communication, also known as complete exchange, is one of the most dense communication patterns in parallel computing. In this paper, we propose new indirect algorithms for complete exchange on all-port ring and torus. The new algorithms fully utilize all communication links and transmit messages along shortest paths to completely achieve the theoretical lower bounds on message transmission, which have not be achieved among other existing indirect algorithms. For 2D r × c ( r % c ) all-port torus, the algorithm has time complexities of optimal transmission cost and O(c) message startup cost. In addition, the proposed algorithms accommodate non-power-of-two tori where the number of nodes in each dimension needs not be power-of-two or square. Finally, the algorithms are conceptually simple and symmetrical for every message and every node so that they can be easily implemented and achieve the optimum in practice.

An Evaluation of Requirements Management and Traceability Tools

Requirements management is critical to software delivery success and project lifecycle. Requirements management and their traceability provide assistance for many software engineering activities like impact analysis, coverage analysis, requirements validation and regression testing. In addition requirements traceability is the recognized component of many software process improvement initiatives. Requirements traceability also helps to control and manage evolution of a software system. This paper aims to provide an evaluation of current requirements management and traceability tools. Management and test managers require an appropriate tool for the software under test. We hope, evaluation identified here will help to select the efficient and effective tool.

Mimicking Morphogenesis for Robust Behaviour of Cellular Architectures

Morphogenesis is the process that underpins the selforganised development and regeneration of biological systems. The ability to mimick morphogenesis in artificial systems has great potential for many engineering applications, including production of biological tissue, design of robust electronic systems and the co-ordination of parallel computing. Previous attempts to mimick these complex dynamics within artificial systems have relied upon the use of evolutionary algorithms that have limited their size and complexity. This paper will present some insight into the underlying dynamics of morphogenesis, then show how to, without the assistance of evolutionary algorithms, design cellular architectures that converge to complex patterns.

A New Approach of Wireless Network Traffic on VPN

This work presents a new approach of securing a wireless network. The configuration is focused on securing & Protecting wireless network traffic for a small network such as a home or dorm room. The security Mechanism provided both authentication, allowing only known authorized users access to the wireless network, and encryption, preventing anyone from reading the wireless traffic. The mentioned solution utilizes the open source free S/WAN software which implements the Internet Protocol Security –IPSEC. In addition to wireless components, wireless NIC in PC and wireless access point needs a machine running Linux to act as security gateway. While the current configuration assumes that the wireless PC clients are running Linux, Windows XP/VISTA/7 based machines equipped with VPN software which will allow to interface with this configuration.

A Strategy to Optimize the SPC Scheme for Mass Production of HDD Arm with ClusteringTechnique and Three-Way Control Chart

Consider a mass production of HDD arms where hundreds of CNC machines are used to manufacturer the HDD arms. According to an overwhelming number of machines and models of arm, construction of separate control chart for monitoring each HDD arm model by each machine is not feasible. This research proposed a strategy to optimize the SPC management on shop floor. The procedure started from identifying the clusters of the machine with similar manufacturing performance using clustering technique. The three way control chart ( I - MR - R ) is then applied to each clustered group of machine. This proposed research has advantageous to the manufacturer in terms of not only better performance of the SPC but also the quality management paradigm.

Development of a Catchment Water Quality Model for Continuous Simulations of Pollutants Build-up and Wash-off

Estimation of runoff water quality parameters is required to determine appropriate water quality management options. Various models are used to estimate runoff water quality parameters. However, most models provide event-based estimates of water quality parameters for specific sites. The work presented in this paper describes the development of a model that continuously simulates the accumulation and wash-off of water quality pollutants in a catchment. The model allows estimation of pollutants build-up during dry periods and pollutants wash-off during storm events. The model was developed by integrating two individual models; rainfall-runoff model, and catchment water quality model. The rainfall-runoff model is based on the time-area runoff estimation method. The model allows users to estimate the time of concentration using a range of established methods. The model also allows estimation of the continuing runoff losses using any of the available estimation methods (i.e., constant, linearly varying or exponentially varying). Pollutants build-up in a catchment was represented by one of three pre-defined functions; power, exponential, or saturation. Similarly, pollutants wash-off was represented by one of three different functions; power, rating-curve, or exponential. The developed runoff water quality model was set-up to simulate the build-up and wash-off of total suspended solids (TSS), total phosphorus (TP) and total nitrogen (TN). The application of the model was demonstrated using available runoff and TSS field data from road and roof surfaces in the Gold Coast, Australia. The model provided excellent representation of the field data demonstrating the simplicity yet effectiveness of the proposed model.

A Retrospective of High-Lift Device Technology

The present paper deals with the most adopted technical solutions for the enhancement of the lift force of a wing. In fact, during several flight conditions (such as take off and landing), the lift force needs to be dramatically enhanced. Both trailing edge devices (such as flaps) and leading edge ones (such as slats) are described. Finally, the most advanced aerodynamic solutions to avoid the separation of the boundary layer from aircraft wings at high angles of attack are reviewed.

A Fuzzy Time Series Forecasting Model for Multi-Variate Forecasting Analysis with Fuzzy C-Means Clustering

In this study, a fuzzy integrated logical forecasting method (FILF) is extended for multi-variate systems by using a vector autoregressive model. Fuzzy time series forecasting (FTSF) method was recently introduced by Song and Chissom [1]-[2] after that Chen improved the FTSF method. Rather than the existing literature, the proposed model is not only compared with the previous FTS models, but also with the conventional time series methods such as the classical vector autoregressive model. The cluster optimization is based on the C-means clustering method. An empirical study is performed for the prediction of the chartering rates of a group of dry bulk cargo ships. The root mean squared error (RMSE) metric is used for the comparing of results of methods and the proposed method has superiority than both traditional FTS methods and also the classical time series methods.

Effect of Heat Treatment on the Portevin-Le Chatelier Effect of Al-2.5%Mg Alloy

An experimental study is presented on the effect of microstructural change on the Portevin-Le Chatelier effect behaviour of Al-2.5%Mg alloy. Tensile tests are performed on the as received and heat treated (at 400 ºC for 16 hours) samples for a wide range of strain rates. The serrations observed in the stress-time curve are investigated from statistical analysis point of view. Microstructures of the samples are characterized by optical metallography and X-ray diffraction. It is found that the excess vacancy generated due to heat treatment leads to decrease in the strain rate sensitivity and the increase in the number of stress drop occurrences per unit time during the PLC effect. The microstructural parameters like domain size, dislocation density have no appreciable effect on the PLC effect as far as the statistical behavior of the serrations is considered.

Block Activity in Metric Neural Networks

The model of neural networks on the small-world topology, with metric (local and random connectivity) is investigated. The synaptic weights are random, driving the network towards a chaotic state for the neural activity. An ordered macroscopic neuron state is induced by a bias in the network connections. When the connections are mainly local, the network emulates a block-like structure. It is found that the topology and the bias compete to influence the network to evolve into a global or a block activity ordering, according to the initial conditions.

Development of Monitoring and Simulation System of Human Tracking System Based On Mobile Agent Technologies

In recent years, the number of the cases of information leaks is increasing. Companies and Research Institutions make various actions against information thefts and security accidents. One of the actions is adoption of the crime prevention system, including the monitoring system by surveillance cameras. In order to solve difficulties of multiple cameras monitoring, we develop the automatic human tracking system using mobile agents through multiple surveillance cameras to track target persons. In this paper, we develop the monitor which confirms mobile agents tracing target persons, and the simulator of video picture analysis to construct the tracking algorithm.

A New Technique for Solar Activity Forecasting Using Recurrent Elman Networks

In this paper we present an efficient approach for the prediction of two sunspot-related time series, namely the Yearly Sunspot Number and the IR5 Index, that are commonly used for monitoring solar activity. The method is based on exploiting partially recurrent Elman networks and it can be divided into three main steps: the first one consists in a “de-rectification" of the time series under study in order to obtain a new time series whose appearance, similar to a sum of sinusoids, can be modelled by our neural networks much better than the original dataset. After that, we normalize the derectified data so that they have zero mean and unity standard deviation and, finally, train an Elman network with only one input, a recurrent hidden layer and one output using a back-propagation algorithm with variable learning rate and momentum. The achieved results have shown the efficiency of this approach that, although very simple, can perform better than most of the existing solar activity forecasting methods.

Odor Discrimination Using Neural Decoding of Olfactory Bulbs in Rats

This paper presents a novel method for inferring the odor based on neural activities observed from rats- main olfactory bulbs. Multi-channel extra-cellular single unit recordings were done by micro-wire electrodes (tungsten, 50μm, 32 channels) implanted in the mitral/tufted cell layers of the main olfactory bulb of anesthetized rats to obtain neural responses to various odors. Neural response as a key feature was measured by substraction of neural firing rate before stimulus from after. For odor inference, we have developed a decoding method based on the maximum likelihood (ML) estimation. The results have shown that the average decoding accuracy is about 100.0%, 96.0%, 84.0%, and 100.0% with four rats, respectively. This work has profound implications for a novel brain-machine interface system for odor inference.

An Automatic Tool for Checking Consistency between Data Flow Diagrams (DFDs)

System development life cycle (SDLC) is a process uses during the development of any system. SDLC consists of four main phases: analysis, design, implement and testing. During analysis phase, context diagram and data flow diagrams are used to produce the process model of a system. A consistency of the context diagram to lower-level data flow diagrams is very important in smoothing up developing process of a system. However, manual consistency check from context diagram to lower-level data flow diagrams by using a checklist is time-consuming process. At the same time, the limitation of human ability to validate the errors is one of the factors that influence the correctness and balancing of the diagrams. This paper presents a tool that automates the consistency check between Data Flow Diagrams (DFDs) based on the rules of DFDs. The tool serves two purposes: as an editor to draw the diagrams and as a checker to check the correctness of the diagrams drawn. The consistency check from context diagram to lower-level data flow diagrams is embedded inside the tool to overcome the manual checking problem.

Conceptual Frameworks of Carbon Credit Registry System for Thailand

This research explores on the development of the structure of Carbon Credit Registry System those accords to the need of future events in Thailand. This research also explores the big picture of every connected system by referring to the design of each system, the Data Flow Diagram, and the design in term of the system-s data using DES standard. The purpose of this paper is to show how to design the model of each system. Furthermore, this paper can serve as guideline for designing an appropriate Carbon Credit Registry System.

Dynamic Time Warping in Gait Classificationof Motion Capture Data

The method of gait identification based on the nearest neighbor classification technique with motion similarity assessment by the dynamic time warping is proposed. The model based kinematic motion data, represented by the joints rotations coded by Euler angles and unit quaternions is used. The different pose distance functions in Euler angles and quaternion spaces are considered. To evaluate individual features of the subsequent joints movements during gait cycle, joint selection is carried out. To examine proposed approach database containing 353 gaits of 25 humans collected in motion capture laboratory is used. The obtained results are promising. The classifications, which takes into consideration all joints has accuracy over 91%. Only analysis of movements of hip joints allows to correctly identify gaits with almost 80% precision.

Wind Speed Data Analysis using Wavelet Transform

Renewable energy systems are becoming a topic of great interest and investment in the world. In recent years wind power generation has experienced a very fast development in the whole world. For planning and successful implementations of good wind power plant projects, wind potential measurements are required. In these projects, of great importance is the effective choice of the micro location for wind potential measurements, installation of the measurement station with the appropriate measuring equipment, its maintenance and analysis of the gained data on wind potential characteristics. In this paper, a wavelet transform has been applied to analyze the wind speed data in the context of insight in the characteristics of the wind and the selection of suitable locations that could be the subject of a wind farm construction. This approach shows that it can be a useful tool in investigation of wind potential.

Presenting a Combinatorial Feature to Estimate Depth of Anesthesia

Determining depth of anesthesia is a challenging problem in the context of biomedical signal processing. Various methods have been suggested to determine a quantitative index as depth of anesthesia, but most of these methods suffer from high sensitivity during the surgery. A novel method based on energy scattering of samples in the wavelet domain is suggested to represent the basic content of electroencephalogram (EEG) signal. In this method, first EEG signal is decomposed into different sub-bands, then samples are squared and energy of samples sequence is constructed through each scale and time, which is normalized and finally entropy of the resulted sequences is suggested as a reliable index. Empirical Results showed that applying the proposed method to the EEG signals can classify the awake, moderate and deep anesthesia states similar to BIS.

Adaptive Motion Estimator Based on Variable Block Size Scheme

This paper presents an adaptive motion estimator that can be dynamically reconfigured by the best algorithm depending on the variation of the video nature during the lifetime of an application under running. The 4 Step Search (4SS) and the Gradient Search (GS) algorithms are integrated in the estimator in order to be used in the case of rapid and slow video sequences respectively. The Full Search Block Matching (FSBM) algorithm has been also integrated in order to be used in the case of the video sequences which are not real time oriented. In order to efficiently reduce the computational cost while achieving better visual quality with low cost power, the proposed motion estimator is based on a Variable Block Size (VBS) scheme that uses only the 16x16, 16x8, 8x16 and 8x8 modes. Experimental results show that the adaptive motion estimator allows better results in term of Peak Signal to Noise Ratio (PSNR), computational cost, FPGA occupied area, and dissipated power relatively to the most popular variable block size schemes presented in the literature.