An Implementation of EURORADIO Protocol for ERTMS Systems

European Rail Traffic Management System (ERTMS) is the European reference for interoperable and safer signaling systems to efficiently manage trains running. If implemented, it allows trains cross seamlessly intra-European national borders. ERTMS has defined a secure communication protocol, EURORADIO, based on open communication networks. Its RadioInfill function can improve the reaction of the signaling system to changes in line conditions, avoiding unnecessary braking: its advantages in terms of power saving and travel time has been analyzed. In this paper a software implementation of the EURORADIO protocol with RadioInfill for ERTMS Level 1 using GSM-R is illustrated as part of the SR-Secure Italian project. In this building-blocks architecture the EURORADIO layers communicates together through modular Application Programm Interfaces. Security coding rules and railway industry requirements specified by EN 50128 standard have been respected. The proposed implementation has successfully passed conformity tests and has been tested on a computer-based simulator.

Enabling Remote Desktop in a Virtualized Environment for Cloud Services

Cloud computing is the innovative and leading information technology model for enabling convenient, on-demand network access to a shared pool of configurable computing resources that can be rapidly provisioned and released with minimal management effort. This paper presents our development on enabling an individual user's desktop in a virtualized environment, which is stored on a remote virtual machine rather than locally. We present the initial work on the integration of virtual desktop and application sharing with virtualization technology. Given the development of remote desktop virtualization, this proposed effort has the potential to positively provide an efficient, resilience and elastic environment for online cloud service. Users no longer need to burden the cost of software licenses and platform maintenances. Moreover, this development also helps boost user productivity by promoting a flexible model that lets users access their desktop environments from virtually anywhere.

Research of Dynamics Picking Mechanism of Sulzer Projectile Loom

One of the main and responsible units of Sulzer projectile loom is picking mechanism. It is specifically designed to accelerate projectile to speed of 25 m / s. Initial speed projectile of Sulzer projectile loom is independent of speed loom and determined the potential energy torsion rod. This paper investigates the dynamics picking mechanism of Sulzer projectile loom during its discharge. A result of calculation model, we obtain the law of motion lever of picking mechanism during its discharge. Construction of dynamic model the picking mechanism of Sulzer projectile loom on software complex SimulationX can make calculations for different thickness of torsion rods taking into account the backlashes in the connections, the dissipative forces and resistance forces

Local Linear Model Tree (LOLIMOT) Reconfigurable Parallel Hardware

Local Linear Neuro-Fuzzy Models (LLNFM) like other neuro- fuzzy systems are adaptive networks and provide robust learning capabilities and are widely utilized in various applications such as pattern recognition, system identification, image processing and prediction. Local linear model tree (LOLIMOT) is a type of Takagi-Sugeno-Kang neuro fuzzy algorithm which has proven its efficiency compared with other neuro fuzzy networks in learning the nonlinear systems and pattern recognition. In this paper, a dedicated reconfigurable and parallel processing hardware for LOLIMOT algorithm and its applications are presented. This hardware realizes on-chip learning which gives it the capability to work as a standalone device in a system. The synthesis results on FPGA platforms show its potential to improve the speed at least 250 of times faster than software implemented algorithms.

The Calculation of Electromagnetic Fields (EMF) in Substations of Shopping Centers

In nature, electromagnetic fields always appear like atmosphere static electric field, the earth's static magnetic field and the wide-rang frequency electromagnetic field caused by lightening. However, besides natural electromagnetic fields (EMF), today human beings are mostly exposed to artificial electromagnetic fields due to technology progress and outspread use of electrical devices. To evaluate nuisance of EMF, it is necessary to know field intensity for every frequency which appears and compare it with allowed values. Low frequency EMF-s around transmission and distribution lines are time-varying quasi-static electromagnetic fields which have conservative component of low frequency electrical field caused by charges and eddy component of low frequency magnetic field caused by currents. Displacement current or field delay are negligible, so energy flow in quasi-static EMF involves diffusion, analog like heat transfer. Electrical and magnetic field can be analyzed separately. This paper analysis the numerical calculations in ELF-400 software of EMF in distribution substation in shopping center. Analyzing the results it is possible to specify locations exposed to the fields and give useful suggestion to eliminate electromagnetic effect or reduce it on acceptable level within the non-ionizing radiation norms and norms of protection from EMF.

Benchmarking: Performance on ALPS and Formosa Clusters

This paper presents the benchmarking results and performance evaluation of differentclustersbuilt atthe National Center for High-Performance Computingin Taiwan. Performance of processor, memory subsystem andinterconnect is a critical factor in the overall performance of high performance computing platforms. The evaluation compares different system architecture and software platforms. Most supercomputer used HPL to benchmark their system performance, in accordance with the requirement of the TOP500 List. In this paper we consider system memory access factors that affect benchmark performance, such as processor and memory performance.We hope these works will provide useful information for future development and construct cluster system.

A Software-Supported Methodology for Designing General-Purpose Interconnection Networks for Reconfigurable Architectures

Modern applications realized onto FPGAs exhibit high connectivity demands. Throughout this paper we study the routing constraints of Virtex devices and we propose a systematic methodology for designing a novel general-purpose interconnection network targeting to reconfigurable architectures. This network consists of multiple segment wires and SB patterns, appropriately selected and assigned across the device. The goal of our proposed methodology is to maximize the hardware utilization of fabricated routing resources. The derived interconnection scheme is integrated on a Virtex style FPGA. This device is characterized both for its high-performance, as well as for its low-energy requirements. Due to this, the design criterion that guides our architecture selections was the minimal Energy×Delay Product (EDP). The methodology is fully-supported by three new software tools, which belong to MEANDER Design Framework. Using a typical set of MCNC benchmarks, extensive comparison study in terms of several critical parameters proves the effectiveness of the derived interconnection network. More specifically, we achieve average Energy×Delay Product reduction by 63%, performance increase by 26%, reduction in leakage power by 21%, reduction in total energy consumption by 11%, at the expense of increase of channel width by 20%.

Moving From Problem Space to Solution Space

Extracting and elaborating software requirements and transforming them into viable software architecture are still an intricate task. This paper defines a solution architecture which is based on the blurred amalgamation of problem space and solution space. The dependencies between domain constraints, requirements and architecture and their importance are described that are to be considered collectively while evolving from problem space to solution space. This paper proposes a revised version of Twin Peaks Model named Win Peaks Model that reconciles software requirements and architecture in more consistent and adaptable manner. Further the conflict between stakeholders- win-requirements is resolved by proposed Voting methodology that is simple adaptation of win-win requirements negotiation model and QARCC.

Reduction of Power Losses in Distribution Systems

Losses reduction initiatives in distribution systems have been activated due to the increasing cost of supplying electricity, the shortage in fuel with ever-increasing cost to produce more power, and the global warming concerns. These initiatives have been introduced to the utilities in shape of incentives and penalties. Recently, the electricity distribution companies in Oman have been incentivized to reduce the distribution technical and non-technical losses with an equal annual reduction rate for 6 years. In this paper, different techniques for losses reduction in Mazoon Electricity Company (MZEC) are addressed. In this company, high numbers of substation and feeders were found to be non-compliant with the Distribution System Security Standard (DSSS). Therefore, 33 projects have been suggested to bring non-complying 29 substations and 28 feeders to meet the planed criteria and to comply with the DSSS. The largest part of MZEC-s network (South Batinah region) was modeled by ETAP software package. The model has been extended to implement the proposed projects and to examine their effects on losses reduction. Simulation results have shown that the implementation of these projects leads to a significant improvement in voltage profile, and reduction in the active and the reactive power losses. Finally, the economical analysis has revealed that the implementation of the proposed projects in MZEC leads to an annual saving of about US$ 5 million.

Cost Optimized CO2 Pipeline Transportation Grid: A Case Study from Italian Industries

This paper presents the feasibility study of CO2 sequestration from the sources to the sinks in the prospective of Italian Industries. CO2 produced at these sources captured, compressed to supercritical pressures, transported via pipelines and stored in underground geologic formations such as depleted oil and natural gas reservoirs, un-minable coal seams and deep saline aquifers. In this work, we present the optimized pipeline infrastructure for the CO2 with appropriate constraints to find lower cost system by the use of nonlinear optimization software LINGO 11.0. This study was conducted on CO2 transportation complex network of Italian Industries, to find minimum cost network for transporting the CO2 from sources to the sinks.

Controlling of Load Elevators by the Fuzzy Logic Method

In this study, a fuzzy-logic based control system was designed to ensure that time and energy is saved during the operation of load elevators which are used during the construction of tall buildings. In the control system that was devised, for the load elevators to work more efficiently, the energy interval where the motor worked was taken as the output variable whereas the amount of load and the building height were taken as input variables. The most appropriate working intervals depending on the characteristics of these variables were defined by the help of an expert. Fuzzy expert system software was formed using Delphi programming language. In this design, mamdani max-min inference mechanism was used and the centroid method was employed in the clarification procedure. In conclusion, it is observed that the system that was designed is feasible and this is supported by statistical analyses..

A New Scheme for Improving the Quality of Service in Heterogeneous Wireless Network for Data Stream Sending

In this paper, we first consider the quality of service problems in heterogeneous wireless networks for sending the video data, which their problem of being real-time is pronounced. At last, we present a method for ensuring the end-to-end quality of service at application layer level for adaptable sending of the video data at heterogeneous wireless networks. To do this, mechanism in different layers has been used. We have used the stop mechanism, the adaptation mechanism and the graceful degrade at the application layer, the multi-level congestion feedback mechanism in the network layer and connection cutting off decision mechanism in the link layer. At the end, the presented method and the achieved improvement is simulated and presented in the NS-2 software.

The Functionality and Usage of CRM Systems

Modern information and communication technologies offer a variety of support options for the efficient handling of customer relationships. CRM systems have been developed, which are designed to support the processes in the areas of marketing, sales and service. Along with technological progress, CRM systems are constantly changing, i.e. the systems are continually enhanced by new functions. However, not all functions are suitable for every company because of different frameworks and business processes. In this context the question arises whether or not CRM systems are widely used in Austrian companies and which business processes are most frequently supported by CRM systems. This paper aims to shed light on the popularity of CRM systems in Austrian companies in general and the use of different functions to support their daily business. First of all, the paper provides a theoretical overview of the structure of modern CRM systems and proposes a categorization of currently available software functionality for collaborative, operational and analytical CRM processes, which provides the theoretical background for the empirical study. Apart from these theoretical considerations, the paper presents the empirical results of a field survey on the use of CRM systems in Austrian companies and analyzes its findings.

The Advent of Electronic Logbook Technology - Reducing Cost and Risk to Both Marine Resources and the Fishing Industry

Fisheries management all around the world is hampered by the lack, or poor quality, of critical data on fish resources and fishing operations. The main reasons for the chronic inability to collect good quality data during fishing operations is the culture of secrecy common among fishers and the lack of modern data gathering technology onboard most fishing vessels. In response, OLRAC-SPS, a South African company, developed fisheries datalogging software (eLog in short) and named it Olrac. The Olrac eLog solution is capable of collecting, analysing, plotting, mapping, reporting, tracing and transmitting all data related to fishing operations. Olrac can be used by skippers, fleet/company managers, offshore mariculture farmers, scientists, observers, compliance inspectors and fisheries management authorities. The authors believe that using eLog onboard fishing vessels has the potential to revolutionise the entire process of data collection and reporting during fishing operations and, if properly deployed and utilised, could transform the entire commercial fleet to a provider of good quality data and forever change the way fish resources are managed. In addition it will make it possible to trace catches back to the actual individual fishing operation, to improve fishing efficiency and to dramatically improve control of fishing operations and enforcement of fishing regulations.

Investigation of Time Delay Factors in Global Software Development

Global Software Development (GSD) projects are passing through different boundaries of a company, country and even in other continents where time zone differs between both sites. Beside many benefits of such development, research declared plenty of negative impacts on these GSD projects. It is important to understand problems which may lie during the execution of GSD project with different time zones. This research project discussed and provided different issues related to time delays in GSD projects. In this paper, authors investigated some of the time delay factors which usually lie in GSD projects with different time zones. This investigation is done through systematic review of literature. Furthermore, the practices to overcome these delay factors which have already been reported in literature and GSD organizations are also explored through literature survey and case studies.

Theoretical Considerations for Software Component Metrics

We have defined two suites of metrics, which cover static and dynamic aspects of component assembly. The static metrics measure complexity and criticality of component assembly, wherein complexity is measured using Component Packing Density and Component Interaction Density metrics. Further, four criticality conditions namely, Link, Bridge, Inheritance and Size criticalities have been identified and quantified. The complexity and criticality metrics are combined to form a Triangular Metric, which can be used to classify the type and nature of applications. Dynamic metrics are collected during the runtime of a complete application. Dynamic metrics are useful to identify super-component and to evaluate the degree of utilisation of various components. In this paper both static and dynamic metrics are evaluated using Weyuker-s set of properties. The result shows that the metrics provide a valid means to measure issues in component assembly. We relate our metrics suite with McCall-s Quality Model and illustrate their impact on product quality and to the management of component-based product development.

Similarity Detection in Collaborative Development of Object-Oriented Formal Specifications

The complexity of today-s software systems makes collaborative development necessary to accomplish tasks. Frameworks are necessary to allow developers perform their tasks independently yet collaboratively. Similarity detection is one of the major issues to consider when developing such frameworks. It allows developers to mine existing repositories when developing their own views of a software artifact, and it is necessary for identifying the correspondences between the views to allow merging them and checking their consistency. Due to the importance of the requirements specification stage in software development, this paper proposes a framework for collaborative development of Object- Oriented formal specifications along with a similarity detection approach to support the creation, merging and consistency checking of specifications. The paper also explores the impact of using additional concepts on improving the matching results. Finally, the proposed approach is empirically evaluated.

A Two-Phase Mechanism for Agent's Action Selection in Soccer Simulation

Soccer simulation is an effort to motivate researchers and practitioners to do artificial and robotic intelligence research; and at the same time put into practice and test the results. Many researchers and practitioners throughout the world are continuously working to polish their ideas and improve their implemented systems. At the same time, new groups are forming and they bring bright new thoughts to the field. The research includes designing and executing robotic soccer simulation algorithms. In our research, a soccer simulation player is considered to be an intelligent agent that is capable of receiving information from the environment, analyze it and to choose the best action from a set of possible ones, for its next move. We concentrate on developing a two-phase method for the soccer player agent to choose its best next move. The method is then implemented into our software system called Nexus simulation team of Ferdowsi University. This system is based on TsinghuAeolus[1] team that was the champion of the world RoboCup soccer simulation contest in 2001 and 2002.

The Wijma Delivery Expectancy/Experience Questionnaire (W-DEQ) with Turkish Sample: Confirmatory and Exploratory Factor Analysis

The propose of this study is to investigate the factor structures of the W-DEQ, originally developed on UK and Swedish women, were confirmed in Turkish samples, and to obtain a new modified factor structure appropriate to Turkish culture. Statistical analyses of the data obtained were performed using SPSS© for Windows version 13.0 and the SAS statistical software Version 9.1. Both confirmatory and exploratory factor analysis of W-DEQ were performed in the study. Factor analysis yielded four factors related to hope, fear, lack of positive anticipation and riskiness. The alpha estimates of the total W-DEQ score were somewhat higher, being 0.92 for the parous and 0.90 for the nulliparous sample. These are well above the accepted limit of 0.70 and indicate excellent levels of internal reliability, thus showing that the questions were appropriate to the Turkish culture and useful scale for the evaluation of fear of childbirth in Turkish pregnants.

Numerical Study of Oxygen Enrichment on NO Pollution Spread in a Combustion Chamber

In this study, a 3D combustion chamber was simulated using FLUENT 6.32. Aim to obtain detailed information on combustion characteristics and _ nitrogen oxides in the furnace and the effect of oxygen enrichment in a combustion process. Oxygenenriched combustion is an effective way to reduce emissions. This paper analyzes NO emission, including thermal NO and prompt NO. Flow rate ratio of air to fuel is varied as 1.3, 3.2 and 5.1 and the oxygen enriched flow rates are 28, 54 and 68 lit/min. The 3D Reynolds Averaged Navier Stokes (RANS) equations with standard k-ε turbulence model are solved together by Fluent 6.32 software. First order upwind scheme is used to model governing equations and the SIMPLE algorithm is used as pressure velocity coupling. Results show that for AF=1.3, increase the oxygen flow rate of oxygen reduction in NO emissions is Lance. Moreover, in a fixed oxygen enrichment condition, increasing the air to fuel ratio will increase the temperature peak, but not the NO emission rate. As a result, oxygen enrichment can reduce the NO emission at this kind of furnace in low air to fuel rates.