High-Speed Train Planning in France, Lessons from Mediterranean TGV-Line

To fight against the economic crisis, French Government, like many others in Europe, has decided to give a boost to high-speed line projects. This paper explores the implementation and decision-making process in TGV projects, their evolutions, especially since the Mediterranean TGV-line. This project was probably the most controversial, but paradoxically represents today a huge success for all the actors involved. What kind of lessons we can learn from this experience? How to evaluate the impact of this project on TGV-line planning? How can we characterize this implementation and decision-making process regards to the sustainability challenges? The construction of Mediterranean TGV-line was the occasion to make several innovations: to introduce more dialog into the decisionmaking process, to take into account the environment, to introduce a new project management and technological innovations. That-s why this project appears today as an example in terms of integration of sustainable development. In this paper we examine the different kinds of innovations developed in this project, by using concepts from sociology of innovation to understand how these solutions emerged in a controversial situation. Then we analyze the lessons which were drawn from this decision-making process (in the immediacy and a posteriori) and the way in which procedures evolved: creation of new tools and devices (public consultation, project management...). Finally we try to highlight the impact of this evolution on TGV projects governance. In particular, new methods of implementation and financing involve a reconfiguration of the system of actors. The aim of this paper is to define the impact of this reconfiguration on negotiations between stakeholders.

Comanche – A Compiler-Driven I/O Management System

Most scientific programs have large input and output data sets that require out-of-core programming or use virtual memory management (VMM). Out-of-core programming is very error-prone and tedious; as a result, it is generally avoided. However, in many instance, VMM is not an effective approach because it often results in substantial performance reduction. In contrast, compiler driven I/O management will allow a program-s data sets to be retrieved in parts, called blocks or tiles. Comanche (COmpiler MANaged caCHE) is a compiler combined with a user level runtime system that can be used to replace standard VMM for out-of-core programs. We describe Comanche and demonstrate on a number of representative problems that it substantially out-performs VMM. Significantly our system does not require any special services from the operating system and does not require modification of the operating system kernel.

Evaluation on Recent Committed Crypt Analysis Hash Function

This paper describes the study of cryptographic hash functions, one of the most important classes of primitives used in recent techniques in cryptography. The main aim is the development of recent crypt analysis hash function. We present different approaches to defining security properties more formally and present basic attack on hash function. We recall Merkle-Damgard security properties of iterated hash function. The Main aim of this paper is the development of recent techniques applicable to crypt Analysis hash function, mainly from SHA family. Recent proposed attacks an MD5 & SHA motivate a new hash function design. It is designed not only to have higher security but also to be faster than SHA-256. The performance of the new hash function is at least 30% better than that of SHA-256 in software. And it is secure against any known cryptographic attacks on hash functions.

Design and Construction of Microcontroller-Based Telephone Exchange System

This paper demonstrates design and construction of microcontroller-based telephone exchange system and the aims of this paper is to study telecommunication, connection with PIC16F877A and DTMF MT8870D. In microcontroller system, PIC 16F877 microcontroller is used to control the call processing. Dial tone, busy tone and ring tone are provided during call progress. Instead of using ready made tone generator IC, oscillator based tone generator is used. The results of this telephone exchange system are perfect for homes and small businesses needing the extensions. It requires the phone operation control system, the analog interface circuit and the switching circuit. This exchange design will contain eight channels. It is the best low cost, good quality telephone exchange for today-s telecommunication needs. It offers the features available in much more expensive PBX units without using high-priced phones. It is for long distance telephone services.

Energy Saving Suction Hood

Public awareness towards green energy are on the rise and this can be prove by many product being manufactured or prerequired to be made as energy saving devices mainly to save consumer from spending more on utility billing. These schemes are popular nowadays and many homemade appliances are turned into energy saving gadget which attracts the attention of consumers. Knowing the public demands and pattern towards purchasing home appliances thus the idea of “energy saving suction hood (ESSH)" is proposed. The ESSH can be used in many places that require smoke ventilation or even to reduce the room temperature as many conventional suction hoods (CSH) do, but this device works automatically by the usage of sensors that detects the smoke/temperature and automatically spins the exhaust fan. As it turns, the mechanical rotation rotates the AC generator which is coupled together with the fan and then charges the battery. The innovation of this product is, it does not rely on the utility supply as it is also hook up with a solar panel which also charges the battery, Secondly, it generates energy as the exhaust fan mechanically rotates. Thirdly, an energy loop back feature is introduced to this system which will supply for the ventilator fan. Another major innovation is towards interfacing this device with an in house production of generator. This generator is produced by proper design on stator as well as rotor to reduce the losses. A comparison is made between the ESSH and the CSH and result shows that the ESSH saves 172.8kWh/year of utility supply which is used by CSH. This amount of energy can save RM 3.14 from monthly utility bill and a total of RM 37.67 per year. In fact this product can generate 175 Watt of power from generator(75W) and solar panel(100W) that can be used either to supply other household appliances and/or to loop back to supply the fans motor. The innovation of this system is essential for future production of other equipment by using the loopback power method and turning most equipment into a standalone system.

Metaphor in Terminology: Visualization as a Way to Term Perception

Metaphor has recently gained extensive interest most probably due to developments in cognitive sciences and the study of language as the reflection of humans- world perception. Metaphor is no longer reckoned as solely literary expressive means. Nowadays it is studied in a whole number of discourses, such as politics, law, medicine, sports, etc. with the purpose of the analysis and determining its role. The scientific language is not an exception. It might seem that metaphor cannot suit it; we would dare to draw a hypothesis that metaphor has indeed found its stable place in terminology. In comprehension of metaphorically represented terms the stage of visualization plays a significant role. We proceeded on the assumption that this stage is the main in provision of better term comprehension and would try to exemplify it with metaphoricallyoriented terms.

Entrepreneurial Activity - Indicator of Regional Development in Croatia

Given that entrepreneurship is a very significant factor of regional development, it is necessary to approach systematically the development with measures of regional politics. According to international classification The Nomenclature of Territorial Units for Statistics (NUTS II), there are three regions in Croatia. The indicators of entrepreneurial activities on the national level of Croatia are analyzed in the paper, taking into consideration the results of referent research. The level of regional development is shown based on the analysis of entrepreneurs- operations. The results of the analysis show a very unfavorable situation in entrepreneurial activities on the national level of Croatia. The origin of this situation is to be found in the surroundings with an expressed inequality of regional development, which is caused by the non-existence of a strategically directed regional policy. In this paper recommendations which could contribute to the reduction of regional inequality in Croatia, have been made.

Belief Theory-Based Classifiers Comparison for Static Human Body Postures Recognition in Video

This paper presents various classifiers results from a system that can automatically recognize four different static human body postures in video sequences. The considered postures are standing, sitting, squatting, and lying. The three classifiers considered are a naïve one and two based on the belief theory. The belief theory-based classifiers use either a classic or restricted plausibility criterion to make a decision after data fusion. The data come from the people 2D segmentation and from their face localization. Measurements consist in distances relative to a reference posture. The efficiency and the limits of the different classifiers on the recognition system are highlighted thanks to the analysis of a great number of results. This system allows real-time processing.

Analysis of Gamma-Ray Spectra Using Levenberg-Marquardt Method

Levenberg-Marquardt method (LM) was proposed to be applied as a non-linear least-square fitting in the analysis of a natural gamma-ray spectrum that was taken by the Hp (Ge) detector. The Gaussian function that composed of three components, main Gaussian, a step background function and tailing function in the lowenergy side, has been suggested to describe each of the y-ray lines mathematically in the spectrum. The whole spectrum has been analyzed by determining the energy and relative intensity for the strong y-ray lines.

Optimization of Reaction Rate Parameters in Modeling of Heavy Paraffins Dehydrogenation

In the present study, a procedure was developed to determine the optimum reaction rate constants in generalized Arrhenius form and optimized through the Nelder-Mead method. For this purpose, a comprehensive mathematical model of a fixed bed reactor for dehydrogenation of heavy paraffins over Pt–Sn/Al2O3 catalyst was developed. Utilizing appropriate kinetic rate expressions for the main dehydrogenation reaction as well as side reactions and catalyst deactivation, a detailed model for the radial flow reactor was obtained. The reactor model composed of a set of partial differential equations (PDE), ordinary differential equations (ODE) as well as algebraic equations all of which were solved numerically to determine variations in components- concentrations in term of mole percents as a function of time and reactor radius. It was demonstrated that most significant variations observed at the entrance of the bed and the initial olefin production obtained was rather high. The aforementioned method utilized a direct-search optimization algorithm along with the numerical solution of the governing differential equations. The usefulness and validity of the method was demonstrated by comparing the predicted values of the kinetic constants using the proposed method with a series of experimental values reported in the literature for different systems.

Identification of Critical Success Factors in Non-Formal Service Sector Using Delphi Technique

The purpose of this study is to identify the critical success factors (CSFs) for the effective implementation of Six Sigma in non-formal service Sectors. Based on the survey of literature, the critical success factors (CSFs) for Six Sigma have been identified and are assessed for their importance in Non-formal service sector using Delphi Technique. These selected CSFs were put forth to the panel of expert to cluster them and prepare cognitive map to establish their relationship. All the critical success factors examined and obtained from the review of literature have been assessed for their importance with respect to their contribution to Six Sigma effectiveness in non formal service sector. The study is limited to the non-formal service sectors involved in the organization of religious festival only. However, the similar exercise can be conducted for broader sample of other non-formal service sectors like temple/ashram management, religious tours management etc. The research suggests an approach to identify CSFs of Six Sigma for Non-formal service sector. All the CSFs of the formal service sector will not be applicable to Non-formal services, hence opinion of experts was sought to add or delete the CSFs. In the first round of Delphi, the panel of experts has suggested, two new CSFs-“competitive benchmarking (F19) and resident’s involvement (F28)”, which were added for assessment in the next round of Delphi.  One of the CSFs-“fulltime six sigma personnel (F15)” has been omitted in proposed clusters of CSFs for non-formal organization, as it is practically impossible to deploy full time trained Six Sigma recruits.

GridNtru: High Performance PKCS

Cryptographic algorithms play a crucial role in the information society by providing protection from unauthorized access to sensitive data. It is clear that information technology will become increasingly pervasive, Hence we can expect the emergence of ubiquitous or pervasive computing, ambient intelligence. These new environments and applications will present new security challenges, and there is no doubt that cryptographic algorithms and protocols will form a part of the solution. The efficiency of a public key cryptosystem is mainly measured in computational overheads, key size and bandwidth. In particular the RSA algorithm is used in many applications for providing the security. Although the security of RSA is beyond doubt, the evolution in computing power has caused a growth in the necessary key length. The fact that most chips on smart cards can-t process key extending 1024 bit shows that there is need for alternative. NTRU is such an alternative and it is a collection of mathematical algorithm based on manipulating lists of very small integers and polynomials. This allows NTRU to high speeds with the use of minimal computing power. NTRU (Nth degree Truncated Polynomial Ring Unit) is the first secure public key cryptosystem not based on factorization or discrete logarithm problem. This means that given sufficient computational resources and time, an adversary, should not be able to break the key. The multi-party communication and requirement of optimal resource utilization necessitated the need for the present day demand of applications that need security enforcement technique .and can be enhanced with high-end computing. This has promoted us to develop high-performance NTRU schemes using approaches such as the use of high-end computing hardware. Peer-to-peer (P2P) or enterprise grids are proven as one of the approaches for developing high-end computing systems. By utilizing them one can improve the performance of NTRU through parallel execution. In this paper we propose and develop an application for NTRU using enterprise grid middleware called Alchemi. An analysis and comparison of its performance for various text files is presented.

Maxwell-Cattaneo Regularization of Heat Equation

This work focuses on analysis of classical heat transfer equation regularized with Maxwell-Cattaneo transfer law. Computer simulations are performed in MATLAB environment. Numerical experiments are first developed on classical Fourier equation, then Maxwell-Cattaneo law is considered. Corresponding equation is regularized with a balancing diffusion term to stabilize discretizing scheme with adjusted time and space numerical steps. Several cases including a convective term in model equations are discussed, and results are given. It is shown that limiting conditions on regularizing parameters have to be satisfied in convective case for Maxwell-Cattaneo regularization to give physically acceptable solutions. In all valid cases, uniform convergence to solution of initial heat equation with Fourier law is observed, even in nonlinear case.

Traffic Load based Performance Analysis of DSR and STAR Routing Protocol

The wireless adhoc network is comprised of wireless node which can move freely and are connected among themselves without central infrastructure. Due to the limited transmission range of wireless interfaces, in most cases communication has to be relayed over intermediate nodes. Thus, in such multihop network each node (also called router) is independent, self-reliant and capable to route the messages over the dynamic network topology. Various protocols are reported in this field and it is very difficult to decide the best one. A key issue in deciding which type of routing protocol is best for adhoc networks is the communication overhead incurred by the protocol. In this paper STAR a table driven and DSR on demand protocols based on IEEE 802.11 are analyzed for their performance on different performance measuring metrics versus varying traffic CBR load using QualNet 5.0.2 network simulator.

An Automatic Pipeline Monitoring System Based on PCA and SVM

This paper proposes a novel system for monitoring the health of underground pipelines. Some of these pipelines transport dangerous contents and any damage incurred might have catastrophic consequences. However, most of these damage are unintentional and usually a result of surrounding construction activities. In order to prevent these potential damages, monitoring systems are indispensable. This paper focuses on acoustically recognizing road cutters since they prelude most construction activities in modern cities. Acoustic recognition can be easily achieved by installing a distributed computing sensor network along the pipelines and using smart sensors to “listen" for potential threat; if there is a real threat, raise some form of alarm. For efficient pipeline monitoring, a novel monitoring approach is proposed. Principal Component Analysis (PCA) was studied and applied. Eigenvalues were regarded as the special signature that could characterize a sound sample, and were thus used for the feature vector for sound recognition. The denoising ability of PCA could make it robust to noise interference. One class SVM was used for classifier. On-site experiment results show that the proposed PCA and SVM based acoustic recognition system will be very effective with a low tendency for raising false alarms.

Three-Dimensional Simulation of Free Electron Laser with Prebunching and Efficiency Enhancement

Three-dimensional simulation of harmonic up generation in free electron laser amplifier operating simultaneously with a cold and relativistic electron beam is presented in steady-state regime where the slippage of the electromagnetic wave with respect to the electron beam is ignored. By using slowly varying envelope approximation and applying the source-dependent expansion to wave equations, electromagnetic fields are represented in terms of the Hermit Gaussian modes which are well suited for the planar wiggler configuration. The electron dynamics is described by the fully threedimensional Lorentz force equation in presence of the realistic planar magnetostatic wiggler and electromagnetic fields. A set of coupled nonlinear first-order differential equations is derived and solved numerically. The fundamental and third harmonic radiation of the beam is considered. In addition to uniform beam, prebunched electron beam has also been studied. For this effect of sinusoidal distribution of entry times for the electron beam on the evolution of radiation is compared with uniform distribution. It is shown that prebunching reduces the saturation length substantially. For efficiency enhancement the wiggler is set to decrease linearly when the radiation of the third harmonic saturates. The optimum starting point of tapering and the slope of radiation in the amplitude of wiggler are found by successive run of the code.

Experimental Comparison of Combustion Characteristic and Pollutant Emission of Gas Oil and Biodiesel

The increasing industrialization and motorization of the world has led to a steep rise for the demand of petroleum-based fuels. Petroleum-based fuels are obtained from limited reserves. These finite reserves are highly concentrated in certain regions of the world. Therefore, those countries not having these resources are facing energy/foreign exchange crisis, mainly due to the import of crude petroleum. Hence, it is necessary to look for alternative fuels which can be produced from resources available locally within the country such as alcohol, biodiesel, vegetable oils etc. Biodiesel is a renewable, domestically produced fuel that has been shown to reduce particulate, hydrocarbon, and carbon monoxide emissions from combustion. In the present study an experimental investigation on emission characteristic of a liquid burner system operating on several percentage of biodiesel and gas oil is carried out. Samples of exhaust gas are analysed with Testo 350 Xl. The results show that biodiesel can lower some pollutant such as CO, CO2 and particulate matter emissions while NOx emission would increase in comparison with gas oil. The results indicate there may be benefits to using biodiesel in industrial processes.

Event Information Extraction System (EIEE): FSM vs HMM

Automatic Extraction of Event information from social text stream (emails, social network sites, blogs etc) is a vital requirement for many applications like Event Planning and Management systems and security applications. The key information components needed from Event related text are Event title, location, participants, date and time. Emails have very unique distinctions over other social text streams from the perspective of layout and format and conversation style and are the most commonly used communication channel for broadcasting and planning events. Therefore we have chosen emails as our dataset. In our work, we have employed two statistical NLP methods, named as Finite State Machines (FSM) and Hidden Markov Model (HMM) for the extraction of event related contextual information. An application has been developed providing a comparison among the two methods over the event extraction task. It comprises of two modules, one for each method, and works for both bulk as well as direct user input. The results are evaluated using Precision, Recall and F-Score. Experiments show that both methods produce high performance and accuracy, however HMM was good enough over Title extraction and FSM proved to be better for Venue, Date, and time.

Protein Secondary Structure Prediction Using Parallelized Rule Induction from Coverings

Protein 3D structure prediction has always been an important research area in bioinformatics. In particular, the prediction of secondary structure has been a well-studied research topic. Despite the recent breakthrough of combining multiple sequence alignment information and artificial intelligence algorithms to predict protein secondary structure, the Q3 accuracy of various computational prediction algorithms rarely has exceeded 75%. In a previous paper [1], this research team presented a rule-based method called RT-RICO (Relaxed Threshold Rule Induction from Coverings) to predict protein secondary structure. The average Q3 accuracy on the sample datasets using RT-RICO was 80.3%, an improvement over comparable computational methods. Although this demonstrated that RT-RICO might be a promising approach for predicting secondary structure, the algorithm-s computational complexity and program running time limited its use. Herein a parallelized implementation of a slightly modified RT-RICO approach is presented. This new version of the algorithm facilitated the testing of a much larger dataset of 396 protein domains [2]. Parallelized RTRICO achieved a Q3 score of 74.6%, which is higher than the consensus prediction accuracy of 72.9% that was achieved for the same test dataset by a combination of four secondary structure prediction methods [2].