Planning of Road Infrastructure Financing: Computational Finance Viewpoint

Lack of resources for road infrastructure financing is a problem that currently affects not only eastern European economies but also many other countries especially in relation to the impact of global financial crisis. In this context, we are talking about the socalled short-investment problem as a result of long-term lack of investment resources. Based on an analysis of road infrastructure financing in the Czech Republic this article points out at weaknesses of current system and proposes a long-term planning methodology supported by system approach. Within this methodology and using created system dynamic model the article predicts the development of short-investment problem in the Country and in reaction on the downward trend of certain sources the article presents various scenarios resulting from the change of the structure of financial sources. In the discussion the article focuses more closely on the possibility of introduction of tax on vehicles instead of taxes with declining revenue streams and estimates its approximate price in relation to reaching various solutions of short-investment in time.

A Multiple-Objective Environmental Rationalization and Optimization for Material Substitution in the Production of Stone-Washed Jeans- Garments

As the Textile Industry is the second largest industry in Egypt and as small and medium-sized enterprises (SMEs) make up a great portion of this industry therein it is essential to apply the concept of Cleaner Production for the purpose of reducing pollution. In order to achieve this goal, a case study concerned with ecofriendly stone-washing of jeans-garments was investigated. A raw material-substitution option was adopted whereby the toxic potassium permanganate and sodium sulfide were replaced by the environmentally compatible hydrogen peroxide and glucose respectively where the concentrations of both replaced chemicals together with the operating time were optimized. In addition, a process-rationalization option involving four additional processes was investigated. By means of criteria such as product quality, effluent analysis, mass and heat balance; and cost analysis with the aid of a statistical model, a process optimization treatment revealed that the superior process optima were 50%, 0.15% and 50min for H2O2 concentration, glucose concentration and time, respectively. With these values the superior process ought to reduce the annual cost by about EGP 105 relative to the currently used conventional method.

Learning Spatio-Temporal Topology of a Multi-Camera Network by Tracking Multiple People

This paper presents a novel approach for representing the spatio-temporal topology of the camera network with overlapping and non-overlapping fields of view (FOVs). The topology is determined by tracking moving objects and establishing object correspondence across multiple cameras. To track people successfully in multiple camera views, we used the Merge-Split (MS) approach for object occlusion in a single camera and the grid-based approach for extracting the accurate object feature. In addition, we considered the appearance of people and the transition time between entry and exit zones for tracking objects across blind regions of multiple cameras with non-overlapping FOVs. The main contribution of this paper is to estimate transition times between various entry and exit zones, and to graphically represent the camera topology as an undirected weighted graph using the transition probabilities.

Development of 3D Coordinates and Damaged Point Detection System for Ducts using IMU

Recently, as the scale of construction projects has increases, more ground excavation for foundations is carried out than ever before. Consequently, damage to underground ducts (gas, water/sewage or oil pipelines, communication cables or power cable ducts) or superannuated pipelines frequently cause serious accidents resulting in damage to life and property. (In Korea, the total length of city water pipelines was approximately 2,000 km as of the end of 2009.) In addition, large amounts of damage caused by fractures, water and gas leakage caused by superannuation or damage to underground ducts in construction has been reported. Therefore, a system is required to precisely detect defects and deterioration in underground pipelines and the locations of such defects, for timely and accurate maintenance or replacement of the ducts. In this study, a system was developed which can locate underground structures (gas and water pipelines, power cable ducts, etc.) in 3D-coordinates and monitor the degree and position of defects using an Inertial Measurement Unit (IMU) sensing technique. The system can prevent damage to underground ducts and superannuated pipelines during construction, and provide reliable data for maintenance. The utility of the IMU sensing technique used in aircraft and ships in civil applications was verified.

The Suitability of GPS Receivers Update Rates for Navigation Applications

Navigation is the processes of monitoring and controlling the movement of an object from one place to another. Currently, Global Positioning System (GPS) is the main navigation system used all over the world for navigation applications. GPS receiver receives signals from at least three satellites to locate and display itself. Displayed positioning information is updated continuously. Update rate is the number of times per second that a display is illuminated. The speed of update is governed by receiver update rate. A higher update rate decreases display lag time and improves distance measurements and tracking especially when moving on a curvy route. The majority of GPS receivers used nowadays are updated every second continuously. This period is considered reasonable for some applications while it is long relatively for high speed applications. In this paper, the suitability and feasibility of GPS receiver with different update rates will be evaluated for various applications according to the level of speed and update rate needed for particular applications.

A Simulation Software for DNA Computing Algorithms Implementation

The capturing of gel electrophoresis image represents the output of a DNA computing algorithm. Before this image is being captured, DNA computing involves parallel overlap assembly (POA) and polymerase chain reaction (PCR) that is the main of this computing algorithm. However, the design of the DNA oligonucleotides to represent a problem is quite complicated and is prone to errors. In order to reduce these errors during the design stage before the actual in-vitro experiment is carried out; a simulation software capable of simulating the POA and PCR processes is developed. This simulation software capability is unlimited where problem of any size and complexity can be simulated, thus saving cost due to possible errors during the design process. Information regarding the DNA sequence during the computing process as well as the computing output can be extracted at the same time using the simulation software.

Using Suffix Tree Document Representation in Hierarchical Agglomerative Clustering

In text categorization problem the most used method for documents representation is based on words frequency vectors called VSM (Vector Space Model). This representation is based only on words from documents and in this case loses any “word context" information found in the document. In this article we make a comparison between the classical method of document representation and a method called Suffix Tree Document Model (STDM) that is based on representing documents in the Suffix Tree format. For the STDM model we proposed a new approach for documents representation and a new formula for computing the similarity between two documents. Thus we propose to build the suffix tree only for any two documents at a time. This approach is faster, it has lower memory consumption and use entire document representation without using methods for disposing nodes. Also for this method is proposed a formula for computing the similarity between documents, which improves substantially the clustering quality. This representation method was validated using HAC - Hierarchical Agglomerative Clustering. In this context we experiment also the stemming influence in the document preprocessing step and highlight the difference between similarity or dissimilarity measures to find “closer" documents.

A Super-Efficiency Model for Evaluating Efficiency in the Presence of Time Lag Effect

In many cases, there are some time lag between the consumption of inputs and the production of outputs. This time lag effect should be considered in evaluating the performance of organizations. Recently, a couple of DEA models were developed for considering time lag effect in efficiency evaluation of research activities. Multi-periods input(MpI) and Multi-periods output(MpO) models are integrate models to calculate simple efficiency considering time lag effect. However, these models can’t discriminate efficient DMUs because of the nature of basic DEA model in which efficiency scores are limited to ‘1’. That is, efficient DMUs can’t be discriminated because their efficiency scores are same. Thus, this paper suggests a super-efficiency model for efficiency evaluation under the consideration of time lag effect based on the MpO model. A case example using a long term research project is given to compare the suggested model with the MpO model.

The Application of an Ensemble of Boosted Elman Networks to Time Series Prediction: A Benchmark Study

In this paper, the application of multiple Elman neural networks to time series data regression problems is studied. An ensemble of Elman networks is formed by boosting to enhance the performance of the individual networks. A modified version of the AdaBoost algorithm is employed to integrate the predictions from multiple networks. Two benchmark time series data sets, i.e., the Sunspot and Box-Jenkins gas furnace problems, are used to assess the effectiveness of the proposed system. The simulation results reveal that an ensemble of boosted Elman networks can achieve a higher degree of generalization as well as performance than that of the individual networks. The results are compared with those from other learning systems, and implications of the performance are discussed.

Development of a Methodology for Processing of Drilling Operations

Drilling is the most common machining operation and it forms the highest machining cost in many manufacturing activities including automotive engine production. The outcome of this operation depends upon many factors including utilization of proper cutting tool geometry, cutting tool material and the type of coating used to improve hardness and resistance to wear, and also cutting parameters. With the availability of a large array of tool geometries, materials and coatings, is has become a challenging task to select the best tool and cutting parameters that would result in the lowest machining cost or highest profit rate. This paper describes an algorithm developed to help achieve good performances in drilling operations by automatically determination of proper cutting tools and cutting parameters. It also helps determine machining sequences resulting in minimum tool changes that would eventually reduce machining time and cost where multiple tools are used.

Surveillance of Super-Extended Objects: Bimodal Approach

This paper describes an effective solution to the task of a remote monitoring of super-extended objects (oil and gas pipeline, railways, national frontier). The suggested solution is based on the principle of simultaneously monitoring of seismoacoustic and optical/infrared physical fields. The principle of simultaneous monitoring of those fields is not new but in contrast to the known solutions the suggested approach allows to control super-extended objects with very limited operational costs. So-called C-OTDR (Coherent Optical Time Domain Reflectometer) systems are used to monitor the seismoacoustic field. Far-CCTV systems are used to monitor the optical/infrared field. A simultaneous data processing provided by both systems allows effectively detecting and classifying target activities, which appear in the monitored objects vicinity. The results of practical usage had shown high effectiveness of the suggested approach.

Motor Imagery Signal Classification for a Four State Brain Machine Interface

Motor imagery classification provides an important basis for designing Brain Machine Interfaces [BMI]. A BMI captures and decodes brain EEG signals and transforms human thought into actions. The ability of an individual to control his EEG through imaginary mental tasks enables him to control devices through the BMI. This paper presents a method to design a four state BMI using EEG signals recorded from the C3 and C4 locations. Principle features extracted through principle component analysis of the segmented EEG are analyzed using two novel classification algorithms using Elman recurrent neural network and functional link neural network. Performance of both classifiers is evaluated using a particle swarm optimization training algorithm; results are also compared with the conventional back propagation training algorithm. EEG motor imagery recorded from two subjects is used in the offline analysis. From overall classification performance it is observed that the BP algorithm has higher average classification of 93.5%, while the PSO algorithm has better training time and maximum classification. The proposed methods promises to provide a useful alternative general procedure for motor imagery classification

A New Effective Local Search Heuristic for the Maximum Clique Problem

An edge based local search algorithm, called ELS, is proposed for the maximum clique problem (MCP), a well-known combinatorial optimization problem. ELS is a two phased local search method effectively £nds the near optimal solutions for the MCP. A parameter ’support’ of vertices de£ned in the ELS greatly reduces the more number of random selections among vertices and also the number of iterations and running times. Computational results on BHOSLIB and DIMACS benchmark graphs indicate that ELS is capable of achieving state-of-the-art-performance for the maximum clique with reasonable average running times.

On Constructing Approximate Convex Hull

The algorithms of convex hull have been extensively studied in literature, principally because of their wide range of applications in different areas. This article presents an efficient algorithm to construct approximate convex hull from a set of n points in the plane in O(n + k) time, where k is the approximation error control parameter. The proposed algorithm is suitable for applications preferred to reduce the computation time in exchange of accuracy level such as animation and interaction in computer graphics where rapid and real-time graphics rendering is indispensable.

Constructing Distinct Kinds of Solutions for the Time-Dependent Coefficients Coupled Klein-Gordon-Schrödinger Equation

We seek exact solutions of the coupled Klein-Gordon-Schrödinger equation with variable coefficients with the aid of Lie classical approach. By using the Lie classical method, we are able to derive symmetries that are used for reducing the coupled system of partial differential equations into ordinary differential equations. From reduced differential equations we have derived some new exact solutions of coupled Klein-Gordon-Schrödinger equations involving some special functions such as Airy wave functions, Bessel functions, Mathieu functions etc.

iDENTM Phones Automated Stress Testing

System testing is actually done to the entire system against the Functional Requirement Specification and/or the System Requirement Specification. Moreover, it is an investigatory testing phase, where the focus is to have almost a destructive attitude and test not only the design, but also the behavior and even the believed expectations of the customer. It is also intended to test up to and beyond the bounds defined in the software/hardware requirements specifications. In Motorola®, Automated Testing is one of the testing methodologies uses by GSG-iSGT (Global Software Group - iDEN TM Subcriber Group-Test) to increase the testing volume, productivity and reduce test cycle-time in iDEN TM phones testing. Testing is able to produce more robust products before release to the market. In this paper, iHopper is proposed as a tool to perform stress test on iDEN TM phonse. We will discuss the value that automation has brought to iDEN TM Phone testing such as improving software quality in the iDEN TM phone together with some metrics. We will also look into the advantages of the proposed system and some discussion of the future work as well.

Mathematical Model of Smoking Time Temperature Effect on Ribbed Smoked Sheets Quality

The quality of Ribbed Smoked Sheets (RSS) primarily based on color, dryness, and the presence or absence of fungus and bubbles. This quality is strongly influenced by the drying and fumigation process namely smoking process. Smoking that is held in high temperature long time will result scorched dark brown sheets, whereas if the temperature is too low or slow drying rate would resulted in less mature sheets and growth of fungus. Therefore need to find the time and temperature for optimum quality of sheets. Enhance, unmonitored heat and mass transfer during smoking process lead to high losses of energy balance. This research aims to generate simple empirical mathematical model describing the effect of smoking time and temperature to RSS quality of color, water content, fungus and bubbles. The second goal of study was to analyze energy balance during smoking process. Experimental study was conducted by measuring temperature, residence time and quality parameters of 16 sheets sample in smoking rooms. Data for energy consumption balance such as mass of fuel wood, mass of sheets being smoked, construction temperature, ambient temperature and relative humidity were taken directly along the smoking process. It was found that mathematical model correlating smoking temperature and time with color is Color = -169 - 0.184 T4 - 0.193 T3 - 0.160 0.405 T1 + T2 + 0.388 t1 +3.11 t2 + 3.92t3 + 0.215 t4 with R square 50.8% and with moisture is Moisture = -1.40-0.00123 T4 + 0.00032 T3 + 0.00260 T2 - 0.00292 T1 - 0.0105 t1 + 0.0290 t2 + 0.0452 t3 + 0.00061 t4 with R square of 49.9%. Smoking room energy analysis found useful energy was 27.8%. The energy stored in the material construction 7.3%. Lost of energy in conversion of wood combustion, ventilation and others were 16.6%. The energy flowed out through the contact of material construction with the ambient air was found to be the highest contribution to energy losses, it reached 48.3%.

Design of Extremum Seeking Control with PD Accelerator and its Application to Monod and Williams-Otto Models

In this paper, we are concerned with the design and its simulation studies of a modified extremum seeking control for nonlinear systems. A standard extremum seeking control has a simple structure, but it takes a long time to reach an optimal operating point. We consider a modification of the standard extremum seeking control which is aimed to reach the optimal operating point more speedily than the standard one. In the modification, PD acceleration term is added before an integrator making a principal control, so that it enables the objects to be regulated to the optimal point smoothly. This proposed method is applied to Monod and Williams-Otto models to investigate its effectiveness. Numerical simulation results show that this modified method can improve the time response to the optimal operating point more speedily than the standard one.

Network Application Identification Based on Communication Characteristics of Application Messages

A person-to-person information sharing is easily realized by P2P networks in which servers are not essential. Leakage of information, which are caused by malicious accesses for P2P networks, has become a new social issues. To prevent information leakage, it is necessary to detect and block traffics of P2P software. Since some P2P softwares can spoof port numbers, it is difficult to detect the traffics sent from P2P softwares by using port numbers. It is more difficult to devise effective countermeasures for detecting the software because their protocol are not public. In this paper, a discriminating method of network applications based on communication characteristics of application messages without port numbers is proposed. The proposed method is based on an assumption that there can be some rules about time intervals to transmit messages in application layer and the number of necessary packets to send one message. By extracting the rule from network traffic, the proposed method can discriminate applications without port numbers.

Streamflow Modeling for a Small Watershed Using Limited Hydrological Data

This research was conducted in the Pua Watershed whereas located in the Upper Nan River Basin in Nan province, Thailand. Nan River basin originated in Nan province that comprises of many tributary streams to produce as inflow to the Sirikit dam provided huge reservoir with the storage capacity of 9510 million cubic meters. The common problems of most watersheds were found i.e. shortage water supply for consumption and agriculture utilizations, deteriorate of water quality, flood and landslide including debris flow, and unstable of riverbank. The Pua Watershed is one of several small river basins that flow through the Nan River Basin. The watershed includes 404 km2 representing the Pua District, the Upper Nan Basin, or the whole Nan River Basin, of 61.5%, 18.2% or 1.2% respectively. The Pua River is a main stream producing all year streamflow supplying the Pua District and an inflow to the Upper Nan Basin. Its length approximately 56.3 kilometers with an average slope of the channel by 1.9% measured. A diversion weir namely Pua weir bound the plain and mountainous areas with a very steep slope of the riverbed to 2.9% and drainage area of 149 km2 as upstream watershed while a mild slope of the riverbed to 0.2% found in a river reach of 20.3 km downstream of this weir, which considered as a gauged basin. However, the major branch streams of the Pua River are ungauged catchments namely: Nam Kwang and Nam Koon with the drainage area of 86 and 35 km2 respectively. These upstream watersheds produce runoff through the 3-streams downstream of Pua weir, Jao weir, and Kang weir, with an averaged annual runoff of 578 million cubic meters. They were analyzed using both statistical data at Pua weir and simulated data resulted from the hydrologic modeling system (HEC–HMS) which applied for the remaining ungauged basins. Since the Kwang and Koon catchments were limited with lack of hydrological data included streamflow and rainfall. Therefore, the mathematical modeling: HEC-HMS with the Snyder-s hydrograph synthesized and transposed methods were applied for those areas using calibrated hydrological parameters from the upstream of Pua weir with continuously daily recorded of streamflow and rainfall data during 2008-2011. The results showed that the simulated daily streamflow and sum up as annual runoff in 2008, 2010, and 2011 were fitted with observed annual runoff at Pua weir using the simple linear regression with the satisfied correlation R2 of 0.64, 062, and 0.59, respectively. The sensitivity of simulation results were come from difficulty using calibrated parameters i.e. lag-time, coefficient of peak flow, initial losses, uniform loss rates, and missing some daily observed data. These calibrated parameters were used to apply for the other 2-ungauged catchments and downstream catchments simulated.