Faster FPGA Routing Solution using DNA Computing

There are many classical algorithms for finding routing in FPGA. But Using DNA computing we can solve the routes efficiently and fast. The run time complexity of DNA algorithms is much less than other classical algorithms which are used for solving routing in FPGA. The research in DNA computing is in a primary level. High information density of DNA molecules and massive parallelism involved in the DNA reactions make DNA computing a powerful tool. It has been proved by many research accomplishments that any procedure that can be programmed in a silicon computer can be realized as a DNA computing procedure. In this paper we have proposed two tier approaches for the FPGA routing solution. First, geometric FPGA detailed routing task is solved by transforming it into a Boolean satisfiability equation with the property that any assignment of input variables that satisfies the equation specifies a valid routing. Satisfying assignment for particular route will result in a valid routing and absence of a satisfying assignment implies that the layout is un-routable. In second step, DNA search algorithm is applied on this Boolean equation for solving routing alternatives utilizing the properties of DNA computation. The simulated results are satisfactory and give the indication of applicability of DNA computing for solving the FPGA Routing problem.

Static Voltage Stability Assessment Considering the Power System Contingencies using Continuation Power Flow Method

According to the increasing utilization in power system, the transmission lines and power plants often operate in stability boundary and system probably lose its stable condition by over loading or occurring disturbance. According to the reasons that are mentioned, the prediction and recognition of voltage instability in power system has particular importance and it makes the network security stronger.This paper, by considering of power system contingencies based on the effects of them on Mega Watt Margin (MWM) and maximum loading point is focused in order to analyse the static voltage stability using continuation power flow method. The study has been carried out on IEEE 14-Bus Test System using Matlab and Psat softwares and results are presented.

Action Potential Propagation in Inhomogeneous 2D Mouse Ventricular Tissue Model

Heterogeneous repolarization causes dispersion of the T-wave and has been linked to arrhythmogenesis. Such heterogeneities appear due to differential expression of ionic currents in different regions of the heart, both in healthy and diseased animals and humans. Mice are important animals for the study of heart diseases because of the ability to create transgenic animals. We used our previously reported model of mouse ventricular myocytes to develop 2D mouse ventricular tissue model consisting of 14,000 cells (apical or septal ventricular myocytes) and to study the stability of action potential propagation and Ca2+ dynamics. The 2D tissue model was implemented as a FORTRAN program code for highperformance multiprocessor computers that runs on 36 processors. Our tissue model is able to simulate heterogeneities not only in action potential repolarization, but also heterogeneities in intracellular Ca2+ transients. The multicellular model reproduced experimentally observed velocities of action potential propagation and demonstrated the importance of incorporation of realistic Ca2+ dynamics for action potential propagation. The simulations show that relatively sharp gradients of repolarization are predicted to exist in 2D mouse tissue models, and they are primarily determined by the cellular properties of ventricular myocytes. Abrupt local gradients of channel expression can cause alternans at longer pacing basic cycle lengths than gradual changes, and development of alternans depends on the site of stimulation.

Structural Sustainability Techniques for RC High Rise Buildings

Over the early years of the 21st century, cities throughout the Middle East, particularly in the Gulf region have expanded more rapidly than ever before. Given the presence of a large volume of high-rise buildings allover the region, the local authority aims to set a new standard for sustainable development; with an integrated approach to maintain a balance between economy, quality, environmental protection and safety of life. In the very near future, as mandatory requirements, sustainability will be the criteria that should be included in all building projects. It is well known in the building sustainability topics that structural design engineers do not have a key role in this matter. In addition, the LEED (Leadership in Energy and Environmental Design) has looked almost exclusively on the environmental components and materials specifications. The objective of this paper is to focus and establish groundwork for sustainability techniques and applications related to the RC high-rise buildings design, from the structural point of view. A set of recommendations related to local conditions, structural modeling and analysis is given, and some helpful suggestions for structural design team work are addressed. This paper attempts to help structural engineers in identifying the building sustainability design, in order to meet local needs and achieve alternative solutions at an early stage of project design.

Context Modeling and Context-Aware Service Adaptation for Pervasive Computing Systems

Devices in a pervasive computing system (PCS) are characterized by their context-awareness. It permits them to provide proactively adapted services to the user and applications. To do so, context must be well understood and modeled in an appropriate form which enhance its sharing between devices and provide a high level of abstraction. The most interesting methods for modeling context are those based on ontology however the majority of the proposed methods fail in proposing a generic ontology for context which limit their usability and keep them specific to a particular domain. The adaptation task must be done automatically and without an explicit intervention of the user. Devices of a PCS must acquire some intelligence which permits them to sense the current context and trigger the appropriate service or provide a service in a better suitable form. In this paper we will propose a generic service ontology for context modeling and a context-aware service adaptation based on a service oriented definition of context.

Efficiency in Urban Governance towards Sustainability and Competitiveness of City : A Case Study of Kuala Lumpur

Malaysia has successfully applied economic planning to guide the development of the country from an economy of agriculture and mining to a largely industrialised one. Now, with its sights set on attaining the economic level of a fully developed nation by 2020, the planning system must be made even more efficient and focused. It must ensure that every investment made in the country, contribute towards creating the desirable objective of a strong, modern, internationally competitive, technologically advanced, post-industrial economy. Cities in Malaysia must also be fully aware of the enormous competition it faces in a region with rapidly expanding and modernising economies, all contending for the same pool of potential international investments. Efficiency of urban governance is also fundamental issue in development characterized by sustainability, subsidiarity, equity, transparency and accountability, civic engagement and citizenship, and security. As described above, city competitiveness is harnessed through 'city marketing and city management'. High technology and high skilled industries, together with finance, transportation, tourism, business, information and professional services shopping and other commercial activities, are the principal components of the nation-s economy, which must be developed to a level well beyond where it is now. In this respect, Kuala Lumpur being the premier city must play the leading role.

Influence of Heterogeneous Traffic on the Roadside Fine (PM2.5 and PM1) and Coarse(PM10) Particulate Matter Concentrations in Chennai City, India

In this paper the influence of heterogeneous traffic on the temporal variation of ambient PM10, PM2.5 and PM1 concentrations at a busy arterial route (Sardar Patel Road) in the Chennai city has been analyzed. The hourly PM concentration, traffic counts and average speed of the vehicles have been monitored at the study site for one week (19th-25th January 2009). Results indicated that the concentrations of coarse (PM10) and fine PM (PM2.5 and PM1) concentrations at SP road are having similar trend during peak and non-peak hours, irrespective of the days. The PM concentrations showed daily two peaks corresponding to morning (8 to 10 am) and evening (7 to 9 pm) peak hour traffic flow. The PM10 concentration is dominated by fine particles (53% of PM2.5 and 45% of PM1). The high PM2.5/PM10 ratio indicates that the majority of PM10 particles originate from re-suspension of road dust. The analysis of traffic flow at the study site showed that 2W, 3W and 4W are having similar diurnal trend as PM concentrations. This confirms that the 2W, 3W and 4W are the main emission source contributing to ambient PM concentration at SP road. The speed measurement at SP road showed that the average speed of 2W, 3W, 4W, LCV and HCV are 38, 40, 38, 40 and 38 km/hr and 43, 41, 42, 40 and 41 km/hr respectively for the weekdays and weekdays.

Measuring Creativity in Die Products for Technological Education

Creative design requires new approaches to assessment in vocational and technological education. To date, there has been little discussion on instruments used to evaluate dies produced by students in vocational and technological education. Developing a generic instrument has been very difficult due to the diversity of creative domains, the specificity of content, and the subjectivity involved in judgment. This paper presents an instrument for measuring the creativity in the design of products by expanding the Consensual Assessment Technique (CAT). The content-based scale was evaluated for content validity by 5 experts. The scale comprises 5 criteria: originality; practicability; precision; aesthetics; and exchangeability. Nine experts were invited to evaluate the dies produced by 38 college students who enrolled in a Product Design and Development course. To further explore the degree of rater agreement, inter-rater reliability was calculated for each dimension using Kendall's coefficient of concordance test. The inter-judge reliability scores achieved significance, with coefficients ranging from 0.53 to 0.71.

Good Practices in the Development of the Erasmus Mundus Master program in Color in Informatics and Media Technology

The main objective of this paper is to identify and disseminate good practice in quality assurance and enhancement as well as in teaching and learning at master level. This paper focuses on the experience of the Erasmus Mundus Master program CIMET (Color in Informatics and Media Technology). Amongst topics covered, we discuss the adjustments necessary to a curriculum designed for excellent international students and their preparation for a global labor market.

Lease Agreement in the European Countries

This paper present lease agreement regulations in selected European countries. The lease agreement has a long history and now is one of the main ways to manage agricultural lands in Europe. The analysis of individual regulations, which has been done, indicates that this agreement is very important to build social relations in agriculture and society. This article provides an analysis of the legal regulations concerning the lease in France, Spain, Switzerland, Ukraine and Italy. Article is example of study of the legal regulations and can be used for legal changes in individual countries.

Effective Defect Prevention Approach in Software Process for Achieving Better Quality Levels

Defect prevention is the most vital but habitually neglected facet of software quality assurance in any project. If functional at all stages of software development, it can condense the time, overheads and wherewithal entailed to engineer a high quality product. The key challenge of an IT industry is to engineer a software product with minimum post deployment defects. This effort is an analysis based on data obtained for five selected projects from leading software companies of varying software production competence. The main aim of this paper is to provide information on various methods and practices supporting defect detection and prevention leading to thriving software generation. The defect prevention technique unearths 99% of defects. Inspection is found to be an essential technique in generating ideal software generation in factories through enhanced methodologies of abetted and unaided inspection schedules. On an average 13 % to 15% of inspection and 25% - 30% of testing out of whole project effort time is required for 99% - 99.75% of defect elimination. A comparison of the end results for the five selected projects between the companies is also brought about throwing light on the possibility of a particular company to position itself with an appropriate complementary ratio of inspection testing.

A Frugal Bidding Procedure for Replicating WWW Content

Fine-grained data replication over the Internet allows duplication of frequently accessed data objects, as opposed to entire sites, to certain locations so as to improve the performance of largescale content distribution systems. In a distributed system, agents representing their sites try to maximize their own benefit since they are driven by different goals such as to minimize their communication costs, latency, etc. In this paper, we will use game theoretical techniques and in particular auctions to identify a bidding mechanism that encapsulates the selfishness of the agents, while having a controlling hand over them. In essence, the proposed game theory based mechanism is the study of what happens when independent agents act selfishly and how to control them to maximize the overall performance. A bidding mechanism asks how one can design systems so that agents- selfish behavior results in the desired system-wide goals. Experimental results reveal that this mechanism provides excellent solution quality, while maintaining fast execution time. The comparisons are recorded against some well known techniques such as greedy, branch and bound, game theoretical auctions and genetic algorithms.

Applications of Prediction and Identification Using Adaptive DCMAC Neural Networks

An adaptive dynamic cerebellar model articulation controller (DCMAC) neural network used for solving the prediction and identification problem is proposed in this paper. The proposed DCMAC has superior capability to the conventional cerebellar model articulation controller (CMAC) neural network in efficient learning mechanism, guaranteed system stability and dynamic response. The recurrent network is embedded in the DCMAC by adding feedback connections in the association memory space so that the DCMAC captures the dynamic response, where the feedback units act as memory elements. The dynamic gradient descent method is adopted to adjust DCMAC parameters on-line. Moreover, the analytical method based on a Lyapunov function is proposed to determine the learning-rates of DCMAC so that the variable optimal learning-rates are derived to achieve most rapid convergence of identifying error. Finally, the adaptive DCMAC is applied in two computer simulations. Simulation results show that accurate identifying response and superior dynamic performance can be obtained because of the powerful on-line learning capability of the proposed DCMAC.

Audio Watermarking Using Spectral Modifications

In this paper, we present a non-blind technique of adding the watermark to the Fourier spectral components of audio signal in a way such that the modified amplitude does not exceed the maximum amplitude spread (MAS). This MAS is due to individual Discrete fourier transform (DFT) coefficients in that particular frame, which is derived from the Energy Spreading function given by Schroeder. Using this technique one can store double the information within a given frame length i.e. overriding the watermark on the host of equal length with least perceptual distortion. The watermark is uniformly floating on the DFT components of original signal. This helps in detecting any intentional manipulations done on the watermarked audio. Also, the scheme is found robust to various signal processing attacks like presence of multiple watermarks, Additive white gaussian noise (AWGN) and mp3 compression.

Coordination for Synchronous Cooperative Systems Based on Fuzzy Causal Relations

Synchronous cooperative systems (SCS) bring together users that are geographically distributed and connected through a network to carry out a task. Examples of SCS include Tele- Immersion and Tele-Conferences. In SCS, the coordination is the core of the system, and it has been defined as the act of managing interdependencies between activities performed to achieve a goal. Some of the main problems that SCS present deal with the management of constraints between simultaneous activities and the execution ordering of these activities. In order to resolve these problems, orderings based on Lamport-s happened-before relation have been used, namely, causal, Δ-causal, and causal-total orderings. They mainly differ in the degree of asynchronous execution allowed. One of the most important orderings is the causal order, which establishes that the events must be seen in the cause-effect order as they occur in the system. In this paper we show that for certain SCS (e.g. videoconferences, tele-immersion) where some degradation of the system is allowed, ensuring the causal order is still rigid, which can render negative affects to the system. In this paper, we illustrate how a more relaxed ordering, which we call Fuzzy Causal Order (FCO), is useful for such kind of systems by allowing a more asynchronous execution than the causal order. The benefit of the FCO is illustrated by applying it to a particular scenario of intermedia synchronization of an audio-conference system.

Sonic Localization Cues for Classrooms: A Structural Model Proposal

We investigate sonic cues for binaural sound localization within classrooms and present a structural model for the same. Two of the primary cues for localization, interaural time difference (ITD) and interaural level difference (ILD) created between the two ears by sounds from a particular point in space, are used. Although these cues do not lend any information about the elevation of a sound source, the torso, head, and outer ear carry out elevation dependent spectral filtering of sounds before they reach the inner ear. This effect is commonly captured in head related transfer function (HRTF) which aids in resolving the ambiguity from the ITDs and ILDs alone and helps localize sounds in free space. The proposed structural model of HRTF produces well controlled horizontal as well as vertical effects. The implemented HRTF is a signal processing model which tries to mimic the physical effects of the sounds interacting with different parts of the body. The effectiveness of the method is tested by synthesizing spatial audio, in MATLAB, for use in listening tests with human subjects and is found to yield satisfactory results in comparison with existing models.

Reform-Oriented Teaching of Introductory Statistics in the Health, Social and Behavioral Sciences – Historical Context and Rationale

There is widespread emphasis on reform in the teaching of introductory statistics at the college level. Underpinning this reform is a consensus among educators and practitioners that traditional curricular materials and pedagogical strategies have not been effective in promoting statistical literacy, a competency that is becoming increasingly necessary for effective decision-making and evidence-based practice. This paper explains the historical context of, and rationale for reform-oriented teaching of introductory statistics (at the college level) in the health, social and behavioral sciences (evidence-based disciplines). A firm understanding and appreciation of the basis for change in pedagogical approach is important, in order to facilitate commitment to reform, consensus building on appropriate strategies, and adoption and maintenance of best practices. In essence, reform-oriented pedagogy, in this context, is a function of the interaction among content, pedagogy, technology, and assessment. The challenge is to create an appropriate balance among these domains.

A Preliminary Study on the Suitability of Data Driven Approach for Continuous Water Level Modeling

Reliable water level forecasts are particularly important for warning against dangerous flood and inundation. The current study aims at investigating the suitability of the adaptive network based fuzzy inference system for continuous water level modeling. A hybrid learning algorithm, which combines the least square method and the back propagation algorithm, is used to identify the parameters of the network. For this study, water levels data are available for a hydrological year of 2002 with a sampling interval of 1-hour. The number of antecedent water level that should be included in the input variables is determined by two statistical methods, i.e. autocorrelation function and partial autocorrelation function between the variables. Forecasting was done for 1-hour until 12-hour ahead in order to compare the models generalization at higher horizons. The results demonstrate that the adaptive networkbased fuzzy inference system model can be applied successfully and provide high accuracy and reliability for river water level estimation. In general, the adaptive network-based fuzzy inference system provides accurate and reliable water level prediction for 1-hour ahead where the MAPE=1.15% and correlation=0.98 was achieved. Up to 12-hour ahead prediction, the model still shows relatively good performance where the error of prediction resulted was less than 9.65%. The information gathered from the preliminary results provide a useful guidance or reference for flood early warning system design in which the magnitude and the timing of a potential extreme flood are indicated.

The Shaping of a Triangle Steel Plate into an Equilateral Vertical Steel by Finite-Element Modeling

The orthogonal processes to shape the triangle steel plate into a equilateral vertical steel are examined by an incremental elasto-plastic finite-element method based on an updated Lagrangian formulation. The highly non-linear problems due to the geometric changes, the inelastic constitutive behavior and the boundary conditions varied with deformation are taken into account in an incremental manner. On the contact boundary, a modified Coulomb friction mode is specially considered. A weighting factor r-minimum is employed to limit the step size of loading increment to linear relation. In particular, selective reduced integration was adopted to formulate the stiffness matrix. The simulated geometries of verticality could clearly demonstrate the vertical processes until unloading. A series of experiments and simulations were performed to validate the formulation in the theory, leading to the development of the computer codes. The whole deformation history and the distribution of stress, strain and thickness during the forming process were obtained by carefully considering the moving boundary condition in the finite-element method. Therefore, this modeling can be used for judging whether a equilateral vertical steel can be shaped successfully. The present work may be expected to improve the understanding of the formation of the equilateral vertical steel.

Approach to Implementation of Power Management with Load Prioritizations in Modern Civil Aircraft

Any use of energy in industrial productive activities is combined with various environment impacts. Withintransportation, this fact was not only found among land transport, railways and maritime transport, but also in the air transport industry. An effective climate protection requires strategies and measures for reducing all greenhouses gas emissions, in particular carbon dioxide, and must take into account the economic, ecologic and social aspects. It seem simperative now to develop and manufacture environmentally friendly products and systems, to reduce consumption and use less resource, and to save energy and power. Today-sproducts could better serve these requirements taking into account the integration of a power management system into the electrical power system.This paper gives an overview of an approach ofpower management with load prioritization in modernaircraft. Load dimensioning and load management strategies on current civil aircraft will be presented and used as a basis for the proposed approach.