Voice Command Recognition System Based on MFCC and VQ Algorithms

The goal of this project is to design a system to recognition voice commands. Most of voice recognition systems contain two main modules as follow “feature extraction" and “feature matching". In this project, MFCC algorithm is used to simulate feature extraction module. Using this algorithm, the cepstral coefficients are calculated on mel frequency scale. VQ (vector quantization) method will be used for reduction of amount of data to decrease computation time. In the feature matching stage Euclidean distance is applied as similarity criterion. Because of high accuracy of used algorithms, the accuracy of this voice command system is high. Using these algorithms, by at least 5 times repetition for each command, in a single training session, and then twice in each testing session zero error rate in recognition of commands is achieved.

Multi-Agent Systems Applied in the Modeling and Simulation of Biological Problems: A Case Study in Protein Folding

Multi-agent system approach has proven to be an effective and appropriate abstraction level to construct whole models of a diversity of biological problems, integrating aspects which can be found both in "micro" and "macro" approaches when modeling this type of phenomena. Taking into account these considerations, this paper presents the important computational characteristics to be gathered into a novel bioinformatics framework built upon a multiagent architecture. The version of the tool presented herein allows studying and exploring complex problems belonging principally to structural biology, such as protein folding. The bioinformatics framework is used as a virtual laboratory to explore a minimalist model of protein folding as a test case. In order to show the laboratory concept of the platform as well as its flexibility and adaptability, we studied the folding of two particular sequences, one of 45-mer and another of 64-mer, both described by an HP model (only hydrophobic and polar residues) and coarse grained 2D-square lattice. According to the discussion section of this piece of work, these two sequences were chosen as breaking points towards the platform, in order to determine the tools to be created or improved in such a way to overcome the needs of a particular computation and analysis of a given tough sequence. The backwards philosophy herein is that the continuous studying of sequences provides itself important points to be added into the platform, to any time improve its efficiency, as is demonstrated herein.

Financing - Scheduling Optimization for Construction Projects by using Genetic Algorithms

Investment in a constructed facility represents a cost in the short term that returns benefits only over the long term use of the facility. Thus, the costs occur earlier than the benefits, and the owners of facilities must obtain the capital resources to finance the costs of construction. A project cannot proceed without an adequate financing, and the cost of providing an adequate financing can be quite large. For these reasons, the attention to the project finance is an important aspect of project management. Finance is also a concern to the other organizations involved in a project such as the general contractor and material suppliers. Unless an owner immediately and completely covers the costs incurred by each participant, these organizations face financing problems of their own. At a more general level, the project finance is the only one aspect of the general problem of corporate finance. If numerous projects are considered and financed together, then the net cash flow requirements constitute the corporate financing problem for capital investment. Whether project finance is performed at the project or at the corporate level does not alter the basic financing problem .In this paper, we will first consider facility financing from the owner's perspective, with due consideration for its interaction with other organizations involved in a project. Later, we discuss the problems of construction financing which are crucial to the profitability and solvency of construction contractors. The objective of this paper is to present the steps utilized to determine the best combination of minimum project financing. The proposed model considers financing; schedule and maximum net area .The proposed model is called Project Financing and Schedule Integration using Genetic Algorithms "PFSIGA". This model intended to determine more steps (maximum net area) for any project with a subproject. An illustrative example will demonstrate the feature of this technique. The model verification and testing are put into consideration.

A New Extended Group Mutual Exclusion Algorithm with Low Message Complexity in Distributed Systems

The group mutual exclusion (GME) problem is an interesting generalization of the mutual exclusion problem. In the group mutual exclusion, multiple processes can enter a critical section simultaneously if they belong to the same group. In the extended group mutual exclusion, each process is a member of multiple groups at the same time. As a result, after the process by selecting a group enter critical section, other processes can select the same group with its belonging group and can enter critical section at the moment, so that it avoids their unnecessary blocking. This paper presents a quorum-based distributed algorithm for the extended group mutual exclusion problem. The message complexity of our algorithm is O(4Q ) in the best case and O(5Q) in the worst case, where Q is a quorum size.

Reconfigurable Autonomous Mini Robot Design using CPLD's

This paper explains a project based learning method where autonomous mini-robots are developed for research, education and entertainment purposes. In case of remote systems wireless sensors are developed in critical areas, which would collect data at specific time intervals, send the data to the central wireless node based on certain preferred information would make decisions to turn on or off a switch or control unit. Such information transfers hardly sums up to a few bytes and hence low data rates would suffice for such implementations. As a robot is a multidisciplinary platform, the interfacing issues involved are discussed in this paper. The paper is mainly focused on power supply, grounding and decoupling issues.

Survey of Impact of Production and Adoption of Nanocrops on Food Security

Perspective of food security in 21 century showed shortage of food that production is faced to vital problem. Food security strategy is applied longtime method to assess required food. Meanwhile, nanotechnology revolution changes the world face. Nanotechnology is adequate method utilize of its characteristics to decrease environmental problems and possible further access to food for small farmers. This article will show impact of production and adoption of nanocrops on food security. Population is researchers of agricultural research center of Esfahan province. The results of study show that there was a relationship between uses, conversion, distribution, and production of nanocrops, operative human resources, operative circumstance, and constrains of usage of nanocrops and food security. Multivariate regression analysis by enter model shows that operative circumstance, use, production and constrains of usage of nanocrops had positive impact on food security and they determine in four steps 20 percent of it.

Augmenting Use Case View for Modeling

Mathematical, graphical and intuitive models are often constructed in the development process of computational systems. The Unified Modeling Language (UML) is one of the most popular modeling languages used by practicing software engineers. This paper critically examines UML models and suggests an augmented use case view with the addition of new constructs for modeling software. It also shows how a use case diagram can be enhanced. The improved modeling constructs are presented with examples for clarifying important design and implementation issues.

A Novel Estimation Method for Integer Frequency Offset in Wireless OFDM Systems

Ren et al. presented an efficient carrier frequency offset (CFO) estimation method for orthogonal frequency division multiplexing (OFDM), which has an estimation range as large as the bandwidth of the OFDM signal and achieves high accuracy without any constraint on the structure of the training sequence. However, its detection probability of the integer frequency offset (IFO) rapidly varies according to the fractional frequency offset (FFO) change. In this paper, we first analyze the Ren-s method and define two criteria suitable for detection of IFO. Then, we propose a novel method for the IFO estimation based on the maximum-likelihood (ML) principle and the detection criteria defined in this paper. The simulation results demonstrate that the proposed method outperforms the Ren-s method in terms of the IFO detection probability irrespective of a value of the FFO.

An Integrated Natural Language Processing Approach for Conversation System

The main aim of this research is to investigate a novel technique for implementing a more natural and intelligent conversation system. Conversation systems are designed to converse like a human as much as their intelligent allows. Sometimes, we can think that they are the embodiment of Turing-s vision. It usually to return a predetermined answer in a predetermined order, but conversations abound with uncertainties of various kinds. This research will focus on an integrated natural language processing approach. This approach includes an integrated knowledge-base construction module, a conversation understanding and generator module, and a state manager module. We discuss effectiveness of this approach based on an experiment.

Turbulent Mixing and its Effects on Thermal Fatigue in Nuclear Reactors

The turbulent mixing of coolant streams of different temperature and density can cause severe temperature fluctuations in piping systems in nuclear reactors. In certain periodic contraction cycles these conditions lead to thermal fatigue. The resulting aging effect prompts investigation in how the mixing of flows over a sharp temperature/density interface evolves. To study the fundamental turbulent mixing phenomena in the presence of density gradients, isokinetic (shear-free) mixing experiments are performed in a square channel with Reynolds numbers ranging from 2-500 to 60-000. Sucrose is used to create the density difference. A Wire Mesh Sensor (WMS) is used to determine the concentration map of the flow in the cross section. The mean interface width as a function of velocity, density difference and distance from the mixing point are analyzed based on traditional methods chosen for the purposes of atmospheric/oceanic stratification analyses. A definition of the mixing layer thickness more appropriate to thermal fatigue and based on mixedness is devised. This definition shows that the thermal fatigue risk assessed using simple mixing layer growth can be misleading and why an approach that separates the effects of large scale (turbulent) and small scale (molecular) mixing is necessary.

Effects of Microwave Heating on Biogas Production, Chemical Oxygen Demand and Volatile Solids Solubilization of Food Residues

This paper presents the results of the preliminary investigation of microwave (MW) irradiation pretreatments on the anaerobic digestion of food residues using biochemical methane potential (BMP) assays. Low solids systems with a total solids (TS) content ranging from 5.0-10.0% were analyzed. The inoculum to bulk mass of substrates to water ratio was 1:2:2 (mass basis). The experimental conditions for pretreatments were as follows: a control (no MW irradiation), two runs with MW irradiation for 15 and 30 minutes at 320 W, and another two runs with MW irradiation at 528 W for 30 and 60 minutes. The cumulative biogas production were 6.3 L and 8.7 L for 15min/320 W and 30min/320 W MW irradiation conditions, respectively, and 10.5 L and 11.4 L biogas for 30min/528 W and 60min/528 W, respectively, as compared to the control giving 5.8 L biogas. Both an increase in exposure time of irradiation and power of MW had increased the rate and yield of biogas. Singlefactor ANOVA tests (p

The Relationship between Spatial Planning and Transportation Planning in Southern Africa and its Consequences for Human Settlement

The paper reviews the relationship between spatial and transportation planning in the Southern African Development Community (SADC) region of Sub-Saharan Africa. It argues that most urbanisation in the region has largely occurred subsequent to the 1950s and, accordingly, urban development has been profoundly and negatively affected by the (misguided) spatial and institutional tenets of modernism. It demonstrates how a considerable amount of the poor performance of these settlements can be directly attributed to this. Two factors in particular about the planning systems are emphasized: the way in which programmatic land-use planning lies at the heart of both spatial and transportation planning; and the way on which transportation and spatial planning have been separated into independent processes. In the final section, the paper identifies ways of improving the planning system. Firstly, it identifies the performance qualities which Southern African settlements should be seeking to achieve. Secondly, it focuses on two necessary arenas of change: the need to replace programmatic land-use planning practices with structuralspatial approaches; and it makes a case for making urban corridors a spatial focus of integrated planning, as a way of beginning the restructuring and intensification of settlements which are currently characterised by sprawl, fragmentation and separation

Environmental Management System According to ISO 14001 as a Source of Eco-Innovations in Enterprises - A Case of Podkarpackie Voivodeship

This paper presents results of empirical studies that were conducted in enterprises from Podkarpackie Voivodeship (Poland). It shows the experiences of those enterprises resulting from implementing and improving the eco-innovativeness management that is formal Environmental Management System (EMS). This study shows the expected and obtained internal benefits which are the effects of a functioning EMS. The aim of this paper is to determine whether the information included in international theoretical studies concerning the benefits of implementing, functioning and improving formal EMS (which is based on the international standard ISO 14001) are confirmed by the effects of the enterprises- activities.

Challenges on Adopting Scrum for Distributed Teams in Home Office Environments

This paper describes the two actual tendencies in the software development process usage: 'Scrum' and 'work in home office'. It-s exposed the four main challenges to adopt Scrum framework for distributed teams in this cited kind of work. The challenges are mainly based on the communication problems due distances since the Scrum encourages the team to work together in the same room, and this is not possible when people work distributed in their homes.

Simplified Models to Determine Nodal Voltagesin Problems of Optimal Allocation of Capacitor Banks in Power Distribution Networks

This paper presents two simplified models to determine nodal voltages in power distribution networks. These models allow estimating the impact of the installation of reactive power compensations equipments like fixed or switched capacitor banks. The procedure used to develop the models is similar to the procedure used to develop linear power flow models of transmission lines, which have been widely used in optimization problems of operation planning and system expansion. The steady state non-linear load flow equations are approximated by linear equations relating the voltage amplitude and currents. The approximations of the linear equations are based on the high relationship between line resistance and line reactance (ratio R/X), which is valid for power distribution networks. The performance and accuracy of the models are evaluated through comparisons with the exact results obtained from the solution of the load flow using two test networks: a hypothetical network with 23 nodes and a real network with 217 nodes.

Performance of Histogram-Based Skin Colour Segmentation for Arms Detection in Human Motion Analysis Application

Arms detection is one of the fundamental problems in human motion analysis application. The arms are considered as the most challenging body part to be detected since its pose and speed varies in image sequences. Moreover, the arms are usually occluded with other body parts such as the head and torso. In this paper, histogram-based skin colour segmentation is proposed to detect the arms in image sequences. Six different colour spaces namely RGB, rgb, HSI, TSL, SCT and CIELAB are evaluated to determine the best colour space for this segmentation procedure. The evaluation is divided into three categories, which are single colour component, colour without luminance and colour with luminance. The performance is measured using True Positive (TP) and True Negative (TN) on 250 images with manual ground truth. The best colour is selected based on the highest TN value followed by the highest TP value.

High Performance Computing Using Out-of- Core Sparse Direct Solvers

In-core memory requirement is a bottleneck in solving large three dimensional Navier-Stokes finite element problem formulations using sparse direct solvers. Out-of-core solution strategy is a viable alternative to reduce the in-core memory requirements while solving large scale problems. This study evaluates the performance of various out-of-core sequential solvers based on multifrontal or supernodal techniques in the context of finite element formulations for three dimensional problems on a Windows platform. Here three different solvers, HSL_MA78, MUMPS and PARDISO are compared. The performance of these solvers is evaluated on a 64-bit machine with 16GB RAM for finite element formulation of flow through a rectangular channel. It is observed that using out-of-core PARDISO solver, relatively large problems can be solved. The implementation of Newton and modified Newton's iteration is also discussed.

European Ecological Network Natura 2000 - Opportunities and Threats

The research objective of the project and article “European Ecological Network Natura 2000 – opportunities and threats” Natura 2000 sites constitute a form of environmental protection, several legal problems are likely to result. Most controversially, certain sites will be subject to two regimes of protection: as national parks and as Natura 2000 sites. This dualism of the legal regulation makes it difficult to perform certain legal obligations related to the regimes envisaged under each form of environmental protection. Which regime and which obligations resulting from the particular form of environmental protection have priority and should prevail? What should be done if these obligations are contradictory? Furthermore, an institutional problem consists in that no public administration authority has the power to resolve legal conflicts concerning the application of a particular regime on a given site. There are also no criteria to decide priority and superiority of one form of environmental protection over the other. Which regulations are more important, those that pertain to national parks or to Natura 2000 sites? In the light of the current regulations, it is impossible to give a decisive answer to these questions. The internal hierarchy of forms of environmental protection has not been determined, and all such forms should be treated equally.

Genetic Algorithms with Oracle for the Traveling Salesman Problem

By introducing the concept of Oracle we propose an approach for improving the performance of genetic algorithms for large-scale asymmetric Traveling Salesman Problems. The results have shown that the proposed approach allows overcoming some traditional problems for creating efficient genetic algorithms.

Linear Elasticity Problems Solved by Using the Fictitious Domain Method and Total - FETI Domain Decomposition

The main goal of this paper is to show a possibility, how to solve numerically elliptic boundary value problems arising in 2D linear elasticity by using the fictitious domain method (FDM) and the Total-FETI domain decomposition method. We briefly mention the theoretical background of these methods and demonstrate their performance on a benchmark.