IEEE 802.11 b and g WLAN Propagation Model using Power Density Measurements at ESPOL

This paper describes the development of a WLAN propagation model, using Spectral Analyzer measurements. The signal is generated by two Access Points (APs) on the base floor at the administrative Communication School of ESPOL building. In general, users do not have a Q&S reference about a wireless network; however, this depends on the level signal as a function of frequency, distance and other path conditions between receiver and transmitter. Then, power density of the signal decrease as it propagates through space and data transfer rate is affected. This document evaluates and implements empirical mathematical formulation for the characterization of WLAN radio wave propagation on two aisles of the building base floor.

Equilibrium Modeling of Carbon Dioxide Adsorption on Zeolites

High pressure adsorption of carbon dioxide on zeolite 13X was investigated in the pressure range (0 to 4) Mpa and temperatures 298, 308 and 323K. The data fitting is accomplished with the Toth, UNILAN, Dubinin-Astakhov and virial adsorption models which are generally used for micro porous adsorbents such as zeolites. Comparison with experimental data from the literature indicated that the virial model would best determine results. These results may be partly attributed to the flexibility of the virial model which can accommodate as many constants as the data warrants.

Swarm Navigation in a Complex Environment

This paper proposes a solution to the motion planning and control problem of car-like mobile robots which is required to move safely to a designated target in a priori known workspace cluttered with swarm of boids exhibiting collective emergent behaviors. A generalized algorithm for target convergence and swarm avoidance is proposed that will work for any number of swarms. The control laws proposed in this paper also ensures practical stability of the system. The effectiveness of the proposed control laws are demonstrated via computer simulations of an emergent behavior.

A Logic Based Framework for Planning for Mobile Agents

The objective of the paper is twofold. First, to develop a formal framework for planning for mobile agents. A logical language based on a temporal logic is proposed that can express a type of tasks which often arise in network management. Second, to design a planning algorithm for such tasks. The aim of this paper is to study the importance of finding plans for mobile agents. Although there has been a lot of research in mobile agents, not much work has been done to incorporate planning ideas for such agents. This paper makes an attempt in this direction. A theoretical study of finding plans for mobile agents is undertaken. A planning algorithm (based on the paradigm of mobile computing) is proposed and its space, time, and communication complexity is analyzed. The algorithm is illustrated by working out an example in detail.

Identification of Nonlinear Predictor and Simulator Models of a Cement Rotary Kiln by Locally Linear Neuro-Fuzzy Technique

One of the most important parts of a cement factory is the cement rotary kiln which plays a key role in quality and quantity of produced cement. In this part, the physical exertion and bilateral movement of air and materials, together with chemical reactions take place. Thus, this system has immensely complex and nonlinear dynamic equations. These equations have not worked out yet. Only in exceptional case; however, a large number of the involved parameter were crossed out and an approximation model was presented instead. This issue caused many problems for designing a cement rotary kiln controller. In this paper, we presented nonlinear predictor and simulator models for a real cement rotary kiln by using nonlinear identification technique on the Locally Linear Neuro- Fuzzy (LLNF) model. For the first time, a simulator model as well as a predictor one with a precise fifteen minute prediction horizon for a cement rotary kiln is presented. These models are trained by LOLIMOT algorithm which is an incremental tree-structure algorithm. At the end, the characteristics of these models are expressed. Furthermore, we presented the pros and cons of these models. The data collected from White Saveh Cement Company is used for modeling.

Physiological and Pathology Demographics of Veteran Rugby Athletes: Golden Oldies Rugby Festival

Recently, the health of retired National Football League players, particularly lineman has been investigated. A number of studies have reported increased cardiometabolic risk, premature ardiovascular disease and incidence of type 2 diabetes. Rugby union players have somatotypes very similar to National Football league players which suggest that rugby players may have similar health risks. The International Golden Oldies World Rugby Festival (GORF) provided a unique opportunity to investigate the demographics of veteran rugby players. METHODOLOGIES: A cross-sectional, observational study was completed using an online web-based questionnaire that consisted of medical history and physiological measures. Data analysis was completed using a one sample t-test (50yrs) and Chi-square test. RESULTS: A total of 216 veteran rugby competitors (response rate = 6.8%) representing 10 countries, aged 35-72 yrs (mean 51.2, S.D. ±8.0), participated in the online survey. As a group, the incidence of current smokers was low at 8.8% (avg 72.4 cigs/wk) whilst the percentage consuming alcohol was high (93.1% (avg 11.2 drinks/wk). Competitors reported the following top six chronic diseases/disorders; hypertension (18.6%), arthritis (OA/RA, 11.5%), asthma (9.3%), hyperlipidemia (8.2%), diabetes (all types, 7.5%) and gout (6%), there were significant differences between groups with regard to cancer (all types) and migraines. When compared to the Australian general population (Australian Bureau of Statistics data, n=18,000), GORF competitors had a Climstein Mike, Walsh Joe (corresponding author) and Burke Stephen School of Exercise Science, Australian Catholic University, 25A Barker Road, Strathfield, Sydney, NSW, 2016, Australia (e-mail: [email protected], [email protected], [email protected]). John Best is with Orthosports, 160 Belmore Rd., Randwick, Sydney,NSW 2031, Australia (e-mail: [email protected]). Heazlewood, Ian Timothy is with School of Environmental and Life Sciences, Faculty Education, Health and Science, Charles Darwin University, Precinct Yellow Building 2, Charles Darwin University, NT 0909, Australia (e-mail: [email protected]). Kettunen Jyrki Arcada University of Applied Sciences, Jan-Magnus Janssonin aukio 1, FI-00550, Helsinki, Finland (e-mail: [email protected]). Adams Kent is with California State University Monterey Bay, Kinesiology Department, 100 Campus Center, Seaside, CA., 93955, USA (email: [email protected]). DeBeliso Mark is with Department of Physical Education and Human Performance, Southern Utah University, 351 West University Blvd, Cedar City, Utah, USA (e-mail: [email protected]). significantly lower incidence of anxiety (p

Iterative Process to Improve Simple Adaptive Subdivision Surfaces Method with Butterfly Scheme

Subdivision surfaces were applied to the entire meshes in order to produce smooth surfaces refinement from coarse mesh. Several schemes had been introduced in this area to provide a set of rules to converge smooth surfaces. However, to compute and render all the vertices are really inconvenient in terms of memory consumption and runtime during the subdivision process. It will lead to a heavy computational load especially at a higher level of subdivision. Adaptive subdivision is a method that subdivides only at certain areas of the meshes while the rest were maintained less polygons. Although adaptive subdivision occurs at the selected areas, the quality of produced surfaces which is their smoothness can be preserved similar as well as regular subdivision. Nevertheless, adaptive subdivision process burdened from two causes; calculations need to be done to define areas that are required to be subdivided and to remove cracks created from the subdivision depth difference between the selected and unselected areas. Unfortunately, the result of adaptive subdivision when it reaches to the higher level of subdivision, it still brings the problem with memory consumption. This research brings to iterative process of adaptive subdivision to improve the previous adaptive method that will reduce memory consumption applied on triangular mesh. The result of this iterative process was acceptable better in memory and appearance in order to produce fewer polygons while it preserves smooth surfaces.

Optimal One Bit Time Reversal For UWB Impulse Radio In Multi-User Wireless Communications

In this paper, with the purpose of further reducing the complexity of the system, while keeping its temporal and spatial focusing performance, we investigate the possibility of using optimal one bit time reversal (TR) system for impulse radio ultra wideband multi-user wireless communications. The results show that, by optimally selecting the number of used taps in the pre-filter the optimal one bit TR system can outperform the full one bit TR system. In some cases, the temporal and spatial focusing performance of the optimal one bit TR system appears to be compatible with that of the original TR system. This is a significant result as the overhead cost is much lower than it is required in the original TR system.

Ornament as a Universal Language of Peace (Based On Comparative Analysis of Cultures of proto-Turkic Peoples and Indian Tribes of North America)

In this article, the authors reviewed and analyzed the survey materials similarities ornament proto-Turkic and northern Indians. The study examined the materials scientists - geneticists, archaeologists, anthropologists. Numerous studies of scientists from different directions once again prove the relevance of the topic. The authors approached the subject from an artistic side. The study authors have made the appropriate conclusions. This publication is based on the proceedings of the investigation.

Identification of MIMO Systems Using Neuro-Fuzzy Models with a Shuffled Frog Leaping Algorithm

In this paper, a TSK-type Neuro-fuzzy Inference System that combines the features of fuzzy sets and neural networks has been applied for the identification of MIMO systems. The procedure of adapting parameters in TSK model employs a Shuffled Frog Leaping Algorithm (SFLA) which is inspired from the memetic evolution of a group of frogs when seeking for food. To demonstrate the accuracy and effectiveness of the proposed controller, two nonlinear systems have been considered as the MIMO plant, and results have been compared with other learning methods based on Particle Swarm Optimization algorithm (PSO) and Genetic Algorithm (GA).

Density Clustering Based On Radius of Data (DCBRD)

Clustering algorithms are attractive for the task of class identification in spatial databases. However, the application to large spatial databases rises the following requirements for clustering algorithms: minimal requirements of domain knowledge to determine the input parameters, discovery of clusters with arbitrary shape and good efficiency on large databases. The well-known clustering algorithms offer no solution to the combination of these requirements. In this paper, a density based clustering algorithm (DCBRD) is presented, relying on a knowledge acquired from the data by dividing the data space into overlapped regions. The proposed algorithm discovers arbitrary shaped clusters, requires no input parameters and uses the same definitions of DBSCAN algorithm. We performed an experimental evaluation of the effectiveness and efficiency of it, and compared this results with that of DBSCAN. The results of our experiments demonstrate that the proposed algorithm is significantly efficient in discovering clusters of arbitrary shape and size.

Deployment of Service Quality Characteristics

This work discusses an innovative methodology for deployment of service quality characteristics. Four groups of organizational features that may influence the quality of services are identified: human resource, technology, planning, and organizational relationships. A House of Service Quality (HOSQ) matrix is built to extract the desired improvement in the service quality characteristics and to translate them into a hierarchy of important organizational features. The Mean Square Error (MSE) criterion enables the pinpointing of the few essential service quality characteristics to be improved as well as selection of the vital organizational features. The method was implemented in an engineering supply enterprise and provides useful information on its vital service dimensions.

A Study of Computational Organizational Narrative Generation for Decision Support

Narratives are invaluable assets of human lives. Due to the distinct features of narratives, they are useful for supporting human reasoning processes. However, many useful narratives become residuals in organizations or human minds nowadays. Researchers have contributed effort to investigate and improve narrative generation processes. This paper attempts to contemplate essential components in narratives and explore a computational approach to acquire and extract knowledge to generate narratives. The methodology and significant benefit for decision support are presented.

Proposing Enterprise Wide Information Systems Business Performance Model

Enterprise Wide Information Systems (EWIS) implementation involves the entire business and will require changes throughout the firm. Because of the scope, complexity and continuous nature of ERP, the project-based approach to managing the implementation process resulted in failure rates of between 60% and 80%. In recent years ERP systems have received much attention. The organizational relevance and risk of ERP projects make it important for organizations to focus on ways to make ERP implementation successful. Once these systems are in place, however, their performance depends on the identified macro variables viz. 'Business Process', 'Decision Making' and 'Individual / Group working'. The questionnaire was designed and administered. The responses from 92 organizations were compiled. The relationship of these variables with EWIS performance is analyzed using inferential statistical measurements. The study helps to understand the performance of model presented. The study suggested in keeping away from the calamities and thereby giving the necessary competitive edge. Whenever some discrepancy is identified during the process of performance appraisal care has to be taken to draft necessary preventive measures. If all these measures are taken care off then the EWIS performance will definitely deliver the results.

Design and Development of Ferroelectric Material for Microstrip Patch Array Antenna

This paper presents the utilizing of ferroelectric material on antenna application. There are two different ferroelectric had been used on the proposed antennas which include of Barium Strontium Titanate (BST) and Bismuth Titanate (BiT), suitable for Access Points operating in the WLAN IEEE 802.11 b/g and WiMAX IEEE 802.16 within the range of 2.3 GHz to 2.5 GHz application. BST, which had been tested to own a dielectric constant of εr = 15 while BiT has a dielectric constant that higher than BST which is εr = 21 and both materials are in rectangular shaped. The influence of various parameters on antenna characteristics were investigated extensively using commercial electromagnetic simulations software by Communication Simulation Technology (CST). From theoretical analysis and simulation results, it was demonstrated that ferroelectric material used have not only improved the directive emission but also enhanced the radiation efficiency.

Standard Deviation of Mean and Variance of Rows and Columns of Images for CBIR

This paper describes a novel and effective approach to content-based image retrieval (CBIR) that represents each image in the database by a vector of feature values called “Standard deviation of mean vectors of color distribution of rows and columns of images for CBIR". In many areas of commerce, government, academia, and hospitals, large collections of digital images are being created. This paper describes the approach that uses contents as feature vector for retrieval of similar images. There are several classes of features that are used to specify queries: colour, texture, shape, spatial layout. Colour features are often easily obtained directly from the pixel intensities. In this paper feature extraction is done for the texture descriptor that is 'variance' and 'Variance of Variances'. First standard deviation of each row and column mean is calculated for R, G, and B planes. These six values are obtained for one image which acts as a feature vector. Secondly we calculate variance of the row and column of R, G and B planes of an image. Then six standard deviations of these variance sequences are calculated to form a feature vector of dimension six. We applied our approach to a database of 300 BMP images. We have determined the capability of automatic indexing by analyzing image content: color and texture as features and by applying a similarity measure Euclidean distance.

Adaptive Anisotropic Diffusion for Ultrasonic Image Denoising and Edge Enhancement

Utilizing echoic intension and distribution from different organs and local details of human body, ultrasonic image can catch important medical pathological changes, which unfortunately may be affected by ultrasonic speckle noise. A feature preserving ultrasonic image denoising and edge enhancement scheme is put forth, which includes two terms: anisotropic diffusion and edge enhancement, controlled by the optimum smoothing time. In this scheme, the anisotropic diffusion is governed by the local coordinate transformation and the first and the second order normal derivatives of the image, while the edge enhancement is done by the hyperbolic tangent function. Experiments on real ultrasonic images indicate effective preservation of edges, local details and ultrasonic echoic bright strips on denoising by our scheme.

Implementation of Adder-Subtracter Design with VerilogHDL

According to the density of the chips, designers are trying to put so any facilities of computational and storage on single chips. Along with the complexity of computational and storage circuits, the designing, testing and debugging become more and more complex and expensive. So, hardware design will be built by using very high speed hardware description language, which is more efficient and cost effective. This paper will focus on the implementation of 32-bit ALU design based on Verilog hardware description language. Adder and subtracter operate correctly on both unsigned and positive numbers. In ALU, addition takes most of the time if it uses the ripple-carry adder. The general strategy for designing fast adders is to reduce the time required to form carry signals. Adders that use this principle are called carry look- ahead adder. The carry look-ahead adder is to be designed with combination of 4-bit adders. The syntax of Verilog HDL is similar to the C programming language. This paper proposes a unified approach to ALU design in which both simulation and formal verification can co-exist.

Analysis and Research of Two-Level Scheduling Profile for Open Real-Time System

In an open real-time system environment, the coexistence of different kinds of real-time and non real-time applications makes the system scheduling mechanism face new requirements and challenges. One two-level scheduling scheme of the open real-time systems is introduced, and points out that hard and soft real-time applications are scheduled non-distinctively as the same type real-time applications, the Quality of Service (QoS) cannot be guaranteed. It has two flaws: The first, it can not differentiate scheduling priorities of hard and soft real-time applications, that is to say, it neglects characteristic differences between hard real-time applications and soft ones, so it does not suit a more complex real-time environment. The second, the worst case execution time of soft real-time applications cannot be predicted exactly, so it is not worth while to cost much spending in order to assure all soft real-time applications not to miss their deadlines, and doing that may cause resource wasting. In order to solve this problem, a novel two-level real-time scheduling mechanism (including scheduling profile and scheduling algorithm) which adds the process of dealing with soft real-time applications is proposed. Finally, we verify real-time scheduling mechanism from two aspects of theory and experiment. The results indicate that our scheduling mechanism can achieve the following objectives. (1) It can reflect the difference of priority when scheduling hard and soft real-time applications. (2) It can ensure schedulability of hard real-time applications, that is, their rate of missing deadline is 0. (3) The overall rate of missing deadline of soft real-time applications can be less than 1. (4) The deadline of a non-real-time application is not set, whereas the scheduling algorithm that server 0 S uses can avoid the “starvation" of jobs and increase QOS. By doing that, our scheduling mechanism is more compatible with different types of applications and it will be applied more widely.

Evaluation of Optimum Performance of Lateral Intakes

In designing river intakes and diversion structures, it is paramount that the sediments entering the intake are minimized or, if possible, completely separated. Due to high water velocity, sediments can significantly damage hydraulic structures especially when mechanical equipment like pumps and turbines are used. This subsequently results in wasting water, electricity and further costs. Therefore, it is prudent to investigate and analyze the performance of lateral intakes affected by sediment control structures. Laboratory experiments, despite their vast potential and benefits, can face certain limitations and challenges. Some of these include: limitations in equipment and facilities, space constraints, equipment errors including lack of adequate precision or mal-operation, and finally, human error. Research has shown that in order to achieve the ultimate goal of intake structure design – which is to design longlasting and proficient structures – the best combination of sediment control structures (such as sill and submerged vanes) along with parameters that increase their performance (such as diversion angle and location) should be determined. Cost, difficulty of execution and environmental impacts should also be included in evaluating the optimal design. This solution can then be applied to similar problems in the future. Subsequently, the model used to arrive at the optimal design requires high level of accuracy and precision in order to avoid improper design and execution of projects. Process of creating and executing the design should be as comprehensive and applicable as possible. Therefore, it is important that influential parameters and vital criteria is fully understood and applied at all stages of choosing the optimal design. In this article, influential parameters on optimal performance of the intake, advantages and disadvantages, and efficiency of a given design are studied. Then, a multi-criterion decision matrix is utilized to choose the optimal model that can be used to determine the proper parameters in constructing the intake.