Use of Heliox during Spontaneous Ventilation: Model Study

The study deals with the modelling of the gas flow during heliox therapy. A special model has been developed to study the effect of the helium upon the gas flow in the airways during the spontaneous breathing. Lower density of helium compared with air decreases the Reynolds number and it allows improving the flow during the spontaneous breathing. In the cases, where the flow becomes turbulent while the patient inspires air the flow is still laminar when the patient inspires heliox. The use of heliox decreases the work of breathing and improves ventilation. It allows in some cases to prevent the intubation of the patients.

64 bit Computer Architectures for Space Applications – A study

The more recent satellite projects/programs makes extensive usage of real – time embedded systems. 16 bit processors which meet the Mil-Std-1750 standard architecture have been used in on-board systems. Most of the Space Applications have been written in ADA. From a futuristic point of view, 32 bit/ 64 bit processors are needed in the area of spacecraft computing and therefore an effort is desirable in the study and survey of 64 bit architectures for space applications. This will also result in significant technology development in terms of VLSI and software tools for ADA (as the legacy code is in ADA). There are several basic requirements for a special processor for this purpose. They include Radiation Hardened (RadHard) devices, very low power dissipation, compatibility with existing operational systems, scalable architectures for higher computational needs, reliability, higher memory and I/O bandwidth, predictability, realtime operating system and manufacturability of such processors. Further on, these may include selection of FPGA devices, selection of EDA tool chains, design flow, partitioning of the design, pin count, performance evaluation, timing analysis etc. This project deals with a brief study of 32 and 64 bit processors readily available in the market and designing/ fabricating a 64 bit RISC processor named RISC MicroProcessor with added functionalities of an extended double precision floating point unit and a 32 bit signal processing unit acting as co-processors. In this paper, we emphasize the ease and importance of using Open Core (OpenSparc T1 Verilog RTL) and Open “Source" EDA tools such as Icarus to develop FPGA based prototypes quickly. Commercial tools such as Xilinx ISE for Synthesis are also used when appropriate.

Direct Democracy and Social Contract in Ancient Athens

In the present essay, a model of choice by actors is analysedby utilizing the theory of chaos to explain how change comes about. Then, by using ancient and modern sources of literature, the theory of the social contract is analysed as a historical phenomenon that first appeared during the period of Classical Greece. Then, based on the findings of this analysis, the practice of direct democracy and public choice in ancient Athens is analysed, through two historical cases: Eubulus and Lycurgus political program in the second half of the 4th century. The main finding of this research is that these policies can be interpreted as an implementation of a social contract, through which citizens were taking decisions based on rational choice according to economic considerations.

A Linearization and Decomposition Based Approach to Minimize the Non-Productive Time in Transfer Lines

We address the balancing problem of transfer lines in this paper to find the optimal line balancing that minimizes the nonproductive time. We focus on the tool change time and face orientation change time both of which influence the makespane. We consider machine capacity limitations and technological constraints associated with the manufacturing process of auto cylinder heads. The problem is represented by a mixed integer programming model that aims at distributing the design features to workstations and sequencing the machining processes at a minimum non-productive time. The proposed model is solved by an algorithm established using linearization schemes and Benders- decomposition approach. The experiments show the efficiency of the algorithm in reaching the exact solution of small and medium problem instances at reasonable time.

Iterative Joint Power Control and Partial Crosstalk Cancellation in Upstream VDSL

Crosstalk is the major limiting issue in very high bit-rate digital subscriber line (VDSL) systems in terms of bit-rate or service coverage. At the central office side, joint signal processing accompanied by appropriate power allocation enables complex multiuser processors to provide near capacity rates. Unfortunately complexity grows with the square of the number of lines within a binder, so by taking into account that there are only a few dominant crosstalkers who contribute to main part of crosstalk power, the canceller structure can be simplified which resulted in a much lower run-time complexity. In this paper, a multiuser power control scheme, namely iterative waterfilling, is combined with previously proposed partial crosstalk cancellation approaches to demonstrate the best ever achieved performance which is verified by simulation results.

A Distributed Algorithm for Intrinsic Cluster Detection over Large Spatial Data

Clustering algorithms help to understand the hidden information present in datasets. A dataset may contain intrinsic and nested clusters, the detection of which is of utmost importance. This paper presents a Distributed Grid-based Density Clustering algorithm capable of identifying arbitrary shaped embedded clusters as well as multi-density clusters over large spatial datasets. For handling massive datasets, we implemented our method using a 'sharednothing' architecture where multiple computers are interconnected over a network. Experimental results are reported to establish the superiority of the technique in terms of scale-up, speedup as well as cluster quality.

Comparing Transformational Leadership in Successful and Unsuccessful Companies

In this article, while it is attempted to describe the problem and its importance, transformational leadership is studied by considering leadership theories. Issues such as the definition of transformational leadership and its aspects are compared on the basis of the ideas of various connoisseurs and then it (transformational leadership) is examined in successful and unsuccessful companies. According to the methodology, the method of research, hypotheses, population and statistical sample are investigated and research findings are analyzed by using descriptive and inferential statistical methods in the framework of analytical tables. Finally, our conclusion is provided by considering the results of statistical tests. The final result shows that transformational leadership is significantly higher in successful companies than unsuccessful ones P

An Agent Oriented Approach to Operational Profile Management

Software reliability, defined as the probability of a software system or application functioning without failure or errors over a defined period of time, has been an important area of research for over three decades. Several research efforts aimed at developing models to improve reliability are currently underway. One of the most popular approaches to software reliability adopted by some of these research efforts involves the use of operational profiles to predict how software applications will be used. Operational profiles are a quantification of usage patterns for a software application. The research presented in this paper investigates an innovative multiagent framework for automatic creation and management of operational profiles for generic distributed systems after their release into the market. The architecture of the proposed Operational Profile MAS (Multi-Agent System) is presented along with detailed descriptions of the various models arrived at following the analysis and design phases of the proposed system. The operational profile in this paper is extended to comprise seven different profiles. Further, the criticality of operations is defined using a new composed metrics in order to organize the testing process as well as to decrease the time and cost involved in this process. A prototype implementation of the proposed MAS is included as proof-of-concept and the framework is considered as a step towards making distributed systems intelligent and self-managing.

Surrogate based Evolutionary Algorithm for Design Optimization

Optimization is often a critical issue for most system design problems. Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, finding optimal solution to complex high dimensional, multimodal problems often require highly computationally expensive function evaluations and hence are practically prohibitive. The Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model presented in our earlier work [14] reduced computation time by controlled use of meta-models to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the meta-model are generated from a single uniform model. Situations like model formation involving variable input dimensions and noisy data certainly can not be covered by this assumption. In this paper we present an enhanced version of DAFHEA that incorporates a multiple-model based learning approach for the SVM approximator. DAFHEA-II (the enhanced version of the DAFHEA framework) also overcomes the high computational expense involved with additional clustering requirements of the original DAFHEA framework. The proposed framework has been tested on several benchmark functions and the empirical results illustrate the advantages of the proposed technique.

Shadow Detection for Increased Accuracy of Privacy Enhancing Methods in Video Surveillance Edge Devices

Shadow detection is still considered as one of the potential challenges for intelligent automated video surveillance systems. A pre requisite for reliable and accurate detection and tracking is the correct shadow detection and classification. In such a landscape of conditions, privacy issues add more and more complexity and require reliable shadow detection. In this work the intertwining between security, accuracy, reliability and privacy is analyzed and, accordingly, a novel architecture for Privacy Enhancing Video Surveillance (PEVS) is introduced. Shadow detection and masking are dealt with through the combination of two different approaches simultaneously. This results in a unique privacy enhancement, without affecting security. Subsequently, the methodology was employed successfully in a large-scale wireless video surveillance system; privacy relevant information was stored and encrypted on the unit, without transferring it over an un-trusted network.

Hybrid Intelligent Intrusion Detection System

Intrusion Detection Systems are increasingly a key part of systems defense. Various approaches to Intrusion Detection are currently being used, but they are relatively ineffective. Artificial Intelligence plays a driving role in security services. This paper proposes a dynamic model Intelligent Intrusion Detection System, based on specific AI approach for intrusion detection. The techniques that are being investigated includes neural networks and fuzzy logic with network profiling, that uses simple data mining techniques to process the network data. The proposed system is a hybrid system that combines anomaly, misuse and host based detection. Simple Fuzzy rules allow us to construct if-then rules that reflect common ways of describing security attacks. For host based intrusion detection we use neural-networks along with self organizing maps. Suspicious intrusions can be traced back to its original source path and any traffic from that particular source will be redirected back to them in future. Both network traffic and system audit data are used as inputs for both.

Endeavoring Innovation via Research and Development Management: A Case of Iranian Industrial Sector

This study aims at investigating factors in research and development (R&D) growth and exploring the role of R&D management in enhancing social innovation and productivity improvement in Iran-s industrial sector. It basically explores the common types of R&D activities and the industries which benefited the most from active R&D units in Iran. The researchers generated qualitative analyses obtained from primary and secondary data. The primary data have been retrieved through interviews with five key players (Managing Director, Internal Manager, General Manager, Executive Manager, and Project Manager) in the industrial sector. The secondary data acquired from an investigation on Mazandaran, a province of northern Iran. The findings highlight Iran-s focuses of R & D on cost reduction and upgrading productivity. Industries that have benefited the most from active R&D units are metallic, machinery and equipment design, and automotive. We rank order the primary effects of R&D on productivity improvement as follows, industry improvement, economic growth, using professional human resources, generating productivity and creativity culture, creating a competitive and innovative environment, and increasing people-s knowledge. Generally, low budget dedication and insufficient supply of highly skilled scientists and engineers are two important obstacles for R&D in Iran. Whereas, R&D has resulted in improvement in Iranian society, transfer of contemporary knowledge into the international market is still lacking.

Dosimetric Comparison of aSi1000 EPID and ImatriXX 2-D Array System for Volumetric Modulated Arc and Intensity Modulated Radiotherapy Patient Specific Quality Assurance

Prior to the use of detectors, characteristics comparison study was performed and baseline established. In patient specific QA, the portal dosimetry mean values of area gamma, average gamma and maximum gamma were 1.02, 0.31 and 1.31 with standard deviation of 0.33, 0.03 and 0.14 for IMRT and the corresponding values were 1.58, 0.48 and 1.73 with standard deviation of 0.31, 0.06 and 0.66 for VMAT. With ImatriXX 2-D array system, on an average 99.35% of the pixels passed the criteria of 3%-3 mm gamma with standard deviation of 0.24 for dynamic IMRT. For VMAT, the average value was 98.16% with a standard deviation of 0.86. The results showed that both the systems can be used in patient specific QA measurements for IMRT and VMAT. The values obtained with the portal dosimetry system were found to be relatively more consistent compared to those obtained with ImatriXX 2-D array system.

Histological Study of Postmortem Juvenile Green Sea Turtle (Chelonia mydas) from Royal Thai Navy Sea Turtle Nursery, Phang-nga, Thailand

The problem on the conservation programme of the Royal Thai Navy Sea Turtle Nursery, Phang-nga Province, Thailand is high mortality rate of juvenile green sea turtle (Cheloniamydas) on nursing period. So, during May to October 2012, postmortem examinations of juvenile green sea turtle were performed to determine the causes of dead. Fresh tissues of postmortem of 15 juvenile green sea turtles (1-3 months old) were investigated using paraffin section technique. The results showed normal ultrastructure of all tissue organs. These instances reviewed the health and stability of the environments in which juvenile green sea turtles live and concern for the survival rate. The present article also provides guidance for a review of the biology, guidelines for appropriate postmortem tissue, normal histology and sampling collection and procedures. The data also provides information for conservation of this endangered species in term of acknowledging and encouraging people to protect the animals and their habitats in nature.

Supply Chain Model of Catfish Production and Trade in Yogyakarta, Indonesia

Currently, the demand for marine and fisheries commodity in Yogyakarta, Indonesia continues to increase. The existing condition shows that the aquaculture supply cannot be supplied by Yogyakarta region itself, but still need to be supported by regions outside Yogyakarta. The effort to optimize the market is initiated by reviewing and designing the supply chain of production and trade of aquaculture commodity in order to create the implementation of aquaculture production and trade commodity optimally. This formulated supply chain model indicates 4 performance indicators of measurable success in terms of: (1) efficiency; (2) flexibility; (3) responsiveness; and (4) quality. These indicators had been exercised as the success benchmarks for priority marketing management in local level as well as national level. The result of this research indicates that if the catfish fishery system is managed as business as usual then the catfish demand in Yogyakarta region will experience to increase in the future. The increase of demand is inline with the increase of number of people in Yogyakarta and also the fluctuation of catfish consumption per capita. The highest production of catfish will experience in the third year approximately 30,118 tons. Other result of the research indicates that the catfish demand in Yogyakarta region cannot be supplied yet from the local region. Therefore, to fulfill the supply from outside Yogyakarta region, the local farmers should improve the supply through land extension. The fluctuation of commodity price will experience in the future annually and the catfish supply from outside Yogyakarta region will be lowering the price in the market.

Performance Analysis of Load Balancing Algorithms

Load balancing is the process of improving the performance of a parallel and distributed system through a redistribution of load among the processors [1] [5]. In this paper we present the performance analysis of various load balancing algorithms based on different parameters, considering two typical load balancing approaches static and dynamic. The analysis indicates that static and dynamic both types of algorithm can have advancements as well as weaknesses over each other. Deciding type of algorithm to be implemented will be based on type of parallel applications to solve. The main purpose of this paper is to help in design of new algorithms in future by studying the behavior of various existing algorithms.

Aroma Composition and Polyphenol Content of Ciders Available in Latvian Market

Aroma forming volatiles are important components of fermented beverages. The aim of current research is to evaluate the volatile compounds and phenolic compounds of commercial ciders. Volatile aroma compounds and TPC of seven commercial ciders were determined. Extraction of aroma compounds was performed using solid phase microextraction (DVB/Car/PDMS fibre). Analysis of volatile aroma compounds was made using a Perkin Elmer Clarus 500 GC/MS. Total phenol content (TPC) was determined according to the Folin-Ciocalteu spectrophotometric method and results were expressed as gallic acid equivalents. The highest volatile compounds were in apple ciders with pear flavor. The highest TPC and lower content of volatile compounds were detected in French ciders.

Evaluation of Groundwater Trend of Arsanjan Plain

Groundwater resources in Arsanjan plain provide water for agriculture, industry, and human consumption. Continued agricultural development in this area needs to additional groundwater resources for, particularly during of drought periods, and effects on the quantity and quality of ground water available. The purpose of this study is to evaluate water level changes in the aquifer of Arsanjan plain in the Fars province in order to determine the areas of greatest depletion and the causes of depletion. In this plain, farmers and other users are pumping groundwater faster than its natural replenishment rate, causing a continuous drop in groundwater tables and depletion of this resource. In this research variation of groundwater level, their effects and ways to help control groundwater levels in aquifer of the Arsanjan plains were evaluated .Excessive exploitation of groundwater in this aquifer caused the groundwater levels fall too fast or to unacceptable levels. The average drawdown of the groundwater level in this plain were 19.66 meters during 1996 to 2003.

Direct Measurement of Electromagnetic Thrust of Electrodeless Helicon Plasma Thruster Using Magnetic Nozzle

In order to realize long-lived electric propulsion systems, we have been investigating an electrodeless plasma thruster. In our concept, a helicon plasma is accelerated by the magnetic nozzle for the thrusts production. In addition, the electromagnetic thrust can be enhanced by the additional radio-frequency rotating electric field (REF) power in the magnetic nozzle. In this study, a direct measurement of the electromagnetic thrust and a probe measurement have been conducted using a laboratory model of the thruster under the condition without the REF power input. Fromthrust measurement, it is shown that the thruster produces a sub-milli-newton order electromagnetic thrust force without the additional REF power. The thrust force and the density jump are observed due to the discharge mode transition from the inductive coupled plasma to the helicon wave excited plasma. The thermal thrust is theoretically estimated, and the total thrust force, which is a sum of the electromagnetic and the thermal thrust force and specific impulse are calculated to be up to 650 μN (plasma production power of 400 W, Ar gas mass flow rate of 1.0 mg/s) and 210 s (plasma production power of 400 W, Ar gas mass flow rate of 0.2 mg/s), respectively.

Analysis of Event-related Response in Human Visual Cortex with fMRI

Functional Magnetic Resonance Imaging(fMRI) is a noninvasive imaging technique that measures the hemodynamic response related to neural activity in the human brain. Event-related functional magnetic resonance imaging (efMRI) is a form of functional Magnetic Resonance Imaging (fMRI) in which a series of fMRI images are time-locked to a stimulus presentation and averaged together over many trials. Again an event related potential (ERP) is a measured brain response that is directly the result of a thought or perception. Here the neuronal response of human visual cortex in normal healthy patients have been studied. The patients were asked to perform a visual three choice reaction task; from the relative response of each patient corresponding neuronal activity in visual cortex was imaged. The average number of neurons in the adult human primary visual cortex, in each hemisphere has been estimated at around 140 million. Statistical analysis of this experiment was done with SPM5(Statistical Parametric Mapping version 5) software. The result shows a robust design of imaging the neuronal activity of human visual cortex.