Design and Implementation of an Intelligent System for Detection of Hazardous Gases using PbPc Sensor Array

The voltage/current characteristics and the effect of NO2 gas on the electrical conductivity of a PbPc gas sensor array is investigated. The gas sensor is manufactured using vacuum deposition of gold electrodes on sapphire substrate with the leadphathalocyanine vacuum sublimed on the top of the gold electrodes. Two versions of the PbPc gas sensor array are investigated. The tested types differ in the gap sizes between the deposited gold electrodes. The sensors are tested at different temperatures to account for conductivity changes as the molecular adsorption/desorption rate is affected by heat. The obtained results found to be encouraging as the sensors shoed stability and sensitivity towards low concentration of applied NO2 gas.

DAMQ-Based Approach for Efficiently Using the Buffer Spaces of a NoC Router

In this paper we present high performance dynamically allocated multi-queue (DAMQ) buffer schemes for fault tolerance systems on chip applications that require an interconnection network. Two virtual channels shared the same buffer space. Fault tolerant mechanisms for interconnection networks are becoming a critical design issue for large massively parallel computers. It is also important to high performance SoCs as the system complexity keeps increasing rapidly. On the message switching layer, we make improvement to boost system performance when there are faults involved in the components communication. The proposed scheme is when a node or a physical channel is deemed as faulty, the previous hop node will terminate the buffer occupancy of messages destined to the failed link. The buffer usage decisions are made at switching layer without interactions with higher abstract layer, thus buffer space will be released to messages destined to other healthy nodes quickly. Therefore, the buffer space will be efficiently used in case fault occurs at some nodes.

Combining Molecular Statics with Heat Transfer Finite Difference Method for Analysis of Nanoscale Orthogonal Cutting of Single-Crystal Silicon Temperature Field

This paper uses quasi-steady molecular statics model and diamond tool to carry out simulation temperature rise of nanoscale orthogonal cutting single-crystal silicon. It further qualitatively analyzes temperature field of silicon workpiece without considering heat transfer and considering heat transfer. This paper supposes that the temperature rise of workpiece is mainly caused by two heat sources: plastic deformation heat and friction heat. Then, this paper develops a theoretical model about production of the plastic deformation heat and friction heat during nanoscale orthogonal cutting. After the increased temperature produced by these two heat sources are added up, the acquired total temperature rise at each atom of the workpiece is substituted in heat transfer finite difference equation to carry out heat transfer and calculates the temperature field in each step and makes related analysis.

The Effect of the Direct Contact Heat Exchanger on Steam Power Plant

An actual power plant, which is the power plant of Iron and Steel Factory at Misurata city in Libya , has been modeled using Matlab in order to compare its results to the actual results of the actual cycle. This paper concentrates on two factors: a- The comparison between exergy losses in the actual cycle and the modeled cycle. b- The effect of extracting pressure on temperature water at boiler inlet. Closed heat exchangers used in this plant have been substituted by open heat exchangers in the current study of the modeled power plant and the required changes in the pressure have been considered. In the following investigation the two points mentioned above are taken in consideration.

The Predictability and Abstractness of Language: A Study in Understanding and Usage of the English Language through Probabilistic Modeling and Frequency

Accounts of language acquisition differ significantly in their treatment of the role of prediction in language learning. In particular, nativist accounts posit that probabilistic learning about words and word sequences has little to do with how children come to use language. The accuracy of this claim was examined by testing whether distributional probabilities and frequency contributed to how well 3-4 year olds repeat simple word chunks. Corresponding chunks were the same length, expressed similar content, and were all grammatically acceptable, yet the results of the study showed marked differences in performance when overall distributional frequency varied. It was found that a distributional model of language predicted the empirical findings better than a number of other models, replicating earlier findings and showing that children attend to distributional probabilities in an adult corpus. This suggested that language is more prediction-and-error based, rather than on abstract rules which nativist camps suggest.

DACS3:Embedding Individual Ant Behavior in Ant Colony System

Ants are fascinating creatures that demonstrate the ability to find food and bring it back to their nest. Their ability as a colony, to find paths to food sources has inspired the development of algorithms known as Ant Colony Systems (ACS). The principle of cooperation forms the backbone of such algorithms, commonly used to find solutions to problems such as the Traveling Salesman Problem (TSP). Ants communicate to each other through chemical substances called pheromones. Modeling individual ants- ability to manipulate this substance can help an ACS find the best solution. This paper introduces a Dynamic Ant Colony System with threelevel updates (DACS3) that enhance an existing ACS. Experiments were conducted to observe single ant behavior in a colony of Malaysian House Red Ants. Such behavior was incorporated into the DACS3 algorithm. We benchmark the performance of DACS3 versus DACS on TSP instances ranging from 14 to 100 cities. The result shows that the DACS3 algorithm can achieve shorter distance in most cases and also performs considerably faster than DACS.

Cold Hardiness in Near Isogenic Lines of Bread Wheat (Triticum Aestivum L. em. Thell.)

Low temperature (LT) is one of the most abiotic stresses causing loss of yield in wheat (T. aestivum). Four major genes in wheat (Triticum aestivum L.) with the dominant alleles designated Vrn–A1,Vrn–B1,Vrn–D1 and Vrn4, are known to have large effects on the vernalization response, but the effects on cold hardiness are ambiguous. Poor cold tolerance has restricted winter wheat production in regions of high winter stress [9]. It was known that nearly all wheat chromosomes [5] or at least 10 chromosomes of 21 chromosome pairs are important in winter hardiness [15]. The objective of present study was to clarify the role of each chromosome in cold tolerance. With this purpose we used 20 isogenic lines of wheat. In each one of these isogenic lines only a chromosome from ‘Bezostaya’ variety (a winter habit cultivar) was substituted to ‘Capple desprez’ variety. The plant materials were planted in controlled conditions with 20º C and 16 h day length in moderately cold areas of Iran at Karaj Agricultural Research Station in 2006-07 and the acclimation period was completed for about 4 weeks in a cold room with 4º C. The cold hardiness of these isogenic lines was measured by LT50 (the temperature in which 50% of the plants are killed by freezing stress).The experimental design was completely randomized block design (RCBD)with three replicates. The results showed that chromosome 5A had a major effect on freezing tolerance, and then chromosomes 1A and 4A had less effect on this trait. Further studies are essential to understanding the importance of each chromosome in controlling cold hardiness in wheat.

The Problem of Power and Management in the Information Society

Modern civilization has come in recent decades into a new phase in its development, called the information society. The concept of "information society" has become one of the most common. Therefore, the attempt to understand what exactly the society we live in, what are its essential features, and possible future scenarios, is important to the social and philosophical analysis. At the heart of all these deep transformations is more increasing, almost defining role knowledge and information as play substrata of «information society». The mankind opened for itself and actively exploits a new resource – information. Information society puts forward on the arena new type of the power, at the heart of which activity – mastering by a new resource: information and knowledge. The password of the new power – intelligence as synthesis of knowledge, information and communications, the strength of mind, fundamental sociocultural values. In a postindustrial society, the power of knowledge and information is crucial in the management of the company, pushing into the background the influence of money and state coercion.

The Risk Assessment of Nano-particles and Investigation of Their Environmental Impact

Nanotechnology is the science of creating, using and manipulating objects which have at least one dimension in range of 0.1 to 100 nanometers. In other words, nanotechnology is reconstructing a substance using its individual atoms and arranging them in a way that is desirable for our purpose. The main reason that nanotechnology has been attracting attentions is the unique properties that objects show when they are formed at nano-scale. These differing characteristics that nano-scale materials show compared to their nature-existing form is both useful in creating high quality products and dangerous when being in contact with body or spread in environment. In order to control and lower the risk of such nano-scale particles, the main following three topics should be considered: 1) First of all, these materials would cause long term diseases that may show their effects on body years after being penetrated in human organs and since this science has become recently developed in industrial scale not enough information is available about their hazards on body. 2) The second is that these particles can easily spread out in environment and remain in air, soil or water for very long time, besides their high ability to penetrate body skin and causing new kinds of diseases. 3) The third one is that to protect body and environment against the danger of these particles, the protective barriers must be finer than these small objects and such defenses are hard to accomplish. This paper will review, discuss and assess the risks that human and environment face as this new science develops at a high rate.

More Realistic Model for Simulating Min Protein Dynamics: Lattice Boltzmann Method Incorporating the Role of Nucleoids

The dynamics of Min proteins plays a center role in accurate cell division. Although the nucleoids may presumably play an important role in prokaryotic cell division, there is a lack of models to account for its participation. In this work, we apply the lattice Boltzmann method to investigate protein oscillation based on a mesoscopic model that takes into account the nucleoid-s role. We found that our numerical results are in reasonably good agreement with the previous experimental results On comparing with the other computational models without the presence of nucleoids, the highlight of our finding is that the local densities of MinD and MinE on the cytoplasmic membrane increases, especially along the cell width, when the size of the obstacle increases, leading to a more distinct cap-like structure at the poles. This feature indicated the realistic pattern and reflected the combination of Min protein dynamics and nucleoid-s role.

A Propagator Method like Algorithm for Estimation of Multiple Real-Valued Sinusoidal Signal Frequencies

In this paper a novel method for multiple one dimensional real valued sinusoidal signal frequency estimation in the presence of additive Gaussian noise is postulated. A computationally simple frequency estimation method with efficient statistical performance is attractive in many array signal processing applications. The prime focus of this paper is to combine the subspace-based technique and a simple peak search approach. This paper presents a variant of the Propagator Method (PM), where a collaborative approach of SUMWE and Propagator method is applied in order to estimate the multiple real valued sine wave frequencies. A new data model is proposed, which gives the dimension of the signal subspace is equal to the number of frequencies present in the observation. But, the signal subspace dimension is twice the number of frequencies in the conventional MUSIC method for estimating frequencies of real-valued sinusoidal signal. The statistical analysis of the proposed method is studied, and the explicit expression of asymptotic (large-sample) mean-squared-error (MSE) or variance of the estimation error is derived. The performance of the method is demonstrated, and the theoretical analysis is substantiated through numerical examples. The proposed method can achieve sustainable high estimation accuracy and frequency resolution at a lower SNR, which is verified by simulation by comparing with conventional MUSIC, ESPRIT and Propagator Method.

Glutamic Acid Production from Potato by Brevibacterium linens

In this study, the possibility of using potato as a substrate for glutamic acid production by Brevibacterium linens was investigated. For preparation of fermentation medium, potato was hydrolyzed by hydrochloridric acid. The medium contained potato hydrolysate, tween 80, mineral solution, glucose, and potassium hydrogen phosphate. The initial pH of the medium was adjusted to 7-7.5. For achieving the optimum time with maximum yield, the beakers containing the medium and the inoculums were incubated in a rotary water bath flask shaker for one to five days. Thin layer choromatography was used for quantitative and qualitative assay of the glutamic acid produced. The results revealed that as fermentation time increased, pH of the fermentation medium significantly decreased (P

Design of Composite Risers for Minimum Weight

The use of composite materials in offshore engineering for deep sea oil production riser systems has drawn considerable interest due to the potential weight savings and improvement in durability. The design of composite risers consists of two stages: (1) local design based on critical local load cases, and (2) global analysis of the full length composite riser under global loads and assessment of critical locations. In the first stage, eight different material combinations were selected and their laminate configurations optimised under local load considerations. Stage two includes a final local stress analysis of the critical sections of the riser under the combined loads determined in the global analysis. This paper describes two design methodologies of the composite riser to provide minimum structural weight and shows that the use of off angle fibre orientations in addition to axial and hoop reinforcements offer substantial weight savings and ensure the structural capacity.

In vitro Anti-tubercular Screening of Newly Synthesized Benzimidazole Derivatives

A series of 1-(1H-benzimidazol-2-yl)-3-(substituted phenyl)-2-propen-1-one were allowed to react with hydrazine hydrate and phenyl hydrazine in submitted reactions to get pyrazoline and phenyl pyrazoline derivatives. All the compounds entered for screening at the Tuberculosis Antimicrobial Acquisition and Coordinating Facility (TAACF) for their in vitro antibacterial activity against Mycobacterium tuberculosis H37Rv strain (ATCC 27294) using Microplate Alamar Blue Assay (MABA) susceptibility test. The results expressed as MIC (minimum inhibitory concentration) in μg/mL. Among the fifteen compounds, eight compounds were found to have MIC values less than 10 μg/mL. These were subjected for cytotoxicity assay in VERO cells to determine CC50 (cytotoxic concentration 50%) values and finally SI (Selectivity Index) were calculated. Compound (XV) 2-[5-(4- fluorophenyl)-1-phenyl-4,5-dihydro-1H-3-pyrazolyl]-1Hbenzimidazole was considered the best candidate of the series that could be a good starting point to develop new lead compounds in the fight against tuberculosis.

Influence of Metakaolin on the Performance of Mortars and Concretes

The use of additions in cement in manufacturing, mortar and concrete offers economic and ecological advantages. Cements with additions such as limestone, slag and natural pouzzolana are produced in cement factories in Algeria. Several studies analyzed the effect of these additions on the physical and mechanical properties as well as the durability of concrete. However, few studies were conducted on the effect of local metakaolin on mechanical properties and durability of concrete. The main purpose of this paper is to analyze the performance of mortar and concrete with local metakaolin. The preparation of the metakaolin was carried out by calcination of kaolin at a temperature of 850 °C for a period of 3 hours. The experimental results have shown that the rates of substitutions of 10 and 15% metakaolin increases the compressive strength and flexural strength at both early age and long term. The durability and the permeability were also improved by reducing the coefficient of sorptivity.

Active Cyber Defense within the Concept of NATO’s Protection of Critical Infrastructures

Cyber attacks pose a serious threat to all states. Therefore, states constantly seek for various methods to encounter those threats. In addition, recent changes in the nature of cyber attacks and their more complicated methods have created a new concept: active cyber defense (ACD). This article tries to answer firstly why ACD is important to NATO and find out the viewpoint of NATO towards ACD. Secondly, infrastructure protection is essential to cyber defense. Critical infrastructure protection with ACD means is even more important. It is assumed that by implementing active cyber defense, NATO may not only be able to repel the attacks but also be deterrent. Hence, the use of ACD has a direct positive effect in all international organizations’ future including NATO.

Mechanical-Physical Characteristics Affecting the Durability of Fibre Reinforced Concrete with Recycled Aggregate

The article presents findings from the study and analysis of the results of an experimental programme focused on the production of concrete and fibre reinforced concrete in which natural aggregate has been substituted with brick or concrete recyclate. The research results are analyzed to monitor the effect of mechanicalphysical characteristics on the durability properties of tested cementitious composites. The key parts of the fibre reinforced concrete mix are the basic components: aggregates – recyclate, cement, fly ash, water and fibres. Their specific ratios and the properties of individual components principally affect the resulting behaviour of fresh fibre reinforced concrete and the characteristics of the final product. The article builds on the sources dealing with the use of recycled aggregates from construction and demolition waste in the production of fibre reinforced concrete. The implemented procedure of testing the composite contributes to the building sustainability in environmental engineering.

A Middleware Transparent Framework for Applying MDA to SOA

Although Model Driven Architecture has taken successful steps toward model-based software development, this approach still faces complex situations and ambiguous questions while applying to real world software systems. One of these questions - which has taken the most interest and focus - is how model transforms between different abstraction levels, MDA proposes. In this paper, we propose an approach based on Story Driven Modeling and Aspect Oriented Programming to ease these transformations. Service Oriented Architecture is taken as the target model to test the proposed mechanism in a functional system. Service Oriented Architecture and Model Driven Architecture [1] are both considered as the frontiers of their own domain in the software world. Following components - which was the greatest step after object oriented - SOA is introduced, focusing on more integrated and automated software solutions. On the other hand - and from the designers' point of view - MDA is just initiating another evolution. MDA is considered as the next big step after UML in designing domain.

Stochastic Simulation of Reaction-Diffusion Systems

Reactiondiffusion systems are mathematical models that describe how the concentration of one or more substances distributed in space changes under the influence of local chemical reactions in which the substances are converted into each other, and diffusion which causes the substances to spread out in space. The classical representation of a reaction-diffusion system is given by semi-linear parabolic partial differential equations, whose general form is ÔêétX(x, t) = DΔX(x, t), where X(x, t) is the state vector, D is the matrix of the diffusion coefficients and Δ is the Laplace operator. If the solute move in an homogeneous system in thermal equilibrium, the diffusion coefficients are constants that do not depend on the local concentration of solvent and of solutes and on local temperature of the medium. In this paper a new stochastic reaction-diffusion model in which the diffusion coefficients are function of the local concentration, viscosity and frictional forces of solvent and solute is presented. Such a model provides a more realistic description of the molecular kinetics in non-homogenoeus and highly structured media as the intra- and inter-cellular spaces. The movement of a molecule A from a region i to a region j of the space is described as a first order reaction Ai k- → Aj , where the rate constant k depends on the diffusion coefficient. Representing the diffusional motion as a chemical reaction allows to assimilate a reaction-diffusion system to a pure reaction system and to simulate it with Gillespie-inspired stochastic simulation algorithms. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the specific speed of reaction and diffusion events. Redi is the software tool, developed to implement the model of reaction-diffusion kinetics and dynamics. It is a free software, that can be downloaded from http://www.cosbi.eu. To demonstrate the validity of the new reaction-diffusion model, the simulation results of the chaperone-assisted protein folding in cytoplasm obtained with Redi are reported. This case study is redrawing the attention of the scientific community due to current interests on protein aggregation as a potential cause for neurodegenerative diseases.

A Neighborhood Condition for Fractional k-deleted Graphs

Abstract–Let k ≥ 3 be an integer, and let G be a graph of order n with n ≥ 9k +3- 42(k - 1)2 + 2. Then a spanning subgraph F of G is called a k-factor if dF (x) = k for each x ∈ V (G). A fractional k-factor is a way of assigning weights to the edges of a graph G (with all weights between 0 and 1) such that for each vertex the sum of the weights of the edges incident with that vertex is k. A graph G is a fractional k-deleted graph if there exists a fractional k-factor after deleting any edge of G. In this paper, it is proved that G is a fractional k-deleted graph if G satisfies δ(G) ≥ k + 1 and |NG(x) ∪ NG(y)| ≥ 1 2 (n + k - 2) for each pair of nonadjacent vertices x, y of G.