Trace Emergence of Ants- Traffic Flow, based upon Exclusion Process

Biological evolution has generated a rich variety of successful solutions; from nature, optimized strategies can be inspired. One interesting example is the ant colonies, which are able to exhibit a collective intelligence, still that their dynamic is simple. The emergence of different patterns depends on the pheromone trail, leaved by the foragers. It serves as positive feedback mechanism for sharing information. In this paper, we use the dynamic of TASEP as a model of interaction at a low level of the collective environment in the ant-s traffic flow. This work consists of modifying the movement rules of particles “ants" belonging to the TASEP model, so that it adopts with the natural movement of ants. Therefore, as to respect the constraints of having no more than one particle per a given site, and in order to avoid collision within a bidirectional circulation, we suggested two strategies: decease strategy and waiting strategy. As a third work stage, this is devoted to the study of these two proposed strategies- stability. As a final work stage, we applied the first strategy to the whole environment, in order to get to the emergence of traffic flow, which is a way of learning.

Biodegradation of PCP by the Rhizobacteria Isolated from Pentachlorophenol-tolerant Crop Species

Pentachlorophenol (PCP) is a polychlorinated aromatic compound that is widespread in industrial effluents and is considered to be a serious pollutant. Among the variety of industrial effluents encountered, effluents from tanning industry are very important and have a serious pollution potential. PCP is also formed unintentionally in effluents of paper and pulp industries. It is highly persistent in soils and is lethal to a wide variety of beneficial microorganisms and insects, human beings and animals. The natural processes that breakdown toxic chemicals in the environment have become the focus of much attention to develop safe and environmentfriendly deactivation technologies. Microbes and plants are among the most important biological agents that remove and degrade waste materials to enable their recycling in the environment. The present investigation was carried out with the aim of developing a microbial system for bioremediation of PCP polluted soils. A number of plant species were evaluated for their ability to tolerate different concentrations of pentachlorophenol (PCP) in the soil. The experiment was conducted for 30 days under pot culture conditions. The toxic effect of PCP on plants was studied by monitoring seed germination, plant growth and biomass. As the concentration of PCP was increased to 50 ppm, the inhibition of seed germination, plant growth and biomass was also increased. Although PCP had a negative effect on all plant species tested, maize and groundnut showed the maximum tolerance to PCP. Other tolerating crops included wheat, safflower, sunflower, and soybean. From the rhizosphere soil of the tolerant seedlings, as many as twenty seven PCP tolerant bacteria were isolated. From soybean, 8; sunflower, 3; safflower 8; maize 2; groundnut and wheat, 3 each isolates were made. They were screened for their PCP degradation potentials. HPLC analyses of PCP degradation revealed that the isolate MAZ-2 degraded PCP completely. The isolate MAZ-1 was the next best isolate with 90 per cent PCP degradation. These strains hold promise to be used in the bioremediation of PCP polluted soils.

Use of Detectors Technology for Gamma Ray Issued from Radioactive Isotopes and its Impact on Knowledge of Behavior of the Stationary Case of Solid Phase Holdup

For gamma radiation detection, assemblies having scintillation crystals and a photomultiplier tube, also there is a preamplifier connected to the detector because the signals from photomultiplier tube are of small amplitude. After pre-amplification the signals are sent to the amplifier and then to the multichannel analyser. The multichannel analyser sorts all incoming electrical signals according to their amplitudes and sorts the detected photons in channels covering small energy intervals. The energy range of each channel depends on the gain settings of the multichannel analyser and the high voltage across the photomultiplier tube. The exit spectrum data of the two main isotopes studied ,putting data in biomass program ,process it by Matlab program to get the solid holdup image (solid spherical nuclear fuel)

Biometric Technology in Securing the Internet Using Large Neural Network Technology

The article examines the methods of protection of citizens' personal data on the Internet using biometric identity authentication technology. It`s celebrated their potential danger due to the threat of loss of base biometric templates. To eliminate the threat of compromised biometric templates is proposed to use neural networks large and extra-large sizes, which will on the one hand securely (Highly reliable) to authenticate a person by his biometrics, and on the other hand make biometrics a person is not available for observation and understanding. This article also describes in detail the transformation of personal biometric data access code. It`s formed the requirements for biometrics converter code for his work with the images of "Insider," "Stranger", all the "Strangers". It`s analyzed the effect of the dimension of neural networks on the quality of converters mystery of biometrics in access code.

A New Divide and Conquer Software Process Model

The software system goes through a number of stages during its life and a software process model gives a standard format for planning, organizing and running a project. The article presents a new software development process model named as “Divide and Conquer Process Model", based on the idea first it divides the things to make them simple and then gathered them to get the whole work done. The article begins with the backgrounds of different software process models and problems in these models. This is followed by a new divide and conquer process model, explanation of its different stages and at the end edge over other models is shown.

Large-Dimensional Shells under Mining Tremors from Various Mining Regions in Poland

In the paper a detailed analysis of the dynamic response of a cooling tower shell to mining tremors originated from two main regions of mining activity in Poland (Upper Silesian Coal Basin and Legnica-Glogow Copper District) was presented. The representative time histories registered in the both regions were used as ground motion data in calculations of the dynamic response of the structure. It was proved that the dynamic response of the shell is strongly dependent not only on the level of vibration amplitudes but on the dominant frequency range of the mining shock typical for the mining region as well. Also a vertical component of vibrations occurred to have considerable influence on the total dynamic response of the shell. Finally, it turned out that non-uniformity of kinematic excitation resulting from spatial variety of ground motion plays a significant role in dynamic analysis of large-dimensional shells under mining shocks.

Extraction of Temporal Relation by the Creation of Historical Natural Disaster Archive

In historical science and social science, the influence of natural disaster upon society is a matter of great interest. In recent years, some archives are made through many hands for natural disasters, however it is inefficiency and waste. So, we suppose a computer system to create a historical natural disaster archive. As the target of this analysis, we consider newspaper articles. The news articles are considered to be typical examples that prescribe the temporal relations of affairs for natural disaster. In order to do this analysis, we identify the occurrences in newspaper articles by some index entries, considering the affairs which are specific to natural disasters, and show the temporal relation between natural disasters. We designed and implemented the automatic system of “extraction of the occurrences of natural disaster" and “temporal relation table for natural disaster."

Computational Analysis of the MembraneTargeting Domains of Plant-specific PRAF Proteins

The PRAF family of proteins is a plant specific family of proteins with distinct domain architecture and various unique sequence/structure traits. We have carried out an extensive search of the Arabidopsis genome using an automated pipeline and manual methods to verify previously known and identify unknown instances of PRAF proteins, characterize their sequence and build 3D structures of their individual domains. Integrating the sequence, structure and whatever little known experimental details for each of these proteins and their domains, we present a comprehensive characterization of the different domains in these proteins and their variant properties.

Considering Assembly Operations and Product Structure for Manufacturing Cell Formation

This paper considers the integration of assembly operations and product structure to Cellular Manufacturing System (CMS) design so that to correct the drawbacks of previous researches in the literature. For this purpose, a new mathematical model is developed which dedicates machining and assembly operations to manufacturing cells while the objective function is to minimize the intercellular movements resulting due to both of them. A linearization method is applied to achieve optimum solution through solving aforementioned nonlinear model by common programming language such as Lingo. Then, using different examples and comparing the results, the importance of integrating assembly considerations is demonstrated.

Simulations of Routing Protocols of Wireless Sensor Networks

Wireless Sensor Network is widely used in electronics. Wireless sensor networks are now used in many applications including military, environmental, healthcare applications, home automation and traffic control. We will study one area of wireless sensor networks, which is the routing protocol. Routing protocols are needed to send data between sensor nodes and the base station. In this paper, we will discuss two routing protocols, such as datacentric and hierarchical routing protocol. We will show the output of the protocols using the NS-2 simulator. This paper will compare the simulation output of the two routing protocol using Nam. We will simulate using Xgraph to find the throughput and delay of the protocol.

Extended Well-Founded Semantics in Bilattices

One of the most used assumptions in logic programming and deductive databases is the so-called Closed World Assumption (CWA), according to which the atoms that cannot be inferred from the programs are considered to be false (i.e. a pessimistic assumption). One of the most successful semantics of conventional logic programs based on the CWA is the well-founded semantics. However, the CWA is not applicable in all circumstances when information is handled. That is, the well-founded semantics, if conventionally defined, would behave inadequately in different cases. The solution we adopt in this paper is to extend the well-founded semantics in order for it to be based also on other assumptions. The basis of (default) negative information in the well-founded semantics is given by the so-called unfounded sets. We extend this concept by considering optimistic, pessimistic, skeptical and paraconsistent assumptions, used to complete missing information from a program. Our semantics, called extended well-founded semantics, expresses also imperfect information considered to be missing/incomplete, uncertain and/or inconsistent, by using bilattices as multivalued logics. We provide a method of computing the extended well-founded semantics and show that Kripke-Kleene semantics is captured by considering a skeptical assumption. We show also that the complexity of the computation of our semantics is polynomial time.

Investigation of Shear Thickening Liquid Protection Fibrous Material

The stab resistance performance of newly developed fabric composites composed of hexagonal paper honeycombs, filled with shear thickening fluid (STF), and woven Kevlar® fabric or UHMPE was investigated in this study. The STF was prepared by dispersing submicron SiO2 particles into polyethylene glycol (PEG). Our results indicate that the STF-Kevlar composite possessed lower penetration depth than that of neat Kevlar. In other words, the STF-Kevlar composite can attain the same energy level in stab-resistance test with fewer layers of Kevlar fabrics than that of the neat Kevlar fabrics. It also indicates that STF can be used for the fabrication of flexible body armors and can provide improved protection against stab threats. We found that the stab resistance of the STF-Kevlar composite increases with the increase of SiO2 concentration in STF. Moreover, the silica particles functionalized with silane coupling agent can further improve the stab resistance.

Optimum Radio Capacity Estimation of a Single-Cell Spread Spectrum MIMO System under Rayleigh Fading Conditions

In this paper, the problem of estimating the optimal radio capacity of a single-cell spread spectrum (SS) multiple-inputmultiple- output (MIMO) system operating in a Rayleigh fading environment is examined. The optimisation between the radio capacity and the theoretically achievable average channel capacity (in the sense of information theory) per user of a MIMO single-cell SS system operating in a Rayleigh fading environment is presented. Then, the spectral efficiency is estimated in terms of the achievable average channel capacity per user, during the operation over a broadcast time-varying link, and leads to a simple novel-closed form expression for the optimal radio capacity value based on the maximization of the achieved spectral efficiency. Numerical results are presented to illustrate the proposed analysis.

Inverse Problem Methodology for the Measurement of the Electromagnetic Parameters Using MLP Neural Network

This paper presents an approach which is based on the use of supervised feed forward neural network, namely multilayer perceptron (MLP) neural network and finite element method (FEM) to solve the inverse problem of parameters identification. The approach is used to identify unknown parameters of ferromagnetic materials. The methodology used in this study consists in the simulation of a large number of parameters in a material under test, using the finite element method (FEM). Both variations in relative magnetic permeability and electrical conductivity of the material under test are considered. Then, the obtained results are used to generate a set of vectors for the training of MLP neural network. Finally, the obtained neural network is used to evaluate a group of new materials, simulated by the FEM, but not belonging to the original dataset. Noisy data, added to the probe measurements is used to enhance the robustness of the method. The reached results demonstrate the efficiency of the proposed approach, and encourage future works on this subject.

An Incomplete Factorization Preconditioner for LMS Adaptive Filter

In this paper an efficient incomplete factorization preconditioner is proposed for the Least Mean Squares (LMS) adaptive filter. The proposed preconditioner is approximated from a priori knowledge of the factors of input correlation matrix with an incomplete strategy, motivated by the sparsity patter of the upper triangular factor in the QRD-RLS algorithm. The convergence properties of IPLMS algorithm are comparable with those of transform domain LMS(TDLMS) algorithm. Simulation results show efficiency and robustness of the proposed algorithm with reduced computational complexity.

Enhancement of m-FISH Images using Spectral Unmixing

Breast carcinoma is the most common form of cancer in women. Multicolour fluorescent in-situ hybridisation (m-FISH) is a common method for staging breast carcinoma. The interpretation of m-FISH images is complicated due to two effects: (i) Spectral overlap in the emission spectra of fluorochrome marked DNA probes and (ii) tissue autofluorescence. In this paper hyper-spectral images of m-FISH samples are used and spectral unmixing is applied to produce false colour images with higher contrast and better information content than standard RGB images. The spectral unmixing is realised by combinations of: Orthogonal Projection Analysis (OPA), Alterating Least Squares (ALS), Simple-to-use Interactive Self-Modeling Mixture Analysis (SIMPLISMA) and VARIMAX. These are applied on the data to reduce tissue autofluorescence and resolve the spectral overlap in the emission spectra. The results show that spectral unmixing methods reduce the intensity caused by tissue autofluorescence by up to 78% and enhance image contrast by algorithmically reducing the overlap of the emission spectra.

Featured based Segmentation of Color Textured Images using GLCM and Markov Random Field Model

In this paper, we propose a new image segmentation approach for colour textured images. The proposed method for image segmentation consists of two stages. In the first stage, textural features using gray level co-occurrence matrix(GLCM) are computed for regions of interest (ROI) considered for each class. ROI acts as ground truth for the classes. Ohta model (I1, I2, I3) is the colour model used for segmentation. Statistical mean feature at certain inter pixel distance (IPD) of I2 component was considered to be the optimized textural feature for further segmentation. In the second stage, the feature matrix obtained is assumed to be the degraded version of the image labels and modeled as Markov Random Field (MRF) model to model the unknown image labels. The labels are estimated through maximum a posteriori (MAP) estimation criterion using ICM algorithm. The performance of the proposed approach is compared with that of the existing schemes, JSEG and another scheme which uses GLCM and MRF in RGB colour space. The proposed method is found to be outperforming the existing ones in terms of segmentation accuracy with acceptable rate of convergence. The results are validated with synthetic and real textured images.

Metallographic Analysis of Laser and Mechanically Formed HSLA Steel

This research was conducted to develop a correlation between microstructure of HSLA steel and the mechanical properties that occur as a result of both laser and mechanical forming processes of the metal. The technique of forming flat metals by applying laser beams is a relatively new concept in the manufacturing industry. However, the effects of laser energy on the stability of metal alloy phases have not yet been elucidated in terms of phase transformations and microhardness. In this work, CO2 laser source was used to irradiate the surface of a flat metal then the microstructure and microhardness of the metal were studied on the formed specimen. The extent to which the microstructure changed depended on the heat inputs of up to 1000 J/cm2 with cooling rates of about 4.8E+02 K/s. Experimental results revealed that the irradiated surface of a HSLA steel had transformed to austenitic structure during the heating process.

New Product-Type Estimators for the Population Mean Using Quartiles of the Auxiliary Variable

In this paper, we suggest new product-type estimators for the population mean of the variable of interest exploiting the first or the third quartile of the auxiliary variable. We obtain mean square error equations and the bias for the estimators. We study the properties of these estimators using simple random sampling (SRS) and ranked set sampling (RSS) methods. It is found that, SRS and RSS produce approximately unbiased estimators of the population mean. However, the RSS estimators are more efficient than those obtained using SRS based on the same number of measured units for all values of the correlation coefficient.

Weight-Based Query Optimization System Using Buffer

Fast retrieval of data has been a need of user in any database application. This paper introduces a buffer based query optimization technique in which queries are assigned weights according to their number of execution in a query bank. These queries and their optimized executed plans are loaded into the buffer at the start of the database application. For every query the system searches for a match in the buffer and executes the plan without creating new plans.