Biometric Technology in Securing the Internet Using Large Neural Network Technology

The article examines the methods of protection of citizens' personal data on the Internet using biometric identity authentication technology. It`s celebrated their potential danger due to the threat of loss of base biometric templates. To eliminate the threat of compromised biometric templates is proposed to use neural networks large and extra-large sizes, which will on the one hand securely (Highly reliable) to authenticate a person by his biometrics, and on the other hand make biometrics a person is not available for observation and understanding. This article also describes in detail the transformation of personal biometric data access code. It`s formed the requirements for biometrics converter code for his work with the images of "Insider," "Stranger", all the "Strangers". It`s analyzed the effect of the dimension of neural networks on the quality of converters mystery of biometrics in access code.

Extraction of Temporal Relation by the Creation of Historical Natural Disaster Archive

In historical science and social science, the influence of natural disaster upon society is a matter of great interest. In recent years, some archives are made through many hands for natural disasters, however it is inefficiency and waste. So, we suppose a computer system to create a historical natural disaster archive. As the target of this analysis, we consider newspaper articles. The news articles are considered to be typical examples that prescribe the temporal relations of affairs for natural disaster. In order to do this analysis, we identify the occurrences in newspaper articles by some index entries, considering the affairs which are specific to natural disasters, and show the temporal relation between natural disasters. We designed and implemented the automatic system of “extraction of the occurrences of natural disaster" and “temporal relation table for natural disaster."

Considering Assembly Operations and Product Structure for Manufacturing Cell Formation

This paper considers the integration of assembly operations and product structure to Cellular Manufacturing System (CMS) design so that to correct the drawbacks of previous researches in the literature. For this purpose, a new mathematical model is developed which dedicates machining and assembly operations to manufacturing cells while the objective function is to minimize the intercellular movements resulting due to both of them. A linearization method is applied to achieve optimum solution through solving aforementioned nonlinear model by common programming language such as Lingo. Then, using different examples and comparing the results, the importance of integrating assembly considerations is demonstrated.

A Wind Farm Reduced Order Model Using Integral Manifold Theory

Due to the increasing penetration of wind energy, it is necessary to possess design tools that are able to simulate the impact of these installations in utility grids. In order to provide a net contribution to this issue a detailed wind park model has been developed and is briefly presented. However, the computational costs associated with the performance of such a detailed model in describing the behavior of a wind park composed by a considerable number of units may render its practical application very difficult. To overcome this problem integral manifolds theory has been applied to reduce the order of the detailed wind park model, and therefore create the conditions for the development of a dynamic equivalent which is able to retain the relevant dynamics with respect to the existing a.c. system. In this paper integral manifold method has been introduced for order reduction. Simulation results of the proposed method represents that integral manifold method results fit the detailed model results with a higher precision than singular perturbation method.

Simulations of Routing Protocols of Wireless Sensor Networks

Wireless Sensor Network is widely used in electronics. Wireless sensor networks are now used in many applications including military, environmental, healthcare applications, home automation and traffic control. We will study one area of wireless sensor networks, which is the routing protocol. Routing protocols are needed to send data between sensor nodes and the base station. In this paper, we will discuss two routing protocols, such as datacentric and hierarchical routing protocol. We will show the output of the protocols using the NS-2 simulator. This paper will compare the simulation output of the two routing protocol using Nam. We will simulate using Xgraph to find the throughput and delay of the protocol.

Extended Well-Founded Semantics in Bilattices

One of the most used assumptions in logic programming and deductive databases is the so-called Closed World Assumption (CWA), according to which the atoms that cannot be inferred from the programs are considered to be false (i.e. a pessimistic assumption). One of the most successful semantics of conventional logic programs based on the CWA is the well-founded semantics. However, the CWA is not applicable in all circumstances when information is handled. That is, the well-founded semantics, if conventionally defined, would behave inadequately in different cases. The solution we adopt in this paper is to extend the well-founded semantics in order for it to be based also on other assumptions. The basis of (default) negative information in the well-founded semantics is given by the so-called unfounded sets. We extend this concept by considering optimistic, pessimistic, skeptical and paraconsistent assumptions, used to complete missing information from a program. Our semantics, called extended well-founded semantics, expresses also imperfect information considered to be missing/incomplete, uncertain and/or inconsistent, by using bilattices as multivalued logics. We provide a method of computing the extended well-founded semantics and show that Kripke-Kleene semantics is captured by considering a skeptical assumption. We show also that the complexity of the computation of our semantics is polynomial time.

Investigation of Shear Thickening Liquid Protection Fibrous Material

The stab resistance performance of newly developed fabric composites composed of hexagonal paper honeycombs, filled with shear thickening fluid (STF), and woven Kevlar® fabric or UHMPE was investigated in this study. The STF was prepared by dispersing submicron SiO2 particles into polyethylene glycol (PEG). Our results indicate that the STF-Kevlar composite possessed lower penetration depth than that of neat Kevlar. In other words, the STF-Kevlar composite can attain the same energy level in stab-resistance test with fewer layers of Kevlar fabrics than that of the neat Kevlar fabrics. It also indicates that STF can be used for the fabrication of flexible body armors and can provide improved protection against stab threats. We found that the stab resistance of the STF-Kevlar composite increases with the increase of SiO2 concentration in STF. Moreover, the silica particles functionalized with silane coupling agent can further improve the stab resistance.

Inverse Problem Methodology for the Measurement of the Electromagnetic Parameters Using MLP Neural Network

This paper presents an approach which is based on the use of supervised feed forward neural network, namely multilayer perceptron (MLP) neural network and finite element method (FEM) to solve the inverse problem of parameters identification. The approach is used to identify unknown parameters of ferromagnetic materials. The methodology used in this study consists in the simulation of a large number of parameters in a material under test, using the finite element method (FEM). Both variations in relative magnetic permeability and electrical conductivity of the material under test are considered. Then, the obtained results are used to generate a set of vectors for the training of MLP neural network. Finally, the obtained neural network is used to evaluate a group of new materials, simulated by the FEM, but not belonging to the original dataset. Noisy data, added to the probe measurements is used to enhance the robustness of the method. The reached results demonstrate the efficiency of the proposed approach, and encourage future works on this subject.

Development of a Simulator for Explaining Organic Chemical Reactions Based on Qualitative Process Theory

This paper discusses the development of a qualitative simulator (abbreviated QRiOM) for predicting the behaviour of organic chemical reactions. The simulation technique is based on the qualitative process theory (QPT) ontology. The modelling constructs of QPT embody notions of causality which can be used to explain the behaviour of a chemical system. The major theme of this work is that, in a qualitative simulation environment, students are able to articulate his/her knowledge through the inspection of explanations generated by software. The implementation languages are Java and Prolog. The software produces explanation in various forms that stresses on the causal theories in the chemical system which can be effectively used to support learning.

Recycling Poultry Feathers for Pb Removal from Wastewater: Kinetic and Equilibrium Studies

Chicken feathers were used as biosorbent for Pb removal from aqueous solution. In this paper, the kinetics and equilibrium studies at several pH, temperature, and metal concentration values are reported. For tested conditions, the Pb sorption capacity of this poultry waste ranged from 0.8 to 8.3 mg/g. Optimal conditions for Pb removal by chicken feathers have been identified. Pseudo-first order and pseudo-second order equations were used to analyze the experimental data. In addition, the sorption isotherms were fitted to classical Langmuir and Freundlich models. Finally, thermodynamic parameters for the sorption process have been determined. In summary, the results showed that chicken feathers are an alternative and promising sorbent for the treatment of effluents polluted by Pb ions.

Featured based Segmentation of Color Textured Images using GLCM and Markov Random Field Model

In this paper, we propose a new image segmentation approach for colour textured images. The proposed method for image segmentation consists of two stages. In the first stage, textural features using gray level co-occurrence matrix(GLCM) are computed for regions of interest (ROI) considered for each class. ROI acts as ground truth for the classes. Ohta model (I1, I2, I3) is the colour model used for segmentation. Statistical mean feature at certain inter pixel distance (IPD) of I2 component was considered to be the optimized textural feature for further segmentation. In the second stage, the feature matrix obtained is assumed to be the degraded version of the image labels and modeled as Markov Random Field (MRF) model to model the unknown image labels. The labels are estimated through maximum a posteriori (MAP) estimation criterion using ICM algorithm. The performance of the proposed approach is compared with that of the existing schemes, JSEG and another scheme which uses GLCM and MRF in RGB colour space. The proposed method is found to be outperforming the existing ones in terms of segmentation accuracy with acceptable rate of convergence. The results are validated with synthetic and real textured images.

Metallographic Analysis of Laser and Mechanically Formed HSLA Steel

This research was conducted to develop a correlation between microstructure of HSLA steel and the mechanical properties that occur as a result of both laser and mechanical forming processes of the metal. The technique of forming flat metals by applying laser beams is a relatively new concept in the manufacturing industry. However, the effects of laser energy on the stability of metal alloy phases have not yet been elucidated in terms of phase transformations and microhardness. In this work, CO2 laser source was used to irradiate the surface of a flat metal then the microstructure and microhardness of the metal were studied on the formed specimen. The extent to which the microstructure changed depended on the heat inputs of up to 1000 J/cm2 with cooling rates of about 4.8E+02 K/s. Experimental results revealed that the irradiated surface of a HSLA steel had transformed to austenitic structure during the heating process.

Measuring the Development Level of Chinese Regional Service Industry: An Empirical Analysis based on Entropy Weight and TOPSIS

Using entropy weight and TOPSIS method, a comprehensive evaluation is done on the development level of Chinese regional service industry in this paper. Firstly, based on existing research results, an evaluation index system is constructed from the scale of development, the industrial structure and the economic benefits. An evaluation model is then built up based on entropy weight and TOPSIS, and an empirical analysis is conducted on the development level of service industries in 31 Chinese provinces during 2006 and 2009 from the two dimensions or time series and cross section, which provides new idea for assessing regional service industry. Furthermore, the 31 provinces are classified into four categories based on the evaluation results, and deep analysis is carried out on the evaluation results.

Weight-Based Query Optimization System Using Buffer

Fast retrieval of data has been a need of user in any database application. This paper introduces a buffer based query optimization technique in which queries are assigned weights according to their number of execution in a query bank. These queries and their optimized executed plans are loaded into the buffer at the start of the database application. For every query the system searches for a match in the buffer and executes the plan without creating new plans.

Research on Weakly Hard Real-Time Constraints and Their Boolean Combination to Support Adaptive QoS

Advances in computing applications in recent years have prompted the demand for more flexible scheduling models for QoS demand. Moreover, in practical applications, partly violated temporal constraints can be tolerated if the violation meets certain distribution. So we need extend the traditional Liu and Lanland model to adapt to these circumstances. There are two extensions, which are the (m, k)-firm model and Window-Constrained model. This paper researches on weakly hard real-time constraints and their combination to support QoS. The fact that a practical application can tolerate some violations of temporal constraint under certain distribution is employed to support adaptive QoS on the open real-time system. The experiment results show these approaches are effective compared to traditional scheduling algorithms.

Distributed Relay Selection and Channel Choice in Cognitive Radio Network

In this paper, we study the cooperative communications where multiple cognitive radio (CR) transmit-receive pairs competitive maximize their own throughputs. In CR networks, the influences of primary users and the spectrum availability are usually different among CR users. Due to the existence of multiple relay nodes and the different spectrum availability, each CR transmit-receive pair should not only select the relay node but also choose the appropriate channel. For this distributed problem, we propose a game theoretic framework to formulate this problem and we apply a regret-matching learning algorithm which is leading to correlated equilibrium. We further formulate a modified regret-matching learning algorithm which is fully distributed and only use the local information of each CR transmit-receive pair. This modified algorithm is more practical and suitable for the cooperative communications in CR network. Simulation results show the algorithm convergence and the modified learning algorithm can achieve comparable performance to the original regretmatching learning algorithm.

Nickel on Inner Surface and Stainless Steel on Outer Surface for Functionally Graded Cylindrical Shell

Study is on the vibration of thin cylindrical shells made of a functionally gradient material (FGM) composed of stainless steel and nickel is presented. The effects of the FGM configuration are studied by studying the frequencies of FG cylindrical shells. In this case FG cylindrical shell has Nickel on its inner surface and stainless steel on its outer surface. The study is carried out based on third order shear deformation shell theory. The objective is to study the natural frequencies, the influence of constituent volume fractions and the effects of configurations of the constituent materials on the frequencies. The properties are graded in the thickness direction according to the volume fraction power-law distribution. Results are presented on the frequency characteristics, the influence of the constituent various volume fractions on the frequencies.

A New Derivative-Free Quasi-Secant Algorithm For Solving Non-Linear Equations

Most of the nonlinear equation solvers do not converge always or they use the derivatives of the function to approximate the root of such equations. Here, we give a derivative-free algorithm that guarantees the convergence. The proposed two-step method, which is to some extent like the secant method, is accompanied with some numerical examples. The illustrative instances manifest that the rate of convergence in proposed algorithm is more than the quadratically iterative schemes.

Improving Injection Moulding Processes Using Experimental Design

Moulded parts contribute to more than 70% of components in products. However, common defects particularly in plastic injection moulding exist such as: warpage, shrinkage, sink marks, and weld lines. In this paper Taguchi experimental design methods are applied to reduce the warpage defect of thin plate Acrylonitrile Butadiene Styrene (ABS) and are demonstrated in two levels; namely, orthogonal arrays of Taguchi and the Analysis of Variance (ANOVA). Eight trials have been run in which the optimal parameters that can minimize the warpage defect in factorial experiment are obtained. The results obtained from ANOVA approach analysis with respect to those derived from MINITAB illustrate the most significant factors which may cause warpage in injection moulding process. Moreover, ANOVA approach in comparison with other approaches like S/N ratio is more accurate and with the interaction of factors it is possible to achieve higher and the better outcomes.

Automatic Authentication of Handwritten Documents via Low Density Pixel Measurements

We introduce an effective approach for automatic offline au- thentication of handwritten samples where the forgeries are skillfully done, i.e., the true and forgery sample appearances are almost alike. Subtle details of temporal information used in online verification are not available offline and are also hard to recover robustly. Thus the spatial dynamic information like the pen-tip pressure characteristics are considered, emphasizing on the extraction of low density pixels. The points result from the ballistic rhythm of a genuine signature which a forgery, however skillful that may be, always lacks. Ten effective features, including these low density points and den- sity ratio, are proposed to make the distinction between a true and a forgery sample. An adaptive decision criteria is also derived for better verification judgements.