Visual Cryptography by Random Grids with Identifiable Shares

This paper proposes a visual cryptography by random grids scheme with identifiable shares. The method encodes an image O in two shares that exhibits the following features: (1) each generated share has the same scale as O, (2) any share singly has noise-like appearance that reveals no secret information on O, (3) the secrets can be revealed by superimposing the two shares, (4) folding a share up can disclose some identification patterns, and (5) both of the secret information and the designated identification patterns are recognized by naked eye without any computation. The property to show up identification patterns on folded shares establishes a simple and friendly interface for users to manage the numerous shares created by VC schemes.

Two States Mapping Based Neural Network Model for Decreasing of Prediction Residual Error

The objective of this paper is to design a model of human vital sign prediction for decreasing prediction error by using two states mapping based time series neural network BP (back-propagation) model. Normally, lot of industries has been applying the neural network model by training them in a supervised manner with the error back-propagation algorithm for time series prediction systems. However, it still has a residual error between real value and prediction output. Therefore, we designed two states of neural network model for compensation of residual error which is possible to use in the prevention of sudden death and metabolic syndrome disease such as hypertension disease and obesity. We found that most of simulations cases were satisfied by the two states mapping based time series prediction model compared to normal BP. In particular, small sample size of times series were more accurate than the standard MLP model. We expect that this algorithm can be available to sudden death prevention and monitoring AGENT system in a ubiquitous homecare environment.

What is the Key Element for the Territory's State of Development?

The result of process of territory-s development is the territory-s state of development (TSoD), which is pointed towards the provision and improvement of people-s life conditions. The authors offer to measure the TSoD according to their own developed model. Using the available statistical data regarding the values of model-s elements, the authors empirically show which element mainly determines the TSoD. The findings of the research showed that the key elements of the TSoD are the “Material welfare of people" and “People-s health". Performing a deeper statistical analysis of correlation between these elements, it turned out that it is not so necessary for a country to be bent on trying to increase the material growth of a territory, because a relatively high index of life expectancy at birth could be ensured also by much more modest material resources. On the other hand, the economical feedback of longer lifespan within countries with lower material performance is also relatively low.

The Preservation of Cultural Heritage: Continuity and Memory

Contemporary science and technologies largely widen the gap between the spiritual and rational of the society. Industrial and technological breakthroughs might radically affect most processes in the society, thus losing the cultural heritage. The thinkers recognized the dangers of the decadence in the first place. In the present article the ways of preserving cultural heritage have been investigated. Memory has always been a necessary condition for selfidentification, - continuity is based on this. The authors have supported the hypothesis that continuity and ethnic memory are the very mechanisms that preserve cultural heritage. Such problemformulating will facilitate another, new look at the material, spiritual and arts spheres of the cultural heritage of numerous ethnic groups. The fundamental works by major European and Kazakh scientists have been taken as a basis for the research done.

Failure Analysis of Methanol Evaporator

Thermal water hammer is a special type of water hammer which rarely occurs in heat exchangers. In biphasic fluids, if steam bubbles are surrounded by condensate, regarding lower condensate temperature than steam, they will suddenly collapse. As a result, the vacuum caused by an extreme change in volume lead to movement of the condensates in all directions and their collision the force produced by this collision leads to a severe stress in the pipe wall. This phenomenon is a special type of water hammer. According to fluid mechanics, this phenomenon is a particular type of transient flows during which abrupt change of fluid leads to sudden pressure change inside the tube. In this paper, the mechanism of abrupt failure of 80 tubes of 481 tubes of a methanol heat exchanger is discussed. Initially, due to excessive temperature differences between heat transfer fluids and simultaneous failure of 80 tubes, thermal shock was presupposed as the reason of failure. Deeper investigation on cross-section of failed tubes showed that failure was, ductile type of failure, so the first hypothesis was rejected. Further analysis and more accurate experiments revealed that failure of tubes caused by thermal water hammer. Finally, the causes of thermal water hammer and various solutions to avoid such mechanism are discussed.

Intervention of Sambucus Nigra Polyphenolic Extract in Experimental Arterial Hypertension

The research focuses on the effects of polyphenols extracted from Sambucus nigra fruit, using an experimental arterial hypertension pattern, as well as their influence on the oxidative stress. The results reveal the normalization of the reduced glutathion concentration, as well as a considerable reduction in the malondialdehide serum concentration by the polyphenolic protection. The rat blood pressure values were recorded using a CODATM system, which uses a non-invasive blood pressure measuring method. All the measured blood pressure components revealed a biostatistically significant (p

Adaptive Weighted Averaging Filter Using the Appropriate Number of Consecutive Frames

In this paper, we propose a novel adaptive spatiotemporal filter that utilizes image sequences in order to remove noise. The consecutive frames include: current, previous and next noisy frames. The filter proposed in this paper is based upon the weighted averaging pixels intensity and noise variance in image sequences. It utilizes the Appropriate Number of Consecutive Frames (ANCF) based on the noisy pixels intensity among the frames. The number of consecutive frames is adaptively calculated for each region in image and its value may change from one region to another region depending on the pixels intensity within the region. The weights are determined by a well-defined mathematical criterion, which is adaptive to the feature of spatiotemporal pixels of the consecutive frames. It is experimentally shown that the proposed filter can preserve image structures and edges under motion while suppressing noise, and thus can be effectively used in image sequences filtering. In addition, the AWA filter using ANCF is particularly well suited for filtering sequences that contain segments with abruptly changing scene content due to, for example, rapid zooming and changes in the view of the camera.

Rural Connectivity Technologies Cost Analysis

Rural areas of Tanzania are still disadvantaged in terms of diffusion of IP-based services; this is due to lack of Information and Communication Technology (ICT) infrastructures, especially lack of connectivity. One of the limitations for connectivity problems in rural areas of Tanzania is the high cost to establish infrastructures for IP-based services [1-2]. However the cost of connectivity varies from one technology to the other and at the same time, the cost is also different from one operator (service provider) to another within the country. This paper presents development of software system to calculate cost of connectivity to rural areas of Tanzania. The system is developed to make an easy access of connectivity cost from different technologies and different operators. The development of the calculator follows the V-model software development lifecycle. The calculator is used to evaluate the economic viability of different technologies considered as being potential candidates to provide rural connectivity. In this paper, the evaluation is based on the techno-economic analysis approach.

Influence of Jerusalem Artichoke Powder on the Nutritional Value of Pastry Products

From year to year, the incidence of different diseases is increasing in humans, and the cause is inadequate intake of dietary fibre, vitamins, and minerals. One of the possibilities to take care of your health preventively is including in the diet products with increased dietary fibre, vitamin, and mineral content.Jerusalem artichoke powder (JAP) made from Jerusalem artichoke (Helianthus tuberosus L) roots is a valuable product. By adding it to pastry goods, we can obtain a fibre-rich food that could be healthier and an excellent alternative to the classical pastry products of this kind.Experiments were carried out at the Faculty of Food Technology of Latvia University of Agriculture (LLU). Results of experiments showed that addition of Jerusalem artichoke powder has significant impact on all the studied pastry products nutritional value (p

Towards Self-ware via Swarm-Array Computing

The work reported in this paper proposes Swarm-Array computing, a novel technique inspired by swarm robotics, and built on the foundations of autonomic and parallel computing. The approach aims to apply autonomic computing constructs to parallel computing systems and in effect achieve the self-ware objectives that describe self-managing systems. The constitution of swarm-array computing comprising four constituents, namely the computing system, the problem/task, the swarm and the landscape is considered. Approaches that bind these constituents together are proposed. Space applications employing FPGAs are identified as a potential area for applying swarm-array computing for building reliable systems. The feasibility of a proposed approach is validated on the SeSAm multi-agent simulator and landscapes are generated using the MATLAB toolkit.

Design of the Large Dimension Cold Shield Cooled by G-M Cryocooler

The design of methods of the 20 K large dimension cold shield used for infrared radiation demarcating in space environment simulation test were introduced in this paper. The cold shield were cooled by five G-M cryocoolers , and the dimension of the cold shield is the largest in our country.Cold shield installation and distribution and compensator for contraction on cooling were introduced detailedly. The temperature distribution and cool-down time of cold shield surface were also calculated and analysed in this paper. The design of cold shield resolves the difficulty of compensator for contraction on cooling successfully. Test results show that the actual technical performance indicators of cold shield met and exceeded the design requirements.

Preparation and Characterization of M. × Piperita L. Oil Based Gel Formulation

The essential oil of M. × piperita L. was formulated into a topical gel. The prepared gel was characterized for its pH, viscosity, spreadiblity, consistency and extrudiblity, while its stability was evaluated under different temperature conditions. The prepared M. × piperita oil gel was clear and transparent. The pH value of developed gel was 6.6, while its viscosity was 1200 cP. Spreadability and consistency of the M. × piperita oil gel was 10.7 g.cm/sec and 7 mm, respectively. The prepared gel showed good extrudiblity. During the stability studies, no significant change in pH and viscosity as a function of time for gel was observed, indicating stability of prepared formulation. The gel developed in this study is expected to forward the usage of M. × piperita essential towards commercial application.

Mechanical Modeling Issues in Optimization of Dynamic Behavior of RF MEMS Switches

This paper details few mechanical modeling and design issues of RF MEMS switches. We concentrate on an electrostatically actuated broad side series switch; surface micromachined with a crab leg membrane. The same results are extended to any complex structure. With available experimental data and fabrication results, we present the variation in dynamic performance and compliance of the switch with reference to few design issues, which we find are critical in deciding the dynamic behavior of the switch, without compromise on the RF characteristics. The optimization of pull in voltage, transient time and resonant frequency with regard to these critical design parameters are also presented.

A Blind SLM Scheme for Reduction of PAPR in OFDM Systems

In this paper we propose a blind algorithm for peakto- average power ratio (PAPR) reduction in OFDM systems, based on selected mapping (SLM) algorithm as a distortionless method. The main drawback of the conventional SLM technique is the need for transmission of several side information bits, for each data block, which results in loss in data rate transmission. In the proposed method some special number of carriers in the OFDM frame is reserved to be rotated with one of the possible phases according to the number of phase sequence blocks in SLM algorithm. Reserving some limited number of carriers wont effect the reduction in PAPR of OFDM signal. Simulation results show using ML criteria at the receiver will lead to the same system-performance as the conventional SLM algorithm, while there is no need to send any side information to the receiver.

Iris Localization using Circle and Fuzzy Circle Detection Method

Iris localization is a very important approach in biometric identification systems. Identification process usually is implemented in three levels: iris localization, feature extraction, and pattern matching finally. Accuracy of iris localization as the first step affects all other levels and this shows the importance of iris localization in an iris based biometric system. In this paper, we consider Daugman iris localization method as a standard method, propose a new method in this field and then analyze and compare the results of them on a standard set of iris images. The proposed method is based on the detection of circular edge of iris, and improved by fuzzy circles and surface energy difference contexts. Implementation of this method is so easy and compared to the other methods, have a rather high accuracy and speed. Test results show that the accuracy of our proposed method is about Daugman method and computation speed of it is 10 times faster.

Comparing Autoregressive Moving Average (ARMA) Coefficients Determination using Artificial Neural Networks with Other Techniques

Autoregressive Moving average (ARMA) is a parametric based method of signal representation. It is suitable for problems in which the signal can be modeled by explicit known source functions with a few adjustable parameters. Various methods have been suggested for the coefficients determination among which are Prony, Pade, Autocorrelation, Covariance and most recently, the use of Artificial Neural Network technique. In this paper, the method of using Artificial Neural network (ANN) technique is compared with some known and widely acceptable techniques. The comparisons is entirely based on the value of the coefficients obtained. Result obtained shows that the use of ANN also gives accurate in computing the coefficients of an ARMA system.

The Ethics of Dissent: The Case of David Kelly

In this paper, we rely on the story of the late British weapons inspector David Kelly to illustrate how sensemaking can inform the study of the ethics of suppression of dissent. Using archival data, we reconstruct Dr. Kelly-s key responsibilities as a weapons inspector and government employee. We begin by clarifying the concept of dissent and how it is a useful organizational process. We identify the various ways that dissent has been discussed in the organizational literature and reconsider the process of sensemaking. We conclude that suppression of opinions that deviate from the majority is part of the identity maintenance of the sensemaking process. We illustrate the prevention of dissent in organizations consists of a set of unsatisfactory trade-offs.

A Direct Down-conversion Receiver for Low-power Wireless Sensor Networks

A direct downconversion receiver implemented in 0.13 μm 1P8M process is presented. The circuit is formed by a single-end LNA, an active balun for conversion into balanced mode, a quadrature double-balanced passive switch mixer and a quadrature voltage-controlled oscillator. The receiver operates in the 2.4 GHz ISM band and complies with IEEE 802.15.4 (ZigBee) specifications. The circuit exhibits a very low noise figure of only 2.27 dB and dissipates only 14.6 mW with a 1.2 V supply voltage and is hence suitable for low-power applications.

A Fragile Watermarking Scheme for Color Image Authentication

In this paper, a fragile watermarking scheme is proposed for color image specified object-s authentication. The color image is first transformed from RGB to YST color space, suitable for watermarking the color media. The T channel corresponds to the chrominance component of a color image andYS ÔèÑ T , therefore selected for embedding the watermark. The T channel is first divided into 2×2 non-overlapping blocks and the two LSBs are set to zero. The object that is to be authenticated is also divided into 2×2 nonoverlapping blocks and each block-s intensity mean is computed followed by eight bit encoding. The generated watermark is then embedded into T channel randomly selected 2×2 block-s LSBs using 2D-Torus Automorphism. Selection of block size is paramount for exact localization and recovery of work. The proposed scheme is blind, efficient and secure with ability to detect and locate even minor tampering applied to the image with full recovery of original work. The quality of watermarked media is quite high both subjectively and objectively. The technique is suitable for class of images with format such as gif, tif or bitmap.

Globally Convergent Edge-preserving Reconstruction with Contour-line Smoothing

The standard approach to image reconstruction is to stabilize the problem by including an edge-preserving roughness penalty in addition to faithfulness to the data. However, this methodology produces noisy object boundaries and creates a staircase effect. The existing attempts to favor the formation of smooth contour lines take the edge field explicitly into account; they either are computationally expensive or produce disappointing results. In this paper, we propose to incorporate the smoothness of the edge field in an implicit way by means of an additional penalty term defined in the wavelet domain. We also derive an efficient half-quadratic algorithm to solve the resulting optimization problem, including the case when the data fidelity term is non-quadratic and the cost function is nonconvex. Numerical experiments show that our technique preserves edge sharpness while smoothing contour lines; it produces visually pleasing reconstructions which are quantitatively better than those obtained without wavelet-domain constraints.