Heat Stress Monitor by Using Low-Cost Temperature and Humidity Sensors

The aim of this study is to develop a cost-effective WBGT heat stress monitor which provides precise heat stress measurement. The proposed device employs SHT15 and DS18B20 as a temperature and humidity sensors, respectively, incorporating with ATmega328 microcontroller. The developed heat stress monitor was calibrated and adjusted to that of the standard temperature and humidity sensors in the laboratory. The results of this study illustrated that the mean percentage error and the standard deviation from the measurement of the globe temperature was 2.33 and 2.71 respectively, while 0.94 and 1.02 were those of the dry bulb temperature, 0.79 and 0.48 were of the wet bulb temperature, and 4.46 and 1.60 were of the relative humidity sensor. This device is relatively low-cost and the measurement error is acceptable.

A Watermarking Scheme for MP3 Audio Files

In this work, we present for the first time in our perception an efficient digital watermarking scheme for mpeg audio layer 3 files that operates directly in the compressed data domain, while manipulating the time and subband/channel domain. In addition, it does not need the original signal to detect the watermark. Our scheme was implemented taking special care for the efficient usage of the two limited resources of computer systems: time and space. It offers to the industrial user the capability of watermark embedding and detection in time immediately comparable to the real music time of the original audio file that depends on the mpeg compression, while the end user/audience does not face any artifacts or delays hearing the watermarked audio file. Furthermore, it overcomes the disadvantage of algorithms operating in the PCMData domain to be vulnerable to compression/recompression attacks, as it places the watermark in the scale factors domain and not in the digitized sound audio data. The strength of our scheme, that allows it to be used with success in both authentication and copyright protection, relies on the fact that it gives to the users the enhanced capability their ownership of the audio file not to be accomplished simply by detecting the bit pattern that comprises the watermark itself, but by showing that the legal owner knows a hard to compute property of the watermark.

Managing a Manufacturing System with Integration of Walking Worker and Lean Thinking

A product goes through various processes in a production flow which is also known as assembly line in manufacturing process management. Toyota created a new concept which is known as lean concept in manufacturing industry. Today it is the leading model in manufacturing plants through the globe. The linear walking worker assembly line is a flexible assembly system where each worker travels down the line carrying out each assembly task at each station; and each worker accomplishes the assembly of a unit from start to finish. This paper attempts to combine the flexibility of the walking worker and lean in order to quantify the benefits from applying the shop floor principles of lean management.

Analyzing Multi-Labeled Data Based on the Roll of a Concept against a Semantic Range

Classifying data hierarchically is an efficient approach to analyze data. Data is usually classified into multiple categories, or annotated with a set of labels. To analyze multi-labeled data, such data must be specified by giving a set of labels as a semantic range. There are some certain purposes to analyze data. This paper shows which multi-labeled data should be the target to be analyzed for those purposes, and discusses the role of a label against a set of labels by investigating the change when a label is added to the set of labels. These discussions give the methods for the advanced analysis of multi-labeled data, which are based on the role of a label against a semantic range.

Filteristic Soft Lattice Implication Algebras

Applying the idea of soft set theory to lattice implication algebras, the novel concept of (implicative) filteristic soft lattice implication algebras which related to (implicative) filter(for short, (IF-)F-soft lattice implication algebras) are introduced. Basic properties of (IF-)F-soft lattice implication algebras are derived. Two kinds of fuzzy filters (i.e.(2, 2 _qk)((2, 2 _ qk))-fuzzy (implicative) filter) of L are introduced, which are generalizations of fuzzy (implicative) filters. Some characterizations for a soft set to be a (IF-)F-soft lattice implication algebra are provided. Analogously, this idea can be used in other types of filteristic lattice implication algebras (such as fantastic (positive implicative) filteristic soft lattice implication algebras).

A Neuro Adaptive Control Strategy for Movable Power Source of Proton Exchange Membrane Fuel Cell Using Wavelets

Movable power sources of proton exchange membrane fuel cells (PEMFC) are the important research done in the current fuel cells (FC) field. The PEMFC system control influences the cell performance greatly and it is a control system for industrial complex problems, due to the imprecision, uncertainty and partial truth and intrinsic nonlinear characteristics of PEMFCs. In this paper an adaptive PI control strategy using neural network adaptive Morlet wavelet for control is proposed. It is based on a single layer feed forward neural networks with hidden nodes of adaptive morlet wavelet functions controller and an infinite impulse response (IIR) recurrent structure. The IIR is combined by cascading to the network to provide double local structure resulting in improving speed of learning. The proposed method is applied to a typical 1 KW PEMFC system and the results show the proposed method has more accuracy against to MLP (Multi Layer Perceptron) method.

Computer Models of the Vestibular Head Tilt Response, and Their Relationship to EVestG and Meniere's Disease

This paper attempts to explain response components of Electrovestibulography (EVestG) using a computer simulation of a three-canal model of the vestibular system. EVestG is a potentially new diagnostic method for Meniere's disease. EVestG is a variant of Electrocochleography (ECOG), which has been used as a standard method for diagnosing Meniere's disease - it can be used to measure the SP/AP ratio, where an SP/AP ratio greater than 0.4-0.5 is indicative of Meniere-s Disease. In EVestG, an applied head tilt replaces the acoustic stimulus of ECOG. The EVestG output is also an SP/AP type plot, where SP is the summing potential, and AP is the action potential amplitude. AP is thought of as being proportional to the size of a population of afferents in an excitatory neural firing state. A simulation of the fluid volume displacement in the vestibular labyrinth in response to various types of head tilts (ipsilateral, backwards and horizontal rotation) was performed, and a simple neural model based on these simulations developed. The simple neural model shows that the change in firing rate of the utricle is much larger in magnitude than the change in firing rates of all three semi-circular canals following a head tilt (except in a horizontal rotation). The data suggests that the change in utricular firing rate is a minimum 2-3 orders of magnitude larger than changes in firing rates of the canals during ipsilateral/backward tilts. Based on these results, the neural response recorded by the electrode in our EVestG recordings is expected to be dominated by the utricle in ipsilateral/backward tilts (It is important to note that the effect of the saccule and efferent signals were not taken into account in this model). If the utricle response dominates the EVestG recordings as the modeling results suggest, then EVestG has the potential to diagnose utricular hair cell damage due to a viral infection (which has been cited as one possible cause of Meniere's Disease).

A Step-wise Zoom Technique for Exploring Image-based Virtual Reality Applications

Existing image-based virtual reality applications allow users to view image-based 3D virtual environment in a more interactive manner. User could “walkthrough"; looks left, right, up and down and even zoom into objects in these virtual worlds of images. However what the user sees during a “zoom in" is just a close-up view of the same image which was taken from a distant. Thus, this does not give the user an accurate view of the object from the actual distance. In this paper, a simple technique for zooming in an object in a virtual scene is presented. The technique is based on the 'hotspot' concept in existing application. Instead of navigation between two different locations, the hotspots are used to focus into an object in the scene. For each object, several hotspots are created. A different picture is taken for each hotspot. Each consecutive hotspot created will take the user closer to the object. This will provide the user with a correct of view of the object based on his proximity to the object. Implementation issues and the relevance of this technique in potential application areas are highlighted.

Mapping Semantic Networks to Undirected Networks

There exists an injective, information-preserving function that maps a semantic network (i.e a directed labeled network) to a directed network (i.e. a directed unlabeled network). The edge label in the semantic network is represented as a topological feature of the directed network. Also, there exists an injective function that maps a directed network to an undirected network (i.e. an undirected unlabeled network). The edge directionality in the directed network is represented as a topological feature of the undirected network. Through function composition, there exists an injective function that maps a semantic network to an undirected network. Thus, aside from space constraints, the semantic network construct does not have any modeling functionality that is not possible with either a directed or undirected network representation. Two proofs of this idea will be presented. The first is a proof of the aforementioned function composition concept. The second is a simpler proof involving an undirected binary encoding of a semantic network.

Statistical Modeling of Accelerated Pavement Failure Using Response Surface Methodology

Rutting is one of the major load-related distresses in airport flexible pavements. Rutting in paving materials develop gradually with an increasing number of load applications, usually appearing as longitudinal depressions in the wheel paths and it may be accompanied by small upheavals to the sides. Significant research has been conducted to determine the factors which affect rutting and how they can be controlled. Using the experimental design concepts, a series of tests can be conducted while varying levels of different parameters, which could be the cause for rutting in airport flexible pavements. If proper experimental design is done, the results obtained from these tests can give a better insight into the causes of rutting and the presence of interactions and synergisms among the system variables which have influence on rutting. Although traditionally, laboratory experiments are conducted in a controlled fashion to understand the statistical interaction of variables in such situations, this study is an attempt to identify the critical system variables influencing airport flexible pavement rut depth from a statistical DoE perspective using real field data from a full-scale test facility. The test results do strongly indicate that the response (rut depth) has too much noise in it and it would not allow determination of a good model. From a statistical DoE perspective, two major changes proposed for this experiment are: (1) actual replication of the tests is definitely required, (2) nuisance variables need to be identified and blocked properly. Further investigation is necessary to determine possible sources of noise in the experiment.

Review of Scouring on Integral Bridge and its Possible Protection

The purpose of this paper is to summarize the following protection of scouring countermeasures by using Bentonite-Enhanced Sand (BES) mixtures. The concept of underground improvement is being used in this study to reduce the void of the sand. The sand bentonite mixture was used to bond the ground soil conditions surrounding the pile of integral bridge. The right composition of sand bentonite mixture was proposed based on previous findings. The swelling effect of bentonite also was investigated to ensure there is no adverse impact to the structure of the integral bridge. ScourScour, another name for severe erosion, occurs when the erosive capacity of water resulting from natural and manmade events exceeds the ability of earth materials to resist its effects. According to AASHTO LRFD Specifications (Section C3.7.5), scour is the most common reason for the collapse of highway bridges in the United States

Maya Semantic Technique: A Mathematical Technique Used to Determine Partial Semantics for Declarative Sentences

This research uses computational linguistics, an area of study that employs a computer to process natural language, and aims at discerning the patterns that exist in declarative sentences used in technical texts. The approach is mathematical, and the focus is on instructional texts found on web pages. The technique developed by the author and named the MAYA Semantic Technique is used here and organized into four stages. In the first stage, the parts of speech in each sentence are identified. In the second stage, the subject of the sentence is determined. In the third stage, MAYA performs a frequency analysis on the remaining words to determine the verb and its object. In the fourth stage, MAYA does statistical analysis to determine the content of the web page. The advantage of the MAYA Semantic Technique lies in its use of mathematical principles to represent grammatical operations which assist processing and accuracy if performed on unambiguous text. The MAYA Semantic Technique is part of a proposed architecture for an entire web-based intelligent tutoring system. On a sample set of sentences, partial semantics derived using the MAYA Semantic Technique were approximately 80% accurate. The system currently processes technical text in one domain, namely Cµ programming. In this domain all the keywords and programming concepts are known and understood.

PI Controller for Automatic Generation Control Based on Performance Indices

The optimal design of PI controller for Automatic Generation Control in two area is presented in this paper. The concept of Dual mode control is applied in the PI controller, such that the proportional mode is made active when the rate of change of the error is sufficiently larger than a specified limit otherwise switched to the integral mode. A digital simulation is used in conjunction with the Hooke-Jeeve’s optimization technique to determine the optimum parameters (individual gain of proportional and integral controller) of the PI controller. Integrated Square of the Error (ISE), Integrated Time multiplied by Absolute Error(ITAE) , and Integrated Absolute Error(IAE) performance indices are considered to measure the appropriateness of the designed controller.  The proposed controller are tested for a two area single nonreheat thermal system considering the practical aspect of the problem such as Deadband and Generation Rate Constraint(GRC). Simulation results show that  dual mode with optimized values of the gains improved the control performance than the commonly used Variable Structure .

A Grid-based Neural Network Framework for Multimodal Biometrics

Recent scientific investigations indicate that multimodal biometrics overcome the technical limitations of unimodal biometrics, making them ideally suited for everyday life applications that require a reliable authentication system. However, for a successful adoption of multimodal biometrics, such systems would require large heterogeneous datasets with complex multimodal fusion and privacy schemes spanning various distributed environments. From experimental investigations of current multimodal systems, this paper reports the various issues related to speed, error-recovery and privacy that impede the diffusion of such systems in real-life. This calls for a robust mechanism that caters to the desired real-time performance, robust fusion schemes, interoperability and adaptable privacy policies. The main objective of this paper is to present a framework that addresses the abovementioned issues by leveraging on the heterogeneous resource sharing capacities of Grid services and the efficient machine learning capabilities of artificial neural networks (ANN). Hence, this paper proposes a Grid-based neural network framework for adopting multimodal biometrics with the view of overcoming the barriers of performance, privacy and risk issues that are associated with shared heterogeneous multimodal data centres. The framework combines the concept of Grid services for reliable brokering and privacy policy management of shared biometric resources along with a momentum back propagation ANN (MBPANN) model of machine learning for efficient multimodal fusion and authentication schemes. Real-life applications would be able to adopt the proposed framework to cater to the varying business requirements and user privacies for a successful diffusion of multimodal biometrics in various day-to-day transactions.

A Web Pages Automatic Filtering System

This article describes a Web pages automatic filtering system. It is an open and dynamic system based on multi agents architecture. This system is built up by a set of agents having each a quite precise filtering task of to carry out (filtering process broken up into several elementary treatments working each one a partial solution). New criteria can be added to the system without stopping its execution or modifying its environment. We want to show applicability and adaptability of the multi-agents approach to the networks information automatic filtering. In practice, most of existing filtering systems are based on modular conception approaches which are limited to centralized applications which role is to resolve static data flow problems. Web pages filtering systems are characterized by a data flow which varies dynamically.

The Conceptualization of Integrated Consumer Health Informatics Utilization Framework

The purpose of this paper is to propose an integrated consumer health informatics utilization framework that can be used to gauge the online health information needs and usage patterns among Malaysian women. The proposed framework was developed based on four different theories/models: Use and Gratification Theory, Technology Acceptance 3 Model, Health Belief Model, and Multi-level Model of Information Seeking. The relevant constructs and research hypotheses are also presented in this paper. The framework will be tested in order for it to be used successfully to identify Malaysian women-s preferences of online health information resources and health information seeking activities.

Knowledge Impact on Measurement: A Conceptual Metric for Evaluating Performance Improvement (PI) at the Kuwait Institute for Scientific Research (KISR)

Research and development R&D work involves enormous amount of work that has to do with data measurement and collection. This process evolves as new information is fed, new technologies are utilized, and eventually new knowledge is created by the stakeholders i.e., researchers, clients, and end-users. When new knowledge is created, procedures of R&D work should evolve and produce better results within improved research skills and improved methods of data measurements and collection. This measurement improvement should then be benchmarked against a metric that should be developed at the organization. In this paper, we are suggesting a conceptual metric for R&D work performance improvement (PI) at the Kuwait Institute for Scientific Research (KISR). This PI is to be measured against a set of variables in the suggested metric, which are more closely correlated to organizational output, as opposed to organizational norms. The paper also mentions and discusses knowledge creation and management as an addedvalue to R&D work and measurement improvement. The research methodology followed in this work is qualitative in nature, based on a survey that was distributed to researchers and interviews held with senior researchers at KISR. Research and analyses in this paper also include looking at and analyzing KISR-s literature.

Embedded Semi-Fragile Signature Based Scheme for Ownership Identification and Color Image Authentication with Recovery

In this paper, a novel scheme is proposed for Ownership Identification and Color Image Authentication by deploying Cryptography & Digital Watermarking. The color image is first transformed from RGB to YST color space exclusively designed for watermarking. Followed by color space transformation, each channel is divided into 4×4 non-overlapping blocks with selection of central 2×2 sub-blocks. Depending upon the channel selected two to three LSBs of each central 2×2 sub-block are set to zero to hold the ownership, authentication and recovery information. The size & position of sub-block is important for correct localization, enhanced security & fast computation. As YS ÔèÑ T so it is suitable to embed the recovery information apart from the ownership and authentication information, therefore 4×4 block of T channel along with ownership information is then deployed by SHA160 to compute the content based hash that is unique and invulnerable to birthday attack or hash collision instead of using MD5 that may raise the condition i.e. H(m)=H(m'). For recovery, intensity mean of 4x4 block of each channel is computed and encoded upto eight bits. For watermark embedding, key based mapping of blocks is performed using 2DTorus Automorphism. Our scheme is oblivious, generates highly imperceptible images with correct localization of tampering within reasonable time and has the ability to recover the original work with probability of near one.

A Review of Quality Relationship between IT Processes, IT Products and IT Services

Producing IT products/services required carefully designed. IT development process is intangible and labour intensive. Making optimal use of available resources, both soft (knowledge, skill-set etc.) and hard (computer system, ancillary equipment etc.), is vital if IT development is to achieve sensible economical advantages. Apart from the norm of Project Life Cycle and System Development Life Cycle (SDLC), there is an urgent need to establish a general yet widely acceptable guideline on the most effective and efficient way to precede an IT project in the broader view of Product Life Cycle. The current paper proposes such a framework with two major areas of concern: (1) an integration of IT Products and IT Services within an existing IT Process architecture and; (2) how IT Product and IT Services are built into the framework of Product Life Cycle, Project Life Cycle and SDLC.

Using Quality Models to Evaluate National ID systems: the Case of the UAE

This paper presents findings from the evaluation study carried out to review the UAE national ID card software. The paper consults the relevant literature to explain many of the concepts and frameworks explained herein. The findings of the evaluation work that was primarily based on the ISO 9126 standard for system quality measurement highlighted many practical areas that if taken into account is argued to more likely increase the success chances of similar system implementation projects.