Analytical Model Based Evaluation of Human Machine Interfaces Using Cognitive Modeling

Cognitive models allow predicting some aspects of utility and usability of human machine interfaces (HMI), and simulating the interaction with these interfaces. The action of predicting is based on a task analysis, which investigates what a user is required to do in terms of actions and cognitive processes to achieve a task. Task analysis facilitates the understanding of the system-s functionalities. Cognitive models are part of the analytical approaches, that do not associate the users during the development process of the interface. This article presents a study about the evaluation of a human machine interaction with a contextual assistant-s interface using ACTR and GOMS cognitive models. The present work shows how these techniques may be applied in the evaluation of HMI, design and research by emphasizing firstly the task analysis and secondly the time execution of the task. In order to validate and support our results, an experimental study of user performance is conducted at the DOMUS laboratory, during the interaction with the contextual assistant-s interface. The results of our models show that the GOMS and ACT-R models give good and excellent predictions respectively of users performance at the task level, as well as the object level. Therefore, the simulated results are very close to the results obtained in the experimental study.

An Owl Ontology for Commonkads Template Knowledge Models

This paper gives an overview of how an OWL ontology has been created to represent template knowledge models defined in CML that are provided by CommonKADS. CommonKADS is a mature knowledge engineering methodology which proposes the use of template knowledge model for knowledge modelling. The aim of developing this ontology is to present the template knowledge model in a knowledge representation language that can be easily understood and shared in the knowledge engineering community. Hence OWL is used as it has become a standard for ontology and also it already has user friendly tools for viewing and editing.

Multi Switched Split Vector Quantization of Narrowband Speech Signals

Vector quantization is a powerful tool for speech coding applications. This paper deals with LPC Coding of speech signals which uses a new technique called Multi Switched Split Vector Quantization (MSSVQ), which is a hybrid of Multi, switched, split vector quantization techniques. The spectral distortion performance, computational complexity, and memory requirements of MSSVQ are compared to split vector quantization (SVQ), multi stage vector quantization(MSVQ) and switched split vector quantization (SSVQ) techniques. It has been proved from results that MSSVQ has better spectral distortion performance, lower computational complexity and lower memory requirements when compared to all the above mentioned product code vector quantization techniques. Computational complexity is measured in floating point operations (flops), and memory requirements is measured in (floats).

Maximum Norm Analysis of a Nonmatching Grids Method for Nonlinear Elliptic Boundary Value Problem −Δu = f(u)

We provide a maximum norm analysis of a finite element Schwarz alternating method for a nonlinear elliptic boundary value problem of the form -Δu = f(u), on two overlapping sub domains with non matching grids. We consider a domain which is the union of two overlapping sub domains where each sub domain has its own independently generated grid. The two meshes being mutually independent on the overlap region, a triangle belonging to one triangulation does not necessarily belong to the other one. Under a Lipschitz assumption on the nonlinearity, we establish, on each sub domain, an optimal L∞ error estimate between the discrete Schwarz sequence and the exact solution of the boundary value problem.

Coping with the Rapidity of Information Technology Changes – A Comparison Reviewon Current Practices

Information technology managers nowadays are facing with tremendous pressure to plan, implement, and adopt new technology solution due to the rapidity of technology changes. Resulted from a lack of study that have been done in this topic, the aim of this paper is to provide a comparison review on current tools that are currently being used in order to respond to technological changes. The study is based on extensive literature review of published works with majority of them are ranging from 2000 to the first part of 2011. The works were gathered from journals, books, and other information sources available on the Web. Findings show that, each tools has different focus and none of the tools are providing a framework in holistic view, which should include technical, people, process, and business environment aspect. Hence, this result provides potential information about current available tools that IT managers could use to manage changes in technology. Further, the result reveals a research gap in the area where the industries a short of such framework.

Development of NOx Emission Model for a Tangentially Fired Acid Incinerator

This paper aims to develop a NOx emission model of an acid gas incinerator using Nelder-Mead least squares support vector regression (LS-SVR). Malaysia DOE is actively imposing the Clean Air Regulation to mandate the installation of analytical instrumentation known as Continuous Emission Monitoring System (CEMS) to report emission level online to DOE . As a hardware based analyzer, CEMS is expensive, maintenance intensive and often unreliable. Therefore, software predictive technique is often preferred and considered as a feasible alternative to replace the CEMS for regulatory compliance. The LS-SVR model is built based on the emissions from an acid gas incinerator that operates in a LNG Complex. Simulated Annealing (SA) is first used to determine the initial hyperparameters which are then further optimized based on the performance of the model using Nelder-Mead simplex algorithm. The LS-SVR model is shown to outperform a benchmark model based on backpropagation neural networks (BPNN) in both training and testing data.

Experimental Analysis on Electrical and Photometric Performances of Commercially Available Integrated Compact Fluorescent Lamp

Lighting upgrades involve relatively lower costs which allow the benefits to be spread more widely than is possible with any other energy efficiency measure. In order to popularize the adoption of CFL in Taiwan, the authority proposes to implement a new energy efficient lamp comparative label system. The current study was accordingly undertaken to investigate the factors affecting the performance and the deviation of actual and labeled performance of commercially available integrated CFLs. In this paper, standard test methods to determine the electrical and photometric performances of CFL were developed based on CIE 84-1989 and CIE 60901-1987, then 55 selected CFLs from market were tested. The results show that with higher color temperature of CFLs lower efficacy are achieved. It was noticed that the most packaging of CFL often lack the information of Color Rendering Index. Also, there was no correlation between price and performance of the CFLs was indicated in this work. The results of this paper might help consumers to make more informed CFL-purchasing decisions.

Analysis of the Islands Tourists, Destination Information Sources and Service Satisfaction

The purpose of this study is to analyze the islands tourist travel information sources, as well as for the satisfaction of the tourist destination services. This study used questionnaires to the island of Taiwan to the Penghu Islands to engage in tourism activities tourist adopt the designated convenience sampling method, a total of 889 valid questionnaires were collected. After statistical analysis, this study found that: 1. tourists to the Penghu Islands travel information source for “friends and family came to Penghu". 2. Tourists feel the service of the outlying islands of Penghu, the highest feelings of “friendly local residents". 3. There are different demographic variables affect the tourist travel information source and service satisfaction. Based on the findings of this study not only for Penghu's tourism industry with the unit in charge of the proposed operating and suggestions for future research to other researchers.

Handling Mobility using Virtual Grid in Static Wireless Sensor Networks

Querying a data source and routing data towards sink becomes a serious challenge in static wireless sensor networks if sink and/or data source are mobile. Many a times the event to be observed either moves or spreads across wide area making maintenance of continuous path between source and sink a challenge. Also, sink can move while query is being issued or data is on its way towards sink. In this paper, we extend our already proposed Grid Based Data Dissemination (GBDD) scheme which is a virtual grid based topology management scheme restricting impact of movement of sink(s) and event(s) to some specific cells of a grid. This obviates the need for frequent path modifications and hence maintains continuous flow of data while minimizing the network energy consumptions. Simulation experiments show significant improvements in network energy savings and average packet delay for a packet to reach at sink.

Implementation of Watch Dog Timer for Fault Tolerant Computing on Cluster Server

In today-s new technology era, cluster has become a necessity for the modern computing and data applications since many applications take more time (even days or months) for computation. Although after parallelization, computation speeds up, still time required for much application can be more. Thus, reliability of the cluster becomes very important issue and implementation of fault tolerant mechanism becomes essential. The difficulty in designing a fault tolerant cluster system increases with the difficulties of various failures. The most imperative obsession is that the algorithm, which avoids a simple failure in a system, must tolerate the more severe failures. In this paper, we implemented the theory of watchdog timer in a parallel environment, to take care of failures. Implementation of simple algorithm in our project helps us to take care of different types of failures; consequently, we found that the reliability of this cluster improves.

Seismic Alert System based on Artificial Neural Networks

We board the problem of creating a seismic alert system, based upon artificial neural networks, trained by using the well-known back-propagation and genetic algorithms, in order to emit the alarm for the population located into a specific city, about an eminent earthquake greater than 4.5 Richter degrees, and avoiding disasters and human loses. In lieu of using the propagation wave, we employed the magnitude of the earthquake, to establish a correlation between the recorded magnitudes from a controlled area and the city, where we want to emit the alarm. To measure the accuracy of the posed method, we use a database provided by CIRES, which contains the records of 2500 quakes incoming from the State of Guerrero and Mexico City. Particularly, we performed the proposed method to generate an issue warning in Mexico City, employing the magnitudes recorded in the State of Guerrero.

Development of an Immunoassay Platform for Diagnosis of Acute Kidney Injury

Acute kidney injury (AKI) is a new worldwide public health problem. A diagnosis of this disease using creatinine is still a problem in clinical practice. Therefore, a measurement of biomarkers responsible for AKI has received much attention in the past couple years. Cytokine interleukin-18 (IL-18) was reported as one of the early biomarkers for AKI. The most commonly used method to detect this biomarker is an immunoassay. This study used a planar platform to perform an immunoassay using fluorescence for detection. In this study, anti-IL-18 antibody was immobilized onto a microscope slide using a covalent binding method. Make-up samples were diluted at the concentration between 10 to 1000 pg/ml to create a calibration curve. The precision of the system was determined using a coefficient of variability (CV), which was found to be less than 10%. The performance of this immunoassay system was compared with the measurement from ELISA.

Fast Depth Estimation with Filters

Fast depth estimation from binocular vision is often desired for autonomous vehicles, but, most algorithms could not easily be put into practice because of the much time cost. We present an image-processing technique that can fast estimate depth image from binocular vision images. By finding out the lines which present the best matched area in the disparity space image, the depth can be estimated. When detecting these lines, an edge-emphasizing filter is used. The final depth estimation will be presented after the smooth filter. Our method is a compromise between local methods and global optimization.

Power System Contingency Analysis Using Multiagent Systems

The demand of the energy management systems (EMS) set forth by modern power systems requires fast energy management systems. Contingency analysis is among the functions in EMS which is time consuming. In order to handle this limitation, this paper introduces agent based technology in the contingency analysis. The main function of agents is to speed up the performance. Negotiations process in decision making is explained and the issue set forth is the minimization of the operating costs. The IEEE 14 bus system and its line outage have been used in the research and simulation results are presented.

Development of Mathematical Model for Overall Oxygen Transfer Coefficient of an Aerator and Comparison with CFD Modeling

The value of overall oxygen transfer Coefficient (KLa), which is the best measure of oxygen transfer in water through aeration, is obtained by a simple approach, which sufficiently explains the utility of the method to eliminate the discrepancies due to inaccurate assumption of saturation dissolved oxygen concentration. The rate of oxygen transfer depends on number of factors like intensity of turbulence, which in turns depends on the speed of rotation, size, and number of blades, diameter and immersion depth of the rotor, and size and shape of aeration tank, as well as on physical, chemical, and biological characteristic of water. An attempt is made in this paper to correlate the overall oxygen transfer Coefficient (KLa), as an independent parameter with other influencing parameters mentioned above. It has been estimated that the simulation equation developed predicts the values of KLa and power with an average standard error of estimation of 0.0164 and 7.66 respectively and with R2 values of 0.979 and 0.989 respectively, when compared with experimentally determined values. The comparison of this model is done with the model generated using Computational fluid dynamics (CFD) and both the models were found to be in good agreement with each other.

An Idea About How to Teach OO-Programming to Students

Object-oriented programming is a wonderful way to make programming of huge real life tasks much easier than by using procedural languages. In order to teach those ideas to students, it is important to find a good task that shows the advantages of OOprogramming very naturally. This paper gives an example, the game Battleship, which seems to work excellent for teaching the OO ideas (using Java, [1], [2], [3], [4]). A three-step task is presented for how to teach OO-programming using just one example suitable to convey many of the OO ideas. Observations are given at the end and conclusions about how the whole teaching course worked out.

Effect of Physical Contact (Hand-Holding) on Heart Rate Variability

Heart-s electric field can be measured anywhere on the surface of the body (ECG). When individuals touch, one person-s ECG signal can be registered in other person-s EEG and elsewhere on his body. Now, the aim of this study was to test the hypothesis that physical contact (hand-holding) of two persons changes their heart rate variability. Subjects were sixteen healthy female (age: 20- 26) which divided into eight sets. In each sets, we had two friends that they passed intimacy test of J.sternberg. ECG of two subjects (each set) acquired for 5 minutes before hand-holding (as control group) and 5 minutes during they held their hands (as experimental group). Then heart rate variability signals were extracted from subjects' ECG and analyzed in linear feature space (time and frequency domain) and nonlinear feature space. Considering the results, we conclude that physical contact (hand-holding of two friends) increases parasympathetic activity, as indicate by increase SD1, SD1/SD2, HF and MF power (p

Partial Oxidation of Methane in the Pulsed Compression Reactor: Experiments and Simulation

The Pulsed Compression Reactor promises to be a compact, economical and energy efficient alternative to conventional chemical reactors. In this article, the production of synthesis gas using the Pulsed Compression Reactor is investigated. This is done experimentally as well as with simulations. The experiments are done by means of a single shot reactor, which replicates a representative, single reciprocation of the Pulsed Compression Reactor with great control over the reactant composition, reactor temperature and pressure and temperature history. Simulations are done with a relatively simple method, which uses different models for the chemistry and thermodynamic properties of the species in the reactor. Simulation results show very good agreement with the experimental data, and give great insight into the reaction processes that occur within the cycle.

Data Acquisition from Cell Phone using Logical Approach

Cell phone forensics to acquire and analyze data in the cellular phone is nowadays being used in a national investigation organization and a private company. In order to collect cellular phone flash memory data, we have two methods. Firstly, it is a logical method which acquires files and directories from the file system of the cell phone flash memory. Secondly, we can get all data from bit-by-bit copy of entire physical memory using a low level access method. In this paper, we describe a forensic tool to acquire cell phone flash memory data using a logical level approach. By our tool, we can get EFS file system and peek memory data with an arbitrary region from Korea CDMA cell phone.

A Comparative Study on Eastern and Western Wedding Ceremonies in Korean Films and Hollywood Films

As an adult man and woman love each other and come to have faith in each other as their spouse, they marry each other. Recently people-s economic life has become individualized and women are enjoying a high education level and increased participation in social activities, and these changes are creating environment favorable for single life. Thus, an increasing number of people are choosing celibacy, and many people prefer cohabitation to marriage. Nevertheless, marriage is still regarded as a must-to-do in our thought. Most of people throughout the world admit marriage as one of natural processes of life, and is an important passage rite in life that all people experience as we can see everywhere in the world despite the diversity of lifestyles. With regard to wedding ceremony, however, each country and culture has its own unique tradition and style of festival. It is not just a congratulatory ceremony but contains multiple concepts representing the age, country or culture. Moreover, the form and contents of wedding ceremony changes over time, and such features of wedding ceremony are well represented in films. This study took note of the fact that films reflect and reproduce each country-s historicity, culturality and analyzed four films, which are believed to show differences between Eastern and Western wedding ceremonies. The selected films are: A Perfect Match (2002), Marriage Is a Crazy Thing (2001), Bride Wars (2009) and 27 Dresses (2008). The author attempted to examine wedding ceremonies described in the four films, differences between the East and the West suggested by the films, and changes in their societies.