Response Quality Evaluation in Heterogeneous Question Answering System: A Black-box Approach

The evaluation of the question answering system is a major research area that needs much attention. Before the rise of domain-oriented question answering systems based on natural language understanding and reasoning, evaluation is never a problem as information retrieval-based metrics are readily available for use. However, when question answering systems began to be more domains specific, evaluation becomes a real issue. This is especially true when understanding and reasoning is required to cater for a wider variety of questions and at the same time achieve higher quality responses The research in this paper discusses the inappropriateness of the existing measure for response quality evaluation and in a later part, the call for new standard measures and the related considerations are brought forward. As a short-term solution for evaluating response quality of heterogeneous systems, and to demonstrate the challenges in evaluating systems of different nature, this research presents a black-box approach using observation, classification scheme and a scoring mechanism to assess and rank three example systems (i.e. AnswerBus, START and NaLURI).

Numerical Investigation on Latent Heat Storage Unit of Different Configurations

The storage of thermal energy as a latent heat of phase change material (PCM) has created considerable interest among researchers in recent times. Here, an attempt is made to carry out numerical investigations to analyze the performance of latent heat storage units (LHSU) employing phase change material. The mathematical model developed is based on an enthalpy formulation. Freezing time of PCM packed in three different shaped containers viz. rectangular, cylindrical and cylindrical shell is compared. The model is validated with the results available in the literature. Results show that for the same mass of PCM and surface area of heat transfer, cylindrical shell container takes the least time for freezing the PCM and this geometric effect is more pronounced with an increase in the thickness of the shell than that of length of the shell.

Treatment of Paper and Pulp Mill Effluent by Coagulation

The pulp and paper mill effluent is one of the high polluting effluent amongst the effluents obtained from polluting industries. All the available methods for treatment of pulp and paper mill effluent have certain drawbacks. The coagulation is one of the cheapest process for treatment of various organic effluents. Thus, the removal of chemical oxygen demand (COD) and colour of paper mill effluent is studied using coagulation process. The batch coagulation process was performed using various coagulants like: aluminium chloride, poly aluminium chloride and copper sulphate. The initial pH of the effluent (Coagulation pH) has tremendous effect on COD and colour removal. Poly aluminium chloride (PAC) as coagulant reduced COD to 84 % and 92 % of colour was removed at an optimum pH 5 and coagulant dose of 8 ml l-1. With aluminium chloride at an optimum pH = 4 and coagulant dose of 5 g l-1, 74 % COD and 86 % colour removal were observed. The results using copper sulphate as coagulant (a less commercial coagulant) were encouraging. At an optimum pH 6 and mass loading of 5 g l-1, 76 % COD reduction and 78 % colour reduction were obtained. It was also observed that after addition of coagulant, the pH of the effluent decreases. The decrease in pH was highest for AlCl3, which was followed by PAC and CuSO4. Significant amount of COD reductions was obtained by coagulation process. Since the coagulation process is the first stage for treatment of effluent and some of the coagulant cations usually remain in the treated effluents. Thus, cation like copper may be one of the good catalyst for second stage of treatment process like wet oxidation. The copper has been found to be good oxidation catalyst then iron and aluminum.

Support Vector Machine based Intelligent Watermark Decoding for Anticipated Attack

In this paper, we present an innovative scheme of blindly extracting message bits from an image distorted by an attack. Support Vector Machine (SVM) is used to nonlinearly classify the bits of the embedded message. Traditionally, a hard decoder is used with the assumption that the underlying modeling of the Discrete Cosine Transform (DCT) coefficients does not appreciably change. In case of an attack, the distribution of the image coefficients is heavily altered. The distribution of the sufficient statistics at the receiving end corresponding to the antipodal signals overlap and a simple hard decoder fails to classify them properly. We are considering message retrieval of antipodal signal as a binary classification problem. Machine learning techniques like SVM is used to retrieve the message, when certain specific class of attacks is most probable. In order to validate SVM based decoding scheme, we have taken Gaussian noise as a test case. We generate a data set using 125 images and 25 different keys. Polynomial kernel of SVM has achieved 100 percent accuracy on test data.

Improvement of Stator Slot Structure based on Insulation Stresses Analysis in HV Generator

High voltage generators are being subject to higher voltage rating and are being designed to operate in harsh conditions. Stator windings are the main component of generators in which Electrical, magnetical and thermal stresses remain major failures for insulation degradation accelerated aging. A large number of generators failed due to stator winding problems, mainly insulation deterioration. Insulation degradation assessment plays vital role in the asset life management. Mostly the stator failure is catastrophic causing significant damage to the plant. Other than generation loss, stator failure involves heavy repair or replacement cost. Electro thermal analysis is the main characteristic for improvement design of stator slot-s insulation. Dielectric parameters such as insulation thickness, spacing, material types, geometry of winding and slot are major design consideration. A very powerful method available to analyze electro thermal performance is Finite Element Method (FEM) which is used in this paper. The analysis of various stator coil and slot configurations are used to design the better dielectric system to reduce electrical and thermal stresses in order to increase the power of generator in the same volume of core. This paper describes the process used to perform classical design and improvement analysis of stator slot-s insulation.

Roll of Membership functions in Fuzzy Logic for Prediction of Shoot Length of Mustard Plant Based on Residual Analysis

The selection for plantation of a particular type of mustard plant depending on its productivity (pod yield) at the stage of maturity. The growth of mustard plant dependent on some parameters of that plant, these are shoot length, number of leaves, number of roots and roots length etc. As the plant is growing, some leaves may be fall down and some new leaves may come, so it can not gives the idea to develop the relationship with the seeds weight at mature stage of that plant. It is not possible to find the number of roots and root length of mustard plant at growing stage that will be harmful of this plant as roots goes deeper to deeper inside the land. Only the value of shoot length which increases in course of time can be measured at different time instances. Weather parameters are maximum and minimum humidity, rain fall, maximum and minimum temperature may effect the growth of the plant. The parameters of pollution, water, soil, distance and crop management may be dominant factors of growth of plant and its productivity. Considering all parameters, the growth of the plant is very uncertain, fuzzy environment can be considered for the prediction of shoot length at maturity of the plant. Fuzzification plays a greater role for fuzzification of data, which is based on certain membership functions. Here an effort has been made to fuzzify the original data based on gaussian function, triangular function, s-function, Trapezoidal and L –function. After that all fuzzified data are defuzzified to get normal form. Finally the error analysis (calculation of forecasting error and average error) indicates the membership function appropriate for fuzzification of data and use to predict the shoot length at maturity. The result is also verified using residual (Absolute Residual, Maximum of Absolute Residual, Mean Absolute Residual, Mean of Mean Absolute Residual, Median of Absolute Residual and Standard Deviation) analysis.

Comparison of Compression Ability Using DCT and Fractal Technique on Different Imaging Modalities

Image compression is one of the most important applications Digital Image Processing. Advanced medical imaging requires storage of large quantities of digitized clinical data. Due to the constrained bandwidth and storage capacity, however, a medical image must be compressed before transmission and storage. There are two types of compression methods, lossless and lossy. In Lossless compression method the original image is retrieved without any distortion. In lossy compression method, the reconstructed images contain some distortion. Direct Cosine Transform (DCT) and Fractal Image Compression (FIC) are types of lossy compression methods. This work shows that lossy compression methods can be chosen for medical image compression without significant degradation of the image quality. In this work DCT and Fractal Compression using Partitioned Iterated Function Systems (PIFS) are applied on different modalities of images like CT Scan, Ultrasound, Angiogram, X-ray and mammogram. Approximately 20 images are considered in each modality and the average values of compression ratio and Peak Signal to Noise Ratio (PSNR) are computed and studied. The quality of the reconstructed image is arrived by the PSNR values. Based on the results it can be concluded that the DCT has higher PSNR values and FIC has higher compression ratio. Hence in medical image compression, DCT can be used wherever picture quality is preferred and FIC is used wherever compression of images for storage and transmission is the priority, without loosing picture quality diagnostically.

A Dynamic RGB Intensity Based Steganography Scheme

Steganography meaning covered writing. Steganography includes the concealment of information within computer files [1]. In other words, it is the Secret communication by hiding the existence of message. In this paper, we will refer to cover image, to indicate the images that do not yet contain a secret message, while we will refer to stego images, to indicate an image with an embedded secret message. Moreover, we will refer to the secret message as stego-message or hidden message. In this paper, we proposed a technique called RGB intensity based steganography model as RGB model is the technique used in this field to hide the data. The methods used here are based on the manipulation of the least significant bits of pixel values [3][4] or the rearrangement of colors to create least significant bit or parity bit patterns, which correspond to the message being hidden. The proposed technique attempts to overcome the problem of the sequential fashion and the use of stego-key to select the pixels.

Implementation of Response Surface Methodology using in Small Brown Rice Peeling Machine: Part I

Implementation of response surface methodology (RSM) was employed to study the effects of two factor (rubber clearance and round per minute) in brown rice peeling machine of The optimal BROKENS yield (19.02, average of three repeats),.The optimized composition derived from RSM regression was analyzed using Regression analysis and Analysis of Variance (ANOVA). At a significant level α = 0.05, the values of Regression coefficient, R 2 (adj)were 97.35 % and standard deviation were 1.09513. The independent variables are initial rubber clearance, and round per minute parameters namely. The investigating responses are final rubber clearance, and round per minute (RPM). The restriction of the optimization is the designated.

Design and Simulation of Electromagnetic Flow Meter for Circular Pipe Type

Electromagnetic flow meter by measuring the varying of magnetic flux, which is related to the velocity of conductive flow, can measure the rate of fluids very carefully and precisely. Electromagnetic flow meter operation is based on famous Faraday's second Law. In these equipments, the constant magnetostatic field is produced by electromagnet (winding around the tube) outside of pipe and inducting voltage that is due to conductive liquid flow is measured by electrodes located on two end side of the pipe wall. In this research, we consider to 2-dimensional mathematical model that can be solved by numerical finite difference (FD) solution approach to calculate induction potential between electrodes. The fundamental concept to design the electromagnetic flow meter, exciting winding and simulations are come out by using MATLAB and PDE-Tool software. In the last stage, simulations results will be shown for improvement and accuracy of technical provision.

Architecture of Speech-based Registration System

In this era of technology, fueled by the pervasive usage of the internet, security is a prime concern. The number of new attacks by the so-called “bots", which are automated programs, is increasing at an alarming rate. They are most likely to attack online registration systems. Technology, called “CAPTCHA" (Completely Automated Public Turing test to tell Computers and Humans Apart) do exist, which can differentiate between automated programs and humans and prevent replay attacks. Traditionally CAPTCHA-s have been implemented with the challenge involved in recognizing textual images and reproducing the same. We propose an approach where the visual challenge has to be read out from which randomly selected keywords are used to verify the correctness of spoken text and in turn detect the presence of human. This is supplemented with a speaker recognition system which can identify the speaker also. Thus, this framework fulfills both the objectives – it can determine whether the user is a human or not and if it is a human, it can verify its identity.

Influence of Port Geometry on Thrust Transient of Solid Propellant Rockets at Liftoff

Numerical studies have been carried out using a two dimensional code to examine the influence of pressure / thrust transient of solid propellant rockets at liftoff. This code solves unsteady Reynolds-averaged thin-layer Navier–Stokes equations by an implicit LU-factorization time-integration method. The results from the parametric study indicate that when the port is narrow there is a possibility of increase in pressure / thrust-rise rate due to relatively high flame spread rate. Parametric studies further reveal that flame spread rate can be altered by altering the propellant properties, igniter jet characteristics and nozzle closure burst pressure without altering the grain configuration and/or the mission demanding thrust transient. We observed that when the igniter turbulent intensity is relatively low the vehicle could liftoff early due to the early flow choking of the rocket nozzle. We concluded that the high pressurization-rate has structural implications at liftoff in addition to transient burning effect. Therefore prudent selection of the port geometry and the igniter, for meeting the mission requirements, within the given envelop are meaningful objectives for any designer for the smooth liftoff of solid propellant rockets.

Attack Detection through Image Adaptive Self Embedding Watermarking

Now a days, a significant part of commercial and governmental organisations like museums, cultural organizations, libraries, commercial enterprises, etc. invest intensively in new technologies for image digitization, digital libraries, image archiving and retrieval. Hence image authorization, authentication and security has become prime need. In this paper, we present a semi-fragile watermarking scheme for color images. The method converts the host image into YIQ color space followed by application of orthogonal dual domains of DCT and DWT transforms. The DCT helps to separate relevant from irrelevant image content to generate silent image features. DWT has excellent spatial localisation to help aid in spatial tamper characterisation. Thus image adaptive watermark is generated based of image features which allows the sharp detection of microscopic changes to locate modifications in the image. Further, the scheme utilises the multipurpose watermark consisting of soft authenticator watermark and chrominance watermark. Which has been proved fragile to some predefined processing like intentinal fabrication of the image or forgery and robust to other incidental attacks caused in the communication channel.

Approaches and Schemes for Storing DTDIndependent XML Data in Relational Databases

The volume of XML data exchange is explosively increasing, and the need for efficient mechanisms of XML data management is vital. Many XML storage models have been proposed for storing XML DTD-independent documents in relational database systems. Benchmarking is the best way to highlight pros and cons of different approaches. In this study, we use a common benchmarking scheme, known as XMark to compare the most cited and newly proposed DTD-independent methods in terms of logical reads, physical I/O, CPU time and duration. We show the effect of Label Path, extracting values and storing in another table and type of join needed for each method-s query answering.

Application of Artificial Neural Network to Forecast Actual Cost of a Project to Improve Earned Value Management System

This paper presents an application of Artificial Neural Network (ANN) to forecast actual cost of a project based on the earned value management system (EVMS). For this purpose, some projects randomly selected based on the standard data set , and it is produced necessary progress data such as actual cost ,actual percent complete , baseline cost and percent complete for five periods of project. Then an ANN with five inputs and five outputs and one hidden layer is trained to produce forecasted actual costs. The comparison between real and forecasted data show better performance based on the Mean Absolute Percentage Error (MAPE) criterion. This approach could be applicable to better forecasting the project cost and result in decreasing the risk of project cost overrun, and therefore it is beneficial for planning preventive actions.

Framework of TAZ_OPT Model for Ambulance Location and Allocation Problem

Our study is concerned with the development of an Emergency Medical Services (EMS) ambulance location and allocation model called the Time-based Ambulance Zoning Optimization Model (TAZ_OPT). This paper presents the framework of the study. The model is formulated using the goal programming (GP), where the goals are to determine the satellite locations of ambulances and the number of ambulances to be allocated at these locations. The model aims at maximizing the expected demand coverage based on probability of reaching the emergency location within targetted time, and minimizing the ambulance busyness likelihood value. Among the benefits of the model is the increased accessibility and availability of ambulances, thus, enhanced quality of the EMS ambulance services.

Exploring the Professional Competency Contents for International Marketer in Taiwan

The main purpose of this study was to establish Professional Competency Contents for International Marketer in Taiwan. To establish these contents a set of interviews with international marketing managers and three rounds of Delphi Technique surveys were employed. Five international marketing managers were interviewed for discussions on definitions, framework, and items of international marketing competency. A questionnaire for the " Delphi Technique Survey " was developed based on the results acquired from the interviews. The resulting questionnaire was distributed to another group of 30 international marketer of trading companies in Taiwan. After three rounds of Delphi Technique Survey with these participants, the "Contents of Professional Competency for International Marketer " was established. Five dimensions and thirty indicators were identified. It is hoped that the proposed contents could be served as a self-evaluation tool for international marketer as well as the basis for staffing and training programs for international marketer in Taiwan.

Making Businesses Work Smarter with Mobile Business Intelligence

Through the course of this paper we outline how mobile Business Intelligence (m-BI) can help businesses to work smarter and to improve their agility. When we analyze the industry from the usage perspective or how interaction with the enterprise BI system happens via mobile devices, we may easily understand that there are two major types of mobile BI: passive and active. Active mobile BI gives provisions for users to interact with the BI systems on-the-fly. Active mobile business intelligence often works as a combination of both “push and pull" techniques. Some mistakes were done in the up-to-day progress of mobile technologies and mobile BI, as well as some problems that still have to be resolved. We discussed in the paper rather broadly.

Comparison of Evolutionary Algorithms and their Hybrids Applied to MarioAI

Researchers have been applying artificial/ computational intelligence (AI/CI) methods to computer games. In this research field, further researchesare required to compare AI/CI methods with respect to each game application. In thispaper, we report our experimental result on the comparison of evolution strategy, genetic algorithm and their hybrids, applied to evolving controller agents for MarioAI. GA revealed its advantage in our experiment, whereas the expected ability of ES in exploiting (fine-tuning) solutions was not clearly observed. The blend crossover operator and the mutation operator of GA might contribute well to explore the vast search space.

A Markov Chain Approximation for ATS Modeling for the Variable Sampling Interval CCC Control Charts

The cumulative conformance count (CCC) charts are widespread in process monitoring of high-yield manufacturing. Recently, it is found the use of variable sampling interval (VSI) scheme could further enhance the efficiency of the standard CCC charts. The average time to signal (ATS) a shift in defect rate has become traditional measure of efficiency of a chart with the VSI scheme. Determining the ATS is frequently a difficult and tedious task. A simple method based on a finite Markov Chain approach for modeling the ATS is developed. In addition, numerical results are given.