A Methodology for Data Migration between Different Database Management Systems

In present days the area of data migration is very topical. Current tools for data migration in the area of relational database have several disadvantages that are presented in this paper. We propose a methodology for data migration of the database tables and their data between various types of relational database systems (RDBMS). The proposed methodology contains an expert system. The expert system contains a knowledge base that is composed of IFTHEN rules and based on the input data suggests appropriate data types of columns of database tables. The proposed tool, which contains an expert system, also includes the possibility of optimizing the data types in the target RDBMS database tables based on processed data of the source RDBMS database tables. The proposed expert system is shown on data migration of selected database of the source RDBMS to the target RDBMS.

Evolving a Fuzzy Rule-Base for Image Segmentation

A new method for color image segmentation using fuzzy logic is proposed in this paper. Our aim here is to automatically produce a fuzzy system for color classification and image segmentation with least number of rules and minimum error rate. Particle swarm optimization is a sub class of evolutionary algorithms that has been inspired from social behavior of fishes, bees, birds, etc, that live together in colonies. We use comprehensive learning particle swarm optimization (CLPSO) technique to find optimal fuzzy rules and membership functions because it discourages premature convergence. Here each particle of the swarm codes a set of fuzzy rules. During evolution, a population member tries to maximize a fitness criterion which is here high classification rate and small number of rules. Finally, particle with the highest fitness value is selected as the best set of fuzzy rules for image segmentation. Our results, using this method for soccer field image segmentation in Robocop contests shows 89% performance. Less computational load is needed when using this method compared with other methods like ANFIS, because it generates a smaller number of fuzzy rules. Large train dataset and its variety, makes the proposed method invariant to illumination noise

Stability of Functionally Graded Beams with Piezoelectric Layers Based on the First Order Shear Deformation Theory

Stability of functionally graded beams with piezoelectric layers subjected to axial compressive load that is simply supported at both ends is studied in this paper. The displacement field of beam is assumed based on first order shear deformation beam theory. Applying the Hamilton's principle, the governing equation is established. The influences of applied voltage, dimensionless geometrical parameter, functionally graded index and piezoelectric thickness on the critical buckling load of beam are presented. To investigate the accuracy of the present analysis, a compression study is carried out with a known data.

Existence of Solution for Singular Two-point Boundary Value Problem of Second-order Differential Equation

In this paper, by constructing a special set and utilizing fixed point theory in coin, we study the existence of solution of singular two point’s boundary value problem for second-order differential equation, which improved and generalize the result of related paper.

Implementation of On-Line Cutting Stock Problem on NC Machines

Introduction applicability of high-speed cutting stock problem (CSP) is presented in this paper. Due to the orders continued coming in from various on-line ways for a professional cutting company, to stay competitive, such a business has to focus on sustained production at high levels. In others words, operators have to keep the machine running to stay ahead of the pack. Therefore, the continuous stock cutting problem with setup is proposed to minimize the cutting time and pattern changing time to meet the on-line given demand. In this paper, a novel method is proposed to solve the problem directly by using cutting patterns directly. A major advantage of the proposed method in series on-line production is that the system can adjust the cutting plan according to the floating orders. Examples with multiple items are demonstrated. The results show considerable efficiency and reliability in high-speed cutting of CSP.

A Novel Portable Device for Fast Analysis of Energetic Materials in the Environment

Construction of portable device for fast analysis of energetic materials is described in this paper. The developed analytical system consists of two main parts: a miniaturized microcolumn liquid chromatograph of unique construction and original chemiluminescence detector. This novel portable device is able to determine selectively most of nitramine- and nitroester-based explosives as well as inorganic nitrates at trace concentrations in water or soil extracts in less than 8 minutes.

Automated Thickness Measurement of Retinal Blood Vessels for Implementation of Clinical Decision Support Systems in Diagnostic Diabetic Retinopathy

The structure of retinal vessels is a prominent feature, that reveals information on the state of disease that are reflected in the form of measurable abnormalities in thickness and colour. Vascular structures of retina, for implementation of clinical diabetic retinopathy decision making system is presented in this paper. Retinal Vascular structure is with thin blood vessel, whose accuracy is highly dependent upon the vessel segmentation. In this paper the blood vessel thickness is automatically detected using preprocessing techniques and vessel segmentation algorithm. First the capture image is binarized to get the blood vessel structure clearly, then it is skeletonised to get the overall structure of all the terminal and branching nodes of the blood vessels. By identifying the terminal node and the branching points automatically, the main and branching blood vessel thickness is estimated. Results are presented and compared with those provided by clinical classification on 50 vessels collected from Bejan Singh Eye hospital..

Frequency and Amplitude Measurement of a Vibrating Object in Water Using Ultrasonic Speckle Technique

The principle of frequency and amplitude measurement of a vibrating object in water using ultrasonic speckle technique is presented in this paper. Compared with other traditional techniques, the ultrasonic speckle technique can be applied to vibration measurement of a nonmetal object with rough surface in water in a noncontact way. The relationship between speckle movement and object movement was analyzed. Based on this study, an ultrasonic speckle measurement system was set up. With this system the frequency and amplitude of an underwater vibrating cantilever beam was detected. The result shows that the experimental data is in good agreement with the calibrating data.

Human Verification in a Video Surveillance System Using Statistical Features

A human verification system is presented in this paper. The system consists of several steps: background subtraction, thresholding, line connection, region growing, morphlogy, star skelatonization, feature extraction, feature matching, and decision making. The proposed system combines an advantage of star skeletonization and simple statistic features. A correlation matching and probability voting have been used for verification, followed by a logical operation in a decision making stage. The proposed system uses small number of features and the system reliability is convincing.

Experimental and Computational Analysis of Hygrothermal Performance of an Interior Thermal Insulation System

Combined experimental and computational analysis of hygrothermal performance of an interior thermal insulation system applied on a brick wall is presented in the paper. In the experimental part, the functionality of the insulation system is tested at simulated difference climate conditions using a semi-scale device. The measured temperature and relative humidity profiles are used for the calibration of computer code HEMOT that is finally applied for a long-term hygrothermal analysis of the investigated structure.

An Investigation of Short Circuit Analysis in Komag Sarawak Operations (KSO) Factory

Short circuit currents plays a vital role in influencing the design and operation of equipment and power system and could not be avoided despite careful planning and design, good maintenance and thorough operation of the system. This paper discusses the short circuit analysis conducted in KSO briefly comprising of its significances, methods and results. A result sample of the analysis based on a single transformer is detailed in this paper. Furthermore, the results of the analysis and its significances were also discussed and commented.

Potential of GIS to Find Solutions to Space Related Problems in Construction Industry

Geographic Information System (GIS) is a computerbased tool used extensively to solve various engineering problems related to spatial data. In spite of growing popularity of GIS, its complete potential to construction industry has not been realized. In this paper, the summary of up-to-date work on spatial applications of GIS technologies in construction industry is presented. GIS technologies have the potential to solve space related problems of construction industry involving complex visualization, integration of information, route planning, E-commerce, cost estimation, etc. GISbased methodology to handle time and space issues of construction projects scheduling is developed and discussed in this paper.

Linear Stability Characteristics of Wake-Shear Layers in Two-Phase Shallow Flows

Linear stability of wake-shear layers in two-phase shallow flows is analyzed in the present paper. Stability analysis is based on two-dimensional shallow water equations. It is assumed that the fluid contains uniformly distributed solid particles. No dynamic interaction between the carrier fluid and particles is expected in the initial moment. Linear stability curves are obtained for different values of the particle loading parameter, the velocity ratio and the velocity deficit. It is shown that the increase in the velocity ratio destabilizes the flow. The particle loading parameter has a stabilizing effect on the flow. The role of the velocity deficit is also destabilizing: the increase of the velocity deficit leads to less stable flow.

Defining Programming Problems as Learning Objects

Standards for learning objects focus primarily on content presentation. They were already extended to support automatic evaluation but it is limited to exercises with a predefined set of answers. The existing standards lack the metadata required by specialized evaluators to handle types of exercises with an indefinite set of solutions. To address this issue existing learning object standards were extended to the particular requirements of a specialized domain. A definition of programming problems as learning objects, compatible both with Learning Management Systems and with systems performing automatic evaluation of programs, is presented in this paper. The proposed definition includes metadata that cannot be conveniently represented using existing standards, such as: the type of automatic evaluation; the requirements of the evaluation engine; and the roles of different assets - tests cases, program solutions, etc. The EduJudge project and its main services are also presented as a case study on the use of the proposed definition of programming problems as learning objects.

Improvement of Stator Slot Structure based on Electro-Thermal Analysis in HV Generator

High voltage generators are being subject to higher voltage rating and are being designed to operate in harsh conditions. Stator windings are the main component of generators in which Electrical, magnetically and thermal stresses remain major failures for insulation degradation accelerated aging. A large number of generators failed due to stator winding problems, mainly insulation deterioration. Insulation degradation assessment plays vital role in the asset life management. Mostly the stator failure is catastrophic causing significant damage to the plant. Other than generation loss, stator failure involves heavy repair or replacement cost. Electro thermal analysis is the main characteristic for improvement design of stator slot-s insulation. Dielectric parameters such as insulation thickness, spacing, material types, geometry of winding and slot are major design consideration. A very powerful method available to analyze electro thermal performance is Finite Element Method (FEM) which is used in this paper. The analysis of various stator coil and slot configurations are used to design the better dielectric system to reduce electrical and thermal stresses in order to increase the power of generator in the same volume of core. This paper describes the process used to perform classical design and improvement analysis of stator slot-s insulation.

A Novel Compression Algorithm for Electrocardiogram Signals based on Wavelet Transform and SPIHT

Electrocardiogram (ECG) data compression algorithm is needed that will reduce the amount of data to be transmitted, stored and analyzed, but without losing the clinical information content. A wavelet ECG data codec based on the Set Partitioning In Hierarchical Trees (SPIHT) compression algorithm is proposed in this paper. The SPIHT algorithm has achieved notable success in still image coding. We modified the algorithm for the one-dimensional (1-D) case and applied it to compression of ECG data. By this compression method, small percent root mean square difference (PRD) and high compression ratio with low implementation complexity are achieved. Experiments on selected records from the MIT-BIH arrhythmia database revealed that the proposed codec is significantly more efficient in compression and in computation than previously proposed ECG compression schemes. Compression ratios of up to 48:1 for ECG signals lead to acceptable results for visual inspection.

Recycling-Oriented Product Assessment during Design Process with Usage of Agent Technology

In the paper the method of product analysis from recycling point of view has been described. The analysis bases on set of measures that assess a product from the point of view of final stages of its lifecycle. It was assumed that such analysis will be performed at the design phase – in order to conduct such analysis the computer system that aids the designer during the design process has been developed. The structure of the computer tool, based on agent technology, and example results has been also included in the paper.

A Novel Method for Areal Surface Roughness Measurement

An area-integrating method that uses the technique of total integrated light scatter for evaluating the root mean square height of the surface Sq has been presented in the paper. It is based on the measurement of the scatter power using a flat photodiode integrator rather than an optical sphere or a hemisphere. By this means, one can obtain much less expensive and smaller instruments than traditional ones. Thanks to this, they could find their application for surface control purposes, particularly in small and medium size enterprises. A description of the functioning of the measuring unit as well as the impact caused by different factors on its properties is presented first. Next, results of measurements of the Sq values performed for optical, silicon and metal samples have been shown. It has been also proven that they are in a good agreement with the results obtained using the Ulbricht sphere instrument.

Prospects in Waste Oil Shale Ash Sustainable Valorization

An innovative approach utilizing highly alkaline oil shale waste ash and carbon dioxide gas (CO2), associated with power production, as a resource for production of precipitated calcium carbonate (PCC) is introduced in this paper. The specifics and feasibility of the integrated ash valorization and CO2 sequestration process by indirect aqueous carbonation of lime-consisting ash were elaborated and the main parameters established. Detailed description of the formed precipitates was included. Complimentary carbonation experiments with commercial CaO fine powder were conducted for comparative characterization of the final products obtained on the basis of two different raw materials. Finally, the expected CO2 uptake was evaluated.

Named Entity Recognition using Support Vector Machine: A Language Independent Approach

Named Entity Recognition (NER) aims to classify each word of a document into predefined target named entity classes and is now-a-days considered to be fundamental for many Natural Language Processing (NLP) tasks such as information retrieval, machine translation, information extraction, question answering systems and others. This paper reports about the development of a NER system for Bengali and Hindi using Support Vector Machine (SVM). Though this state of the art machine learning technique has been widely applied to NER in several well-studied languages, the use of this technique to Indian languages (ILs) is very new. The system makes use of the different contextual information of the words along with the variety of features that are helpful in predicting the four different named (NE) classes, such as Person name, Location name, Organization name and Miscellaneous name. We have used the annotated corpora of 122,467 tokens of Bengali and 502,974 tokens of Hindi tagged with the twelve different NE classes 1, defined as part of the IJCNLP-08 NER Shared Task for South and South East Asian Languages (SSEAL) 2. In addition, we have manually annotated 150K wordforms of the Bengali news corpus, developed from the web-archive of a leading Bengali newspaper. We have also developed an unsupervised algorithm in order to generate the lexical context patterns from a part of the unlabeled Bengali news corpus. Lexical patterns have been used as the features of SVM in order to improve the system performance. The NER system has been tested with the gold standard test sets of 35K, and 60K tokens for Bengali, and Hindi, respectively. Evaluation results have demonstrated the recall, precision, and f-score values of 88.61%, 80.12%, and 84.15%, respectively, for Bengali and 80.23%, 74.34%, and 77.17%, respectively, for Hindi. Results show the improvement in the f-score by 5.13% with the use of context patterns. Statistical analysis, ANOVA is also performed to compare the performance of the proposed NER system with that of the existing HMM based system for both the languages.