Lexical Database for Multiple Languages: Multilingual Word Semantic Network

Data mining and knowledge engineering have become a tough task due to the availability of large amount of data in the web nowadays. Validity and reliability of data also become a main debate in knowledge acquisition. Besides, acquiring knowledge from different languages has become another concern. There are many language translators and corpora developed but the function of these translators and corpora are usually limited to certain languages and domains. Furthermore, search results from engines with traditional 'keyword' approach are no longer satisfying. More intelligent knowledge engineering agents are needed. To address to these problems, a system known as Multilingual Word Semantic Network is proposed. This system adapted semantic network to organize words according to concepts and relations. The system also uses open source as the development philosophy to enable the native language speakers and experts to contribute their knowledge to the system. The contributed words are then defined and linked using lexical and semantic relations. Thus, related words and derivatives can be identified and linked. From the outcome of the system implementation, it contributes to the development of semantic web and knowledge engineering.

A Basic Study on Ubiquitous Overloaded Vehicles Regulation System

Load managing method on road became necessary since overloaded vehicles occur damage on road facilities and existing systems for preventing this damage still show many problems.Accordingly, efficient managing system for preventing overloaded vehicles could be organized by using the road itself as a scale by applying genetic algorithm to analyze the load and the drive information of vehicles.Therefore, this paper organized Ubiquitous sensor network system for development of intelligent overload vehicle regulation system, also in this study, to use the behavior of road, the transformation was measured by installing underground box type indoor model and indoor experiment was held using genetic algorithm. And we examined wireless possibility of overloaded vehicle regulation system through experiment of the transmission and reception distance.If this system will apply to road and bridge, might be effective for economy and convenience through establishment of U-IT system..

A Study on Multi-Agent Behavior in a Soccer Game Domain

There have been many games developing simulation of soccer games. Many of these games have been designed with highly realistic features to attract more users. Many have also incorporated better artificial intelligent (AI) similar to that in a real soccer game. One of the challenging issues in a soccer game is the cooperation, coordination and negotiation among distributed agents in a multi-agent system. This paper focuses on the incorporation of multi-agent technique in a soccer game domain. The better the cooperation of a multi-agent team, the more intelligent the game will be. Thus, past studies were done on the robotic soccer game because of the better multi-agent system implementation. From this study, a better approach and technique of multi-agent behavior could be select to improve the author-s 2D online soccer game.

Deriving Causal Explanation from Qualitative Model Reasoning

This paper discusses a qualitative simulator QRiOM that uses Qualitative Reasoning (QR) technique, and a process-based ontology to model, simulate and explain the behaviour of selected organic reactions. Learning organic reactions requires the application of domain knowledge at intuitive level, which is difficult to be programmed using traditional approach. The main objective of QRiOM is to help learners gain a better understanding of the fundamental organic reaction concepts, and to improve their conceptual comprehension on the subject by analyzing the multiple forms of explanation generated by the software. This paper focuses on the generation of explanation based on causal theories to explicate various phenomena in the chemistry subject. QRiOM has been tested with three classes problems related to organic chemistry, with encouraging results. This paper also presents the results of preliminary evaluation of QRiOM that reveal its explanation capability and usefulness.

Polyphenolic Profile and Antioxidant Activities of Nigella Sativa Seed Extracts In Vitro and In Vivo

Nigella sativa L. is an aromatic plant belonging to the family Ranunculaceae. It has been used traditionally, especially in the middle East and India, for the treatment of asthma, cough, bronchitis, headache, rheumatism, fever, influenza and eczema. Several biological activities have been reported in Nigella sativa seeds, including antioxidant. In this context we tried to estimate the antioxidant activity of various extracts prepared from Nigella sativa seeds, methanolic extract (ME), chloroformic extract (CE), hexanic extract (HE : fixed oil), ethyl acetate extract (EAE) water extract (WE). The Folin-Ciocalteu assay showed that CE and EAE contained high level of phenolic compounds 81.31 and 72.43μg GAE/mg of extract respectively. Similarly, the CE and EAE exhibited the highest DPPH radical scavenging activity, with IC50 values of 106.56μg/ml and 121.62μg/ml respectively. In addition, CE and HE showed the most scavenging activity against superoxide radical generated in the PMS-NADH-NBT system with respective IC50 values of 361.86 μg/ml and 371.80 μg/ml, which is comparable to the activity of the standard antioxidant BHT (344.59 μg/ml). Ferrous ion chelating capacity assay showed that WE, EAE and ME are the most active with 40.57, 39.70 and 22.02 mg EDTA-E/g of extract. The inhibition of linoleic acid/ß-carotene coupled oxidation was estimated by ßcarotene bleaching assay, this showed a highest relative antioxidant activity with CE and EAE (69.82% of inhibition). The antioxidant activities of the methanolic extract and the fixed oil are confirmed by an in vivo assay in mice, the daily oral administration of methanolic extract (500 and 800 mg/kg/day) and fixed oil (2 and 4 ml/kg/day) during 21 days, resulted in a significant enhancement of the blood total antioxidant capacity (measured by KRL test) and the plasmatic antioxidant capacity towards DPPH radical.

Effects of Upflow Liquid Velocity on Performance of Expanded Granular Sludge Bed (EGSB) System

The effects of upflow liquid velocity (ULV) on performance of expanded granular sludge bed (EGSB) system were investigated. The EGSB reactor, made from galvanized steel pipe 0.10 m diameter and 5 m height, had been used to treat piggery wastewater, after passing through acidification tank. It consisted of 39.3 l working volume in reaction zone and 122 l working volume in sedimentation zone, at the upper part. The reactor was seeded with anaerobically digested sludge and operated at the ULVs of 4, 8, 12 and 16 m/h, consecutively, corresponding to organic loading rates of 9.6 – 13.0 kg COD/ (m3.d). The average COD concentrations in the influent were 9,601 – 13,050 mg/l. The COD removal was not significantly different, i.e. 93.0% - 94.0%, except at ULV 12 m/h where SS in the influent was exceptionally high so that VSS washout had occurred, leading to low COD removal. The FCOD and VFA concentrations in the effluent of all experiments were not much different, indicating the same range of treatment performance. The biogas production decreased at higher ULV and ULV of 4 m/h is suggested as design criterion for EGSB system.

Intelligent Dynamic Decision-making Model Using in Robot's Movement

This work develops a novel intelligent “model of dynamic decision-making" usingcell assemblies network architecture in robot's movement. The “model of dynamic decision-making" simulates human decision-making, and follows commands to make the correct decisions. The cell assemblies approach consisting of fLIF neurons was used to implement tasks for finding targets and avoiding obstacles. Experimental results show that the cell assemblies approach of can be employed to efficiently complete finding targets and avoiding obstacles tasks and can simulate the human thinking and the mode of information transactions.

Numerical Simulation of Heat Exchanger Area of R410A-R23 and R404A-R508B Cascade Refrigeration System at Various Evaporating and Condensing Temperature

Capacity and efficiency of any refrigerating system diminish rapidly as the difference between the evaporating and condensing temperature is increased by reduction in the evaporator temperature. The single stage vapour compression refrigeration system is limited to an evaporator temperature of -40 0C. Below temperature of -40 0C the either cascade refrigeration system or multi stage vapour compression system is employed. Present work describes thermal design of main three heat exchangers namely condenser (HTS), cascade condenser and evaporator (LTS) of R404A-R508B and R410A-R23 cascade refrigeration system. Heat transfer area of condenser (HTS), cascade condenser and evaporator (LTS) for both systems have been compared and the effect of condensing and evaporating temperature on heat-transfer area for both systems have been studied under same operating condition. The results shows that the required heat-transfer area of condenser and cascade condenser for R410A-R23 cascade system is lower than the R404A-R508B cascade system but heat transfer area of evaporator is similar for both the system. The heat transfer area of condenser and cascade condenser decreases with increase in condensing temperature (Tc), whereas the heat transfer area of cascade condenser and evaporator increases with increase in evaporating temperature (Te).

The Study of using Public Participation Geographic Information System in Indigenous Mapping

Current practice of indigenous Mapping production based on GIS, are mostly produced by professional GIS personnel. Given such persons maintain control over data collection and authoring, it is possible to conceive errors due to misrepresentation or cognitive misunderstanding, causing map production inconsistencies. In order to avoid such issues, this research into tribal GIS interface focuses not on customizing interfaces for individual tribes, but rather generalizing the interface and features based on indigenous tribal user needs. The methods employed differs from the traditional expert top-down approach, and instead gaining deeper understanding into indigenous Mappings and user needs, prior to applying mapping techniques and feature development.

Development of an Intelligent Tool for Planning the Operation

Several optimization algorithms specifically applied to the problem of Operation Planning of Hydrothermal Power Systems have been developed and are used. Although providing solutions to various problems encountered, these algorithms have some weaknesses, difficulties in convergence, simplification of the original formulation of the problem, or owing to the complexity of the objective function. Thus, this paper presents the development of a computational tool for solving optimization problem identified and to provide the User an easy handling. Adopted as intelligent optimization technique, Genetic Algorithms and programming language Java. First made the modeling of the chromosomes, then implemented the function assessment of the problem and the operators involved, and finally the drafting of the graphical interfaces for access to the User. The program has managed to relate a coherent performance in problem resolution without the need for simplification of the calculations together with the ease of manipulating the parameters of simulation and visualization of output results.

Perceptual Framework for a Modern Left-Turn Collision Warning System

Most of the collision warning systems currently available in the automotive market are mainly designed to warn against imminent rear-end and lane-changing collisions. No collision warning system is commercially available to warn against imminent turning collisions at intersections, especially for left-turn collisions when a driver attempts to make a left-turn at either a signalized or non-signalized intersection, conflicting with the path of other approaching vehicles traveling on the opposite-direction traffic stream. One of the major factors that lead to left-turn collisions is the human error and misjudgment of the driver of the turning vehicle when perceiving the speed and acceleration of other vehicles traveling on the opposite-direction traffic stream; therefore, using a properly-designed collision warning system will likely reduce, or even eliminate, this type of collisions by reducing human error. This paper introduces perceptual framework for a proposed collision warning system that can detect imminent left-turn collisions at intersections. The system utilizes a commercially-available detection sensor (either a radar sensor or a laser detector) to detect approaching vehicles traveling on the opposite-direction traffic stream and calculate their speeds and acceleration rates to estimate the time-tocollision and compare that time to the time required for the turning vehicle to clear the intersection. When calculating the time required for the turning vehicle to clear the intersection, consideration is given to the perception-reaction time of the driver of the turning vehicle, which is the time required by the driver to perceive the message given by the warning system and react to it by engaging the throttle. A regression model was developed to estimate perception-reaction time based on age and gender of the driver of the host vehicle. Desired acceleration rate selected by the driver of the turning vehicle, when making the left-turn movement, is another human factor that is considered by the system. Another regression model was developed to estimate the acceleration rate selected by the driver of the turning vehicle based on driver-s age and gender as well as on the location and speed of the nearest approaching vehicle along with the maximum acceleration rate provided by the mechanical characteristics of the turning vehicle. By comparing time-to-collision with the time required for the turning vehicle to clear the intersection, the system displays a message to the driver of the turning vehicle when departure is safe. An application example is provided to illustrate the logic algorithm of the proposed system.

On the Need to have an Additional Methodology for the Psychological Product Measurement and Evaluation

Cognitive Science appeared about 40 years ago, subsequent to the challenge of the Artificial Intelligence, as common territory for several scientific disciplines such as: IT, mathematics, psychology, neurology, philosophy, sociology, and linguistics. The new born science was justified by the complexity of the problems related to the human knowledge on one hand, and on the other by the fact that none of the above mentioned sciences could explain alone the mental phenomena. Based on the data supplied by the experimental sciences such as psychology or neurology, models of the human mind operation are built in the cognition science. These models are implemented in computer programs and/or electronic circuits (specific to the artificial intelligence) – cognitive systems – whose competences and performances are compared to the human ones, leading to the psychology and neurology data reinterpretation, respectively to the construction of new models. During these processes if psychology provides the experimental basis, philosophy and mathematics provides the abstraction level utterly necessary for the intermission of the mentioned sciences. The ongoing general problematic of the cognitive approach provides two important types of approach: the computational one, starting from the idea that the mental phenomenon can be reduced to 1 and 0 type calculus operations, and the connection one that considers the thinking products as being a result of the interaction between all the composing (included) systems. In the field of psychology measurements in the computational register use classical inquiries and psychometrical tests, generally based on calculus methods. Deeming things from both sides that are representing the cognitive science, we can notice a gap in psychological product measurement possibilities, regarded from the connectionist perspective, that requires the unitary understanding of the quality – quantity whole. In such approach measurement by calculus proves to be inefficient. Our researches, deployed for longer than 20 years, lead to the conclusion that measuring by forms properly fits to the connectionism laws and principles.

Software Model for a Computer Based Training for an HVDC Control Desk Simulator

With major technological advances and to reduce the cost of training apprentices for real-time critical systems, it was necessary the development of Intelligent Tutoring Systems for training apprentices in these systems. These systems, in general, have interactive features so that the learning is actually more efficient, making the learner more familiar with the mechanism in question. In the home stage of learning, tests are performed to obtain the student's income, a measure on their use. The aim of this paper is to present a framework to model an Intelligent Tutoring Systems using the UML language. The various steps of the analysis are considered the diagrams required to build a general model, whose purpose is to present the different perspectives of its development.

FPGA Implementation of a Vision-Based Blind Spot Warning System

Vision-based intelligent vehicle applications often require large amounts of memory to handle video streaming and image processing, which in turn increases complexity of hardware and software. This paper presents an FPGA implement of a vision-based blind spot warning system. Using video frames, the information of the blind spot area turns into one-dimensional information. Analysis of the estimated entropy of image allows the detection of an object in time. This idea has been implemented in the XtremeDSP video starter kit. The blind spot warning system uses only 13% of its logic resources and 95k bits block memory, and its frame rate is over 30 frames per sec (fps).

An Artificial Intelligent Technique for Robust Digital Watermarking in Multiwavelet Domain

In this paper, an artificial intelligent technique for robust digital image watermarking in multiwavelet domain is proposed. The embedding technique is based on the quantization index modulation technique and the watermark extraction process does not require the original image. We have developed an optimization technique using the genetic algorithms to search for optimal quantization steps to improve the quality of watermarked image and robustness of the watermark. In addition, we construct a prediction model based on image moments and back propagation neural network to correct an attacked image geometrically before the watermark extraction process begins. The experimental results show that the proposed watermarking algorithm yields watermarked image with good imperceptibility and very robust watermark against various image processing attacks.

An Automated High Pressure Differential Thermal Analysis System for Phase Transformation Studies

A piston cylinder based high pressure differential thermal analyzer system is developed to investigate phase transformations, melting, glass transitions, crystallization behavior of inorganic materials, glassy systems etc., at ambient to 4 GPa and at room temperature to 1073 K. The pressure is calibrated by the phase transition of bismuth and ytterbium and temperature is calibrated by using thermocouple data chart. The system developed is calibrated using benzoic acid, ammonium nitrate and it has a pressure and temperature control of ± 8.9 x 10 -4 GPa , ± 2 K respectively. The phase transition of Asx Te100-x chalcogenides, ferrous oxide and strontium boride are studied using the indigenously developed system.

Kinetic and Optimization Studies on Ethanol Production from Corn Flour

Studies on Simultaneous Saccharification and Fermentation (SSF) of corn flour, a major agricultural product as the substrate using starch digesting glucoamylase enzyme derived from Aspergillus niger and non starch digesting and sugar fermenting Saccharomyces cerevisiae in a batch fermentation. Experiments based on Central Composite Design (CCD) were conducted to study the effect of substrate concentration, pH, temperature, enzyme concentration on Ethanol Concentration and the above parameters were optimized using Response Surface Methodology (RSM). The optimum values of substrate concentration, pH, temperature and enzyme concentration were found to be 160 g/l, 5.5, 30°C and 50 IU respectively. The effect of inoculums age on ethanol concentration was also investigated. The corn flour solution equivalent to 16% initial starch concentration gave the highest ethanol concentration of 63.04 g/l after 48 h of fermentation at optimum conditions of pH and temperature. Monod model and Logistic model were used for growth kinetics and Leudeking – Piret model was used for product formation kinetics.

The Use of Dynamically Optimised High Frequency Moving Average Strategies for Intraday Trading

This paper is motivated by the aspect of uncertainty in financial decision making, and how artificial intelligence and soft computing, with its uncertainty reducing aspects can be used for algorithmic trading applications that trade in high frequency. This paper presents an optimized high frequency trading system that has been combined with various moving averages to produce a hybrid system that outperforms trading systems that rely solely on moving averages. The paper optimizes an adaptive neuro-fuzzy inference system that takes both the price and its moving average as input, learns to predict price movements from training data consisting of intraday data, dynamically switches between the best performing moving averages, and performs decision making of when to buy or sell a certain currency in high frequency.

Mining Image Features in an Automatic Two-Dimensional Shape Recognition System

The number of features required to represent an image can be very huge. Using all available features to recognize objects can suffer from curse dimensionality. Feature selection and extraction is the pre-processing step of image mining. Main issues in analyzing images is the effective identification of features and another one is extracting them. The mining problem that has been focused is the grouping of features for different shapes. Experiments have been conducted by using shape outline as the features. Shape outline readings are put through normalization and dimensionality reduction process using an eigenvector based method to produce a new set of readings. After this pre-processing step data will be grouped through their shapes. Through statistical analysis, these readings together with peak measures a robust classification and recognition process is achieved. Tests showed that the suggested methods are able to automatically recognize objects through their shapes. Finally, experiments also demonstrate the system invariance to rotation, translation, scale, reflection and to a small degree of distortion.

Ontology-based Concept Weighting for Text Documents

Documents clustering become an essential technology with the popularity of the Internet. That also means that fast and high-quality document clustering technique play core topics. Text clustering or shortly clustering is about discovering semantically related groups in an unstructured collection of documents. Clustering has been very popular for a long time because it provides unique ways of digesting and generalizing large amounts of information. One of the issues of clustering is to extract proper feature (concept) of a problem domain. The existing clustering technology mainly focuses on term weight calculation. To achieve more accurate document clustering, more informative features including concept weight are important. Feature Selection is important for clustering process because some of the irrelevant or redundant feature may misguide the clustering results. To counteract this issue, the proposed system presents the concept weight for text clustering system developed based on a k-means algorithm in accordance with the principles of ontology so that the important of words of a cluster can be identified by the weight values. To a certain extent, it has resolved the semantic problem in specific areas.