Agent-Based Simulation of Simulating Anticipatory Systems – Classification

The present paper is oriented to classification and application of agent technique in simulation of anticipatory systems, namely those that use simulation models for the aid of anticipation. The main ideas root in the fact that the best way for description of computer simulation models is the technique of describing the simulated system itself (and the translation into the computer code is provided as automatic), and that the anticipation itself is often nested.

Educational Values of Virtual Reality: The Case of Spatial Ability

The use of Virtual Reality (VR) in schools and higher education is proliferating. Due to its interactive and animated features, it is regarded as a promising technology to increase students- spatial ability. Spatial ability is assumed to have a prominent role in science and engineering domains. However, research concerning individual differences such as spatial ability in the context of VR is still at its infancy. Moreover, empirical studies that focus on the features of VR to improve spatial ability are to date rare. Thus, this paper explores the possible educational values of VR in relation to spatial ability to call for more research concerning spatial ability in the context of VR based on studies in computerbased learning. It is believed that the incorporation of state-of-the-art VR technology for educational purposes should be justified by the enhanced benefits for the target learners.

Design and Trajectory Planning of Bipedal Walking Robot with Minimum Sufficient Actuation System

This paper presents a new type of mechanism and trajectory planning strategy for bipedal walking robot. The newly designed mechanism is able to improve the performance of bipedal walking robot in terms of energy efficiency and weight reduction by utilizing minimum number of actuators. The usage of parallelogram mechanism eliminates the needs of having an extra actuator at the knee joint. This mechanism works together with the joint space trajectory planning in order to realize straight legged walking which cannot be achieved by conventional inverse kinematics trajectory planning due to the singularity. The effectiveness of the proposed strategy is confirmed by computer simulation results.

Effects of Operating Conditions on Calcium Carbonate Fouling in a Plate Heat Exchanger

The aim of this work is to investigate on the internalflow patterns in a plate heat exchanger channel, which affect the rate of sedimentation fouling on the heat transfer surface of the plate heat exchanger. The research methodologies were the computer simulation using Computational Fluid Dynamics (CFD) and the experimental works. COMSOL MULTIPHYSICS™ Version 3.3 was used to simulate the velocity flow fields to verify the low and high flow regions. The results from the CFD technique were then compared with the images obtained from the experiments in which the fouling test rig was set up with a singlechannel plate heat exchanger to monitor the fouling of calcium carbonate. Two parameters were varied i.e., the crossing angle of the two plate: 55/55, 10/10, and 55/10 degree, and the fluid flow rate at the inlet: 0.0566, 0.1132 and 0.1698 m/s. The type of plate “GX-12" (the surface area 0.12 m2, the depth 2.9 mm, the width of fluid flow 215 mm and the thickness of stainless plate of 0.5 mm) was used in this study. The results indicated that the velocity distribution for the case of 55/55 degree seems to be very well organized when compared with the others. Also, an increase in the inlet velocity resulted in the reduction of fouling rate on the surface of plate heat exchangers.

Innovation at the Faculty-level Education through Service Learning

The paper presents the service learning project titled DicDucFac (idea-leadership-product), that was planned and conducted by the team of information sciences students. It was planned as a workshop dealing with the application of modern social media (Facebook, YouTube, Gmail) for the purposes of selfpromotion, free advertising via social networks and marketing own ideas and/or products in the virtual world. The workshop was organized for highly-skilled computer literate unemployed youth. These youth, as final beneficiaries, will be able to apply what they learned in this workshop to “the real world“, increasing their chances for employment and self-employment. The results of the project reveal that the basic, active-learning principles embodied in our teaching approach allow students to learn more effectively and gain essential life skills (from computer applications to teamwork) that can only be learned by doing. It also shows that our students received the essentials of professional ethics and citizenship through direct, personal engagement in professional activities and the life of the community.

An E-learning System Architecture based on Cloud Computing

The massive proliferation of affordable computers, Internet broadband connectivity and rich education content has created a global phenomenon in which information and communication technology (ICT) is being used to transform education. Therefore, there is a need to redesign the educational system to meet the needs better. The advent of computers with sophisticated software has made it possible to solve many complex problems very fast and at a lower cost. This paper introduces the characteristics of the current E-Learning and then analyses the concept of cloud computing and describes the architecture of cloud computing platform by combining the features of E-Learning. The authors have tried to introduce cloud computing to e-learning, build an e-learning cloud, and make an active research and exploration for it from the following aspects: architecture, construction method and external interface with the model.

Improving Worm Detection with Artificial Neural Networks through Feature Selection and Temporal Analysis Techniques

Computer worm detection is commonly performed by antivirus software tools that rely on prior explicit knowledge of the worm-s code (detection based on code signatures). We present an approach for detection of the presence of computer worms based on Artificial Neural Networks (ANN) using the computer's behavioral measures. Identification of significant features, which describe the activity of a worm within a host, is commonly acquired from security experts. We suggest acquiring these features by applying feature selection methods. We compare three different feature selection techniques for the dimensionality reduction and identification of the most prominent features to capture efficiently the computer behavior in the context of worm activity. Additionally, we explore three different temporal representation techniques for the most prominent features. In order to evaluate the different techniques, several computers were infected with five different worms and 323 different features of the infected computers were measured. We evaluated each technique by preprocessing the dataset according to each one and training the ANN model with the preprocessed data. We then evaluated the ability of the model to detect the presence of a new computer worm, in particular, during heavy user activity on the infected computers.

Improving Quality of Business Networks for Information Systems

Computer networks are essential part in computerbased information systems. The performance of these networks has a great influence on the whole information system. Measuring the usability criteria and customers satisfaction on small computer network is very important. In this article, an effective approach for measuring the usability of business network in an information system is introduced. The usability process for networking provides us with a flexible and a cost-effective way to assess the usability of a network and its products. In addition, the proposed approach can be used to certify network product usability late in the development cycle. Furthermore, it can be used to help in developing usable interfaces very early in the cycle and to give a way to measure, track, and improve usability. Moreover, a new approach for fast information processing over computer networks is presented. The entire data are collected together in a long vector and then tested as a one input pattern. Proposed fast time delay neural networks (FTDNNs) use cross correlation in the frequency domain between the tested data and the input weights of neural networks. It is proved mathematically and practically that the number of computation steps required for the presented time delay neural networks is less than that needed by conventional time delay neural networks (CTDNNs). Simulation results using MATLAB confirm the theoretical computations.

Semi-Automatic Analyzer to Detect Authorial Intentions in Scientific Documents

Information Retrieval has the objective of studying models and the realization of systems allowing a user to find the relevant documents adapted to his need of information. The information search is a problem which remains difficult because the difficulty in the representing and to treat the natural languages such as polysemia. Intentional Structures promise to be a new paradigm to extend the existing documents structures and to enhance the different phases of documents process such as creation, editing, search and retrieval. The intention recognition of the author-s of texts can reduce the largeness of this problem. In this article, we present intentions recognition system is based on a semi-automatic method of extraction the intentional information starting from a corpus of text. This system is also able to update the ontology of intentions for the enrichment of the knowledge base containing all possible intentions of a domain. This approach uses the construction of a semi-formal ontology which considered as the conceptualization of the intentional information contained in a text. An experiments on scientific publications in the field of computer science was considered to validate this approach.

On the Need to have an Additional Methodology for the Psychological Product Measurement and Evaluation

Cognitive Science appeared about 40 years ago, subsequent to the challenge of the Artificial Intelligence, as common territory for several scientific disciplines such as: IT, mathematics, psychology, neurology, philosophy, sociology, and linguistics. The new born science was justified by the complexity of the problems related to the human knowledge on one hand, and on the other by the fact that none of the above mentioned sciences could explain alone the mental phenomena. Based on the data supplied by the experimental sciences such as psychology or neurology, models of the human mind operation are built in the cognition science. These models are implemented in computer programs and/or electronic circuits (specific to the artificial intelligence) – cognitive systems – whose competences and performances are compared to the human ones, leading to the psychology and neurology data reinterpretation, respectively to the construction of new models. During these processes if psychology provides the experimental basis, philosophy and mathematics provides the abstraction level utterly necessary for the intermission of the mentioned sciences. The ongoing general problematic of the cognitive approach provides two important types of approach: the computational one, starting from the idea that the mental phenomenon can be reduced to 1 and 0 type calculus operations, and the connection one that considers the thinking products as being a result of the interaction between all the composing (included) systems. In the field of psychology measurements in the computational register use classical inquiries and psychometrical tests, generally based on calculus methods. Deeming things from both sides that are representing the cognitive science, we can notice a gap in psychological product measurement possibilities, regarded from the connectionist perspective, that requires the unitary understanding of the quality – quantity whole. In such approach measurement by calculus proves to be inefficient. Our researches, deployed for longer than 20 years, lead to the conclusion that measuring by forms properly fits to the connectionism laws and principles.

A Comparison Study of Inspector's Performance between Regular and Complex Tasks

This research was to study a comparison of inspector-s performance between regular and complex visual inspection task. Visual task was simulated on DVD read control circuit. Inspection task was performed by using computer. Subjects were 10 undergraduate randomly selected and test for 20/20. Then, subjects were divided into two groups, five for regular inspection (control group) and five for complex inspection (treatment group) tasks. Result was showed that performance on regular and complex inspectors was significantly difference at the level of 0.05. Inspector performance on regular inspection was showed high percentage on defects detected by using equal time to complex inspection. This would be indicated that inspector performance was affected by visual inspection task.

A Proposed Information Extraction Technique in Engineering Drawing for Reuse Design

The extensive number of engineering drawing will be referred for planning process and the changes will produce a good engineering design to meet the demand in producing a new model. The advantage in reuse of engineering designs is to allow continuous product development to further improve the quality of product development, thus reduce the development costs. However, to retrieve the existing engineering drawing, it is time consuming, a complex process and are expose to errors. Engineering drawing file searching system will be proposed to solve this problem. It is essential for engineer and designer to have some sort of medium to enable them to search for drawing in the most effective way. This paper lays out the proposed research project under the area of information extraction in engineering drawing.

Data Mining Classification Methods Applied in Drug Design

Data mining incorporates a group of statistical methods used to analyze a set of information, or a data set. It operates with models and algorithms, which are powerful tools with the great potential. They can help people to understand the patterns in certain chunk of information so it is obvious that the data mining tools have a wide area of applications. For example in the theoretical chemistry data mining tools can be used to predict moleculeproperties or improve computer-assisted drug design. Classification analysis is one of the major data mining methodologies. The aim of thecontribution is to create a classification model, which would be able to deal with a huge data set with high accuracy. For this purpose logistic regression, Bayesian logistic regression and random forest models were built using R software. TheBayesian logistic regression in Latent GOLD software was created as well. These classification methods belong to supervised learning methods. It was necessary to reduce data matrix dimension before construct models and thus the factor analysis (FA) was used. Those models were applied to predict the biological activity of molecules, potential new drug candidates.

A Novel FFT-Based Frequency Offset Estimator for OFDM Systems

This paper proposes a novel frequency offset (FO) estimator for orthogonal frequency division multiplexing. Simplicity is most significant feature of this algorithm and can be repeated to achieve acceptable accuracy. Also fractional and integer part of FO is estimated jointly with use of the same algorithm. To do so, instead of using conventional algorithms that usually use correlation function, we use DFT of received signal. Therefore, complexity will be reduced and we can do synchronization procedure by the same hardware that is used to demodulate OFDM symbol. Finally, computer simulation shows that the accuracy of this method is better than other conventional methods.

Software Model for a Computer Based Training for an HVDC Control Desk Simulator

With major technological advances and to reduce the cost of training apprentices for real-time critical systems, it was necessary the development of Intelligent Tutoring Systems for training apprentices in these systems. These systems, in general, have interactive features so that the learning is actually more efficient, making the learner more familiar with the mechanism in question. In the home stage of learning, tests are performed to obtain the student's income, a measure on their use. The aim of this paper is to present a framework to model an Intelligent Tutoring Systems using the UML language. The various steps of the analysis are considered the diagrams required to build a general model, whose purpose is to present the different perspectives of its development.

Research on IBR-Driven Distributed Collaborative Visualization System

Image-based Rendering(IBR) techniques recently reached in broad fields which leads to a critical challenge to build up IBR-Driven visualization platform where meets requirement of high performance, large bounds of distributed visualization resource aggregation and concentration, multiple operators deploying and CSCW design employing. This paper presents an unique IBR-based visualization dataflow model refer to specific characters of IBR techniques and then discusses prominent feature of IBR-Driven distributed collaborative visualization (DCV) system before finally proposing an novel prototype. The prototype provides a well-defined three level modules especially work as Central Visualization Server, Local Proxy Server and Visualization Aid Environment, by which data and control for collaboration move through them followed the previous dataflow model. With aid of this triple hierarchy architecture of that, IBR oriented application construction turns to be easy. The employed augmented collaboration strategy not only achieve convenient multiple users synchronous control and stable processing management, but also is extendable and scalable.

Effective Online Staff Training: Is This Possible?

The purpose of this paper is to consider the introduction of online courses to replace the current classroom-based staff training. The current training is practical, and must be completed before access to the financial computer system is authorized. The long term objective is to measure the efficacy, effectiveness and efficiency of the training, and to establish whether a transfer of knowledge back to the workplace has occurred. This paper begins with an overview explaining the importance of staff training in an evolving, competitive business environment and defines the problem facing this particular organization. A summary of the literature review is followed by a brief discussion of the research methodology and objective. The implementation of the alpha version of the online course is then described. This paper may be of interest to those seeking insights into, or new theory regarding, practical interventions of online learning in the real world.

In-Situ Monitoring the Thermal Forming of Glass and Si Foils for Space X-Ray Telescopes

We developed a non-contact method for the in-situ monitoring of the thermal forming of glass and Si foils to optimize the manufacture of mirrors for high-resolution space x-ray telescopes. Their construction requires precise and light-weight segmented optics with angular resolution better than 5 arcsec. We used 75x25 mm Desag D263 glass foils 0.75 mm thick and 0.6 mm thick Si foils. The glass foils were shaped by free slumping on a frame at viscosities in the range of 109.3-1012 dPa·s, the Si foils by forced slumping above 1000°C. Using a Nikon D80 digital camera, we took snapshots of a foil-s shape every 5 min during its isothermal heat treatment. The obtained results we can use for computer simulations. By comparing the measured and simulated data, we can more precisely define material properties of the foils and optimize the forming technology.

Packing Theory for Natural and Crushed Aggregate to Obtain the Best Mix of Aggregate: Research and Development

Concrete performance is strongly affected by the particle packing degree since it determines the distribution of the cementitious component and the interaction of mineral particles. By using packing theory designers will be able to select optimal aggregate materials for preparing concrete with low cement content, which is beneficial from the point of cost. Optimum particle packing implies minimizing porosity and thereby reducing the amount of cement paste needed to fill the voids between the aggregate particles, taking also the rheology of the concrete into consideration. For reaching good fluidity superplasticizers are required. The results from pilot tests at Luleå University of Technology (LTU) show various forms of the proposed theoretical models, and the empirical approach taken in the study seems to provide a safer basis for developing new, improved packing models.

Improvement of Semen Quality in Holstein Bulls during Heat Stress by Supplementing Omega-3 Fatty Acids

The aim of current study was to investigate the changes in the quality parameters of Holstein bull semen during the heat stress and the effect of feeding a source of omega-3 fatty acids in this period. Samples were obtained from 19 Holstein bulls during the expected time of heat stress in Iran (June to September 2009). Control group (n=10) were fed a standard concentrate feed while treatment group (n=9) had this feed top dressed with 100 g of an omega-3 enriched nutriceutical. Semen quality was assessed on ejaculates collected after 1, 5, 9 and 12 weeks of supplementation. Computer-assisted assessment of sperm motility, viability (eosinnigrosin) and hypo-osmotic swelling test (HOST) were conducted. Heat stress affected sperm quality parameters by week 5 and 9 (p