A Game Design Framework for Vocational Education

Serious games have proven to be a useful instrument to engage learners and increase motivation. Nevertheless, a broadly accepted, practical instructional design approach to serious games does not exist. In this paper, we introduce the use of an instructional design model that has not been applied to serious games yet, and has some advantages compared to other design approaches. We present the case of mechanics mechatronics education to illustrate the close match with timing and role of knowledge and information that the instructional design model prescribes and how this has been translated to a rigidly structured game design. The structured approach answers the learning needs of applicable knowledge within the target group. It combines advantages of simulations with strengths of entertainment games to foster learner-s motivation in the best possible way. A prototype of the game will be evaluated along a well-respected evaluation method within an advanced test setting including test and control group.

The Predictability and Abstractness of Language: A Study in Understanding and Usage of the English Language through Probabilistic Modeling and Frequency

Accounts of language acquisition differ significantly in their treatment of the role of prediction in language learning. In particular, nativist accounts posit that probabilistic learning about words and word sequences has little to do with how children come to use language. The accuracy of this claim was examined by testing whether distributional probabilities and frequency contributed to how well 3-4 year olds repeat simple word chunks. Corresponding chunks were the same length, expressed similar content, and were all grammatically acceptable, yet the results of the study showed marked differences in performance when overall distributional frequency varied. It was found that a distributional model of language predicted the empirical findings better than a number of other models, replicating earlier findings and showing that children attend to distributional probabilities in an adult corpus. This suggested that language is more prediction-and-error based, rather than on abstract rules which nativist camps suggest.

Implementation of Second Order Current- Mode Quadrature Sinusoidal Oscillator with Current Controllability

The realization of current-mode quadrature oscillators using current controlled current conveyor transconductance amplifiers (CCCCTAs) and grounded capacitors is presented. The proposed oscillators can provide 2 sinusoidal output currents with 90º phase difference. It is enabled non-interactive dual-current control for both the condition of oscillation and the frequency of oscillation. High output impedances of the configurations enable the circuit to be cascaded without additional current buffers. The use of only grounded capacitors is ideal for integration. The circuit performances are depicted through PSpice simulations, they show good agreement to theoretical anticipation.

Risk Assessment of Selected Source for Emergency Water Supply Case Study II

The case study deals with the semi-quantitative risk assessment of water resource earmarked for the emergency supply of population with drinking water. The risk analysis has been based on previously identified hazards/sensitivities of the elements of hydrogeological structure and technological equipment of ground water resource as well as on the assessment of the levels of hazard, sensitivity and criticality of individual resource elements in the form of point indexes. The following potential sources of hazard have been considered: natural disasters caused by atmospheric and geological changes, technological hazards, and environmental burdens. The risk analysis has proved that the assessed risks are acceptable and the water resource may be integrated into a crisis plan of a given region.

Statistical Evaluation of Nonlinear Distortion using the Multi-Canonical Monte Carlo Method and the Split Step Fourier Method

In high powered dense wavelength division multiplexed (WDM) systems with low chromatic dispersion, four-wave mixing (FWM) can prove to be a major source of noise. The MultiCanonical Monte Carlo Method (MCMC) and the Split Step Fourier Method (SSFM) are combined to accurately evaluate the probability density function of the decision variable of a receiver, limited by FWM. The combination of the two methods leads to more accurate results, and offers the possibility of adding other optical noises such as the Amplified Spontaneous Emission (ASE) noise.

Building Virtual Reality Environments for Distance Education on the Web: A Case Study in Medical Education

The paper presents an investigation into the role of virtual reality and web technologies in the field of distance education. Within this frame, special emphasis is given on the building of web-based virtual learning environments so as to successfully fulfill their educational objectives. In particular, basic pedagogical methods are studied, focusing mainly on the efficient preparation, approach and presentation of learning content, and specific designing rules are presented considering the hypermedia, virtual and educational nature of this kind of applications. The paper also aims to highlight the educational benefits arising from the use of virtual reality technology in medicine and study the emerging area of web-based medical simulations. Finally, an innovative virtual reality environment for distance education in medicine is demonstrated. The proposed environment reproduces conditions of the real learning process and enhances learning through a real-time interactive simulator.

Fast Forecasting of Stock Market Prices by using New High Speed Time Delay Neural Networks

Fast forecasting of stock market prices is very important for strategic planning. In this paper, a new approach for fast forecasting of stock market prices is presented. Such algorithm uses new high speed time delay neural networks (HSTDNNs). The operation of these networks relies on performing cross correlation in the frequency domain between the input data and the input weights of neural networks. It is proved mathematically and practically that the number of computation steps required for the presented HSTDNNs is less than that needed by traditional time delay neural networks (TTDNNs). Simulation results using MATLAB confirm the theoretical computations.

Stability Issues on an Implemented All-Pass Filter Circuitry

The so-called all-pass filter circuits are commonly used in the field of signal processing, control and measurement. Being connected to capacitive loads, these circuits tend to loose their stability; therefore the elaborate analysis of their dynamic behavior is necessary. The compensation methods intending to increase the stability of such circuits are discussed in this paper, including the socalled lead-lag compensation technique being treated in detail. For the dynamic modeling, a two-port network model of the all-pass filter is being derived. The results of the model analysis show, that effective lead-lag compensation can be achieved, alone by the optimization of the circuit parameters; therefore the application of additional electric components are not needed to fulfill the stability requirement.

Performance Improvements of DSP Applications on a Generic Reconfigurable Platform

Speedups from mapping four real-life DSP applications on an embedded system-on-chip that couples coarsegrained reconfigurable logic with an instruction-set processor are presented. The reconfigurable logic is realized by a 2-Dimensional Array of Processing Elements. A design flow for improving application-s performance is proposed. Critical software parts, called kernels, are accelerated on the Coarse-Grained Reconfigurable Array. The kernels are detected by profiling the source code. For mapping the detected kernels on the reconfigurable logic a prioritybased mapping algorithm has been developed. Two 4x4 array architectures, which differ in their interconnection structure among the Processing Elements, are considered. The experiments for eight different instances of a generic system show that important overall application speedups have been reported for the four applications. The performance improvements range from 1.86 to 3.67, with an average value of 2.53, compared with an all-software execution. These speedups are quite close to the maximum theoretical speedups imposed by Amdahl-s law.

Development of Better Quality Low-Cost Activated Carbon from South African Pine Tree (Pinus patula) Sawdust: Characterization and Comparative Phenol Adsorption

The remediation of water resources pollution in developing countries requires the application of alternative sustainable cheaper and efficient end-of-pipe wastewater treatment technologies. The feasibility of use of South African cheap and abundant pine tree (Pinus patula) sawdust for development of lowcost AC of comparable quality to expensive commercial ACs in the abatement of water pollution was investigated. AC was developed at optimized two-stage N2-superheated steam activation conditions in a fixed bed reactor, and characterized for proximate and ultimate properties, N2-BET surface area, pore size distribution, SEM, pHPZC and FTIR. The sawdust pyrolysis activation energy was evaluated by TGA. Results indicated that the chars prepared at 800oC and 2hrs were suitable for development of better quality AC at 800oC and 47% burn-off having BET surface area (1086m2/g), micropore volume (0.26cm3/g), and mesopore volume (0.43cm3/g) comparable to expensive commercial ACs, and suitable for water contaminants removal. The developed AC showed basic surface functionality at pHPZC at 10.3, and a phenol adsorption capacity that was higher than that of commercial Norit (RO 0.8) AC. Thus, it is feasible to develop better quality low-cost AC from (Pinus patula) sawdust using twostage N2-steam activation in fixed-bed reactor.

Energy Consumption Analysis of Design Patterns

The importance of low power consumption is widely acknowledged due to the increasing use of portable devices, which require minimizing the consumption of energy. Energy dissipation is heavily dependent on the software used in the system. Applying design patterns in object-oriented designs is a common practice nowadays. In this paper we analyze six design patterns and explore the effect of them on energy consumption and performance.

Demand and Price Evolution Forecasting as Tools for Facilitating the RoadMapping Process of the Photonic Component Industry

The photonic component industry is a highly innovative industry with a large value chain. In order to ensure the growth of the industry much effort must be devoted to road mapping activities. In such activities demand and price evolution forecasting tools can prove quite useful in order to help in the roadmap refinement and update process. This paper attempts to provide useful guidelines in roadmapping of optical components and considers two models based on diffusion theory and the extended learning curve for demand and price evolution forecasting.

Integrating Agents and Computational Intelligence Techniques in E-learning Environments

In this contribution a newly developed elearning environment is presented, which incorporates Intelligent Agents and Computational Intelligence Techniques. The new e-learning environment is constituted by three parts, the E-learning platform Front-End, the Student Questioner Reasoning and the Student Model Agent. These parts are distributed geographically in dispersed computer servers, with main focus on the design and development of these subsystems through the use of new and emerging technologies. These parts are interconnected in an interoperable way, using web services for the integration of the subsystems, in order to enhance the user modelling procedure and achieve the goals of the learning process.

Implementation of Vertical Neutron Camera (VNC) for ITER Fusion Plasma Neutron Source Profile Reconstruction

In present work the problem of the ITER fusion plasma neutron source parameter reconstruction using only the Vertical Neutron Camera data was solved. The possibility of neutron source parameter reconstruction was estimated by the numerical simulations and the analysis of adequateness of mathematic model was performed. The neutron source was specified in a parametric form. The numerical analysis of solution stability with respect to data distortion was done. The influence of the data errors on the reconstructed parameters is shown: • is reconstructed with errors less than 4% at all examined values of δ (until 60%); • is determined with errors less than 10% when δ do not overcome 5%; • is reconstructed with relative error more than 10 %; • integral intensity of the neutron source is determined with error 10% while δ error is less than 15%; where -error of signal measurements, (R0,Z0), the plasma center position,- /parameter of neutron source profile.

A Further Improvement on the Resurrected Core-Spreading Vortex Method

In a previously developed fast vortex method, the diffusion of the vortex sheet induced at the solid wall by the no-slip boundary conditions was modeled according to the approximation solution of Koumoutsakos and converted into discrete blobs in the vicinity of the wall. This scheme had been successfully applied to a simulation of the flow induced with an impulsively initiated circular cylinder. In this work, further modifications on this vortex method are attempted, including replacing the approximation solution by the boundary-element-method solution, incorporating a new algorithm for handling the over-weak vortex blobs, and diffusing the vortex sheet circulation in a new way suitable for high-curvature solid bodies. The accuracy is thus largely improved. The predictions of lift and drag coefficients for a uniform flow past a NASA airfoil agree well with the existing literature.

A Software-Supported Methodology for Designing General-Purpose Interconnection Networks for Reconfigurable Architectures

Modern applications realized onto FPGAs exhibit high connectivity demands. Throughout this paper we study the routing constraints of Virtex devices and we propose a systematic methodology for designing a novel general-purpose interconnection network targeting to reconfigurable architectures. This network consists of multiple segment wires and SB patterns, appropriately selected and assigned across the device. The goal of our proposed methodology is to maximize the hardware utilization of fabricated routing resources. The derived interconnection scheme is integrated on a Virtex style FPGA. This device is characterized both for its high-performance, as well as for its low-energy requirements. Due to this, the design criterion that guides our architecture selections was the minimal Energy×Delay Product (EDP). The methodology is fully-supported by three new software tools, which belong to MEANDER Design Framework. Using a typical set of MCNC benchmarks, extensive comparison study in terms of several critical parameters proves the effectiveness of the derived interconnection network. More specifically, we achieve average Energy×Delay Product reduction by 63%, performance increase by 26%, reduction in leakage power by 21%, reduction in total energy consumption by 11%, at the expense of increase of channel width by 20%.

The Self-Energy of an Ellectron Bound in a Coulomb Field

Recent progress in calculation of the one-loop selfenergy of the electron bound in the Coulomb field is summarized. The relativistic multipole expansion is introduced. This expansion is based on a single assumption: except for the part of the time component of the electron four-momentum corresponding to the electron rest mass, the exchange of four-momentum between the virtual electron and photon can be treated perturbatively. For non Sstates and normalized difference n3En −E1 of the S-states this itself yields very accurate results after taking the method to the third order. For the ground state the perturbation treatment of the electron virtual states with very high three-momentum is to be avoided. For these states one can always rearrange the pertinent expression in such a way that free-particle approximation is allowed. Combination of the relativistic multipole expansion and free-particle approximation yields very accurate result after taking the method to the ninth order. These results are in very good agreement with the previous results obtained by the partial wave expansion and definitely exclude the possibility that the uncertainity in determination of the proton radius comes from the uncertainity in the calculation of the one-loop selfenergy.

Using Teager Energy Cepstrum and HMM distancesin Automatic Speech Recognition and Analysis of Unvoiced Speech

In this study, the use of silicon NAM (Non-Audible Murmur) microphone in automatic speech recognition is presented. NAM microphones are special acoustic sensors, which are attached behind the talker-s ear and can capture not only normal (audible) speech, but also very quietly uttered speech (non-audible murmur). As a result, NAM microphones can be applied in automatic speech recognition systems when privacy is desired in human-machine communication. Moreover, NAM microphones show robustness against noise and they might be used in special systems (speech recognition, speech conversion etc.) for sound-impaired people. Using a small amount of training data and adaptation approaches, 93.9% word accuracy was achieved for a 20k Japanese vocabulary dictation task. Non-audible murmur recognition in noisy environments is also investigated. In this study, further analysis of the NAM speech has been made using distance measures between hidden Markov model (HMM) pairs. It has been shown the reduced spectral space of NAM speech using a metric distance, however the location of the different phonemes of NAM are similar to the location of the phonemes of normal speech, and the NAM sounds are well discriminated. Promising results in using nonlinear features are also introduced, especially under noisy conditions.

Concurrency without Locking in Parallel Hash Structures used for Data Processing

Various mechanisms providing mutual exclusion and thread synchronization can be used to support parallel processing within a single computer. Instead of using locks, semaphores, barriers or other traditional approaches in this paper we focus on alternative ways for making better use of modern multithreaded architectures and preparing hash tables for concurrent accesses. Hash structures will be used to demonstrate and compare two entirely different approaches (rule based cooperation and hardware synchronization support) to an efficient parallel implementation using traditional locks. Comparison includes implementation details, performance ranking and scalability issues. We aim at understanding the effects the parallelization schemes have on the execution environment with special focus on the memory system and memory access characteristics.

Ionic Liquid Promoted One-pot Synthesis of Benzo[b][1,4]oxazines

benzo[b][1,4]oxazines have been synthesized in good to excellent yields in the presence of the ionic liquid 1-butyl-3- methylimidazolium bromide [bmim]Br under relatively mild conditions without any added catalyst, the reaction workup is simple and the ionic liquid can be easily separated from the product and reused.