A Learning-Community Recommendation Approach for Web-Based Cooperative Learning

Cooperative learning has been defined as learners working together as a team to solve a problem to complete a task or to accomplish a common goal, which emphasizes the importance of interactions among members to promote the whole learning performance. With the popularity of society networks, cooperative learning is no longer limited to traditional classroom teaching activities. Since society networks facilitate to organize online learners, to establish common shared visions, and to advance learning interaction, the online community and online learning community have triggered the establishment of web-based societies. Numerous research literatures have indicated that the collaborative learning community is a critical issue to enhance learning performance. Hence, this paper proposes a learning community recommendation approach to facilitate that a learner joins the appropriate learning communities, which is based on k-nearest neighbor (kNN) classification. To demonstrate the viability of the proposed approach, the proposed approach is implemented for 117 students to recommend learning communities. The experimental results indicate that the proposed approach can effectively recommend appropriate learning communities for learners.

Iris Localization using Circle and Fuzzy Circle Detection Method

Iris localization is a very important approach in biometric identification systems. Identification process usually is implemented in three levels: iris localization, feature extraction, and pattern matching finally. Accuracy of iris localization as the first step affects all other levels and this shows the importance of iris localization in an iris based biometric system. In this paper, we consider Daugman iris localization method as a standard method, propose a new method in this field and then analyze and compare the results of them on a standard set of iris images. The proposed method is based on the detection of circular edge of iris, and improved by fuzzy circles and surface energy difference contexts. Implementation of this method is so easy and compared to the other methods, have a rather high accuracy and speed. Test results show that the accuracy of our proposed method is about Daugman method and computation speed of it is 10 times faster.

Globally Convergent Edge-preserving Reconstruction with Contour-line Smoothing

The standard approach to image reconstruction is to stabilize the problem by including an edge-preserving roughness penalty in addition to faithfulness to the data. However, this methodology produces noisy object boundaries and creates a staircase effect. The existing attempts to favor the formation of smooth contour lines take the edge field explicitly into account; they either are computationally expensive or produce disappointing results. In this paper, we propose to incorporate the smoothness of the edge field in an implicit way by means of an additional penalty term defined in the wavelet domain. We also derive an efficient half-quadratic algorithm to solve the resulting optimization problem, including the case when the data fidelity term is non-quadratic and the cost function is nonconvex. Numerical experiments show that our technique preserves edge sharpness while smoothing contour lines; it produces visually pleasing reconstructions which are quantitatively better than those obtained without wavelet-domain constraints.

Induction of alpha-Amylase in Wheat Grain Cultivars as an Indicator of Resistance to Pre-harvest Sprouting

The influence of humidity and low temperature on the α- amylase activity and isoenzyme composition of grains of different wheat varieties have been studied. The identified samples of varieties have significant difference in the level of enzyme induction under the impact of high humidity and low temperature. It is proposed to use this methodological approach for testing genotypes and wheat breeding lines for resistance to pre-harvest sprouting (PHS).

Concurrency without Locking in Parallel Hash Structures used for Data Processing

Various mechanisms providing mutual exclusion and thread synchronization can be used to support parallel processing within a single computer. Instead of using locks, semaphores, barriers or other traditional approaches in this paper we focus on alternative ways for making better use of modern multithreaded architectures and preparing hash tables for concurrent accesses. Hash structures will be used to demonstrate and compare two entirely different approaches (rule based cooperation and hardware synchronization support) to an efficient parallel implementation using traditional locks. Comparison includes implementation details, performance ranking and scalability issues. We aim at understanding the effects the parallelization schemes have on the execution environment with special focus on the memory system and memory access characteristics.

Profit Efficiency and Competitiveness of Commercial Banks in Malaysia

This paper attempts to identify the significance of Information and Communications Technology (ICT) and competitiveness to the profit efficiency of commercial banks in Malaysia. The profit efficiency of commercial banks in Malaysia, the dependent variable, was estimated using the Stochastic Frontier Approach (SFA) on a sample of unbalanced panel data, covering 23 commercial banks, between 1995 to 2007. Based on the empirical results, ICT was not found to exert a significant impact on profit efficiency, whereas competitiveness, non ICT stock expenditure and ownership were significant contributors. On the other hand, the size of banks was found to have significantly reduced profit efficiency, opening up for various interpretations of the interrelated role of ICT and competition.

Determination and Assessment of Ground Motion and Spectral Parameters for Iran

Many studies have been conducted for derivation of attenuation relationships worldwide, however few relationships have been developed to use for the seismic region of Iranian plateau and only few of these studies have been conducted for derivation of attenuation relationships for parameters such as uniform duration. Uniform duration is the total time during which the acceleration is larger than a given threshold value (default is 5% of PGA). In this study, the database was same as that used previously by Ghodrati Amiri et al. (2007) with same correction methods for earthquake records in Iran. However in this study, records from earthquakes with MS< 4.0 were excluded from this database, each record has individually filtered afterward, and therefore the dataset has been expanded. These new set of attenuation relationships for Iran are derived based on tectonic conditions with soil classification into rock and soil. Earthquake parameters were chosen to be hypocentral distance and magnitude in order to make it easier to use the relationships for seismic hazard analysis. Tehran is the capital city of Iran wit ha large number of important structures. In this study, a probabilistic approach has been utilized for seismic hazard assessment of this city. The resulting uniform duration against return period diagrams are suggested to be used in any projects in the area.

Calibration of Parallel Multi-View Cameras

This paper focuses on the calibration problem of a multi-view shooting system designed for the production of 3D content for auto-stereoscopic visualization. The considered multiview camera is characterized by coplanar and decentered image sensors regarding to the corresponding optical axis. Based on the Faugéras and Toscani-s calibration approach, a calibration method is herein proposed for the case of multi-view camera with parallel and decentered image sensors. At first, the geometrical model of the shooting system is recalled and some industrial prototypes with some shooting simulations are presented. Next, the development of the proposed calibration method is detailed. Finally, some simulation results are presented before ending with some conclusions about this work.

A Usability Testing Approach to Evaluate User-Interfaces in Business Administration

This interdisciplinary study is an investigation to evaluate user-interfaces in business administration. The study is going to be implemented on two computerized business administration systems with two distinctive user-interfaces, so that differences between the two systems can be determined. Both systems, a commercial and a prototype developed for the purpose of this study, deal with ordering of supplies, tendering procedures, issuing purchase orders, controlling the movement of the stocks against their actual balances on the shelves and editing them on their tabulations. In the second suggested system, modern computer graphics and multimedia issues were taken into consideration to cover the drawbacks of the first system. To highlight differences between the two investigated systems regarding some chosen standard quality criteria, the study employs various statistical techniques and methods to evaluate the users- interaction with both systems. The study variables are divided into two divisions: independent representing the interfaces of the two systems, and dependent embracing efficiency, effectiveness, satisfaction, error rate etc.

Recovering Artifacts from Legacy Systems Using Pattern Matching

Modernizing legacy applications is the key issue facing IT managers today because there's enormous pressure on organizations to change the way they run their business to meet the new requirements. The importance of software maintenance and reengineering is forever increasing. Understanding the architecture of existing legacy applications is the most critical issue for maintenance and reengineering. The artifacts recovery can be facilitated with different recovery approaches, methods and tools. The existing methods provide static and dynamic set of techniques for extracting architectural information, but are not suitable for all users in different domains. This paper presents a simple and lightweight pattern extraction technique to extract different artifacts from legacy systems using regular expression pattern specifications with multiple language support. We used our custom-built tool DRT to recover artifacts from existing system at different levels of abstractions. In order to evaluate our approach a case study is conducted.

Nonlinear Optimal Line-Of-Sight Stabilization with Fuzzy Gain-Scheduling

A nonlinear optimal controller with a fuzzy gain scheduler has been designed and applied to a Line-Of-Sight (LOS) stabilization system. Use of Linear Quadratic Regulator (LQR) theory is an optimal and simple manner of solving many control engineering problems. However, this method cannot be utilized directly for multigimbal LOS systems since they are nonlinear in nature. To adapt LQ controllers to nonlinear systems at least a linearization of the model plant is required. When the linearized model is only valid within the vicinity of an operating point a gain scheduler is required. Therefore, a Takagi-Sugeno Fuzzy Inference System gain scheduler has been implemented, which keeps the asymptotic stability performance provided by the optimal feedback gain approach. The simulation results illustrate that the proposed controller is capable of overcoming disturbances and maintaining a satisfactory tracking performance.

An E-Learning Tool for The Self-Study of Mathematics for the CPE Examination

In this paper, we give an overview of an online elearning tool which has been developed for kids aged from nine to eleven years old in Mauritius for the self-study of Mathematics in order to prepare them for the CPE examination. The software does not intend to render obsolete the existing pedagogical approaches. Nowadays, the teaching-learning process is mainly focused towards the class-room model. Moreover, most of the e-learning platforms that exist are simply static ways of delivering resources using the internet. There is nearly no interaction between the learner and the tool. Our application will enable students to practice exercises online and also work out sample examination papers. Another interesting feature is that the kid will not have to wait for someone to correct the work as the correction will be done online and on the spot. Additional feedback is also provided for some exercises.

Simulation of Snow Covers Area by a Physical based Model

Snow cover is an important phenomenon in hydrology, hence modeling the snow accumulation and melting is an important issue in places where snowmelt significantly contributes to runoff and has significant effect on water balance. The physics-based models are invariably distributed, with the basin disaggregated into zones or grid cells. Satellites images provide valuable data to verify the accuracy of spatially distributed model outputs. In this study a spatially distributed physically based model (WetSpa) was applied to predict snow cover and melting in the Latyan dam watershed in Iran. Snowmelt is simulated based on an energy balance approach. The model is applied and calibrated with one year of observed daily precipitation, air temperature, windspeed, and daily potential evaporation. The predicted snow-covered area is compared with remotely sensed images (MODIS). The results show that simulated snow cover area SCA has a good agreement with satellite image snow cover area SCA from MODIS images. The model performance is also tested by statistical and graphical comparison of simulated and measured discharges entering the Latyan dam reservoir.

Simulation of Water Droplet on Horizontally Smooth and Rough Surfaces Using Quasi-Molecular Modelling

We developed a method based on quasi-molecular modelling to simulate the fall of water drops on horizontally smooth and rough surfaces. Each quasi-molecule was a group of particles that interacted in a fashion entirely analogous to classical Newtonian molecular interactions. When a falling water droplet was simulated at low impact velocity on both smooth and rough surfaces, the droplets moved periodically (i.e. the droplets moved up and down for a certain period, finally they stopped moving and reached a steady state), spreading and recoiling without splash or break-up. Spreading rates of falling water droplets increased rapidly as time increased until the spreading rate reached its steady state at time t ~ 0.25 s for rough surface and t ~ 0.40 s for smooth surface. The droplet height above both surfaces decreased as time increased, remained constant after the droplet diameter attained a maximum value and reached its steady state at time t ~ 0.4 s. However, rough surface had higher spreading rates of falling water droplets and lower height on the surface than smooth one.

Thermodynamic, Structural and Transport Properties of Molten Copper-Thallium Alloys

A self-association model has been used to understand the concentration dependence of free energy of mixing (GM), heat of mixing (HM), entropy of mixing (SM), activity (a) and microscopic structures, such as concentration fluctuation in long wavelength limit (Scc(0)) and Warren-Cowley short range order parameter ( 1 α )for Cu- Tl molten alloys at 1573K. A comparative study of surface tension of the alloys in the liquid state at that temperature has also been carried out theoretically as function of composition in the light of Butler-s model, Prasad-s model and quasi-chemical approach. Most of the computed thermodynamic properties have been found in agreement with the experimental values. The analysis reveals that the Cu-Tl molten alloys at 1573K represent a segregating system at all concentrations with moderate interaction. Surface tensions computed from different approaches have been found to be comparable to each other showing increment with the composition of copper.

A Frame Work for Query Results Refinement in Multimedia Databases

In the current age, retrieval of relevant information from massive amount of data is a challenging job. Over the years, precise and relevant retrieval of information has attained high significance. There is a growing need in the market to build systems, which can retrieve multimedia information that precisely meets the user's current needs. In this paper, we have introduced a framework for refining query results before showing it to the user, using ambient intelligence, user profile, group profile, user location, time, day, user device type and extracted features. A prototypic tool was also developed to demonstrate the efficiency of the proposed approach.

MPPT Operation for PV Grid-connected System using RBFNN and Fuzzy Classification

This paper presents a novel methodology for Maximum Power Point Tracking (MPPT) of a grid-connected 20 kW Photovoltaic (PV) system using neuro-fuzzy network. The proposed method predicts the reference PV voltage guarantying optimal power transfer between the PV generator and the main utility grid. The neuro-fuzzy network is composed of a fuzzy rule-based classifier and three Radial Basis Function Neural Networks (RBFNN). Inputs of the network (irradiance and temperature) are classified before they are fed into the appropriated RBFNN for either training or estimation process while the output is the reference voltage. The main advantage of the proposed methodology, comparing to a conventional single neural network-based approach, is the distinct generalization ability regarding to the nonlinear and dynamic behavior of a PV generator. In fact, the neuro-fuzzy network is a neural network based multi-model machine learning that defines a set of local models emulating the complex and non-linear behavior of a PV generator under a wide range of operating conditions. Simulation results under several rapid irradiance variations proved that the proposed MPPT method fulfilled the highest efficiency comparing to a conventional single neural network.

Finite Element Modeling of two-dimensional Nanoscale Structures with Surface Effects

Nanomaterials have attracted considerable attention during the last two decades, due to their unusual electrical, mechanical and other physical properties as compared with their bulky counterparts. The mechanical properties of nanostructured materials show strong size dependency, which has been explained within the framework of continuum mechanics by including the effects of surface stress. The size-dependent deformations of two-dimensional nanosized structures with surface effects are investigated in the paper by the finite element method. Truss element is used to evaluate the contribution of surface stress to the total potential energy and the Gurtin and Murdoch surface stress model is implemented with ANSYS through its user programmable features. The proposed approach is used to investigate size-dependent stress concentration around a nanosized circular hole and the size-dependent effective moduli of nanoporous materials. Numerical results are compared with available analytical results to validate the proposed modeling approach.

Towards a Suitable and Systematic Approach for Component Based Software Development

Software crisis refers to the situation in which the developers are not able to complete the projects within time and budget constraints and moreover these overscheduled and over budget projects are of low quality as well. Several methodologies have been adopted form time to time to overcome this situation and now in the focus is component based software engineering. In this approach, emphasis is on reuse of already existing software artifacts. But the results can not be achieved just by preaching the principles; they need to be practiced as well. This paper highlights some of the very basic elements of this approach, which has to be in place to get the desired goals of high quality, low cost with shorter time-to-market software products.

Simulated Annealing and Genetic Algorithm in Telecommunications Network Planning

The main goal of this work is to propose a way for combined use of two nontraditional algorithms by solving topological problems on telecommunications concentrator networks. The algorithms suggested are the Simulated Annealing algorithm and the Genetic Algorithm. The Algorithm of Simulated Annealing unifies the well known local search algorithms. In addition - Simulated Annealing allows acceptation of moves in the search space witch lead to decisions with higher cost in order to attempt to overcome any local minima obtained. The Genetic Algorithm is a heuristic approach witch is being used in wide areas of optimization works. In the last years this approach is also widely implemented in Telecommunications Networks Planning. In order to solve less or more complex planning problem it is important to find the most appropriate parameters for initializing the function of the algorithm.