Measuring Cognitive Load - A Solution to Ease Learning of Programming

Learning programming is difficult for many learners. Some researches have found that the main difficulty relates to cognitive load. Cognitive overload happens in programming due to the nature of the subject which is intrinisicly over-bearing on the working memory. It happens due to the complexity of the subject itself. The problem is made worse by the poor instructional design methodology used in the teaching and learning process. Various efforts have been proposed to reduce the cognitive load, e.g. visualization softwares, part-program method etc. Use of many computer based systems have also been tried to tackle the problem. However, little success has been made to alleviate the problem. More has to be done to overcome this hurdle. This research attempts at understanding how cognitive load can be managed so as to reduce the problem of overloading. We propose a mechanism to measure the cognitive load during pre instruction, post instruction and in instructional stages of learning. This mechanism is used to help the instruction. As the load changes the instruction is made to adapt itself to ensure cognitive viability. This mechanism could be incorporated as a sub domain in the student model of various computer based instructional systems to facilitate the learning of programming.

Remarks on Energy Based Control of a Nonlinear, Underactuated, MIMO and Unstable Benchmark

In the last decade, energy based control theory has undergone a significant breakthrough in dealing with underactated mechanical systems with two successful and similar tools, controlled Lagrangians and controlled Hamiltanians (IDA-PBC). However, because of the complexity of these tools, successful case studies are lacking, in particular, MIMO cases. The seminal theoretical paper of controlled Lagrangians proposed by Bloch and his colleagues presented a benchmark example–a 4 d.o.f underactuated pendulum on a cart but a detailed and completed design is neglected. To compensate this ignorance, the note revisit their design idea by addressing explicit control functions for a similar device motivated by a vector thrust body hovering in the air. To the best of our knowledge, this system is the first MIMO, underactuated example that is stabilized by using energy based tools at the courtesy of the original design idea. Some observations are given based on computer simulation.

Towards Cloud Computing Anatomy

Cloud Computing has recently emerged as a compelling paradigm for managing and delivering services over the internet. The rise of Cloud Computing is rapidly changing the landscape of information technology, and ultimately turning the longheld promise of utility computing into a reality. As the development of Cloud Computing paradigm is speedily progressing, concepts, and terminologies are becoming imprecise and ambiguous, as well as different technologies are interfering. Thus, it becomes crucial to clarify the key concepts and definitions. In this paper, we present the anatomy of Cloud Computing, covering its essential concepts, prominent characteristics, its affects, architectural design and key technologies. We differentiate various service and deployment models. Also, significant challenges and risks need are tackled in order to guarantee the long-term success of Cloud Computing. The aim of this paper is to provide a better understanding of the anatomy of Cloud Computing and pave the way for further research in this area.

Performance of Heterogeneous Autoregressive Models of Realized Volatility: Evidence from U.S. Stock Market

This paper deals with heterogeneous autoregressive models of realized volatility (HAR-RV models) on high-frequency data of stock indices in the USA. Its aim is to capture the behavior of three groups of market participants trading on a daily, weekly and monthly basis and assess their role in predicting the daily realized volatility. The benefits of this work lies mainly in the application of heterogeneous autoregressive models of realized volatility on stock indices in the USA with a special aim to analyze an impact of the global financial crisis on applied models forecasting performance. We use three data sets, the first one from the period before the global financial crisis occurred in the years 2006-2007, the second one from the period when the global financial crisis fully hit the U.S. financial market in 2008-2009 years, and the last period was defined over 2010-2011 years. The model output indicates that estimated realized volatility in the market is very much determined by daily traders and in some cases excludes the impact of those market participants who trade on monthly basis.

Axisymmetric Vibration of Pyrocomposite Hollow Cylinder

Axisymmetric vibration of an infinite Pyrocomposite circular hollow cylinder made of inner and outer pyroelectric layer of 6mm-class bonded together by a Linear Elastic Material with Voids (LEMV) layer is studied. The exact frequency equation is obtained for the traction free surfaces with continuity condition at the interfaces. Numerical results in the form of data and dispersion curves for the first and second mode of the axisymmetric vibration of the cylinder BaTio3 / Adhesive / BaTio3 by taking the Adhesive layer as an existing Carbon Fibre Reinforced Polymer (CFRP) are compared with a hypothetical LEMV layer with and without voids and as well with a pyroelectric hollow cylinder. The damping is analyzed through the imaginary parts of the complex frequencies.

Development of Integrated GIS Interface for Characteristics of Regional Daily Flow

The purpose of this paper primarily intends to develop GIS interface for estimating sequences of stream-flows at ungauged stations based on known flows at gauged stations. The integrated GIS interface is composed of three major steps. The first, precipitation characteristics using statistical analysis is the procedure for making multiple linear regression equation to get the long term mean daily flow at ungauged stations. The independent variables in regression equation are mean daily flow and drainage area. Traditionally, mean flow data are generated by using Thissen polygon method. However, method for obtaining mean flow data can be selected by user such as Kriging, IDW (Inverse Distance Weighted), Spline methods as well as other traditional methods. At the second, flow duration curve (FDC) is computing at unguaged station by FDCs in gauged stations. Finally, the mean annual daily flow is computed by spatial interpolation algorithm. The third step is to obtain watershed/topographic characteristics. They are the most important factors which govern stream-flows. In summary, the simulated daily flow time series are compared with observed times series. The results using integrated GIS interface are closely similar and are well fitted each other. Also, the relationship between the topographic/watershed characteristics and stream flow time series is highly correlated.

Exploring Customer Trust in B2C Mobile Payments – A Qualitative Study

Mobile payments have been deployed by businesses for more than a decade. Customers use mobile payments if they trust in this relatively new payment method, have a belief and confidence in, as well as reliance on its services and applications. Despite its potential, the current literature shows that there is lack of customer trust in B2C mobile payments, and a lack of studies that determine the factors that influence their trust in these payments; which make these factors yet to be understood, especially in the Middle East region. Thus, this study aims to explore the factors that influence customer trust in mobile payments. The empirical data for this explorative study was collected by establishing four focus group sessions in the UAE. The results indicate that the explored significant factors can be classified into five main groups: customer characteristics, environmental (social and cultural) influences, provider characteristics, mobile-device characteristics, and perceived risks.

The Evaluation and the Comparison of the Effect of Without Engine Power and Power Mechanical Systems on Rice Weed

In order to study the influence of different methods of controlling weeds such as mechanical weeding and mechanical weeder efficiency analysis in mechanical cultivation conditions, in farming year of 2011 an experiment was done in a farm in coupling and development of technology center in Haraz,Iran. The treatments consisted of (I) control treatment: where no weeding was done, (II) use of mechanical weeding without engine and (III) power mechanical weeding. Results showed that experimental treatments had significantly different effects (p=0.05) on yield traits and number of filled grains per panicle, while treatments had the significant effects on grain weight and dry weight of weeds in the first, second and third weeding methods at 1% of confidence level. Treatment (II) had its most significant effect on number of filled grains per panicle and yield performance standpoint, which was 3705.97 kg ha-1 in its highest peak. Treatment (III) was ranked as second influential with 3559.8 kg ha-1. In addition, under (I) treatments, 2364.73 kg ha-1 of yield produced. The minimum dry weights of weeds in all weeding methods were related to the treatment (II), (III) and (I), respectively. The correlation coefficient analysis showed that total yield had a significant positive correlation with the panicle grain yield per plant (r= 0.55*) and the number of grains per panicle-1 (r= 0.57*) and the number of filled grains (r= 0.63*). Total rice yield also had negative correlation of r= -0. 64* with weed dry weight at second weed sampling time (17 DAT). The weed dry weight at third and fourth sampling times (24 and 40 DAT) had negative correlations of -0.65** and r=-0.61* with rice yield, respectively.

Adaptive Early Packet Discarding Policy Based on Two Traffic Classes

Unlike the best effort service provided by the internet today, next-generation wireless networks will support real-time applications. This paper proposes an adaptive early packet discard (AEPD) policy to improve the performance of the real time TCP traffic over ATM networks and avoid the fragmentation problem. Three main aspects are incorporated in the proposed policy. First, providing quality-of-service (QoS) guaranteed for real-time applications by implementing a priority scheduling. Second, resolving the partially corrupted packets problem by differentiating the buffered cells of one packet from another. Third, adapting a threshold dynamically using Fuzzy logic based on the traffic behavior to maintain a high throughput under a variety of load conditions. The simulation is run for two priority classes of the input traffic: real time and non-real time classes. Simulation results show that the proposed AEPD policy improves throughput and fairness over that using static threshold under the same traffic conditions.

Exergy Analysis of a Cogeneration Plant

Cogeneration may be defined as a system which contains electricity production and regain of the thermo value of exhaust gases simultaneously. The examination is based on the data-s of an active cogeneration plant. This study, it is aimed to determine which component of the system should be revised first to raise the efficiency and decrease the loss of exergy. For this purpose, second law analysis of thermodynamics is applied to each component due to consider the effects of environmental conditions and take the quality of energy into consideration as well as the quantity of it. The exergy balance equations are produced and exergy loss is calculated for each component. 44,44 % loss of exergy in heat exchanger, 29,59 % in combustion chamber, 18,68 % in steam boiler, 5,25 % in gas turbine and 2,03 % in compressor is calculated.

The Knapsack Sharing Problem: A Tree Search Exact Algorithm

In this paper, we study the knapsack sharing problem, a variant of the well-known NP-Hard single knapsack problem. We investigate the use of a tree search for optimally solving the problem. The used method combines two complementary phases: a reduction interval search phase and a branch and bound procedure one. First, the reduction phase applies a polynomial reduction strategy; that is used for decomposing the problem into a series of knapsack problems. Second, the tree search procedure is applied in order to attain a set of optimal capacities characterizing the knapsack problems. Finally, the performance of the proposed optimal algorithm is evaluated on a set of instances of the literature and its runtime is compared to the best exact algorithm of the literature.

An Efficient Protocol for Cyclic Somatic Embryogenesis in Neem (Azadirachta indica A Juss.)

Neem is a highly heterozygous and commercially important perennial plant. Conventionally, it is propagated by seeds which loose viability within two weeks. Strictly cross pollinating nature of the plant causes serious barrier to the genetic improvement by conventional methods. Alternative methods of tree improvement such as somatic hybridization, mutagenesis and genetic transformation require an efficient in vitro plant regeneration system. In this regard, somatic embryogenesis particularly secondary somatic embryogenesis may offer an effective system for large scale plant propagation without affecting the clonal fidelity of the regenerants. It can be used for synthetic seed production, which further bolsters conservation of this tree species which is otherwise very difficult The present report describes the culture conditions necessary to induce and maintain repetitive somatic embryogenesis, for the first time, in neem. Out of various treatments tested, the somatic embryos were induced directly from immature zygotic embryos of neem on MS + TDZ (0.1 μM) + ABA (4 μM), in more than 76 % cultures. Direct secondary somatic embryogenesis occurred from primary somatic embryos on MS + IAA (5 μM) + GA3 (5 μM) in 12.5 % cultures. Embryogenic competence of the explant as well as of the primary embryos was maintained for a long period by repeated subcultures at frequent intervals. A maximum of 10 % of these somatic embryos were converted into plantlets.

A Side-Peak Cancellation Scheme for CBOC Code Acquisition

In this paper, we propose a side-peak cancellation scheme for code acquisition of composite binary offset carrier (CBOC) signals. We first model the family of CBOC signals in a generic form, and then, propose a side-peak cancellation scheme by combining correlation functions between the divided sub-carrier and received signals. From numerical results, it is shown that the proposed scheme removes the side-peak completely, and moreover, the resulting correlation function demonstrates the better power ratio performance than the CBOC autocorrelation.

Stresses in Cast Metal Inlays Restored Molars

Cast metal inlays can be used on molars requiring a class II restoration instead amalgam and offer a durable alternative. Because it is known that class II inlays may increase the susceptibility to fracture, it is important to ensure optimal performance in selection of the adequate preparation design to reduce stresses in teeth structures and also in the restorations. The aim of the study was to investigate the influence of preparation design on stress distribution in molars with different class II preparations and in cast metal inlays. The first step of the study was to achieve 3D models in order to analyze teeth and cast metal class II inlays. The geometry of the intact tooth was obtained by 3D scanning using a manufactured device. With a NURBS modeling program the preparations and the appropriately inlays were designed. 3D models of first upper molars of the same shape and size were created. Inlay cavities designs were created using literature data. The geometrical model was exported and the mesh structure of the solid 3D model was created for structural simulations. Stresses were located around the occlusal contact areas. For the studied cases, the stress values were not significant influenced by the taper of the preparation. it was demonstrated stresses are higher in the cast metal restorations and therefore the strength of the teeth is not affected.

CEO Duality and Firm Performance: An Integration of Institutional Perceptive with Agency Theory

The recommendation of the committee on corporate governance for public companies in Nigeria, that the position of the CEO be separated from board chair has generated serious debate among scholars and practitioners. They have questioned the appropriateness of implementing corporate governance model that is based on Anglo-Saxon agency problem characterized by dispersed ownership structure; where markets for corporate control, legal regulation, and contractual incentives are the key governance mechanisms. This paper strives to resolve the argument by adopting an institutional perspective in testing the agency theory on board duality. The study developed a theoretical and empirical model to better understand how ownership structure influences agency conflict and how such affects firm performance. Hence, the study examines the relationship between CEO duality and firm performance using two institutional ownership structures – dispersed ownership and concentrated ownership structures. The empirical results show that CEO duality is negatively correlated with firm performance in Nigeria irrespective of the firm-s ownership structure. The findings give credence to the recommendation of the Peterside Commission on the need to separate the position of CEO from board chair.

Efficient Feature-Based Registration for CT-M R Images Based on NSCT and PSO

Feature-based registration is an effective technique for clinical use, because it can greatly reduce computational costs. However, this technique, which estimates the transformation by using feature points extracted from two images, may cause misalignments. To handle with this limitation, we propose to extract the salient edges and extracted control points (CP) of medical images by using efficiency of multiresolution representation of data nonsubsampled contourlet transform (NSCT) that finds the best feature points. The MR images were first decomposed using the NSCT, and then Edge and CP were extracted from bandpass directional subband of NSCT coefficients and some proposed rules. After edge and CP extraction, mutual information was adopted for the registration of feature points and translation parameters are calculated by using particle swarm optimization (PSO). The experimental results showed that the proposed method produces totally accurate performance for registration medical CT-MR images.

Wavelet based Image Registration Technique for Matching Dental x-rays

Image registration plays an important role in the diagnosis of dental pathologies such as dental caries, alveolar bone loss and periapical lesions etc. This paper presents a new wavelet based algorithm for registering noisy and poor contrast dental x-rays. Proposed algorithm has two stages. First stage is a preprocessing stage, removes the noise from the x-ray images. Gaussian filter has been used. Second stage is a geometric transformation stage. Proposed work uses two levels of affine transformation. Wavelet coefficients are correlated instead of gray values. Algorithm has been applied on number of pre and post RCT (Root canal treatment) periapical radiographs. Root Mean Square Error (RMSE) and Correlation coefficients (CC) are used for quantitative evaluation. Proposed technique outperforms conventional Multiresolution strategy based image registration technique and manual registration technique.

Multi-Objective Fuzzy Model in Optimal Sitingand Sizing of DG for Loss Reduction

This paper presents a possibilistic (fuzzy) model in optimal siting and sizing of Distributed Generation (DG) for loss reduction and improve voltage profile in power distribution system. Multi-objective problem is developed in two phases. In the first one, the set of non-dominated planning solutions is obtained (with respect to the objective functions of fuzzy economic cost, and exposure) using genetic algorithm. In the second phase, one solution of the set of non-dominated solutions is selected as optimal solution, using a suitable max-min approach. This method can be determined operation-mode (PV or PQ) of DG. Because of considering load uncertainty in this paper, it can be obtained realistic results. The whole process of this method has been implemented in the MATLAB7 environment with technical and economic consideration for loss reduction and voltage profile improvement. Through numerical example the validity of the proposed method is verified.

Query Algebra for Semistuctured Data

With the tremendous growth of World Wide Web (WWW) data, there is an emerging need for effective information retrieval at the document level. Several query languages such as XML-QL, XPath, XQL, Quilt and XQuery are proposed in recent years to provide faster way of querying XML data, but they still lack of generality and efficiency. Our approach towards evolving a framework for querying semistructured documents is based on formal query algebra. Two elements are introduced in the proposed framework: first, a generic and flexible data model for logical representation of semistructured data and second, a set of operators for the manipulation of objects defined in the data model. In additional to accommodating several peculiarities of semistructured data, our model offers novel features such as bidirectional paths for navigational querying and partitions for data transformation that are not available in other proposals.

A Matching Algorithm of Minutiae for Real Time Fingerprint Identification System

A lot of matching algorithms with different characteristics have been introduced in recent years. For real time systems these algorithms are usually based on minutiae features. In this paper we introduce a novel approach for feature extraction in which the extracted features are independent of shift and rotation of the fingerprint and at the meantime the matching operation is performed much more easily and with higher speed and accuracy. In this new approach first for any fingerprint a reference point and a reference orientation is determined and then based on this information features are converted into polar coordinates. Due to high speed and accuracy of this approach and small volume of extracted features and easily execution of matching operation this approach is the most appropriate for real time applications.