Highly Scalable, Reversible and Embedded Image Compression System

A new method for low complexity image coding is presented, that permits different settings and great scalability in the generation of the final bit stream. This coding presents a continuoustone still image compression system that groups loss and lossless compression making use of finite arithmetic reversible transforms. Both transformation in the space of color and wavelet transformation are reversible. The transformed coefficients are coded by means of a coding system in depending on a subdivision into smaller components (CFDS) similar to the bit importance codification. The subcomponents so obtained are reordered by means of a highly configure alignment system depending on the application that makes possible the re-configure of the elements of the image and obtaining different levels of importance from which the bit stream will be generated. The subcomponents of each level of importance are coded using a variable length entropy coding system (VBLm) that permits the generation of an embedded bit stream. This bit stream supposes itself a bit stream that codes a compressed still image. However, the use of a packing system on the bit stream after the VBLm allows the realization of a final highly scalable bit stream from a basic image level and one or several enhance levels.

Effective Keyword and Similarity Thresholds for the Discovery of Themes from the User Web Access Patterns

Clustering techniques have been used by many intelligent software agents to group similar access patterns of the Web users into high level themes which express users intentions and interests. However, such techniques have been mostly focusing on one salient feature of the Web document visited by the user, namely the extracted keywords. The major aim of these techniques is to come up with an optimal threshold for the number of keywords needed to produce more focused themes. In this paper we focus on both keyword and similarity thresholds to generate themes with concentrated themes, and hence build a more sound model of the user behavior. The purpose of this paper is two fold: use distance based clustering methods to recognize overall themes from the Proxy log file, and suggest an efficient cut off levels for the keyword and similarity thresholds which tend to produce more optimal clusters with better focus and efficient size.

Accurate Crosstalk Analysis for RLC On-Chip VLSI Interconnect

This work proposes an accurate crosstalk noise estimation method in the presence of multiple RLC lines for the use in design automation tools. This method correctly models the loading effects of non switching aggressors and aggressor tree branches using resistive shielding effect and realistic exponential input waveforms. Noise peak and width expressions have been derived. The results obtained are at good agreement with SPICE results. Results show that average error for noise peak is 4.7% and for the width is 6.15% while allowing a very fast analysis.

MATLAB/SIMULINK Based Model of Single- Machine Infinite-Bus with TCSC for Stability Studies and Tuning Employing GA

With constraints on data availability and for study of power system stability it is adequate to model the synchronous generator with field circuit and one equivalent damper on q-axis known as the model 1.1. This paper presents a systematic procedure for modelling and simulation of a single-machine infinite-bus power system installed with a thyristor controlled series compensator (TCSC) where the synchronous generator is represented by model 1.1, so that impact of TCSC on power system stability can be more reasonably evaluated. The model of the example power system is developed using MATLAB/SIMULINK which can be can be used for teaching the power system stability phenomena, and also for research works especially to develop generator controllers using advanced technologies. Further, the parameters of the TCSC controller are optimized using genetic algorithm. The non-linear simulation results are presented to validate the effectiveness of the proposed approach.

AGENTMAP: A Conceptual Meta-Model of Interacting Simulations

A straightforward and intuitive combination of single simulations into an aggregated master-simulation is not trivial. There are lots of problems, which trigger-specific difficulties during the modeling and execution of such a simulation. In this paper we identify these problems and aim to solve them by mapping the task to the field of multi agent systems. The solution is a new meta-model named AGENTMAP, which is able to mitigate most of the problems and to support intuitive modeling at the same time. This meta-model will be introduced and explained on basis of an example from the e-commerce domain.

The Patterns of Unemployment and the Geography of Social Housing

During the last few decades in the academic field, the debate has increased on the effects of social geography on the opportunities of socioeconomic integration. On one hand, it has been discussed how the contents of the urban structure and social geography affect not only the way people interact, but also their chances of social and economic integration. On the other hand, it has also been discussed how the urban structure is also constrained and transformed by the action of social actors. Without questioning the powerful influence of structural factors, related to the logic of the production system, labor markets, education and training, the research has shown the role played by place of residence in shaping individual outcomes such as unemployment. In the context of this debate the importance of territory of residence with respect to the problem of unemployment has been highlighted. Although statistics of unemployment have already demonstrated the unequal incidence of the phenomenon in social groups, the issue of uneven territorial impact on the phenomenon at intra-urban level remains relatively unknown. The purpose of this article is to show and to interpret the spatial patterns of unemployment in the city of Porto using GIS (Geographic Information System - GIS) technology. Under this analysis the overlap of the spatial patterns of unemployment with the spatial distribution of social housing, allows the discussion of the relationship that occurs between these patterns and the reasons that might explain the relative immutability of socioeconomic problems in some neighborhoods.

WLAN Positioning Based on Joint TOA and RSS Characteristics

WLAN Positioning has been presented by many approaches in literatures using the characteristics of Received Signal Strength (RSS), Time of Arrival (TOA) or Time Difference of Arrival (TDOA), Angle of Arrival (AOA) and cell ID. Among these, RSS approach is the simplest method to implement because there is no need of modification on both access points and client devices whereas its accuracy is terrible due to physical environments. For TOA or TDOA approach, the accuracy is quite acceptable but most researches have to modify either software or hardware on existing WLAN infrastructure. The scales of modifications are made on only access card up to the changes in protocol of WLAN. Hence, it is an unattractive approach to use TOA or TDOA for positioning system. In this paper, the new concept of merging both RSS and TOA positioning techniques is proposed. In addition, the method to achieve TOA characteristic for positioning WLAN user without any extra modification necessarily appended in the existing system is presented. The measurement results confirm that the proposed technique using both RSS and TOA characteristics provides better accuracy than using only either RSS or TOA approach.

Cash Flow Optimization on Synthetic CDOs

Collateralized Debt Obligations are not as widely used nowadays as they were before 2007 Subprime crisis. Nonetheless there remains an enthralling challenge to optimize cash flows associated with synthetic CDOs. A Gaussian-based model is used here in which default correlation and unconditional probabilities of default are highlighted. Then numerous simulations are performed based on this model for different scenarios in order to evaluate the associated cash flows given a specific number of defaults at different periods of time. Cash flows are not solely calculated on a single bought or sold tranche but rather on a combination of bought and sold tranches. With some assumptions, the simplex algorithm gives a way to find the maximum cash flow according to correlation of defaults and maturities. The used Gaussian model is not realistic in crisis situations. Besides present system does not handle buying or selling a portion of a tranche but only the whole tranche. However the work provides the investor with relevant elements on how to know what and when to buy and sell.

On Use of Semiconductor Detector Arrays on COMPASS Tokamak

Semiconductor detector arrays are widely used in high-temperature plasma diagnostics. They have a fast response, which allows observation of many processes and instabilities in tokamaks. In this paper, there are reviewed several diagnostics based on semiconductor arrays as cameras, AXUV photodiodes (referred often as fast “bolometers") and detectors of both soft X-rays and visible light installed on the COMPASS tokamak recently. Fresh results from both spring and summer campaigns in 2012 are introduced. Examples of the utilization of the detectors are shown on the plasma shape determination, fast calculation of the radiation center, two-dimensional plasma radiation tomography in different spectral ranges, observation of impurity inflow, and also on investigation of MHD activity in the COMPASS tokamak discharges.

Performance of Soft Handover Algorithm in Varied Propagation Environments

CDMA cellular networks support soft handover, which guarantees the continuity of wireless services and enhanced communication quality. Cellular networks support multimedia services under varied propagation environmental conditions. In this paper, we have shown the effect of characteristic parameters of the cellular environments on the soft handover performance. We consider path loss exponent, standard deviation of shadow fading and correlation coefficient of shadow fading as the characteristic parameters of the radio propagation environment. A very useful statistical measure for characterizing the performance of mobile radio system is the probability of outage. It is shown through numerical results that above parameters have decisive effect on the probability of outage and hence the overall performance of the soft handover algorithm.

Investigation of VMAT Algorithms and Dosimetry

Purpose: Planning and dosimetry of different VMAT algorithms (SmartArc, Ergo++, Autobeam) is compared with IMRT for Head and Neck Cancer patients. Modelling was performed to rule out the causes of discrepancies between planned and delivered dose. Methods: Five HNC patients previously treated with IMRT were re-planned with SmartArc (SA), Ergo++ and Autobeam. Plans were compared with each other and against IMRT and evaluated using DVHs for PTVs and OARs, delivery time, monitor units (MU) and dosimetric accuracy. Modelling of control point (CP) spacing, Leaf-end Separation and MLC/Aperture shape was performed to rule out causes of discrepancies between planned and delivered doses. Additionally estimated arc delivery times, overall plan generation times and effect of CP spacing and number of arcs on plan generation times were recorded. Results: Single arc SmartArc plans (SA4d) were generally better than IMRT and double arc plans (SA2Arcs) in terms of homogeneity and target coverage. Double arc plans seemed to have a positive role in achieving improved Conformity Index (CI) and better sparing of some Organs at Risk (OARs) compared to Step and Shoot IMRT (ss-IMRT) and SA4d. Overall Ergo++ plans achieved best CI for both PTVs. Dosimetric validation of all VMAT plans without modelling was found to be lower than ss-IMRT. Total MUs required for delivery were on average 19%, 30%, 10.6% and 6.5% lower than ss-IMRT for SA4d, SA2d (Single arc with 20 Gantry Spacing), SA2Arcs and Autobeam plans respectively. Autobeam was most efficient in terms of actual treatment delivery times whereas Ergo++ plans took longest to deliver. Conclusion: Overall SA single arc plans on average achieved best target coverage and homogeneity for both PTVs. SA2Arc plans showed improved CI and some OARs sparing. Very good dosimetric results were achieved with modelling. Ergo++ plans achieved best CI. Autobeam resulted in fastest treatment delivery times.

An Interactive Tool for Teaching and Learning English at Upper Primary Level for Mauritius

E-learning refers to the specific kind of learning experienced within the domain of educational technology, which can be used in or out of the classroom. In this paper, we give an overview of an e-learning platform 'An Innovative Interactive and Online English Platform for Upper Primary Students' is an interactive web-based application which will serve as an aid to the primary school students in Mauritius. The objectives of this platform are to offer quality learning resources for the English subject at our primary level of education, encourage self-learning and hence promote e-learning. The platform developed consists of several interesting features, for example, the English Verb Conjugation tool, Negative Form tool, Interrogative Form tool and Close Test Generator. Thus, this learning platform will be useful at a time where our country is looking for an alternative to private tuition and also, looking forward to increase the pass rate.

Computational Initial Value Method for Vibration Analysis of Symmetrically Laminated Composite Plate

In the present paper, an improved initial value numerical technique is presented to analyze the free vibration of symmetrically laminated rectangular plate. A combination of the initial value method (IV) and the finite differences (FD) devices is utilized to develop the present (IVFD) technique. The achieved technique is applied to the equation of motion of vibrating laminated rectangular plate under various types of boundary conditions. Three common types of laminated symmetrically cross-ply, orthotropic and isotropic plates are analyzed here. The convergence and accuracy of the presented Initial Value-Finite Differences (IVFD) technique have been examined. Also, the merits and validity of improved technique are satisfied via comparing the obtained results with those available in literature indicating good agreements.

High Precision Draw Bending of Asymmetric Channel Section with Restriction Dies and Axial Tension

In recent years asymmetric cross section aluminum alloy stock has been finding increasing use in various industrial manufacturing areas such as general structures and automotive components. In these areas, components are generally required to have complex curved configuration and, as such, a bending process is required during manufacture. Undesirable deformation in bending processes such as flattening or wrinkling can easily occur when thin-walled sections are bent. Hence, a thorough understanding of the bending behavior of such sections is needed to prevent these undesirable deformations. In this study, the bending behavior of asymmetric channel section was examined using finite element analysis (FEA). Typical methods of preventing undesirable deformation, such as asymmetric laminated elastic mandrels were included in FEA model of draw bending. Additionally, axial tension was applied to prevent wrinkling. By utilizing the FE simulations effect of restriction dies and axial tension on undesirable deformation during the process was clarified.

Mathematical Simulation of Bubble Column Slurry Reactor for Direct Dimethyl Ether Synthesis Process from Syngas

Based on a global kinetics of direct dimethyl ether (DME) synthesis process from syngas, a steady-state one-dimensional mathematical model for the bubble column slurry reactor (BCSR) has been established. It was built on the assumption of plug flow of gas phase, sedimentation-dispersion model of catalyst grains and isothermal chamber regardless of reaction heats and rates for the design of an industrial scale bubble column slurry reactor. The simulation results indicate that higher pressure and lower temperature were favorable to the increase of CO conversion, DME selectivity, products yield and the height of slurry bed, which has a coincidence with the characteristic of DME synthesis reaction system, and that the height of slurry bed is lessen with the increasing of operation temperature in the range of 220-260℃. CO conversion, the optimal operation conditions in BCSR were proposed. 

A Cell-centered Diffusion Finite Volume Scheme and it's Application to Magnetic Flux Compression Generators

A cell-centered finite volume scheme for discretizing diffusion operators on distorted quadrilateral meshes has recently been designed and added to APMFCG to enable that code to be used as a tool for studying explosive magnetic flux compression generators. This paper describes this scheme. Comparisons with analytic results for 2-D test cases are presented, as well as 2-D results from a test of a "realistic" generator configuration.

Post ERP Feral System and use of ‘Feral System as Coping Mechanism

A number of studies highlighted problems related to ERP systems, yet, most of these studies focus on the problems during the project and implementation stages but not during the postimplementation use process. Problems encountered in the process of using ERP would hinder the effective exploitation and the extended and continued use of ERP systems and their value to organisations. This paper investigates the different types of problems users (operational, supervisory and managerial) faced in using ERP and how 'feral system' is used as the coping mechanism. The paper adopts a qualitative method and uses data collected from two cases and 26 interviews, to inductively develop a casual network model of ERP usage problem and its coping mechanism. This model classified post ERP usage problems as data quality, system quality, interface and infrastructure. The model is also categorised the different coping mechanism through use of 'feral system' inclusive of feral information system, feral data and feral use of technology.

Impacts of Global Warming on the World Food Market According to SRES Scenarios

This research examines possible effects of climatic change focusing on global warming and its impacts on world agricultural product markets, by using a world food model developed to consider climate changes. GDP and population for each scenario were constructed by IPCC and climate data for each scenario was reported by the Hadley Center and are used in this research to consider results in different contexts. Production and consumption of primary agriculture crops of the world for each socio-economic scenario are obtained and investigated by using the modified world food model. Simulation results show that crop production in some countries or regions will have different trends depending on the context. These alternative contexts depend on the rate of GDP growth, population, temperature, and rainfall. Results suggest that the development of environment friendly technologies lead to more consumption of food in many developing countries. Relationships among environmental policy, clean energy development, and poverty elimination warrant further investigation.

Algebraic Quantum Error Correction Codes

A systematic and exhaustive method based on the group structure of a unitary Lie algebra is proposed to generate an enormous number of quantum codes. With respect to the algebraic structure, the orthogonality condition, which is the central rule of generating quantum codes, is proved to be fully equivalent to the distinguishability of the elements in this structure. In addition, four types of quantum codes are classified according to the relation of the codeword operators and some initial quantum state. By linking the unitary Lie algebra with the additive group, the classical correspondences of some of these quantum codes can be rendered.

Preparation of Polylactic Acid Graft Polyvinyl Acetate Compatibilizers for 50/50 Starch/PLLA Blending

Polylactic acid-g-polyvinyl acetate (PLLA-g-PVAc) was used as a compatibilizer for 50/50 starch/PLLA blend. PLLA-g- PVAc with different mol% of PVAc contents were prepared by grafting PVAc onto PLLA backbone via free radical polymerization in solution process. Various conditions such as type and the amount of initiator, monomer concentration, polymerization time and temperature were studied. Results showed that the highest mol% of PVAc grafting (16 mol%) was achieved by conducting graft copolymerization in toluene at 110°C for 10 h using DCP as an initiator. Chemical structure of the PVAc grafted PLLA was confirmed by 1H NMR. Blending of modified starch and PLLA in the presence compatibilizer with different amounts and mol% PVAc was acquired using internal mixer at 160°C for 15 min. Effects of PVAc content and the amount of compatibilizer on mechanical properties of polymer blend were studied. Results revealed that tensile strength and tensile modulus of polymer blend with higher PVAc grafting content compatibilizer showed better properties than that of lower PVAc grafting content compatibilizer. The amount of compatibilizer was found optimized in the range of 0.5-1.0 Wt% depending on the mol% PVAc.