Toward Community-Based Personal Cloud Computing

This paper proposes a new of cloud computing for individual computer users to share applications in distributed communities, called community-based personal cloud computing (CPCC). The paper also presents a prototype design and implementation of CPCC. The users of CPCC are able to share their computing applications with other users of the community. Any member of the community is able to execute remote applications shared by other members. The remote applications behave in the same way as their local counterparts, allowing the user to enter input, receive output as well as providing the access to the local data of the user. CPCC provides a peer-to-peer (P2P) environment where each peer provides applications which can be used by the other peers that are connected CPCC.

Systems with Queueing and their Simulation

In the queueing theory, it is assumed that customer arrivals correspond to a Poisson process and service time has the exponential distribution. Using these assumptions, the behaviour of the queueing system can be described by means of Markov chains and it is possible to derive the characteristics of the system. In the paper, these theoretical approaches are presented on several types of systems and it is also shown how to compute the characteristics in a situation when these assumptions are not satisfied

A Novel Method for the Characterization of Synchronization and Coupling in Multichannel EEG and ECoG

In this paper we introduce a novel method for the characterization of synchronziation and coupling effects in multivariate time series that can be used for the analysis of EEG or ECoG signals recorded during epileptic seizures. The method allows to visualize the spatio-temporal evolution of synchronization and coupling effects that are characteristic for epileptic seizures. Similar to other methods proposed for this purpose our method is based on a regression analysis. However, a more general definition of the regression together with an effective channel selection procedure allows to use the method even for time series that are highly correlated, which is commonly the case in EEG/ECoG recordings with large numbers of electrodes. The method was experimentally tested on ECoG recordings of epileptic seizures from patients with temporal lobe epilepsies. A comparision with the results from a independent visual inspection by clinical experts showed an excellent agreement with the patterns obtained with the proposed method.

Project Management Maturity Models and Organizational Project Management Maturity Model (OPM3®): A Critical Morphological Evaluation

There exists a strong correlation between efficient project management and competitive advantage for organizations. Therefore, organizations are striving to standardize and assess the rigor of their project management processes and capabilities i.e. project management maturity. Researchers and standardization organizations have developed several project management maturity models (PMMMs) to assess project management maturity of the organizations. This study presents a critical evaluation of some of the leading PMMMs against OPM3® in a multitude of ways to look at which PMMM is the most comprehensive model - which could assess most aspects of organizations and also help the organizations in gaining competitive advantage over competitors. After a detailed morphological analysis of the models, it is concluded that OPM3® is the most promising maturity model that can really provide a competitive advantage to the organizations due to its unique approach of assessment and improvement strategies.

Evaluation of Algorithms for Sequential Decision in Biosonar Target Classification

A sequential decision problem, based on the task ofidentifying the species of trees given acoustic echo data collectedfrom them, is considered with well-known stochastic classifiers,including single and mixture Gaussian models. Echoes are processedwith a preprocessing stage based on a model of mammalian cochlearfiltering, using a new discrete low-pass filter characteristic. Stoppingtime performance of the sequential decision process is evaluated andcompared. It is observed that the new low pass filter processingresults in faster sequential decisions.

Lagrangian Method for Solving Unsteady Gas Equation

In this paper we propose, a Lagrangian method to solve unsteady gas equation which is a nonlinear ordinary differential equation on semi-infnite interval. This approach is based on Modified generalized Laguerre functions. This method reduces the solution of this problem to the solution of a system of algebraic equations. We also compare this work with some other numerical results. The findings show that the present solution is highly accurate.

Development of a Fiber based Interferometric Sensor for Non-contact Displacement Measurement

In this paper, a fiber based Fabry-Perot interferometer is proposed and demonstrated for a non-contact displacement measurement. A piece of micro-prism which attached to the mechanical vibrator is served as the target reflector. Interference signal is generated from the superposition between the sensing beam and the reference beam within the sensing arm of the fiber sensor. This signal is then converted to the displacement value by using a developed program written in visual Cµ programming with a resolution of λ/8. A classical function generator is operated for controlling the vibrator. By fixing an excitation frequency of 100 Hz and varying the excitation amplitude range of 0.1 – 3 Volts, the output displacements measured by the fiber sensor are obtained from 1.55 μm to 30.225 μm. A reference displacement sensor with a sensitivity of ~0.4 μm is also employed for comparing the displacement errors between both sensors. We found that over the entire displacement range, a maximum and average measurement error are obtained of 0.977% and 0.44% respectively.

Energy Loss at Drops using Neuro Solutions

Energy dissipation in drops has been investigated by physical models. After determination of effective parameters on the phenomenon, three drops with different heights have been constructed from Plexiglas. They have been installed in two existing flumes in the hydraulic laboratory. Several runs of physical models have been undertaken to measured required parameters for determination of the energy dissipation. Results showed that the energy dissipation in drops depend on the drop height and discharge. Predicted relative energy dissipations varied from 10.0% to 94.3%. This work has also indicated that the energy loss at drop is mainly due to the mixing of the jet with the pool behind the jet that causes air bubble entrainment in the flow. Statistical model has been developed to predict the energy dissipation in vertical drops denotes nonlinear correlation between effective parameters. Further an artificial neural networks (ANNs) approach was used in this paper to develop an explicit procedure for calculating energy loss at drops using NeuroSolutions. Trained network was able to predict the response with R2 and RMSE 0.977 and 0.0085 respectively. The performance of ANN was found effective when compared to regression equations in predicting the energy loss.

Unsteady Laminar Boundary Layer Forced Flow in the Region of the Stagnation Point on a Stretching Flat Sheet

This paper analyses the unsteady, two-dimensional stagnation point flow of an incompressible viscous fluid over a flat sheet when the flow is started impulsively from rest and at the same time, the sheet is suddenly stretched in its own plane with a velocity proportional to the distance from the stagnation point. The partial differential equations governing the laminar boundary layer forced convection flow are non-dimensionalised using semi-similar transformations and then solved numerically using an implicit finitedifference scheme known as the Keller-box method. Results pertaining to the flow and heat transfer characteristics are computed for all dimensionless time, uniformly valid in the whole spatial region without any numerical difficulties. Analytical solutions are also obtained for both small and large times, respectively representing the initial unsteady and final steady state flow and heat transfer. Numerical results indicate that the velocity ratio parameter is found to have a significant effect on skin friction and heat transfer rate at the surface. Furthermore, it is exposed that there is a smooth transition from the initial unsteady state flow (small time solution) to the final steady state (large time solution).

Modeling Spatial Distributions of Point and Nonpoint Source Pollution Loadings in the Great Lakes Watersheds

A physically based, spatially-distributed water quality model is being developed to simulate spatial and temporal distributions of material transport in the Great Lakes Watersheds of the U.S. Multiple databases of meteorology, land use, topography, hydrography, soils, agricultural statistics, and water quality were used to estimate nonpoint source loading potential in the study watersheds. Animal manure production was computed from tabulations of animals by zip code area for the census years of 1987, 1992, 1997, and 2002. Relative chemical loadings for agricultural land use were calculated from fertilizer and pesticide estimates by crop for the same periods. Comparison of these estimates to the monitored total phosphorous load indicates that both point and nonpoint sources are major contributors to the total nutrient loads in the study watersheds, with nonpoint sources being the largest contributor, particularly in the rural watersheds. These estimates are used as the input to the distributed water quality model for simulating pollutant transport through surface and subsurface processes to Great Lakes waters. Visualization and GIS interfaces are developed to visualize the spatial and temporal distribution of the pollutant transport in support of water management programs.

Dempster-Shafer Evidence Theory for Image Segmentation: Application in Cells Images

In this paper we propose a new knowledge model using the Dempster-Shafer-s evidence theory for image segmentation and fusion. The proposed method is composed essentially of two steps. First, mass distributions in Dempster-Shafer theory are obtained from the membership degrees of each pixel covering the three image components (R, G and B). Each membership-s degree is determined by applying Fuzzy C-Means (FCM) clustering to the gray levels of the three images. Second, the fusion process consists in defining three discernment frames which are associated with the three images to be fused, and then combining them to form a new frame of discernment. The strategy used to define mass distributions in the combined framework is discussed in detail. The proposed fusion method is illustrated in the context of image segmentation. Experimental investigations and comparative studies with the other previous methods are carried out showing thus the robustness and superiority of the proposed method in terms of image segmentation.

Lessons to Management from the Control Loop Phenomenon

In a none-super-competitive environment the concepts of closed system, management control remains to be the dominant guiding concept to management. The merits of closed loop have been the sources of most of the management literature and culture for many decades. It is a useful exercise to investigate and poke into the dynamics of the control loop phenomenon and draws some lessons to use for refining the practice of management. This paper examines the multitude of lessons abstracted from the behavior of the Input /output /feedback control loop model, which is the core of control theory. There are numerous lessons that can be learned from the insights this model would provide and how it parallels the management dynamics of the organization. It is assumed that an organization is basically a living system that interacts with the internal and external variables. A viable control loop is the one that reacts to the variation in the environment and provide or exert a corrective action. In managing organizations this is reflected in organizational structure and management control practices. This paper will report findings that were a result of examining several abstract scenarios that are exhibited in the design, operation, and dynamics of the control loop and how they are projected on the functioning of the organization. Valuable lessons are drawn in trying to find parallels and new paradigms, and how the control theory science is reflected in the design of the organizational structure and management practices. The paper is structured in a logical and perceptive format. Further research is needed to extend these findings.

Using the Monte Carlo Simulation to Predict the Assembly Yield

Electronics Products that achieve high levels of integrated communications, computing and entertainment, multimedia features in small, stylish and robust new form factors are winning in the market place. Due to the high costs that an industry may undergo and how a high yield is directly proportional to high profits, IC (Integrated Circuit) manufacturers struggle to maximize yield, but today-s customers demand miniaturization, low costs, high performance and excellent reliability making the yield maximization a never ending research of an enhanced assembly process. With factors such as minimum tolerances, tighter parameter variations a systematic approach is needed in order to predict the assembly process. In order to evaluate the quality of upcoming circuits, yield models are used which not only predict manufacturing costs but also provide vital information in order to ease the process of correction when the yields fall below expectations. For an IC manufacturer to obtain higher assembly yields all factors such as boards, placement, components, the material from which the components are made of and processes must be taken into consideration. Effective placement yield depends heavily on machine accuracy and the vision of the system which needs the ability to recognize the features on the board and component to place the device accurately on the pads and bumps of the PCB. There are currently two methods for accurate positioning, using the edge of the package and using solder ball locations also called footprints. The only assumption that a yield model makes is that all boards and devices are completely functional. This paper will focus on the Monte Carlo method which consists in a class of computational algorithms (information processed algorithms) which depends on repeated random samplings in order to compute the results. This method utilized in order to recreate the simulation of placement and assembly processes within a production line.

Towards an Automatic Translation of Colored Petri Nets to Maude Language

Colored Petri Nets (CPN) are very known kind of high level Petri nets. With sound and complete semantics, rewriting logic is one of very powerful logics in description and verification of non-deterministic concurrent systems. Recently, CPN semantics are defined in terms of rewriting logic, allowing us to built models by formal reasoning. In this paper, we propose an automatic translation of CPN to the rewriting logic language Maude. This tool allows graphical editing and simulating CPN. The tool allows the user drawing a CPN graphically and automatic translating the graphical representation of the drawn CPN to Maude specification. Then, Maude language is used to perform the simulation of the resulted Maude specification. It is the first rewriting logic based environment for this category of Petri Nets.

Real-Time Implementation of STANAG 4539 High-Speed HF Modem

High-frequency (HF) communications have been used by military organizations for more than 90 years. The opportunity of very long range communications without the need for advanced equipment makes HF a convenient and inexpensive alternative of satellite communications. Besides the advantages, voice and data transmission over HF is a challenging task, because the HF channel generally suffers from Doppler shift and spread, multi-path, cochannel interference, and many other sources of noise. In constructing an HF data modem, all these effects must be taken into account. STANAG 4539 is a NATO standard for high-speed data transmission over HF. It allows data rates up to 12800 bps over an HF channel of 3 kHz. In this work, an efficient implementation of STANAG 4539 on a single Texas Instruments- TMS320C6747 DSP chip is described. The state-of-the-art algorithms used in the receiver and the efficiency of the implementation enables real-time high-speed data / digitized voice transmission over poor HF channels.

Some Design Issues in Designing of 50KW 50Krpm Permanent Magnet Synchronous Machine

A numbers of important developments have led to an increasing attractiveness for very high speed electrical machines (either motor or generator). Specifically the increasing switching speed of power electronics, high energy magnets, high strength retaining materials, better high speed bearings and improvements in design analysis are the primary drivers in a move to higher speed. The design challenges come in the mechanical design both in terms of strength and resonant modes and in the electromagnetic design particularly in respect of iron losses and ac losses in the various conducting parts including the rotor. This paper describes detailed design work which has been done on a 50,000 rpm, 50kW permanent magnet( PM) synchronous machine. It describes work on electromagnetic and rotor eddy current losses using a variety of methods including both 2D finite element analysis

Improved Technique of Non-viral Gene Delivery into Cancer Cells

Liposomal magnetofection is a simple, highly efficient technology for cell transfection, demonstrating better outcome than a number of other common gene delivery methods. However, aggregate complexes distribution over the cell surface is non-uniform due to the gradient of the permanent magnetic field. The aim of this study was to estimate the efficiency of liposomal magnetofection for prostate carcinoma PC3 cell line using newly designed device, “DynaFECTOR", ensuring magnetofection in a dynamic gradient magnetic field. Liposomal magnetofection in a dynamic gradient magnetic field demonstrated the highest transfection efficiency for PC3 cells – it increased for 21% in comparison with liposomal magnetofection and for 42% in comparison with lipofection alone. The optimal incubation time under dynamic magnetic field for PC3 cell line was 5 minutes and the optimal rotation frequency of magnets – 5 rpm. The new approach also revealed lower cytotoxic effect to cells than liposomal magnetofection.

A Formative Assessment Model within the Competency-Based-Approach for an Individualized E-learning Path

E-learning is not restricted to the use of new technologies for the online content, but also induces the adoption of new approaches to improve the quality of education. This quality depends on the ability of these approaches (technical and pedagogical) to provide an adaptive learning environment. Thus, the environment should include features that convey intentions and meeting the educational needs of learners by providing a customized learning path to acquiring a competency concerned In our proposal, we believe that an individualized learning path requires knowledge of the learner. Therefore, it must pass through a personalization of diagnosis to identify precisely the competency gaps to fill, and reduce the cognitive load To personalize the diagnosis and pertinently measure the competency gap, we suggest implementing the formative assessment in the e-learning environment and we propose the introduction of a pre-regulation process in the area of formative assessment, involving its individualization and implementation in e-learning.

Generalized Mean-field Theory of Phase Unwrapping via Multiple Interferograms

On the basis of Bayesian inference using the maximizer of the posterior marginal estimate, we carry out phase unwrapping using multiple interferograms via generalized mean-field theory. Numerical calculations for a typical wave-front in remote sensing using the synthetic aperture radar interferometry, phase diagram in hyper-parameter space clarifies that the present method succeeds in phase unwrapping perfectly under the constraint of surface- consistency condition, if the interferograms are not corrupted by any noises. Also, we find that prior is useful for extending a phase in which phase unwrapping under the constraint of the surface-consistency condition. These results are quantitatively confirmed by the Monte Carlo simulation.

Exploring the Narrative Communication: Representing Visual Information from Digital Travel Stories

We present the results of a case study aiming to assess the reflection of the tourism community in the Web and its usability to propose new ways to communicate visually. The wealth of information contained in the Web and the clear facilities to communicate personals points of view makes of the social web a new space of exploration. In this way, social web allow the sharing of information between communities with similar interests. However, the tourism community remains unexplored as is the case of the information covered in travel stories. Along the Web, we find multiples sites allowing the users to communicate their experiences and personal points of view of a particular place of the world. This cultural heritage is found in multiple documents, usually very little supplemented with photos, so they are difficult to explore due to the lack of visual information. This paper explores the possibility of analyzing travel stories to display them visually on maps and generate new knowledge such as patterns of travel routes. This way, travel narratives published in electronic formats can be very important especially to the tourism community because of the great amount of knowledge that can be extracted. Our approach is based on the use of a Geoparsing Web Service to extract geographic coordinates from travel narratives in order to draw the geo-positions and link the documents into a map image.