A Context-Aware based Authorization System for Pervasive Grid Computing

This paper describes the authorization system architecture for Pervasive Grid environment. It discusses the characteristics of classical authorization system and requirements of the authorization system in pervasive grid environment as well. Based on our analysis of current systems and taking into account the main requirements of such pervasive environment, we propose new authorization system architecture as an extension of the existing grid authorization mechanisms. This architecture not only supports user attributes but also context attributes which act as a key concept for context-awareness thought. The architecture allows authorization of users dynamically when there are changes in the pervasive grid environment. For this, we opt for hybrid authorization method that integrates push and pull mechanisms to combine the existing grid authorization attributes with dynamic context assertions. We will investigate the proposed architecture using a real testing environment that includes heterogeneous pervasive grid infrastructures mapped over multiple virtual organizations. Various scenarios are described in the last section of the article to strengthen the proposed mechanism with different facilities for the authorization procedure.

Sensitivity of Small Disturbance Angle Stability to the System Parameters of Future Power Networks

The incorporation of renewable energy sources for the sustainable electricity production is undertaking a more prominent role in electric power systems. Thus, it will be an indispensable incident that the characteristics of future power networks, their prospective stability for instance, get influenced by the imposed features of sustainable energy sources. One of the distinctive attributes of the sustainable energy sources is exhibiting the stochastic behavior. This paper investigates the impacts of this stochastic behavior on the small disturbance rotor angle stability in the upcoming electric power networks. Considering the various types of renewable energy sources and the vast variety of system configurations, the sensitivity analysis can be an efficient breakthrough towards generalizing the effects of new energy sources on the concept of stability. In this paper, the definition of small disturbance angle stability for future power systems and the iterative-stochastic way of its analysis are presented. Also, the effects of system parameters on this type of stability are described by performing a sensitivity analysis for an electric power test system.

Endogenous Fantasy – Based Serious Games: Intrinsic Motivation and Learning

Current technological advances pale in comparison to the changes in social behaviors and 'sense of place' that is being empowered since the Internet made it on the scene. Today-s students view the Internet as both a source of entertainment and an educational tool. The development of virtual environments is a conceptual framework that needs to be addressed by educators and it is important that they become familiar with who these virtual learners are and how they are motivated to learn. Massively multiplayer online role playing games (MMORPGs), if well designed, could become the vehicle of choice to deliver learning content. We suggest that these games, in order to accomplish these goals, must begin with well-established instructional design principles that are co-aligned with established principles of video game design. And have the opportunity to provide an instructional model of significant prescriptive power. The authors believe that game designers need to take advantage of the natural motivation player-learners have for playing games by developing them in such a way so as to promote, intrinsic motivation, content learning, transfer of knowledge, and naturalization.

A Family Cars- Life Cycle Cost (LCC)-Oriented Hybrid Modelling Approach Combining ANN and CBR

Design for cost (DFC) is a method that reduces life cycle cost (LCC) from the angle of designers. Multiple domain features mapping (MDFM) methodology was given in DFC. Using MDFM, we can use design features to estimate the LCC. From the angle of DFC, the design features of family cars were obtained, such as all dimensions, engine power and emission volume. At the conceptual design stage, cars- LCC were estimated using back propagation (BP) artificial neural networks (ANN) method and case-based reasoning (CBR). Hamming space was used to measure the similarity among cases in CBR method. Levenberg-Marquardt (LM) algorithm and genetic algorithm (GA) were used in ANN. The differences of LCC estimation model between CBR and artificial neural networks (ANN) were provided. ANN and CBR separately each method has its shortcomings. By combining ANN and CBR improved results accuracy was obtained. Firstly, using ANN selected some design features that affect LCC. Then using LCC estimation results of ANN could raise the accuracy of LCC estimation in CBR method. Thirdly, using ANN estimate LCC errors and correct errors in CBR-s estimation results if the accuracy is not enough accurate. Finally, economically family cars and sport utility vehicle (SUV) was given as LCC estimation cases using this hybrid approach combining ANN and CBR.

Evaluation of Risk Attributes Driven by Periodically Changing System Functionality

Modeling of the distributed systems allows us to represent the whole its functionality. The working system instance rarely fulfils the whole functionality represented by model; usually some parts of this functionality should be accessible periodically. The reporting system based on the Data Warehouse concept seams to be an intuitive example of the system that some of its functionality is required only from time to time. Analyzing an enterprise risk associated with the periodical change of the system functionality, we should consider not only the inaccessibility of the components (object) but also their functions (methods), and the impact of such a situation on the system functionality from the business point of view. In the paper we suggest that the risk attributes should be estimated from risk attributes specified at the requirements level (Use Case in the UML model) on the base of the information about the structure of the model (presented at other levels of the UML model). We argue that it is desirable to consider the influence of periodical changes in requirements on the enterprise risk estimation. Finally, the proposition of such a solution basing on the UML system model is presented.

Application of 0-1 Fuzzy Programming in Optimum Project Selection

In this article, a mathematical programming model for choosing an optimum portfolio of investments is developed. The investments are considered as investment projects. The uncertainties of the real world are associated through fuzzy concepts for coefficients of the proposed model (i. e. initial investment costs, profits, resource requirement, and total available budget). Model has been coded by using LINGO 11.0 solver. The results of a full analysis of optimistic and pessimistic derivative models are promising for selecting an optimum portfolio of projects in presence of uncertainty.

Concepts for Designing Low Power Wireless Sensor Network

Wireless sensor networks have been used in wide areas of application and become an attractive area for researchers in recent years. Because of the limited energy storage capability of sensor nodes, Energy consumption is one of the most challenging aspects of these networks and different strategies and protocols deals with this area. This paper presents general methods for designing low power wireless sensor network. Different sources of energy consumptions in these networks are discussed here and techniques for alleviating the consumption of energy are presented.

On the Mathematical Structure and Algorithmic Implementation of Biochemical Network Models

Modeling and simulation of biochemical reactions is of great interest in the context of system biology. The central dogma of this re-emerging area states that it is system dynamics and organizing principles of complex biological phenomena that give rise to functioning and function of cells. Cell functions, such as growth, division, differentiation and apoptosis are temporal processes, that can be understood if they are treated as dynamic systems. System biology focuses on an understanding of functional activity from a system-wide perspective and, consequently, it is defined by two hey questions: (i) how do the components within a cell interact, so as to bring about its structure and functioning? (ii) How do cells interact, so as to develop and maintain higher levels of organization and functions? In recent years, wet-lab biologists embraced mathematical modeling and simulation as two essential means toward answering the above questions. The credo of dynamics system theory is that the behavior of a biological system is given by the temporal evolution of its state. Our understanding of the time behavior of a biological system can be measured by the extent to which a simulation mimics the real behavior of that system. Deviations of a simulation indicate either limitations or errors in our knowledge. The aim of this paper is to summarize and review the main conceptual frameworks in which models of biochemical networks can be developed. In particular, we review the stochastic molecular modelling approaches, by reporting the principal conceptualizations suggested by A. A. Markov, P. Langevin, A. Fokker, M. Planck, D. T. Gillespie, N. G. van Kampfen, and recently by D. Wilkinson, O. Wolkenhauer, P. S. Jöberg and by the author.

Classifying Biomedical Text Abstracts based on Hierarchical 'Concept' Structure

Classifying biomedical literature is a difficult and challenging task, especially when a large number of biomedical articles should be organized into a hierarchical structure. In this paper, we present an approach for classifying a collection of biomedical text abstracts downloaded from Medline database with the help of ontology alignment. To accomplish our goal, we construct two types of hierarchies, the OHSUMED disease hierarchy and the Medline abstract disease hierarchies from the OHSUMED dataset and the Medline abstracts, respectively. Then, we enrich the OHSUMED disease hierarchy before adapting it to ontology alignment process for finding probable concepts or categories. Subsequently, we compute the cosine similarity between the vector in probable concepts (in the “enriched" OHSUMED disease hierarchy) and the vector in Medline abstract disease hierarchies. Finally, we assign category to the new Medline abstracts based on the similarity score. The results obtained from the experiments show the performance of our proposed approach for hierarchical classification is slightly better than the performance of the multi-class flat classification.

Scheduling a Flexible Flow Shops Problem using DEA

This paper considers a scheduling problem in flexible flow shops environment with the aim of minimizing two important criteria including makespan and cumulative tardiness of jobs. Since the proposed problem is known as an Np-hard problem in literature, we have to develop a meta-heuristic to solve it. We considered general structure of Genetic Algorithm (GA) and developed a new version of that based on Data Envelopment Analysis (DEA). Two objective functions assumed as two different inputs for each Decision Making Unit (DMU). In this paper we focused on efficiency score of DMUs and efficient frontier concept in DEA technique. After introducing the method we defined two different scenarios with considering two types of mutation operator. Also we provided an experimental design with some computational results to show the performance of algorithm. The results show that the algorithm implements in a reasonable time.

OCR for Script Identification of Hindi (Devnagari) Numerals using Error Diffusion Halftoning Algorithm with Neural Classifier

The applications on numbers are across-the-board that there is much scope for study. The chic of writing numbers is diverse and comes in a variety of form, size and fonts. Identification of Indian languages scripts is challenging problems. In Optical Character Recognition [OCR], machine printed or handwritten characters/numerals are recognized. There are plentiful approaches that deal with problem of detection of numerals/character depending on the sort of feature extracted and different way of extracting them. This paper proposes a recognition scheme for handwritten Hindi (devnagiri) numerals; most admired one in Indian subcontinent our work focused on a technique in feature extraction i.e. Local-based approach, a method using 16-segment display concept, which is extracted from halftoned images & Binary images of isolated numerals. These feature vectors are fed to neural classifier model that has been trained to recognize a Hindi numeral. The archetype of system has been tested on varieties of image of numerals. Experimentation result shows that recognition rate of halftoned images is 98 % compared to binary images (95%).

A Universal Approach for the Intuitive Control of Mobile Robots using an AR/VR-based Interface

Mobile robots are used in a large field of scenarios, like exploring contaminated areas, repairing oil rigs under water, finding survivors in collapsed buildings, etc. Currently, there is no unified intuitive user interface (UI) to control such complex mobile robots. As a consequence, some scenarios are done without the exploitation of experience and intuition of human teleoperators. A novel framework has been developed to embed a flexible and modular UI into a complete 3-D virtual reality simulation system. This new approach wants to access maximum benefits of human operators. Sensor information received from the robot is prepared for an intuitive visualization. Virtual reality metaphors support the operator in his decisions. These metaphors are integrated into a real time stereo video stream. This approach is not restricted to any specific type of mobile robot and allows for the operation of different robot types with a consistent concept and user interface.

Analysis of Combustion, Performance and Emission Characteristics of Turbocharged LHR Extended Expansion DI Diesel Engine

The fundamental aim of extended expansion concept is to achieve higher work done which in turn leads to higher thermal efficiency. This concept is compatible with the application of turbocharger and LHR engine. The Low Heat Rejection engine was developed by coating the piston crown, cylinder head inside with valves and cylinder liner with partially stabilized zirconia coating of 0.5 mm thickness. Extended expansion in diesel engines is termed as Miller cycle in which the expansion ratio is increased by reducing the compression ratio by modifying the inlet cam for late inlet valve closing. The specific fuel consumption reduces to an appreciable level and the thermal efficiency of the extended expansion turbocharged LHR engine is improved. In this work, a thermodynamic model was formulated and developed to simulate the LHR based extended expansion turbocharged direct injection diesel engine. It includes a gas flow model, a heat transfer model, and a two zone combustion model. Gas exchange model is modified by incorporating the Miller cycle, by delaying inlet valve closing timing which had resulted in considerable improvement in thermal efficiency of turbocharged LHR engines. The heat transfer model, calculates the convective and radiative heat transfer between the gas and wall by taking into account of the combustion chamber surface temperature swings. Using the two-zone combustion model, the combustion parameters and the chemical equilibrium compositions were determined. The chemical equilibrium compositions were used to calculate the Nitric oxide formation rate by assuming a modified Zeldovich mechanism. The accuracy of this model is scrutinized against actual test results from the engine. The factors which affect thermal efficiency and exhaust emissions were deduced and their influences were discussed. In the final analysis it is seen that there is an excellent agreement in all of these evaluations.

Design and Simulation of Electromagnetic Flow Meter for Circular Pipe Type

Electromagnetic flow meter by measuring the varying of magnetic flux, which is related to the velocity of conductive flow, can measure the rate of fluids very carefully and precisely. Electromagnetic flow meter operation is based on famous Faraday's second Law. In these equipments, the constant magnetostatic field is produced by electromagnet (winding around the tube) outside of pipe and inducting voltage that is due to conductive liquid flow is measured by electrodes located on two end side of the pipe wall. In this research, we consider to 2-dimensional mathematical model that can be solved by numerical finite difference (FD) solution approach to calculate induction potential between electrodes. The fundamental concept to design the electromagnetic flow meter, exciting winding and simulations are come out by using MATLAB and PDE-Tool software. In the last stage, simulations results will be shown for improvement and accuracy of technical provision.

Designing an Irregular Tensegrity as a Monumental Object

A novel and versatile numerical technique to solve a self-stress equilibrium state is adopted herein as a form-finding procedure for an irregular tensegrity structure. The numerical form-finding scheme of a tensegrity structure uses only the connectivity matrix and prototype tension coefficient vector as the initial guess solution. Any information on the symmetrical geometry or other predefined initial structural conditions is not necessary to get the solution in the form-finding process. An eight-node initial condition example is presented to demonstrate the efficiency and robustness of the proposed method in the form-finding of an irregular tensegrity structure. Based on the conception from the form-finding of an eight-node irregular tensegrity structure, a monumental object is designed by considering the real world situation such as self-weight, wind and earthquake loadings.

Optimal Risk Reduction in the Railway Industry by Using Dynamic Programming

The paper suggests for the first time the use of dynamic programming techniques for optimal risk reduction in the railway industry. It is shown that by using the concept ‘amount of removed risk by a risk reduction option’, the problem related to optimal allocation of a fixed budget to achieve a maximum risk reduction in the railway industry can be reduced to an optimisation problem from dynamic programming. For n risk reduction options and size of the available risk reduction budget B (expressed as integer number), the worst-case running time of the proposed algorithm is O (n x (B+1)), which makes the proposed method a very efficient tool for solving the optimal risk reduction problem in the railway industry. 

Characterizations of Ordered Semigroups by (∈,∈ ∨q)-Fuzzy Ideals

Let S be an ordered semigroup. In this paper we first introduce the concepts of (∈,∈ ∨q)-fuzzy ideals, (∈,∈ ∨q)-fuzzy bi-ideals and (∈,∈ ∨q)-fuzzy generalized bi-ideals of an ordered semigroup S, and investigate their related properties. Furthermore, we also define the upper and lower parts of fuzzy subsets of an ordered semigroup S, and investigate the properties of (∈,∈ ∨q)-fuzzy ideals of S. Finally, characterizations of regular ordered semigroups and intra-regular ordered semigroups by means of the lower part of (∈ ,∈ ∨q)-fuzzy left ideals, (∈,∈ ∨q)-fuzzy right ideals and (∈,∈ ∨q)- fuzzy (generalized) bi-ideals are given.

A Study of Visual Attention in Diagnosing Cerebellar Tumours

Visual attention allows user to select the most relevant information to ongoing behaviour. This paper presents a study on; i) the performance of people measurements, ii) accurateness of people measurement of the peaks that correspond to chemical quantities from the Magnetic Resonance Spectroscopy (MRS) graphs and iii) affects of people measurements to the algorithm-based diagnosis. Participant-s eye-movement was recorded using eye-tracker tool (Eyelink II). This experiment involves three participants for examining 20 MRS graphs to estimate the peaks of chemical quantities which indicate the abnormalities associated with Cerebellar Tumours (CT). The status of each MRS is verified by using decision algorithm. Analysis involves determination of humans-s eye movement pattern in measuring the peak of spectrograms, scan path and determining the relationship of distributions of fixation durations with the accuracy of measurement. In particular, the eye-tracking data revealed which aspects of the spectrogram received more visual attention and in what order they were viewed. This preliminary investigation provides a proof of concept for use of the eye tracking technology as the basis for expanded CT diagnosis.

Accurate Optical Flow Based on Spatiotemporal Gradient Constancy Assumption

Variational methods for optical flow estimation are known for their excellent performance. The method proposed by Brox et al. [5] exemplifies the strength of that framework. It combines several concepts into single energy functional that is then minimized according to clear numerical procedure. In this paper we propose a modification of that algorithm starting from the spatiotemporal gradient constancy assumption. The numerical scheme allows to establish the connection between our model and the CLG(H) method introduced in [18]. Experimental evaluation carried out on synthetic sequences shows the significant superiority of the spatial variant of the proposed method. The comparison between methods for the realworld sequence is also enclosed.