Experimental Investigation of a Novel Reaction in Reduction of Sulfates by Natural Gas as a Reducing Agent

In a pilot plant scale of a fluidized bed reactor, a reduction reaction of sodium sulfate by natural gas has been investigated. Natural gas is applied in this study as a reductant. Feed density, feed mass flow rate, natural gas and air flow rate (independent parameters)and temperature of bed and CO concentration in inlet and outlet of reactor (dependent parameters) were monitored and recorded at steady state. The residence time was adjusted close to value of traditional reaction [1]. An artificial neural network (ANN) was established to study dependency of yield and carbon gradient on operating parameters. Resultant 97% accuracy of applied ANN is a good prove that natural gas can be used as a reducing agent. Predicted ANN model for relation between other sources carbon gradient (accuracy 74%) indicates there is not a meaningful relation between other sources carbon variation and reduction process which means carbon in granule does not have significant effect on the reaction yield.

The Optimal Equilibrium Capacity of Information Hiding Based on Game Theory

Game theory could be used to analyze the conflicted issues in the field of information hiding. In this paper, 2-phase game can be used to build the embedder-attacker system to analyze the limits of hiding capacity of embedding algorithms: the embedder minimizes the expected damage and the attacker maximizes it. In the system, the embedder first consumes its resource to build embedded units (EU) and insert the secret information into EU. Then the attacker distributes its resource evenly to the attacked EU. The expected equilibrium damage, which is maximum damage in value from the point of view of the attacker and minimum from the embedder against the attacker, is evaluated by the case when the attacker attacks a subset from all the EU. Furthermore, the optimal equilibrium capacity of hiding information is calculated through the optimal number of EU with the embedded secret information. Finally, illustrative examples of the optimal equilibrium capacity are presented.

Tourist Satisfaction and Repeat Visitation; Toward a New Comprehensive Model

Tourism researchers have recently focused on repeat visitation as a part of destination loyalty. Different models have also considered satisfaction as the main determinant of revisit intention, while findings in many studies show it as a continuous issue. This conceptual paper attempts at evaluating recent empirical studies on satisfaction and revisit intention. Based on limitations and gaps in recent studies, the current paper suggests a new model that would be more comprehensive than those in previous studies. The new model offers new relationships between antecedents (destination image, perceived value, specific novelty seeking, and distance to destination) and both of satisfaction and revisit intention. Revisit intention in turn is suggested to be measured in a temporal approach.

Optimization of Parametric Studies Using Strategies of Sampling Techniques

To improve the efficiency of parametric studies or tests planning the method is proposed, that takes into account all input parameters, but only a few simulation runs are performed to assess the relative importance of each input parameter. For K input parameters with N input values the total number of possible combinations of input values equals NK. To limit the number of runs, only some (totally N) of possible combinations are taken into account. The sampling procedure Updated Latin Hypercube Sampling is used to choose the optimal combinations. To measure the relative importance of each input parameter, the Spearman rank correlation coefficient is proposed. The sensitivity and the influence of all parameters are analyzed within one procedure and the key parameters with the largest influence are immediately identified.

A Predictive Rehabilitation Software for Cerebral Palsy Patients

Young patients suffering from Cerebral Palsy are facing difficult choices concerning heavy surgeries. Diagnosis settled by surgeons can be complex and on the other hand decision for patient about getting or not such a surgery involves important reflection effort. Proposed software combining prediction for surgeries and post surgery kinematic values, and from 3D model representing the patient is an innovative tool helpful for both patients and medicine professionals. Beginning with analysis and classification of kinematics values from Data Base extracted from gait analysis in 3 separated clusters, it is possible to determine close similarity between patients. Prediction surgery best adapted to improve a patient gait is then determined by operating a suitable preconditioned neural network. Finally, patient 3D modeling based on kinematic values analysis, is animated thanks to post surgery kinematic vectors characterizing the closest patient selected from patients clustering.

Secure Secret Recovery by using Weighted Personal Entropy

Authentication plays a vital role in many secure systems. Most of these systems require user to log in with his or her secret password or pass phrase before entering it. This is to ensure all the valuables information is kept confidential guaranteeing also its integrity and availability. However, to achieve this goal, users are required to memorize high entropy passwords or pass phrases. Unfortunately, this sometimes causes difficulty for user to remember meaningless strings of data. This paper presents a new scheme which assigns a weight to each personal question given to the user in revealing the encrypted secrets or password. Concentration of this scheme is to offer fault tolerance to users by allowing them to forget the specific password to a subset of questions and still recover the secret and achieve successful authentication. Comparison on level of security for weight-based and weightless secret recovery scheme is also discussed. The paper concludes with the few areas that requires more investigation in this research.

Renewal of The Swedish Million Dwelling Program, the Public Housing Company and the Local Community, Hindrances and Mutual Aid

Public housing is a vital factor in community development. Successful city, housing and eco system regeneration design is essential in providing positive community development. This concerns work places, nice dwellings, providing premises for child care, care of the elderly, providing qualitative premises for different kinds of commercial service, providing a nice built environment and housing areas and not the least activating tenants. The public housing companies give value to society by stimulating people, renovating socially and economically sustainable as well as being partners to local business and authorities. By their activities the housing companies contribute to sustainable local and regional growth and the identity and reputation of cities. A Social, Economic and Ecological Reputation Effect (SEERE) model for actions to promote housing and community reputation is presented. The model emphasizes regenerative actions to restore natural eco systems as part of housing renewal strategies and to strengthen municipality reputation.

Towards a Unified Approach of Social Justice: Merging Tradition and Modernity in Public Policy Making in India

This paper explores the social and political imperatives in the sphere of public policy relating to social justice. In India, the colonial legacy and post-colonial social and political pressures sustained the appropriation of 'caste' category in allocating public resources to the backward class of citizens. For several reasons, 'economic' category could not be placed in allocating resources. This paper examines the reasons behind the deliberative exercises and formulating policies and seeks an alternative framework in realizing social justice in terms of a unified category. This attempt can be viewed as a reconciliation of traditional and modern values for a viable alternative in public policy making.

Target and Kaizen Costing

increased competition and increased costs of designing made it important for the firms to identify the right products and the right methods for manufacturing the products. Firms should focus on customers and identify customer demands directly to design the right products. Several management methods and techniques that are currently available improve one or more functions or processes in an industry and do not take the complete product life cycle into consideration. On the other hand target costing is a method / philosophy that takes financial, manufacturing and customer aspects into consideration during designing phase and helps firms in making product design decisions to increase the profit / value of the company. It uses various techniques to identify customer demands, to decrease costs of manufacturing and finally to achieve strategic goals. Target Costing forms an integral part of total product design / redesign based on strategic plans.

Performance Evaluation of Qos Parameters in Cognitive Radio Using Genetic Algorithm

The efficient use of available licensed spectrum is becoming more and more critical with increasing demand and usage of the radio spectrum. This paper shows how the use of spectrum as well as dynamic spectrum management can be effectively managed and spectrum allocation schemes in the wireless communication systems be implemented and used, in future. This paper would be an attempt towards better utilization of the spectrum. This research will focus on the decision-making process mainly, with an assumption that the radio environment has already been sensed and the QoS requirements for the application have been specified either by the sensed radio environment or by the secondary user itself. We identify and study the characteristic parameters of Cognitive Radio and use Genetic Algorithm for spectrum allocation. Performance evaluation is done using MATLAB toolboxes.

Control-flow Complexity Measurement of Processes and Weyuker's Properties

Process measurement is the task of empirically and objectively assigning numbers to the properties of business processes in such a way as to describe them. Desirable attributes to study and measure include complexity, cost, maintainability, and reliability. In our work we will focus on investigating process complexity. We define process complexity as the degree to which a business process is difficult to analyze, understand or explain. One way to analyze a process- complexity is to use a process control-flow complexity measure. In this paper, an attempt has been made to evaluate the control-flow complexity measure in terms of Weyuker-s properties. Weyuker-s properties must be satisfied by any complexity measure to qualify as a good and comprehensive one.

Auto Tuning PID Controller based on Improved Genetic Algorithm for Reverse Osmosis Plant

An optimal control of Reverse Osmosis (RO) plant is studied in this paper utilizing the auto tuning concept in conjunction with PID controller. A control scheme composing an auto tuning stochastic technique based on an improved Genetic Algorithm (GA) is proposed. For better evaluation of the process in GA, objective function defined newly in sense of root mean square error has been used. Also in order to achieve better performance of GA, more pureness and longer period of random number generation in operation are sought. The main improvement is made by replacing the uniform distribution random number generator in conventional GA technique to newly designed hybrid random generator composed of Cauchy distribution and linear congruential generator, which provides independent and different random numbers at each individual steps in Genetic operation. The performance of newly proposed GA tuned controller is compared with those of conventional ones via simulation.

Dynamic Modeling of Underplateform Damper used in Turbomachinery

The present work deals with the structural analysis of turbine blades and modeling of turbine blades. A common failure mode for turbine machines is high cycle of fatigue of compressor and turbine blades due to high dynamic stresses caused by blade vibration and resonance within the operation range of the machinery. In this work, proper damping system will be analyzed to reduce the vibrating blade. The main focus of the work is the modeling of under platform damper to evaluate the dynamic analysis of turbine-blade vibrations. The system is analyzed using Bond graph technique. Bond graph is one of the most convenient ways to represent a system from the physical aspect in foreground. It has advantage of putting together multi-energy domains of a system in a single representation in a unified manner. The bond graph model of dry friction damper is simulated on SYMBOLS-shakti® software. In this work, the blades are modeled as Timoshenko beam. Blade Vibrations under different working conditions are being analyzed numerically.

A Comparative Study into Observer based Fault Detection and Diagnosis in DC Motors: Part-I

A model based fault detection and diagnosis technique for DC motor is proposed in this paper. Fault detection using Kalman filter and its different variants are compared. Only incipient faults are considered for the study. The Kalman Filter iterations and all the related computations required for fault detection and fault confirmation are presented. A second order linear state space model of DC motor is used for this work. A comparative assessment of the estimates computed from four different observers and their relative performance is evaluated.

Semantic Spatial Objects Data Structure for Spatial Access Method

Modern spatial database management systems require a unique Spatial Access Method (SAM) in order solve complex spatial quires efficiently. In this case the spatial data structure takes a prominent place in the SAM. Inadequate data structure leads forming poor algorithmic choices and forging deficient understandings of algorithm behavior on the spatial database. A key step in developing a better semantic spatial object data structure is to quantify the performance effects of semantic and outlier detections that are not reflected in the previous tree structures (R-Tree and its variants). This paper explores a novel SSRO-Tree on SAM to the Topo-Semantic approach. The paper shows how to identify and handle the semantic spatial objects with outlier objects during page overflow/underflow, using gain/loss metrics. We introduce a new SSRO-Tree algorithm which facilitates the achievement of better performance in practice over algorithms that are superior in the R*-Tree and RO-Tree by considering selection queries.

Weak Measurement Theory for Discrete Scales

With the increasing spread of computers and the internet among culturally, linguistically and geographically diverse communities, issues of internationalization and localization and becoming increasingly important. For some of the issues such as different scales for length and temperature, there is a well-developed measurement theory. For others such as date formats no such theory will be possible. This paper fills a gap by developing a measurement theory for a class of scales previously overlooked, based on discrete and interval-valued scales such as spanner and shoe sizes. The paper gives a theoretical foundation for a class of data representation problems.

Molecular Docking on Recomposed versus Crystallographic Structures of Zn-Dependent Enzymes and their Natural Inhibitors

Matrix metalloproteinases (MMP) are a class of structural and functional related enzymes involved in altering the natural elements of the extracellular matrix. Most of the MMP structures are cristalographycally determined and published in WorldWide ProteinDataBank, isolated, in full structure or bound to natural or synthetic inhibitors. This study proposes an algorithm to replace missing crystallographic structures in PDB database. We have compared the results of a chosen docking algorithm with a known crystallographic structure in order to validate enzyme sites reconstruction there where crystallographic data are missing.

Artificial Intelligence for Software Quality Improvement

This paper presents a software quality support tool, a Java source code evaluator and a code profiler based on computational intelligence techniques. It is Java prototype software developed by AI Group [1] from the Research Laboratories at Universidad de Palermo: an Intelligent Java Analyzer (in Spanish: Analizador Java Inteligente, AJI). It represents a new approach to evaluate and identify inaccurate source code usage and transitively, the software product itself. The aim of this project is to provide the software development industry with a new tool to increase software quality by extending the value of source code metrics through computational intelligence.

Estimation of the Park-Ang Damage Index for Floating Column Building with Infill Wall

Buildings with floating column are highly undesirable built in seismically active areas. Many urban multi-storey buildings today have floating column buildings which are adopted to accommodate parking at ground floor or reception lobbies in the first storey. The earthquake forces developed at different floor levels in a building need to be brought down along the height to the ground by the shortest path; any deviation or discontinuity in this load transfer path results in poor performance of the building. Floating column buildings are severely damaged during earthquake. Damage on this structure can be reduce by taking the effect of infill wall. This paper presents the effect of stiffness of infill wall to the damage occurred in floating column building when ground shakes. Modelling and analysis are carried out by non linear analysis programme IDARC-2D. Damage occurred in beams, columns, storey are studied by formulating modified Park & Ang model to evaluate damage indices. Overall structural damage indices in buildings due to shaking of ground are also obtained. Dynamic response parameters i.e. lateral floor displacement, storey drift, time period, base shear of buildings are obtained and results are compared with the ordinary moment resisting frame buildings. Formation of cracks, yield, plastic hinge, are also observed during analysis.

Development and Evaluation of a Dynamic Cardiac Phantom for use in Nuclear Medicine

The aim of this study was to develop a dynamic cardiac phantom for quality control in myocardial scintigraphy. The dynamic heart phantom constructed only contained the left ventricle, made of elastic material (latex), comprising two cavities: one internal and one external. The data showed a non-significant variation in the values of left ventricular ejection fraction (LVEF) obtained by varying the heart rate. It was also possible to evaluate the ejection fraction (LVEF) through different arrays of image acquisition and to perform an intercomparison of LVEF by two different scintillation cameras. The results of the quality control tests were satisfactory, showing that they can be used as parameters in future assessments. The new dynamic heart phantom was demonstrated to be effective for use in LVEF measurements. Therefore, the new heart simulator is useful for the quality control of scintigraphic cameras.