Investigation of Inert Gas Injection in Steam Reforming of Methane: Energy

Synthesis gas manufacturing by steam reforming of hydrocarbons is an important industrial process. High endothermic nature of the process makes it one of the most cost and heat intensive processes. In the present work, composite effect of different inert gases on synthesis gas yield, feed gas conversion and temperature distribution along the reactor length has been studied using a heterogeneous model. Mathematical model was developed as a first stage and validated against the existing process models. With the addition of inert gases, a higher yield of synthesis gas is observed. Simultaneously the rector outlet temperature drops to as low as 810 K. It was found that Xenon gives the highest yield and conversion while Helium gives the lowest temperature. Using Xenon inert gas 20 percent reduction in outlet temperature was observed compared to traditional case.

The Synthetic T2 Quality Control Chart and its Multi-Objective Optimization

In some real applications of Statistical Process Control it is necessary to design a control chart to not detect small process shifts, but keeping a good performance to detect moderate and large shifts in the quality. In this work we develop a new quality control chart, the synthetic T2 control chart, that can be designed to cope with this objective. A multi-objective optimization is carried out employing Genetic Algorithms, finding the Pareto-optimal front of non-dominated solutions for this optimization problem.

Estimation of Production Function in Fishery on the Coasts of Caspian Sea

This research was conducted for the first time at the southeastern coasts of the Caspian Sea in order to evaluate the performance of osteichthyes cooperatives through production (catch) function. Using one of the indirect valuation methods in this research, contributory factors in catch were identified and were inserted into the function as independent variables. In order to carry out this research, the performance of 25 Osteichthyes catching cooperatives in the utilization year of 2009 which were involved in fishing in Miankale wildlife refuge region. The contributory factors in catch were divided into groups of economic, ecological and biological factors. In the mentioned function, catch rate of the cooperative were inserted into as the dependant variable and fourteen partial variables in terms of nine general variables as independent variables. Finally, after function estimation, seven variables were rendered significant at 99 percent reliably level. The results of the function estimation indicated that human resource (fisherman quantity) had the greatest positive effect on catch rate with an influence coefficient of 1.7 while weather conditions had the greatest negative effect on the catch rate of cooperatives with an influence coefficient of -2.07. Moreover, factors like member's share, experience and fisherman training and fishing effort played the main roles in the catch rate of cooperative with influence coefficients of 0.81, 0.5 and 0.21, respectively.

Genetic Algorithms with Oracle for the Traveling Salesman Problem

By introducing the concept of Oracle we propose an approach for improving the performance of genetic algorithms for large-scale asymmetric Traveling Salesman Problems. The results have shown that the proposed approach allows overcoming some traditional problems for creating efficient genetic algorithms.

Ranking Genes from DNA Microarray Data of Cervical Cancer by a local Tree Comparison

The major objective of this paper is to introduce a new method to select genes from DNA microarray data. As criterion to select genes we suggest to measure the local changes in the correlation graph of each gene and to select those genes whose local changes are largest. More precisely, we calculate the correlation networks from DNA microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to tumor progression. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth. This indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.

Numerical Simulation of the Liquid-Vapor Interface Evolution with Material Properties

A satured liquid is warmed until boiling in a parallelepipedic boiler. The heat is supplied in a liquid through the horizontal bottom of the boiler, the other walls being adiabatic. During the process of boiling, the liquid evaporates through its free surface by deforming it. This surface which subdivides the boiler into two regions occupied on both sides by the boiled liquid (broth) and its vapor which surmounts it. The broth occupying the region and its vapor the superior region. A two- fluids model is used to describe the dynamics of the broth, its vapor and their interface. In this model, the broth is treated as a monophasic fluid (homogeneous model) and form with its vapor adiphasic pseudo fluid (two-fluid model). Furthermore, the interface is treated as a zone of mixture characterized by superficial void fraction noted α* . The aim of this article is to describe the dynamics of the interface between the boiled fluid and its vapor within a boiler. The resolution of the problem allowed us to show the evolution of the broth and the level of the liquid.

Influence of Hydrocarbons on Plant Cell Ultrastructure and Main Metabolic Enzymes

Influence of octane and benzene on plant cell ultrastructure and enzymes of basic metabolism, such as nitrogen assimilation and energy generation have been studied. Different plants: perennial ryegrass (Lolium perenne) and alfalfa (Medicago sativa); crops- maize (Zea mays L.) and bean (Phaseolus vulgaris); shrubs – privet (Ligustrum sempervirens) and trifoliate orange (Poncirus trifoliate); trees - poplar (Populus deltoides) and white mulberry (Morus alba L.) were exposed to hydrocarbons of different concentrations (1, 10 and 100 mM). Destructive changes in bean and maize leaves cells ultrastructure under the influence of benzene vapour were revealed at the level of photosynthetic and energy generation subcellular organells. Different deviations at the level of subcellular organelles structure and distribution were observed in alfalfa and ryegrass root cells under the influence of benzene and octane, absorbed through roots. The level of destructive changes is concentration dependent. Benzene at low 1 and 10 mM concentration caused the increase in glutamate dehydrogenase (GDH) activity in maize roots and leaves and in poplar and mulberry shoots, though to higher extent in case of lower, 1mM concentration. The induction was more intensive in plant roots. The highest tested 100mM concentration of benzene was inhibitory to the enzyme in all plants. Octane caused induction of GDH in all grassy plants at all tested concentrations; however the rate of induction decreased parallel to increase of the hydrocarbon concentration. Octane at concentration 1 mM caused induction of GDH in privet, trifoliate and white mulberry shoots. The highest, 100mM octane was characterized by inhibitory effect to GDH activity in all plants. Octane had inductive effect on malate dehydrogenase in almost all plants and tested concentrations, indicating the intensification of Trycarboxylic Acid Cycle. The data could be suggested for elaboration of criteria for plant selection for phytoremediation of oil hydrocarbons contaminated soils.

A Comparison between Heterogeneous and Homogeneous Gas Flow Model in Slurry Bubble Column Reactor for Direct Synthesis of DME

In the present study, a heterogeneous and homogeneous gas flow dispersion model for simulation and optimisation of a large-scale catalytic slurry reactor for the direct synthesis of dimethyl ether (DME) from syngas and CO2, using a churn-turbulent regime was developed. In the heterogeneous gas flow model the gas phase was distributed into two bubble phases: small and large, however in the homogeneous one, the gas phase was distributed into only one large bubble phase. The results indicated that the heterogeneous gas flow model was in more agreement with experimental pilot plant data than the homogeneous one.

Assessment of EU Competitiveness Factors by Multivariate Methods

Measurement of competitiveness between countries or regions is an important topic of many economic analysis and scientific papers. In European Union (EU), there is no mainstream approach of competitiveness evaluation and measuring. There are many opinions and methods of measurement and evaluation of competitiveness between states or regions at national and European level. The methods differ in structure of using the indicators of competitiveness and ways of their processing. The aim of the paper is to analyze main sources of competitive potential of the EU Member States with the help of Factor analysis (FA) and to classify the EU Member States to homogeneous units (clusters) according to the similarity of selected indicators of competitiveness factors by Cluster analysis (CA) in reference years 2000 and 2011. The theoretical part of the paper is devoted to the fundamental bases of competitiveness and the methodology of FA and CA methods. The empirical part of the paper deals with the evaluation of competitiveness factors in the EU Member States and cluster comparison of evaluated countries by cluster analysis. 

Soil Improvement using Cement Dust Mixture

Day by day technology increases and problems associated with this technology also increase. Several researches were carried out to investigate the deployment of such material safely in geotechnical engineering in particular and civil engineering in general. However, different types of waste material have such as cement duct, fly ash and slag been proven to be suitable in several applications. In this research cement dust mixed with different percentages of sand will be used in some civil engineering application as will be explained later in this paper throughout filed and laboratory test. The used mixer (waste material with sand) prove high performance, durability to environmental condition, low cost and high benefits. At higher cement dust ratio, small cement ratio is valuable for compressive strength and permeability. Also at small cement dust ratio higher cement ratio is valuable for compressive strength.

A Web Services based Architecture for NGN Services Delivery

The notion of Next Generation Network (NGN) is based on the Network Convergence concept which refers to integration of services (such as IT and communication services) over IP layer. As the most popular implementation of Service Oriented Architecture (SOA), Web Services technology is known to be the base for service integration. In this paper, we present a platform to deliver communication services as web services. We also implement a sample service to show the simplicity of making composite web and communication services using this platform. A Service Logic Execution Environment (SLEE) is used to implement the communication services. The proposed architecture is in agreement with Service Oriented Architecture (SOA) and also can be integrated to an Enterprise Service Bus to make a base for NGN Service Delivery Platform (SDP).

Performance Analysis of HSDPA Systems using Low-Density Parity-Check (LDPC)Coding as Compared to Turbo Coding

HSDPA is a new feature which is introduced in Release-5 specifications of the 3GPP WCDMA/UTRA standard to realize higher speed data rate together with lower round-trip times. Moreover, the HSDPA concept offers outstanding improvement of packet throughput and also significantly reduces the packet call transfer delay as compared to Release -99 DSCH. Till now the HSDPA system uses turbo coding which is the best coding technique to achieve the Shannon limit. However, the main drawbacks of turbo coding are high decoding complexity and high latency which makes it unsuitable for some applications like satellite communications, since the transmission distance itself introduces latency due to limited speed of light. Hence in this paper it is proposed to use LDPC coding in place of Turbo coding for HSDPA system which decreases the latency and decoding complexity. But LDPC coding increases the Encoding complexity. Though the complexity of transmitter increases at NodeB, the End user is at an advantage in terms of receiver complexity and Bit- error rate. In this paper LDPC Encoder is implemented using “sparse parity check matrix" H to generate a codeword at Encoder and “Belief Propagation algorithm "for LDPC decoding .Simulation results shows that in LDPC coding the BER suddenly drops as the number of iterations increase with a small increase in Eb/No. Which is not possible in Turbo coding. Also same BER was achieved using less number of iterations and hence the latency and receiver complexity has decreased for LDPC coding. HSDPA increases the downlink data rate within a cell to a theoretical maximum of 14Mbps, with 2Mbps on the uplink. The changes that HSDPA enables includes better quality, more reliable and more robust data services. In other words, while realistic data rates are only a few Mbps, the actual quality and number of users achieved will improve significantly.

A Comparison among Wolf Pack Search and Four other Optimization Algorithms

The main objective of this paper is applying a comparison between the Wolf Pack Search (WPS) as a newly introduced intelligent algorithm with several other known algorithms including Particle Swarm Optimization (PSO), Shuffled Frog Leaping (SFL), Binary and Continues Genetic algorithms. All algorithms are applied on two benchmark cost functions. The aim is to identify the best algorithm in terms of more speed and accuracy in finding the solution, where speed is measured in terms of function evaluations. The simulation results show that the SFL algorithm with less function evaluations becomes first if the simulation time is important, while if accuracy is the significant issue, WPS and PSO would have a better performance.

Academic Program Administration via Semantic Web – A Case Study

Generally, administrative systems in an academic environment are disjoint and support independent queries. The objective in this work is to semantically connect these independent systems to provide support to queries run on the integrated platform. The proposed framework, by enriching educational material in the legacy systems, provides a value-added semantics layer where activities such as annotation, query and reasoning can be carried out to support management requirements. We discuss the development of this ontology framework with a case study of UAE University program administration to show how semantic web technologies can be used by administration to develop student profiles for better academic program management.

Application of a Systemic Soft Domain-Driven Design Framework

This paper proposes a “soft systems" approach to domain-driven design of computer-based information systems. We propose a systemic framework combining techniques from Soft Systems Methodology (SSM), the Unified Modelling Language (UML), and an implementation pattern known as “Naked Objects". We have used this framework in action research projects that have involved the investigation and modelling of business processes using object-oriented domain models and the implementation of software systems based on those domain models. Within the proposed framework, Soft Systems Methodology (SSM) is used as a guiding methodology to explore the problem situation and to generate a ubiquitous language (soft language) which can be used as the basis for developing an object-oriented domain model. The domain model is further developed using techniques based on the UML and is implemented in software following the “Naked Objects" implementation pattern. We argue that there are advantages from combining and using techniques from different methodologies in this way. The proposed systemic framework is overviewed and justified as multimethodologyusing Mingers multimethodology ideas. This multimethodology approach is being evaluated through a series of action research projects based on real-world case studies. A Peer-Tutoring case study is presented here as a sample of the framework evaluation process

The Banzhaf-Owen Value for Fuzzy Games with a Coalition Structure

In this paper, a generalized form of the Banzhaf-Owen value for cooperative fuzzy games with a coalition structure is proposed. Its axiomatic system is given by extending crisp case. In order to better understand the Banzhaf-Owen value for fuzzy games with a coalition structure, we briefly introduce the Banzhaf-Owen values for two special kinds of fuzzy games with a coalition structure, and give their explicit forms.

An Unified Approach to Thermodynamics of Power Yield in Thermal, Chemical and Electrochemical Systems

This paper unifies power optimization approaches in various energy converters, such as: thermal, solar, chemical, and electrochemical engines, in particular fuel cells. Thermodynamics leads to converter-s efficiency and limiting power. Efficiency equations serve to solve problems of upgrading and downgrading of resources. While optimization of steady systems applies the differential calculus and Lagrange multipliers, dynamic optimization involves variational calculus and dynamic programming. In reacting systems chemical affinity constitutes a prevailing component of an overall efficiency, thus the power is analyzed in terms of an active part of chemical affinity. The main novelty of the present paper in the energy yield context consists in showing that the generalized heat flux Q (involving the traditional heat flux q plus the product of temperature and the sum products of partial entropies and fluxes of species) plays in complex cases (solar, chemical and electrochemical) the same role as the traditional heat q in pure heat engines. The presented methodology is also applied to power limits in fuel cells as to systems which are electrochemical flow engines propelled by chemical reactions. The performance of fuel cells is determined by magnitudes and directions of participating streams and mechanism of electric current generation. Voltage lowering below the reversible voltage is a proper measure of cells imperfection. The voltage losses, called polarization, include the contributions of three main sources: activation, ohmic and concentration. Examples show power maxima in fuel cells and prove the relevance of the extension of the thermal machine theory to chemical and electrochemical systems. The main novelty of the present paper in the FC context consists in introducing an effective or reduced Gibbs free energy change between products p and reactants s which take into account the decrease of voltage and power caused by the incomplete conversion of the overall reaction.

Educational Quiz Board Games for Adaptive E-Learning

Internet computer games turn to be more and more attractive within the context of technology enhanced learning. Educational games as quizzes and quests have gained significant success in appealing and motivating learners to study in a different way and provoke steadily increasing interest in new methods of application. Board games are specific group of games where figures are manipulated in competitive play mode with race conditions on a surface according predefined rules. The article represents a new, formalized model of traditional quizzes, puzzles and quests shown as multimedia board games which facilitates the construction process of such games. Authors provide different examples of quizzes and their models in order to demonstrate the model is quite general and does support not only quizzes, mazes and quests but also any set of teaching activities. The execution process of such models is explained and, as well, how they can be useful for creation and delivery of adaptive e-learning courseware.

An ACO Based Algorithm for Distribution Networks Including Dispersed Generations

With Power system movement toward restructuring along with factors such as life environment pollution, problems of transmission expansion and with advancement in construction technology of small generation units, it is expected that small units like wind turbines, fuel cells, photovoltaic, ... that most of the time connect to the distribution networks play a very essential role in electric power industry. With increase in developing usage of small generation units, management of distribution networks should be reviewed. The target of this paper is to present a new method for optimal management of active and reactive power in distribution networks with regard to costs pertaining to various types of dispersed generations, capacitors and cost of electric energy achieved from network. In other words, in this method it-s endeavored to select optimal sources of active and reactive power generation and controlling equipments such as dispersed generations, capacitors, under load tapchanger transformers and substations in a way that firstly costs in relation to them are minimized and secondly technical and physical constraints are regarded. Because the optimal management of distribution networks is an optimization problem with continuous and discrete variables, the new evolutionary method based on Ant Colony Algorithm has been applied. The simulation results of the method tested on two cases containing 23 and 34 buses exist and will be shown at later sections.

Assessment of Channel Unavailability Effect on the Wireless Networks Teletraffic Modeling and Analysis

Whereas cellular wireless communication systems are subject to short-and long-term fading. The effect of wireless channel has largely been ignored in most of the teletraffic assessment researches. In this paper, a mathematical teletraffic model is proposed to estimate blocking and forced termination probabilities of cellular wireless networks as a result of teletraffic behavior as well as the outage of the propagation channel. To evaluate the proposed teletraffic model, gamma inter-arrival and general service time distributions have been considered based on wireless channel fading effect. The performance is evaluated and compared with the classical model. The proposed model is dedicated and investigated in different operational conditions. These conditions will consider not only the arrival rate process, but also, the different faded channels models.