Using A Hybrid Algorithm to Improve the Quality of Services in Multicast Routing Problem

A hybrid learning automata-genetic algorithm (HLGA) is proposed to solve QoS routing optimization problem of next generation networks. The algorithm complements the advantages of the learning Automato Algorithm(LA) and Genetic Algorithm(GA). It firstly uses the good global search capability of LA to generate initial population needed by GA, then it uses GA to improve the Quality of Service(QoS) and acquiring the optimization tree through new algorithms for crossover and mutation operators which are an NP-Complete problem. In the proposed algorithm, the connectivity matrix of edges is used for genotype representation. Some novel heuristics are also proposed for mutation, crossover, and creation of random individuals. We evaluate the performance and efficiency of the proposed HLGA-based algorithm in comparison with other existing heuristic and GA-based algorithms by the result of simulation. Simulation results demonstrate that this paper proposed algorithm not only has the fast calculating speed and high accuracy but also can improve the efficiency in Next Generation Networks QoS routing. The proposed algorithm has overcome all of the previous algorithms in the literature.

Estimating the Costs of Conservation in Multiple Output Agricultural Setting

Scarcity of resources for biodiversity conservation gives rise to the need of strategic investment with priorities given to the cost of conservation. While the literature provides abundant methodological options for biodiversity conservation; estimating true cost of conservation remains abstract and simplistic, without recognising dynamic nature of the cost. Some recent works demonstrate the prominence of economic theory to inform biodiversity decisions, particularly on the costs and benefits of biodiversity however, the integration of the concept of true cost into biodiversity actions and planning are very slow to come by, and specially on a farm level. Conservation planning studies often use area as a proxy for costs neglecting different land values as well as protected areas. These literature consider only heterogeneous benefits while land costs are considered homogenous. Analysis with the assumption of cost homogeneity results in biased estimation; since not only it doesn’t address the true total cost of biodiversity actions and plans, but also it fails to screen out lands that are more (or less) expensive and/or difficult (or more suitable) for biodiversity conservation purposes, hindering validity and comparability of the results. Economies of scope” is one of the other most neglected aspects in conservation literature. The concept of economies of scope introduces the existence of cost complementarities within a multiple output production system and it suggests a lower cost during the concurrent production of multiple outputs by a given farm. If there are, indeed, economies of scope then simplistic representation of costs will tend to overestimate the true cost of conservation leading to suboptimal outcomes. The aim of this paper, therefore, is to provide first road review of the various theoretical ways in which economies of scope are likely to occur of how they might occur in conservation. Consequently, the paper addresses gaps that have to be filled in future analysis.

Ranking Alternatives in Multi-Criteria Decision Analysis using Common Weights Based on Ideal and Anti-ideal Frontiers

One of the most important issues in multi-criteria decision analysis (MCDA) is to determine the weights of criteria so that all alternatives can be compared based on the collective performance of criteria. In this paper, one of popular methods in data envelopment analysis (DEA) known as common weights (CWs) is used to determine the weights in MCDA. Two frontiers named ideal and anti-ideal frontiers, instead of ideal and anti-ideal alternatives, are defined based on two new proposed CWs models. Ideal and antiideal frontiers are more flexible than that of alternatives. According to the optimal solutions of these two models, the distances of an alternative from the ideal and anti-ideal frontiers are derived. Then, a relative distance is introduced to measure the value of each alternative. The suggested models are linear and despite weight restrictions are feasible. An example is presented for explaining the method and for comparing to the existing literature.

CFD Simulation of Fixed Bed Reactor in Fischer-Tropsch Synthesis of GTL Technology

In this paper 2D Simulation of catalytic Fixed Bed Reactor in Fischer-Tropsch Synthesis of GTL technology has been performed utilizing computational fluid dynamics (CFD). Synthesis gas (a mixture of carbon monoxide and hydrogen) has been used as feedstock. The reactor was modeled and the model equations were solved employing finite volume method. The model was validated against the experimental data reported in literature. The comparison showed a good agreement between simulation results and the experimental data. In addition, the model was applied to predict the concentration contours of the reactants and products along the length of reactor.

Intact and ACL-Deficient Knee MODEL Evaluation

The human knee joint has a three dimensional geometry with multiple body articulations that produce complex mechanical responses under loads that occur in everyday life and sports activities. To produce the necessary joint compliance and stability for optimal daily function various menisci and ligaments are present while muscle forces are used to this effect. Therefore, knowledge of the complex mechanical interactions of these load bearing structures is necessary when treatment of relevant diseases is evaluated and assisting devices are designed. Numerical tools such as finite element analysis are suitable for modeling such joints in order to understand their physics. They have been used in the current study to develop an accurate human knee joint and model its mechanical behavior. To evaluate the efficacy of this articulated model, static load cases were used for comparison purposes with previous experimentally verified modeling works drawn from literature.

A Multi-Phase Methodology for Investigating Localisation Policies within the GCC: The Hotel Industry in the KSA and the UAE

Due to a high unemployment rate among local people and a high reliance on expatriate workers, the governments in the Gulf Co-operation Council (GCC) countries have been implementing programmes of localisation (replacing foreign workers with GCC nationals). These programmes have been successful in the public sector but much less so in the private sector. However, there are now insufficient jobs for locals in the public sector and the onus to provide employment has fallen on the private sector. This paper is concerned with a study, which is a work in progress (certain elements are complete but not the whole study), investigating the effective implementation of localisation policies in four- and five-star hotels in the Kingdom of Saudi Arabia (KSA) and the United Arab Emirates (UAE). The purpose of the paper is to identify the research gap, and to present the need for the research. Further, it will explain how this research was conducted. Studies of localisation in the GCC countries are under-represented in scholarly literature. Currently, the hotel sectors in KSA and UAE play an important part in the countries’ economies. However, the total proportion of Saudis working in the hotel sector in KSA is slightly under 8%, and in the UAE, the hotel sector remains highly reliant on expatriates. There is therefore a need for research on strategies to enhance the implementation of the localisation policies in general and in the hotel sector in particular. Further, despite the importance of the hotel sector to their economies, there remains a dearth of research into the implementation of localisation policies in this sector. Indeed, as far as the researchers are aware, there is no study examining localisation in the hotel sector in KSA, and few in the UAE. This represents a considerable research gap. Regarding how the research was carried out, a multiple case study strategy was used. The four- and five-star hotel sector in KSA is one of the cases, while the four- and five-star hotel sector in the UAE is the other case. Four- and five-star hotels in KSA and the UAE were chosen as these countries have the longest established localisation policies of all the GCC states and there are more hotels of these classifications in these countries than in any of the other Gulf countries. A literature review was carried out to underpin the research. The empirical data were gathered in three phases. In order to gain a pre-understanding of the issues pertaining to the research context, Phase I involved eight unstructured interviews with officials from the Saudi Commission for Tourism and Antiquities (three interviewees); the Saudi Human Resources Development Fund (one); the Abu Dhabi Tourism and Culture Authority (three); and the Abu Dhabi Development Fund (one). In Phase II, a questionnaire was administered to 24 managers and 24 employees in four- and five-star hotels in each country to obtain their beliefs, attitudes, opinions, preferences and practices concerning localisation. Unstructured interviews were carried out in Phase III with six managers in each country in order to allow them to express opinions that may not have been explored in sufficient depth in the questionnaire. The interviews in Phases I and III were analysed using thematic analysis and SPSS will be used to analyse the questionnaire data. It is recommended that future research be undertaken on a larger scale, with a larger sample taken from all over KSA and the UAE rather than from only four cities (i.e., Riyadh and Jeddah in KSA and Abu Dhabi and Sharjah in the UAE), as was the case in this research.

A Study on Flammability of Bio Oil Combustible Vapour Mixtures

Study of fire and explosion is very important mainly in oil and gas industries due to several accidents which have been reported in the past and present. In this work, we have investigated the flammability of bio oil vapour mixtures. This mixture may contribute to fire during the storage and transportation process. Bio oil sample derived from Palm Kernell shell was analysed using Gas Chromatography Mass Spectrometry (GC-MS) to examine the composition of the sample. Mole fractions of 12 selected components in the liquid phase were obtained from the GC-FID data and used to calculate mole fractions of components in the gas phase via modified Raoult-s law. Lower Flammability Limits (LFLs) and Upper Flammability Limits (UFLs) for individual components were obtained from published literature. However, stoichiometric concentration method was used to calculate the flammability limits of some components which their flammability limit values are not available in the literature. The LFL and UFL values for the mixture were calculated using the Le Chatelier equation. The LFLmix and UFLmix values were used to construct a flammability diagram and subsequently used to determine the flammability of the mixture. The findings of this study can be used to propose suitable inherently safer method to prevent the flammable mixture from occurring and to minimizing the loss of properties, business, and life due to fire accidents in bio oil productions.

Numerical Investigation on Latent Heat Storage Unit of Different Configurations

The storage of thermal energy as a latent heat of phase change material (PCM) has created considerable interest among researchers in recent times. Here, an attempt is made to carry out numerical investigations to analyze the performance of latent heat storage units (LHSU) employing phase change material. The mathematical model developed is based on an enthalpy formulation. Freezing time of PCM packed in three different shaped containers viz. rectangular, cylindrical and cylindrical shell is compared. The model is validated with the results available in the literature. Results show that for the same mass of PCM and surface area of heat transfer, cylindrical shell container takes the least time for freezing the PCM and this geometric effect is more pronounced with an increase in the thickness of the shell than that of length of the shell.

Knowledge Flows and Innovative Performances of NTBFs in Gauteng, South Africa: An Attempt to Explain Mixed Findings in Science Park Research

Science parks are often established to drive regional economic growth, especially in countries with emerging economies. However, mixed findings regarding the performances of science park firms are found in the literature. This study tries to explain these mixed findings by taking a relational approach and exploring (un)intended knowledge transfers between new technology-based firms (NTBFs) in the emerging South African economy. Moreover, the innovation outcomes of these NTBFs are examined by using a multi-dimensional construct. Results show that science park location plays a significant role in explaining innovative sales, but is insignificant when a different indicator of innovation outcomes is used. Furthermore, only for innovations that are new to the firms, both science park location and intended knowledge transfer via informal business relationships have a positive impact; whereas social relationships have a negative impact.

Cross Layer Optimization for Fairness Balancing Based on Adaptively Weighted Utility Functions in OFDMA Systems

Cross layer optimization based on utility functions has been recently studied extensively, meanwhile, numerous types of utility functions have been examined in the corresponding literature. However, a major drawback is that most utility functions take a fixed mathematical form or are based on simple combining, which can not fully exploit available information. In this paper, we formulate a framework of cross layer optimization based on Adaptively Weighted Utility Functions (AWUF) for fairness balancing in OFDMA networks. Under this framework, a two-step allocation algorithm is provided as a sub-optimal solution, whose control parameters can be updated in real-time to accommodate instantaneous QoS constrains. The simulation results show that the proposed algorithm achieves high throughput while balancing the fairness among multiple users.

Quantifying the Sustainable Building Criteria Based on Case Studies from Malaysia

In order to encourage the construction of green homes (GH) in Malaysia, a simple and attainable framework for designing and building GHs is needed. This can be achieved by aligning GH principles against Cole-s 'Sustainable Building Criteria' (SBC). This set of considerations was used to categorize the GH features of three case studies from Malaysia. Although the categorization of building features is useful at exploring the presence of sustainability inclinations of each house, the overall impact of building features in each of the five SBCs are unknown. Therefore, this paper explored the possibility of quantifying the impact of building features categorized in SBC1 – “Buildings will have to adapt to the new environment and restore damaged ecology while mitigating resource use" based on existing GH assessment tools and methods and other literature. This process as reported in this paper could lead to a new dimension in green home rating and assessment methods.

MPSO based Model Order Formulation Technique for SISO Continuous Systems

This paper proposes a new version of the Particle Swarm Optimization (PSO) namely, Modified PSO (MPSO) for model order formulation of Single Input Single Output (SISO) linear time invariant continuous systems. In the General PSO, the movement of a particle is governed by three behaviors namely inertia, cognitive and social. The cognitive behavior helps the particle to remember its previous visited best position. In Modified PSO technique split the cognitive behavior into two sections like previous visited best position and also previous visited worst position. This modification helps the particle to search the target very effectively. MPSO approach is proposed to formulate the higher order model. The method based on the minimization of error between the transient responses of original higher order model and the reduced order model pertaining to the unit step input. The results obtained are compared with the earlier techniques utilized, to validate its ease of computation. The proposed method is illustrated through numerical example from literature.

A Novel Reversible Watermarking Method based on Adaptive Thresholding and Companding Technique

Embedding and extraction of a secret information as well as the restoration of the original un-watermarked image is highly desirable in sensitive applications like military, medical, and law enforcement imaging. This paper presents a novel reversible data-hiding method for digital images using integer to integer wavelet transform and companding technique which can embed and recover the secret information as well as can restore the image to its pristine state. The novel method takes advantage of block based watermarking and iterative optimization of threshold for companding which avoids histogram pre and post-processing. Consequently, it reduces the associated overhead usually required in most of the reversible watermarking techniques. As a result, it keeps the distortion small between the marked and the original images. Experimental results show that the proposed method outperforms the existing reversible data hiding schemes reported in the literature.

A Novel Methodology for Synthesis of Fault Trees from MATLAB-Simulink Model

Fault tree analysis is a well-known method for reliability and safety assessment of engineering systems. In the last 3 decades, a number of methods have been introduced, in the literature, for automatic construction of fault trees. The main difference between these methods is the starting model from which the tree is constructed. This paper presents a new methodology for the construction of static and dynamic fault trees from a system Simulink model. The method is introduced and explained in detail, and its correctness and completeness is experimentally validated by using an example, taken from literature. Advantages of the method are also mentioned.

A new Heuristic Algorithm for the Dynamic Facility Layout Problem with Budget Constraint

In this research, we have developed a new efficient heuristic algorithm for the dynamic facility layout problem with budget constraint (DFLPB). This heuristic algorithm combines two mathematical programming methods such as discrete event simulation and linear integer programming (IP) to obtain a near optimum solution. In the proposed algorithm, the non-linear model of the DFLP has been changed to a pure integer programming (PIP) model. Then, the optimal solution of the PIP model has been used in a simulation model that has been designed in a similar manner as the DFLP for determining the probability of assigning a facility to a location. After a sufficient number of runs, the simulation model obtains near optimum solutions. Finally, to verify the performance of the algorithm, several test problems have been solved. The results show that the proposed algorithm is more efficient in terms of speed and accuracy than other heuristic algorithms presented in previous works found in the literature.

The Knapsack Sharing Problem: A Tree Search Exact Algorithm

In this paper, we study the knapsack sharing problem, a variant of the well-known NP-Hard single knapsack problem. We investigate the use of a tree search for optimally solving the problem. The used method combines two complementary phases: a reduction interval search phase and a branch and bound procedure one. First, the reduction phase applies a polynomial reduction strategy; that is used for decomposing the problem into a series of knapsack problems. Second, the tree search procedure is applied in order to attain a set of optimal capacities characterizing the knapsack problems. Finally, the performance of the proposed optimal algorithm is evaluated on a set of instances of the literature and its runtime is compared to the best exact algorithm of the literature.

Effects of Sea Water Level Fluctuations on Seismic Response of Jacket Type Offshore Platforms

To understand the seismic behavior of the offshore structures, the dynamic interaction of the water-structure-soil should be assessed. In this regard the role of the water dynamic properties in magnifying or reducing of the effects of earthquake induced motions on offshore structures haven't been investigated in precise manner in available literature. In this paper the sea water level fluctuations effects on the seismic behavior of a sample of offshore structures has been investigated by emphasizing on the water-structure interaction phenomenon. For this purpose a two dimensional finite element model of offshore structures as well as surrounded water has been developed using ANSYS software. The effect of soil interaction with embedded pile foundation has been imposed by using a series of nonlinear springs in horizontal and vertical directions in soil-piles contact points. In the model, the earthquake induced motions have been applied on springs and consequently the motions propagated upward to the structure and surrounded water. As a result of numerical study, the horizontal deformations of the offshore deck as well as internal force and buckling coefficient in structural elements have been recorded and controlled with and without water presence. In part of study a parametric study has been accomplished on sea water level fluctuations and effect of this parameter has been studied on the aforementioned numerical results.

Using Multi-Objective Particle Swarm Optimization for Bi-objective Multi-Mode Resource-Constrained Project Scheduling Problem

In this paper the multi-mode resource-constrained project scheduling problem with discounted cash flows is considered. Minimizing the makespan and maximization the net present value (NPV) are the two common objectives that have been investigated in the literature. We apply one evolutionary algorithm named multiobjective particle swarm optimization (MOPSO) to find Pareto front solutions. We used standard sets of instances from the project scheduling problem library (PSPLIB). The results are computationally compared respect to different metrics taken from the literature on evolutionary multi-objective optimization.

Low Leakage MUX/XOR Functions Using Symmetric and Asymmetric FinFETs

In this paper, FinFET devices are analyzed with emphasis on sub-threshold leakage current control. This is achieved through proper biasing of the back gate, and through the use of asymmetric work functions for the four terminal FinFET devices. We are also examining different configurations of multiplexers and XOR gates using transistors of symmetric and asymmetric work functions. Based on extensive characterization data for MUX circuits, our proposed configuration using symmetric devices lead to leakage current and delay improvements of 65% and 47% respectively compared to results in the literature. For XOR gates, a 90% improvement in the average leakage current is achieved by using asymmetric devices. All simulations are based on a 25nm FinFET technology using the University of Florida UFDG model.