An Approximate Engineering Method for Aerodynamic Heating Solution around Blunt Body Nose

This paper is devoted to predict laminar and turbulent heating rates around blunt re-entry spacecraft at hypersonic conditions. Heating calculation of a hypersonic body is normally performed during the critical part of its flight trajectory. The procedure is of an inverse method, where a shock wave is assumed, and the body shape that supports this shock, as well as the flowfield between the shock and body, are calculated. For simplicity the normal momentum equation is replaced with a second order pressure relation; this simplification significantly reduces computation time. The geometries specified in this research, are parabola and ellipsoids which may have conical after bodies. An excellent agreement is observed between the results obtained in this paper and those calculated by others- research. Since this method is much faster than Navier-Stokes solutions, it can be used in preliminary design, parametric study of hypersonic vehicles.

Development of a Robust Supply Chain for Dynamic Operating Environment

Development of a Robust Supply Chain for Dynamic Operating Environment as we move further into the twenty first century, organisations are under increasing pressure to deliver a high product variation at a reasonable cost without compromise in quality. In a number of cases this will take the form of a customised or high variety low volume manufacturing system that requires prudent management of resources, among a number of functions, to achieve competitive advantage. Purchasing and Supply Chain management is one of such function and due to the substantial interaction with external elements needs to be strategically managed. This requires a number of primary and supporting tools that will enable the appropriate decisions to be made rapidly. This capability is especially vital in a dynamic environment as it provides a pivotal role in increasing the profit margin of the product. The management of this function can be challenging by itself and even more for Small and Medium Enterprises (SMEs) due to the limited resources and expertise available at their disposal. This paper discusses the development of tools and concepts towards effectively managing the purchasing and supply chain function. The developed tools and concepts will provide a cost effective way of managing this function within SMEs. The paper further shows the use of these tools within Contechs, a manufacturer of luxury boat interiors, and the associated benefits achieved as a result of this implementation. Finally a generic framework towards use in such environments is presented.

Renewable Energies in Spain and Portugal: A Strategic Challenge for the Sustainability

Directive 2009/28/CE establishes, as obligatory objective, a share of renewable energies on energetic consumption of 20%, in European Union, in 2020 However, such European normative gives freedom to member states in the selection of the renewable promotion mechanism that allows them to obtain that objective. In this paper, we analyze the main characteristics of the promotion mechanisms of renewable energy used in the countries that shape the Electricity Iberian Market (Spain and Portugal) and the results in employment. The importance of these countries is given by the great increasing of the renewable energies which suppose a share higher than 30% of the overall generation in 2010. Therefore, this research paper can serve as the basis for the learning of other countries with regard to the main advantages that entail the use of a feed-in tariff system.

A Computer Model of Quantum Field Theory

This paper describes a computer model of Quantum Field Theory (QFT), referred to in this paper as QTModel. After specifying the initial configuration for a QFT process (e.g. scattering) the model generates the possible applicable processes in terms of Feynman diagrams, the equations for the scattering matrix, and evaluates probability amplitudes for the scattering matrix and cross sections. The computations of probability amplitudes are performed numerically. The equations generated by QTModel are provided for demonstration purposes only. They are not directly used as the base for the computations of probability amplitudes. The computer model supports two modes for the computation of the probability amplitudes: (1) computation according to standard QFT, and (2) computation according to a proposed functional interpretation of quantum theory.

A Novel Q-algorithm for EPC Global Class-1 Generation-2 Anti-collision Protocol

This paper provides a scheme to improve the read efficiency of anti-collision algorithm in EPCglobal UHF Class-1 Generation-2 RFID standard. In this standard, dynamic frame slotted ALOHA is specified to solve the anti-collision problem. Also, the Q-algorithm with a key parameter C is adopted to dynamically adjust the frame sizes. In the paper, we split the C parameter into two parameters to increase the read speed and derive the optimal values of the two parameters through simulations. The results indicate our method outperforms the original Q-algorithm.

Syntactic Recognition of Distorted Patterns

In syntactic pattern recognition a pattern can be represented by a graph. Given an unknown pattern represented by a graph g, the problem of recognition is to determine if the graph g belongs to a language L(G) generated by a graph grammar G. The so-called IE graphs have been defined in [1] for a description of patterns. The IE graphs are generated by so-called ETPL(k) graph grammars defined in [1]. An efficient, parsing algorithm for ETPL(k) graph grammars for syntactic recognition of patterns represented by IE graphs has been presented in [1]. In practice, structural descriptions may contain pattern distortions, so that the assignment of a graph g, representing an unknown pattern, to a graph language L(G) generated by an ETPL(k) graph grammar G is rejected by the ETPL(k) type parsing. Therefore, there is a need for constructing effective parsing algorithms for recognition of distorted patterns. The purpose of this paper is to present a new approach to syntactic recognition of distorted patterns. To take into account all variations of a distorted pattern under study, a probabilistic description of the pattern is needed. A random IE graph approach is proposed here for such a description ([2]).

Detection and Classification of Power Quality Disturbances Using S-Transform and Wavelet Algorithm

Detection and classification of power quality (PQ) disturbances is an important consideration to electrical utilities and many industrial customers so that diagnosis and mitigation of such disturbance can be implemented quickly. S-transform algorithm and continuous wavelet transforms (CWT) are time-frequency algorithms, and both of them are powerful in detection and classification of PQ disturbances. This paper presents detection and classification of PQ disturbances using S-transform and CWT algorithms. The results of detection and classification, provides that S-transform is more accurate in detection and classification for most PQ disturbance than CWT algorithm, where as CWT algorithm more powerful in detection in some disturbances like notching

Effect of Eccentricity on Conjugate Natural Convection in Vertical Eccentric Annuli

Combined conduction-free convection heat transfer in vertical eccentric annuli is numerically investigated using a finitedifference technique. Numerical results, representing the heat transfer parameters such as annulus walls temperature, heat flux, and heat absorbed in the developing region of the annulus, are presented for a Newtonian fluid of Prandtl number 0.7, fluid-annulus radius ratio 0.5, solid-fluid thermal conductivity ratio 10, inner and outer wall dimensionless thicknesses 0.1 and 0.2, respectively, and dimensionless eccentricities 0.1, 0.3, 0.5, and 0.7. The annulus walls are subjected to thermal boundary conditions, which are obtained by heating one wall isothermally whereas keeping the other wall at inlet fluid temperature. In the present paper, the annulus heights required to achieve thermal full development for prescribed eccentricities are obtained. Furthermore, the variation in the height of thermal full development as function of the geometrical parameter, i.e., eccentricity is also investigated.

Do Firms Need Strategic Alliances?

This study develops a relation to explore the factors influencing management and technology capabilities in strategic alliances. Alliances between firms are recognizing increasingly popular as a vehicle to create and extract greater value from the market. Firm’s alliance can be described as the collaborative problem solving process to solve problems jointly. This study starts from research questions what factors of firm’s management and technology characteristics affect performance of firms which are formed alliances. In this study, we investigated the effect of strategic alliances on company performance. That is, we try to identify whether firms made an alliance with other organizations are differed by characteristics of management and technology. And we test that alliance type and alliance experiences moderate the relationship between firm’s capabilities and its performance. We employ problem-solving perspective and resource-based view perspective to shed light on this research questions. The empirical work is based on the Survey of Business Activities conducted from2006 to 2008 by Statistics Korea. We verify correlations between to point out that these results contribute new empirical evidence on the effect of strategic alliances on company performance.

Parametric Study of a Vapor Compression Refrigeration Cycle Using a Two-Phase Constant Area Ejector

There are several ways of improving the performance of a vapor compression refrigeration cycle. Use of an ejector as expansion device is one of the alternative ways. The present paper aims at evaluate the performance improvement of a vapor compression refrigeration cycle under a wide range of operating conditions. A numerical model is developed and a parametric study of important parameters such as condensation (30-50°C) and evaporation temperatures (-20-5°C), nozzle and diffuser efficiencies (0.75-0.95), subcooling and superheating degrees (0-15K) are investigated. The model verification gives a good agreement with the literature data. The simulation results revealed that condensation temperature has the highest effect (129%) on the performance improvement ratio while superheating has the lowest one (6.2%). Among ejector efficiencies, the diffuser efficiency has a significant effect on the COP of ejector expansion refrigeration cycle. The COP improvement percentage decreases from 10.9% to 4.6% as subcooling degrees increases by 15K.

Comparison of BER Performances for Conventional and Non-Conventional Mapping Schemes Used in OFDM

Orthogonal Frequency Division Multiplexing (OFDM) is one of the techniques for high speed data rate communication with main consideration for 4G and 5G systems. In OFDM, there are several mapping schemes which provide a way of parallel transmission. In this paper, comparisons of mapping schemes used by some standards have been made and also has been discussed about the performance of the non-conventional modulation technique. The Comparisons of Bit Error Rate (BER) performances for conventional and non-conventional modulation schemes have been done using MATLAB software. Mentioned schemes used in OFDM system can be selected on the basis of the requirement of power or spectrum efficiency and BER analysis.

Ant System with Acoustic Communication

Ant colony optimization is an ant algorithm framework that took inspiration from foraging behavior of ant colonies. Indeed, ACO algorithms use a chemical communication, represented by pheromone trails, to build good solutions. However, ants involve different communication channels to interact. Thus, this paper introduces the acoustic communication between ants while they are foraging. This process allows fine and local exploration of search space and permits optimal solution to be improved.

A Parametric Study of an Inverse Electrostatics Problem (IESP) Using Simulated Annealing, Hooke & Jeeves and Sequential Quadratic Programming in Conjunction with Finite Element and Boundary Element Methods

The aim of the current work is to present a comparison among three popular optimization methods in the inverse elastostatics problem (IESP) of flaw detection within a solid. In more details, the performance of a simulated annealing, a Hooke & Jeeves and a sequential quadratic programming algorithm was studied in the test case of one circular flaw in a plate solved by both the boundary element (BEM) and the finite element method (FEM). The proposed optimization methods use a cost function that utilizes the displacements of the static response. The methods were ranked according to the required number of iterations to converge and to their ability to locate the global optimum. Hence, a clear impression regarding the performance of the aforementioned algorithms in flaw identification problems was obtained. Furthermore, the coupling of BEM or FEM with these optimization methods was investigated in order to track differences in their performance.

An Efficient Run Time Interface for Heterogeneous Architecture of Large Scale Supercomputing System

In this paper we propose a novel Run Time Interface (RTI) technique to provide an efficient environment for MPI jobs on the heterogeneous architecture of PARAM Padma. It suggests an innovative, unified framework for the job management interface system in parallel and distributed computing. This approach employs proxy scheme. The implementation shows that the proposed RTI is highly scalable and stable. Moreover RTI provides the storage access for the MPI jobs in various operating system platforms and improve the data access performance through high performance C-DAC Parallel File System (C-PFS). The performance of the RTI is evaluated by using the standard HPC benchmark suites and the simulation results show that the proposed RTI gives good performance on large scale supercomputing system.

Determination of Stress-Strain Characteristics of Railhead Steel using Image Analysis

True stress-strain curve of railhead steel is required to investigate the behaviour of railhead under wheel loading through elasto-plastic Finite Element (FE) analysis. To reduce the rate of wear, the railhead material is hardened through annealing and quenching. The Australian standard rail sections are not fully hardened and hence suffer from non-uniform distribution of the material property; usage of average properties in the FE modelling can potentially induce error in the predicted plastic strains. Coupons obtained at varying depths of the railhead were, therefore, tested under axial tension and the strains were measured using strain gauges as well as an image analysis technique, known as the Particle Image Velocimetry (PIV). The head hardened steel exhibit existence of three distinct zones of yield strength; the yield strength as the ratio of the average yield strength provided in the standard (σyr=780MPa) and the corresponding depth as the ratio of the head hardened zone along the axis of symmetry are as follows: (1.17 σyr, 20%), (1.06 σyr, 20%-80%) and (0.71 σyr, > 80%). The stress-strain curves exhibit limited plastic zone with fracture occurring at strain less than 0.1.

Wavelet Transform and Support Vector Machine Approach for Fault Location in Power Transmission Line

This paper presents a wavelet transform and Support Vector Machine (SVM) based algorithm for estimating fault location on transmission lines. The Discrete wavelet transform (DWT) is used for data pre-processing and this data are used for training and testing SVM. Five types of mother wavelet are used for signal processing to identify a suitable wavelet family that is more appropriate for use in estimating fault location. The results demonstrated the ability of SVM to generalize the situation from the provided patterns and to accurately estimate the location of faults with varying fault resistance.

Big Bang – Big Crunch Learning Method for Fuzzy Cognitive Maps

Modeling of complex dynamic systems, which are very complicated to establish mathematical models, requires new and modern methodologies that will exploit the existing expert knowledge, human experience and historical data. Fuzzy cognitive maps are very suitable, simple, and powerful tools for simulation and analysis of these kinds of dynamic systems. However, human experts are subjective and can handle only relatively simple fuzzy cognitive maps; therefore, there is a need of developing new approaches for an automated generation of fuzzy cognitive maps using historical data. In this study, a new learning algorithm, which is called Big Bang-Big Crunch, is proposed for the first time in literature for an automated generation of fuzzy cognitive maps from data. Two real-world examples; namely a process control system and radiation therapy process, and one synthetic model are used to emphasize the effectiveness and usefulness of the proposed methodology.

Vertex Configurations and Their Relationship on Orthogonal Pseudo-Polyhedra

Vertex configuration for a vertex in an orthogonal pseudo-polyhedron is an identity of a vertex that is determined by the number of edges, dihedral angles, and non-manifold properties meeting at the vertex. There are up to sixteen vertex configurations for any orthogonal pseudo-polyhedron (OPP). Understanding the relationship between these vertex configurations will give us insight into the structure of an OPP and help us design better algorithms for many 3-dimensional geometric problems. In this paper, 16 vertex configurations for OPP are described first. This is followed by a number of formulas giving insight into the relationship between different vertex configurations in an OPP. These formulas will be useful as an extension of orthogonal polyhedra usefulness on pattern analysis in 3D-digital images.

Massive Lesions Classification using Features based on Morphological Lesion Differences

Purpose of this work is the development of an automatic classification system which could be useful for radiologists in the investigation of breast cancer. The software has been designed in the framework of the MAGIC-5 collaboration. In the automatic classification system the suspicious regions with high probability to include a lesion are extracted from the image as regions of interest (ROIs). Each ROI is characterized by some features based on morphological lesion differences. Some classifiers as a Feed Forward Neural Network, a K-Nearest Neighbours and a Support Vector Machine are used to distinguish the pathological records from the healthy ones. The results obtained in terms of sensitivity (percentage of pathological ROIs correctly classified) and specificity (percentage of non-pathological ROIs correctly classified) will be presented through the Receive Operating Characteristic curve (ROC). In particular the best performances are 88% ± 1 of area under ROC curve obtained with the Feed Forward Neural Network.

Effect of Clustering on Energy Efficiency and Network Lifetime in Wireless Sensor Networks

Wireless Sensor Network is Multi hop Self-configuring Wireless Network consisting of sensor nodes. The deployment of wireless sensor networks in many application areas, e.g., aggregation services, requires self-organization of the network nodes into clusters. Efficient way to enhance the lifetime of the system is to partition the network into distinct clusters with a high energy node as cluster head. The different methods of node clustering techniques have appeared in the literature, and roughly fall into two families; those based on the construction of a dominating set and those which are based solely on energy considerations. Energy optimized cluster formation for a set of randomly scattered wireless sensors is presented. Sensors within a cluster are expected to be communicating with cluster head only. The energy constraint and limited computing resources of the sensor nodes present the major challenges in gathering the data. In this paper we propose a framework to study how partially correlated data affect the performance of clustering algorithms. The total energy consumption and network lifetime can be analyzed by combining random geometry techniques and rate distortion theory. We also present the relation between compression distortion and data correlation.