Proposed a Method for Increasing the Delivery Performance in Dynamic Supply Network

Supply network management adopts a systematic and integrative approach to managing the operations and relationships of various parties in a supply network. The objective of the manufactures in their supply network is to reduce inventory costs and increase customer satisfaction levels. One way of doing that is to synchronize delivery performance. A supply network can be described by nodes representing the companies and the links (relationships) between these nodes. Uncertainty in delivery time depends on type of network relationship between suppliers. The problem is to understand how the individual uncertainties influence the total uncertainty of the network and identify those parts of the network, which has the highest potential for improving the total delivery time uncertainty.

QSI Dynamical Fetch Policy for SMT

A Simultaneous Multithreading (SMT) Processor is capable of executing instructions from multiple threads in the same cycle. SMT in fact was introduced as a powerful architecture to superscalar to increase the throughput of the processor. Simultaneous Multithreading is a technique that permits multiple instructions from multiple independent applications or threads to compete limited resources each cycle. While the fetch unit has been identified as one of the major bottlenecks of SMT architecture, several fetch schemes were proposed by prior works to enhance the fetching efficiency and overall performance. In this paper, we propose a novel fetch policy called queue situation identifier (QSI) which counts some kind of long latency instructions of each thread each cycle then properly selects which threads to fetch next cycle. Simulation results show that in best case our fetch policy can achieve 30% on speedup and also can reduce the data cache level 1 miss rate.

Biomechanical Analysis of the Basic Classical Dance Jump – The Grand Jeté

The aim of this study was to analyse the most important parameters determining the quality of the motion structure of the basic classical dance jump – grand jeté.Research sample consisted of 8 students of the Dance Conservatory in Brno. Using the system Simi motion we performed a 3D kinematic analysis of the jump. On the basis of the comparison of structure quality and measured data of the grand jeté, we defined the optimal values of the relevant parameters determining the quality of the performance. The take-off speed should achieve about 2.4 m·s-1, the optimum take-off angle is 28 - 30º. The take-off leg should swing backward at the beginning of the flight phase with the minimum speed of 3.3 m·s-1.If motor abilities of dancers achieve the level necessary for optimal performance of a classical dance jump, there is room for certain variability of the structure of the dance jump.

Development of Organizational Justice in Incentive Allocation of the Thai Public Sector

An incentive for performance, as one subsystem of a new performance management system, has been implemented in the Thai public sector since 2004. This research investigates the development of organizational justice in the incentive allocation by comparing the roles of distributive and procedural justice on national personnel-s attitudinal outcomes (incentive satisfaction and job performance) between 2 periods, i.e. 2006 and 2008. The data were collected via self-administered questionnaires completed by national government officers and employees. They were stratified using multistage sampling with 2,600 usable samples or 72.0% response rate in 2006, and 1,969 usable samples or 59.3% in 2008. The findings are: (1) There is no difference in means between the two periods relating to distributive justice, procedural justice, incentive satisfaction and job performance. (2) Distributive justice and procedural justice played more important roles in predicting incentive satisfaction and job performance in 2008 than in 2006.

Clinical Decision Support for Disease Classification based on the Tests Association

Until recently, researchers have developed various tools and methodologies for effective clinical decision-making. Among those decisions, chest pain diseases have been one of important diagnostic issues especially in an emergency department. To improve the ability of physicians in diagnosis, many researchers have developed diagnosis intelligence by using machine learning and data mining. However, most of the conventional methodologies have been generally based on a single classifier for disease classification and prediction, which shows moderate performance. This study utilizes an ensemble strategy to combine multiple different classifiers to help physicians diagnose chest pain diseases more accurately than ever. Specifically the ensemble strategy is applied by using the integration of decision trees, neural networks, and support vector machines. The ensemble models are applied to real-world emergency data. This study shows that the performance of the ensemble models is superior to each of single classifiers.

On Simulation based WSN Multi-Parametric Performance Analysis

Optimum communication and performance in Wireless Sensor Networks, constitute multi-facet challenges due to the specific networking characteristics as well as the scarce resource availability. Furthermore, it is becoming increasingly apparent that isolated layer based approaches often do not meet the demands posed by WSNs applications due to omission of critical inter-layer interactions and dependencies. As a counterpart, cross-layer is receiving high interest aiming to exploit these interactions and increase network performance. However, in order to clearly identify existing dependencies, comprehensive performance studies are required evaluating the effect of different critical network parameters on system level performance and behavior.This paper-s main objective is to address the need for multi-parametric performance evaluations considering critical network parameters using a well known network simulator, offering useful and practical conclusions and guidelines. The results reveal strong dependencies among considered parameters which can be utilized by and drive future research efforts, towards designing and implementing highly efficient protocols and architectures.

Effect of Recycle Gas on Activity and Selectivity of Co-Ru/Al2O3 Catalyst in Fischer- Tropsch Synthesis

In industrial scale of Gas to Liquid (GTL) process in Fischer-Tropsch (FT) synthesis, a part of reactor outlet gases such as CO2 and CH4 as side reaction products, is usually recycled. In this study, the influence of CO2 and CH4 on the performance and selectivity of Co-Ru/Al2O3 catalyst is investigated by injection of these gases (0-20 vol. % of feed) to the feed stream. The effect of temperature and feed flow rate, are also inspected. The results show that low amounts of CO2 in the feed stream, doesn`t change the catalyst activity significantly but increasing the amount of CO2 (more than 10 vol. %) cause the CO conversion to decrease and the selectivity of heavy components to increase. Methane acts as an inert gas and doesn`t affect the catalyst performance. Increasing feed flow rate has negative effect on both CO conversion and heavy component selectivity. By raising the temperature, CO conversion will increase but there are more volatile components in the product. The effect of CO2 on the catalyst deactivation is also investigated carefully and a mechanism is suggested to explain the negative influence of CO2 on catalyst deactivation.

Logistics Outsourcing: Performance Models and Financial and Operational Indicators

The growing outsourcing of logistics services resulting from the ongoing current in firms of costs reduction/increased efficiency means that it is becoming more and more important for the companies doing the outsourcing to carry out a proper evaluation. The multiple definitions and measures of logistics service performance found in research on the topic create a certain degree of confusion and do not clear the way towards the proper measurement of their performance. Do a model and a specific set of indicators exist that can be considered appropriate for measuring the performance of logistics services outsourcing in industrial environments? Are said indicators in keeping with the objectives pursued by outsourcing? We aim to answer these and other research questions in the study we have initiated in the field within the framework of the international High Performance Manufacturing (HPM) project of which this paper forms part. As the first stage of this research, this paper reviews articles dealing with the topic published in the last 15 years with the aim of detecting the models most used to make this measurement and determining which performance indicators are proposed as part of said models and which are most used. The first steps are also taken in determining whether these indicators, financial and operational, cover the aims that are being pursued when outsourcing logistics services. The findings show there is a wide variety of both models and indicators used. This would seem to testify to the need to continue with our research in order to try to propose a model and a set of indicators for measuring the performance of logistics services outsourcing in industrial environments.

AHP and Extent Fuzzy AHP Approach for Prioritization of Performance Measurement Attributes

The decision to recruit manpower in an organization requires clear identification of the criteria (attributes) that distinguish successful from unsuccessful performance. The choice of appropriate attributes or criteria in different levels of hierarchy in an organization is a multi-criteria decision problem and therefore multi-criteria decision making (MCDM) techniques can be used for prioritization of such attributes. Analytic Hierarchy Process (AHP) is one such technique that is widely used for deciding among the complex criteria structure in different levels. In real applications, conventional AHP still cannot reflect the human thinking style as precise data concerning human attributes are quite hard to be extracted. Fuzzy logic offers a systematic base in dealing with situations, which are ambiguous or not well defined. This study aims at defining a methodology to improve the quality of prioritization of an employee-s performance measurement attributes under fuzziness. To do so, a methodology based on the Extent Fuzzy Analytic Hierarchy Process is proposed. Within the model, four main attributes such as Subject knowledge and achievements, Research aptitude, Personal qualities and strengths and Management skills with their subattributes are defined. The two approaches conventional AHP approach and the Extent Fuzzy Analytic Hierarchy Process approach have been compared on the same hierarchy structure and criteria set.

Transform-Domain Rate-Distortion Optimization Accelerator for H.264/AVC Video Encoding

In H.264/AVC video encoding, rate-distortion optimization for mode selection plays a significant role to achieve outstanding performance in compression efficiency and video quality. However, this mode selection process also makes the encoding process extremely complex, especially in the computation of the ratedistortion cost function, which includes the computations of the sum of squared difference (SSD) between the original and reconstructed image blocks and context-based entropy coding of the block. In this paper, a transform-domain rate-distortion optimization accelerator based on fast SSD (FSSD) and VLC-based rate estimation algorithm is proposed. This algorithm could significantly simplify the hardware architecture for the rate-distortion cost computation with only ignorable performance degradation. An efficient hardware structure for implementing the proposed transform-domain rate-distortion optimization accelerator is also proposed. Simulation results demonstrated that the proposed algorithm reduces about 47% of total encoding time with negligible degradation of coding performance. The proposed method can be easily applied to many mobile video application areas such as a digital camera and a DMB (Digital Multimedia Broadcasting) phone.

Design and Implementation of Shared Memory based Parallel File System Logging Method for High Performance Computing

I/O workload is a critical and important factor to analyze I/O pattern and file system performance. However tracing I/O operations on the fly distributed parallel file system is non-trivial due to collection overhead and a large volume of data. In this paper, we design and implement a parallel file system logging method for high performance computing using shared memory-based multi-layer scheme. It minimizes the overhead with reduced logging operation response time and provides efficient post-processing scheme through shared memory. Separated logging server can collect sequential logs from multiple clients in a cluster through packet communication. Implementation and evaluation result shows low overhead and high scalability of this architecture for high performance parallel logging analysis.

Analysis of Modified Heap Sort Algorithm on Different Environment

In field of Computer Science and Mathematics, sorting algorithm is an algorithm that puts elements of a list in a certain order i.e. ascending or descending. Sorting is perhaps the most widely studied problem in computer science and is frequently used as a benchmark of a system-s performance. This paper presented the comparative performance study of four sorting algorithms on different platform. For each machine, it is found that the algorithm depends upon the number of elements to be sorted. In addition, as expected, results show that the relative performance of the algorithms differed on the various machines. So, algorithm performance is dependent on data size and there exists impact of hardware also.

Algebraic Specification of Serializability for Partitioned Transactions

The usual correctness condition for a schedule of concurrent database transactions is some form of serializability of the transactions. For general forms, the problem of deciding whether a schedule is serializable is NP-complete. In those cases other approaches to proving correctness, using proof rules that allow the steps of the proof of serializability to be guided manually, are desirable. Such an approach is possible in the case of conflict serializability which is proved algebraically by deriving serial schedules using commutativity of non-conflicting operations. However, conflict serializability can be an unnecessarily strong form of serializability restricting concurrency and thereby reducing performance. In practice, weaker, more general, forms of serializability for extended models of transactions are used. Currently, there are no known methods using proof rules for proving those general forms of serializability. In this paper, we define serializability for an extended model of partitioned transactions, which we show to be as expressive as serializability for general partitioned transactions. An algebraic method for proving general serializability is obtained by giving an initial-algebra specification of serializable schedules of concurrent transactions in the model. This demonstrates that it is possible to conduct algebraic proofs of correctness of concurrent transactions in general cases.

A High Performance Technique in Harmonic Omitting Based on Predictive Current Control of a Shunt Active Power Filter

The perfect operation of common Active Filters is depended on accuracy of identification system distortion. Also, using a suitable method in current injection and reactive power compensation, leads to increased filter performance. Due to this fact, this paper presents a method based on predictive current control theory in shunt active filter applications. The harmonics of the load current is identified by using o–d–q reference frame on load current and eliminating the DC part of d–q components. Then, the rest of these components deliver to predictive current controller as a Threephase reference current by using Park inverse transformation. System is modeled in discreet time domain. The proposed method has been tested using MATLAB model for a nonlinear load (with Total Harmonic Distortion=20%). The simulation results indicate that the proposed filter leads to flowing a sinusoidal current (THD=0.15%) through the source. In addition, the results show that the filter tracks the reference current accurately.

Probabilistic Model Development for Project Performance Forecasting

In this paper, based on the past project cost and time performance, a model for forecasting project cost performance is developed. This study presents a probabilistic project control concept to assure an acceptable forecast of project cost performance. In this concept project activities are classified into sub-groups entitled control accounts. Then obtain the Stochastic S-Curve (SS-Curve), for each sub-group and the project SS-Curve is obtained by summing sub-groups- SS-Curves. In this model, project cost uncertainties are considered through Beta distribution functions of the project activities costs required to complete the project at every selected time sections through project accomplishment, which are extracted from a variety of sources. Based on this model, after a percentage of the project progress, the project performance is measured via Earned Value Management to adjust the primary cost probability distribution functions. Then, accordingly the future project cost performance is predicted by using the Monte-Carlo simulation method.

Oscillation Effect of the Multi-stage Learning for the Layered Neural Networks and Its Analysis

This paper proposes an efficient learning method for the layered neural networks based on the selection of training data and input characteristics of an output layer unit. Comparing to recent neural networks; pulse neural networks, quantum neuro computation, etc, the multilayer network is widely used due to its simple structure. When learning objects are complicated, the problems, such as unsuccessful learning or a significant time required in learning, remain unsolved. Focusing on the input data during the learning stage, we undertook an experiment to identify the data that makes large errors and interferes with the learning process. Our method devides the learning process into several stages. In general, input characteristics to an output layer unit show oscillation during learning process for complicated problems. The multi-stage learning method proposes by the authors for the function approximation problems of classifying learning data in a phased manner, focusing on their learnabilities prior to learning in the multi layered neural network, and demonstrates validity of the multi-stage learning method. Specifically, this paper verifies by computer experiments that both of learning accuracy and learning time are improved of the BP method as a learning rule of the multi-stage learning method. In learning, oscillatory phenomena of a learning curve serve an important role in learning performance. The authors also discuss the occurrence mechanisms of oscillatory phenomena in learning. Furthermore, the authors discuss the reasons that errors of some data remain large value even after learning, observing behaviors during learning.

Solver for a Magnetic Equivalent Circuit and Modeling the Inrush Current of a 3-Phase Transformer

Knowledge about the magnetic quantities in a magnetic circuit is always of great interest. On the one hand, this information is needed for the simulation of a transformer. On the other hand, parameter studies are more reliable, if the magnetic quantities are derived from a well established model. One possibility to model the 3-phase transformer is by using a magnetic equivalent circuit (MEC). Though this is a well known system, it is often not an easy task to set up such a model for a large number of lumped elements which additionally includes the nonlinear characteristic of the magnetic material. Here we show the setup of a solver for a MEC and the results of the calculation in comparison to measurements taken. The equations of the MEC are based on a rearranged system of the nodal analysis. Thus it is possible to achieve a minimum number of equations, and a clear and simple structure. Hence, it is uncomplicated in its handling and it supports the iteration process. Additional helpful tasks are implemented within the solver to enhance the performance. The electric circuit is described by an electric equivalent circuit (EEC). Our results for the 3-phase transformer demonstrate the computational efficiency of the solver, and show the benefit of the application of a MEC.

Takagi-Sugeno Fuzzy Control of Induction Motor

This paper deals with the synthesis of fuzzy state feedback controller of induction motor with optimal performance. First, the Takagi-Sugeno (T-S) fuzzy model is employed to approximate a non linear system in the synchronous d-q frame rotating with electromagnetic field-oriented. Next, a fuzzy controller is designed to stabilise the induction motor and guaranteed a minimum disturbance attenuation level for the closed-loop system. The gains of fuzzy control are obtained by solving a set of Linear Matrix Inequality (LMI). Finally, simulation results are given to demonstrate the controller-s effectiveness.

A Review of Critical Success Factor in Building Maintenance Management Practice for University Sector

Building maintenance plays an important role among other activities in building operation. Building defect and damages are part of the building maintenance 'bread and butter' as their input indicated in the building inspection is very much justified, particularly as to determine the building performance. There will be no escape route or short cut from building maintenance work. This study attempts to identify a competitive performance that translates the Critical Success Factor achievements and satisfactorily meet the university-s expectation. The quality and efficiency of maintenance management operation of building depends, to some extent, on the building condition information, the expectation from the university sector and the works carried out for each maintenance activity. This paper reviews the critical success factor in building maintenance management practice for university sectors from four (4) perspectives which include (1) customer (2) internal processes (3) financial and (4) learning and growth perspective. The enhancement of these perspectives is capable to reach the maintenance management goal for a better living environment in university campus.

Noise Performance Optimization of a Fast Wavelength Calibration Algorithm for OSAs

A new fast correlation algorithm for calibrating the wavelength of Optical Spectrum Analyzers (OSAs) was introduced in [1]. The minima of acetylene gas spectra were measured and correlated with saved theoretical data [2]. So it is possible to find the correct wavelength calibration data using a noisy reference spectrum. First tests showed good algorithmic performance for gas line spectra with high noise. In this article extensive performance tests were made to validate the noise resistance of this algorithm. The filter and correlation parameters of the algorithm were optimized for improved noise performance. With these parameters the performance of this wavelength calibration was simulated to predict the resulting wavelength error in real OSA systems. Long term simulations were made to evaluate the performance of the algorithm over the lifetime of a real OSA.