IPSO Based UPFC Robust Output Feedback Controllers for Damping of Low Frequency Oscillations

On the basis of the linearized Phillips-Herffron model of a single-machine power system, a novel method for designing unified power flow controller (UPFC) based output feedback controller is presented. The design problem of output feedback controller for UPFC is formulated as an optimization problem according to with the time domain-based objective function which is solved by iteration particle swarm optimization (IPSO) that has a strong ability to find the most optimistic results. To ensure the robustness of the proposed damping controller, the design process takes into account a wide range of operating conditions and system configurations. The simulation results prove the effectiveness and robustness of the proposed method in terms of a high performance power system. The simulation study shows that the designed controller by Iteration PSO performs better than Classical PSO in finding the solution.

Processes Simulation Study of Coal to Methanol Based on Gasification Technology

This study presents a simulation model for converting coal to methanol, based on gasification technology with the commercial chemical process simulator, Pro/II® V8.1.1. The methanol plant consists of air separation unit (ASU), gasification unit, gas clean-up unit, and methanol synthetic unit. The clean syngas is produced with the first three operating units, and the model has been verified with the reference data from United States Environment Protection Agency. The liquid phase methanol (LPMEOHTM) process is adopted in the methanol synthetic unit. Clean syngas goes through gas handing section to reach the reaction requirement, reactor loop/catalyst to generate methanol, and methanol distillation to get desired purity over 99.9 wt%. The ratio of the total energy combined with methanol and dimethyl ether to that of feed coal is 78.5% (gross efficiency). The net efficiency is 64.2% with the internal power consumption taken into account, based on the assumption that the efficiency of electricity generation is 40%.

Inference of Stress-Strength Model for a Lomax Distribution

In this paper, the estimation of the stress-strength parameter R = P(Y < X), when X and Y are independent and both are Lomax distributions with the common scale parameters but different shape parameters is studied. The maximum likelihood estimator of R is derived. Assuming that the common scale parameter is known, the bayes estimator and exact confidence interval of R are discussed. Simulation study to investigate performance of the different proposed methods has been carried out.

Effect of Muscle Loss on Hip Muscular Effort during the Swing Phase of Transfemoral Amputee Gait: A Simulation Study

The effect of muscle loss due to transfemoral amputation, on energy expenditure of hip joint and individual residual muscles was simulated. During swing phase of gait, with each muscle as an ideal force generator, the lower extremity was modeled as a two-degree of freedom linkage, for which hip and knee were joints. According to results, muscle loss will not lead to higher energy expenditure of hip joint, as long as other parameters of limb remain unaffected. This finding maybe due to the role of biarticular muscles in hip and knee joints motion. Moreover, if hip flexors are removed from the residual limb, residual flexors, and if hip extensors are removed, residual extensors will do more work. In line with the common practice in transfemoral amputation, this result demonstrates during transfemoral amputation, it is important to maintain the length of residual limb as much as possible.

Inferences on Compound Rayleigh Parameters with Progressively Type-II Censored Samples

This paper considers inference under progressive type II censoring with a compound Rayleigh failure time distribution. The maximum likelihood (ML), and Bayes methods are used for estimating the unknown parameters as well as some lifetime parameters, namely reliability and hazard functions. We obtained Bayes estimators using the conjugate priors for two shape and scale parameters. When the two parameters are unknown, the closed-form expressions of the Bayes estimators cannot be obtained. We use Lindley.s approximation to compute the Bayes estimates. Another Bayes estimator has been obtained based on continuous-discrete joint prior for the unknown parameters. An example with the real data is discussed to illustrate the proposed method. Finally, we made comparisons between these estimators and the maximum likelihood estimators using a Monte Carlo simulation study.

Estimating of the Renewal Function with Heavy-tailed Claims

We develop a new estimator of the renewal function for heavy-tailed claims amounts. Our approach is based on the peak over threshold method for estimating the tail of the distribution with a generalized Pareto distribution. The asymptotic normality of an appropriately centered and normalized estimator is established, and its performance illustrated in a simulation study.

A Comparison of Some Thresholding Selection Methods for Wavelet Regression

In wavelet regression, choosing threshold value is a crucial issue. A too large value cuts too many coefficients resulting in over smoothing. Conversely, a too small threshold value allows many coefficients to be included in reconstruction, giving a wiggly estimate which result in under smoothing. However, the proper choice of threshold can be considered as a careful balance of these principles. This paper gives a very brief introduction to some thresholding selection methods. These methods include: Universal, Sure, Ebays, Two fold cross validation and level dependent cross validation. A simulation study on a variety of sample sizes, test functions, signal-to-noise ratios is conducted to compare their numerical performances using three different noise structures. For Gaussian noise, EBayes outperforms in all cases for all used functions while Two fold cross validation provides the best results in the case of long tail noise. For large values of signal-to-noise ratios, level dependent cross validation works well under correlated noises case. As expected, increasing both sample size and level of signal to noise ratio, increases estimation efficiency.

Markov Chain Monte Carlo Model Composition Search Strategy for Quantitative Trait Loci in a Bayesian Hierarchical Model

Quantitative trait loci (QTL) experiments have yielded important biological and biochemical information necessary for understanding the relationship between genetic markers and quantitative traits. For many years, most QTL algorithms only allowed one observation per genotype. Recently, there has been an increasing demand for QTL algorithms that can accommodate more than one observation per genotypic distribution. The Bayesian hierarchical model is very flexible and can easily incorporate this information into the model. Herein a methodology is presented that uses a Bayesian hierarchical model to capture the complexity of the data. Furthermore, the Markov chain Monte Carlo model composition (MC3) algorithm is used to search and identify important markers. An extensive simulation study illustrates that the method captures the true QTL, even under nonnormal noise and up to 6 QTL.

Comparing Interval Estimators for Reliability in a Dependent Set-up

In this paper some procedures for building confidence intervals for the reliability in stress-strength models are discussed and empirically compared. The particular case of a bivariate normal setup is considered. The confidence intervals suggested are obtained employing approximations or asymptotic properties of maximum likelihood estimators. The coverage and the precision of these intervals are empirically checked through a simulation study. An application to real paired data is also provided.

Trimmed Mean as an Adaptive Robust Estimator of a Location Parameter for Weibull Distribution

One of the purposes of the robust method of estimation is to reduce the influence of outliers in the data, on the estimates. The outliers arise from gross errors or contamination from distributions with long tails. The trimmed mean is a robust estimate. This means that it is not sensitive to violation of distributional assumptions of the data. It is called an adaptive estimate when the trimming proportion is determined from the data rather than being fixed a “priori-. The main objective of this study is to find out the robustness properties of the adaptive trimmed means in terms of efficiency, high breakdown point and influence function. Specifically, it seeks to find out the magnitude of the trimming proportion of the adaptive trimmed mean which will yield efficient and robust estimates of the parameter for data which follow a modified Weibull distribution with parameter λ = 1/2 , where the trimming proportion is determined by a ratio of two trimmed means defined as the tail length. Secondly, the asymptotic properties of the tail length and the trimmed means are also investigated. Finally, a comparison is made on the efficiency of the adaptive trimmed means in terms of the standard deviation for the trimming proportions and when these were fixed a “priori". The asymptotic tail lengths defined as the ratio of two trimmed means and the asymptotic variances were computed by using the formulas derived. While the values of the standard deviations for the derived tail lengths for data of size 40 simulated from a Weibull distribution were computed for 100 iterations using a computer program written in Pascal language. The findings of the study revealed that the tail lengths of the Weibull distribution increase in magnitudes as the trimming proportions increase, the measure of the tail length and the adaptive trimmed mean are asymptotically independent as the number of observations n becomes very large or approaching infinity, the tail length is asymptotically distributed as the ratio of two independent normal random variables, and the asymptotic variances decrease as the trimming proportions increase. The simulation study revealed empirically that the standard error of the adaptive trimmed mean using the ratio of tail lengths is relatively smaller for different values of trimming proportions than its counterpart when the trimming proportions were fixed a 'priori'.

Nonparametric Control Chart Using Density Weighted Support Vector Data Description

In manufacturing industries, development of measurement leads to increase the number of monitoring variables and eventually the importance of multivariate control comes to the fore. Statistical process control (SPC) is one of the most widely used as multivariate control chart. Nevertheless, SPC is restricted to apply in processes because its assumption of data as following specific distribution. Unfortunately, process data are composed by the mixture of several processes and it is hard to estimate as one certain distribution. To alternative conventional SPC, therefore, nonparametric control chart come into the picture because of the strength of nonparametric control chart, the absence of parameter estimation. SVDD based control chart is one of the nonparametric control charts having the advantage of flexible control boundary. However,basic concept of SVDD has been an oversight to the important of data characteristic, density distribution. Therefore, we proposed DW-SVDD (Density Weighted SVDD) to cover up the weakness of conventional SVDD. DW-SVDD makes a new attempt to consider dense of data as introducing the notion of density Weight. We extend as control chart using new proposed SVDD and a simulation study of various distributional data is conducted to demonstrate the improvement of performance.

Evaluating the Response of Rainfed-Chickpea to Population Density in Iran, Using Simulation

The response of growth and yield of rainfed-chickpea to population density should be evaluated based on long-term experiments to include the climate variability. This is achievable just by simulation. In this simulation study, this evaluation was done by running the CYRUS model for long-term daily weather data of five locations in Iran. The tested population densities were 7 to 59 (with interval of 2) stands per square meter. Various functions, including quadratic, segmented, beta, broken linear, and dent-like functions, were tested. Considering root mean square of deviations and linear regression statistics [intercept (a), slope (b), and correlation coefficient (r)] for predicted versus observed variables, the quadratic and broken linear functions appeared to be appropriate for describing the changes in biomass and grain yield, and in harvest index, respectively. Results indicated that in all locations, grain yield tends to show increasing trend with crowding the population, but subsequently decreases. This was also true for biomass in five locations. The harvest index appeared to have plateau state across low population densities, but decreasing trend with more increasing density. The turning point (optimum population density) for grain yield was 30.68 stands per square meter in Isfahan, 30.54 in Shiraz, 31.47 in Kermanshah, 34.85 in Tabriz, and 32.00 in Mashhad. The optimum population density for biomass ranged from 24.6 (in Tabriz) to 35.3 stands per square meter (Mashhad). For harvest index it varied between 35.87 and 40.12 stands per square meter.

Optimal Design of UPFC Based Damping Controller Using Iteration PSO

This paper presents a novel approach for tuning unified power flow controller (UPFC) based damping controller in order to enhance the damping of power system low frequency oscillations. The design problem of damping controller is formulated as an optimization problem according to the eigenvalue-based objective function which is solved using iteration particle swarm optimization (IPSO). The effectiveness of the proposed controller is demonstrated through eigenvalue analysis and nonlinear time-domain simulation studies under a wide range of loading conditions. The simulation study shows that the designed controller by IPSO performs better than CPSO in finding the solution. Moreover, the system performance analysis under different operating conditions show that the δE based controller is superior to the mB based controller.

Impact of Electronic Word-of-Mouth to Consumer Adoption Process in the Online Discussion Forum: A Simulation Study

Web-based technologies have created numerous opportunities for electronic word-of-mouth (eWOM) communication. There are many factors that affect customer adoption and decisionmaking process. However, only a few researches focus on some factors such as the membership time of forum and propensity to trust. Using a discrete-time event simulation to simulate a diffusion model along with a consumer decision model, the study shows the effect of each factor on adoption of opinions on on-line discussion forum. The purpose of this study is to examine the effect of factor affecting information adoption and decision making process. The model is constructed to test quantitative aspects of each factor. The simulation study shows the membership time and the propensity to trust has an effect on information adoption and purchasing decision. The result of simulation shows that the longer the membership time in the communities and the higher propensity to trust could lead to the higher demand rates because consumers find it easier and faster to trust the person in the community and then adopt the eWOM. Other implications for both researchers and practitioners are provided.

On the Performance of Information Criteria in Latent Segment Models

Nevertheless the widespread application of finite mixture models in segmentation, finite mixture model selection is still an important issue. In fact, the selection of an adequate number of segments is a key issue in deriving latent segments structures and it is desirable that the selection criteria used for this end are effective. In order to select among several information criteria, which may support the selection of the correct number of segments we conduct a simulation study. In particular, this study is intended to determine which information criteria are more appropriate for mixture model selection when considering data sets with only categorical segmentation base variables. The generation of mixtures of multinomial data supports the proposed analysis. As a result, we establish a relationship between the level of measurement of segmentation variables and some (eleven) information criteria-s performance. The criterion AIC3 shows better performance (it indicates the correct number of the simulated segments- structure more often) when referring to mixtures of multinomial segmentation base variables.

Simulation Study for Performance Comparison of Routing Protocols in Mobile Adhoc Network

Due to insufficient frequency band and tremendous growth of the mobile users, complex computation is needed for the use of resources. Long distance communication began with the introduction of telegraphs and simple coded pulses, which were used to transmit short messages. Since then numerous advances have rendered reliable transfer of information both easier and quicker. Wireless network refers to any type of computer network that is wireless, and is commonly associated with a telecommunications network whose interconnections between nodes is implemented without the use of wires. Wireless network can be broadly categorized in infrastructure network and infrastructure less network. Infrastructure network is one in which we have a base station to serve the mobile users and in the infrastructure less network is one in which no infrastructure is available to serve the mobile users this kind of networks are also known as mobile Adhoc networks. In this paper we have simulated the result for different scenarios with protocols like AODV and DSR; we simulated the result for throughput, delay and receiving traffic in the given scenario.

On Best Estimation for Parameter Weibull Distribution

The objective of this study is to introduce estimators to the parameters and survival function for Weibull distribution using three different methods, Maximum Likelihood estimation, Standard Bayes estimation and Modified Bayes estimation. We will then compared the three methods using simulation study to find the best one base on MPE and MSE.

Wavelet and K-L Seperability Based Feature Extraction Method for Functional Data Classification

This paper proposes a novel feature extraction method, based on Discrete Wavelet Transform (DWT) and K-L Seperability (KLS), for the classification of Functional Data (FD). This method combines the decorrelation and reduction property of DWT and the additive independence property of KLS, which is helpful to extraction classification features of FD. It is an advanced approach of the popular wavelet based shrinkage method for functional data reduction and classification. A theory analysis is given in the paper to prove the consistent convergence property, and a simulation study is also done to compare the proposed method with the former shrinkage ones. The experiment results show that this method has advantages in improving classification efficiency, precision and robustness.

Artificial Neural Networks Application to Improve Shunt Active Power Filter

Active Power Filters (APFs) are today the most widely used systems to eliminate harmonics compensate power factor and correct unbalanced problems in industrial power plants. We propose to improve the performances of conventional APFs by using artificial neural networks (ANNs) for harmonics estimation. This new method combines both the strategies for extracting the three-phase reference currents for active power filters and DC link voltage control method. The ANNs learning capabilities to adaptively choose the power system parameters for both to compute the reference currents and to recharge the capacitor value requested by VDC voltage in order to ensure suitable transit of powers to supply the inverter. To investigate the performance of this identification method, the study has been accomplished using simulation with the MATLAB Simulink Power System Toolbox. The simulation study results of the new (SAPF) identification technique compared to other similar methods are found quite satisfactory by assuring good filtering characteristics and high system stability.

Computational Simulation of Turbulence Heat Transfer in Multiple Rectangular Ducts

This study comprehensively simulate the use of k-ε model for predicting flow and heat transfer with measured flow field data in a stationary duct with elucidates on the detailed physics encountered in the fully developed flow region, and the sharp 180° bend region. Among the major flow features predicted with accuracy are flow transition at the entrance of the duct, the distribution of mean and turbulent quantities in the developing, fully developed, and sharp 180° bend, the development of secondary flows in the duct cross-section and the sharp 180° bend, and heat transfer augmentation. Turbulence intensities in the sharp 180° bend are found to reach high values and local heat transfer comparisons show that the heat transfer augmentation shifts towards the wall and along the duct. Therefore, understanding of the unsteady heat transfer in sharp 180° bends is important. The design and simulation are related to concept of fluid mechanics, heat transfer and thermodynamics. Simulation study has been conducted on the response of turbulent flow in a rectangular duct in order to evaluate the heat transfer rate along the small scale multiple rectangular duct