Reliability Analysis of Underground Pipelines Using Subset Simulation

An advanced Monte Carlo simulation method, called Subset Simulation (SS) for the time-dependent reliability prediction for underground pipelines has been presented in this paper. The SS can provide better resolution for low failure probability level with efficient investigating of rare failure events which are commonly encountered in pipeline engineering applications. In SS method, random samples leading to progressive failure are generated efficiently and used for computing probabilistic performance by statistical variables. SS gains its efficiency as small probability event as a product of a sequence of intermediate events with larger conditional probabilities. The efficiency of SS has been demonstrated by numerical studies and attention in this work is devoted to scrutinise the robustness of the SS application in pipe reliability assessment. It is hoped that the development work can promote the use of SS tools for uncertainty propagation in the decision-making process of underground pipelines network reliability prediction.

Visualisation and Navigation in Large Scale P2P Service Networks

In Peer-to-Peer service networks, where peers offer any kind of publicly available services or applications, intuitive navigation through all services in the network becomes more difficult as the number of services increases. In this article, a concept is discussed that enables users to intuitively browse and use large scale P2P service networks. The concept extends the idea of creating virtual 3D-environments solely based on Peer-to-Peer technologies. Aside from browsing, users shall have the possibility to emphasize services of interest using their own semantic criteria. The appearance of the virtual world shall intuitively reflect network properties that may be of interest for the user. Additionally, the concept comprises options for load- and traffic-balancing. In this article, the requirements concerning the underlying infrastructure and the graphical user interface are defined. First impressions of the appearance of future systems are presented and the next steps towards a prototypical implementation are discussed.

Study of the Effect of Project Management on Manufacturing and Production Projects

In this article the accumulated results out of the effects and length of the manufacture and production projects in the university and research standard have been settled with the usefulness definition of the process of project management for the accessibility to the proportional pattern in the “time and action" stages. Studies show that many problems confronted by the researchers in these projects are connected to the non-profiting of: 1) autonomous timing for gathering the educational theme, 2) autonomous timing for planning and pattern, presenting before the construction, and 3) autonomous timing for manufacture and sample presentation from the output. The result of this study indicates the division of every manufacture and production projects into three smaller autonomous projects from its kind, budget and autonomous expenditure, shape and order of the stages for the management of these kinds of projects. In this case study real result are compared with theoretical results.

Effects of Xylanase and Cellulase Production during Composting of EFB and POME using Fungi

Empty Fruit Bunches (EFB) and Palm Oil Mill Effluent (POME) are two main wastes from oil palm industries which contain rich lignocellulose. Degradation of EFB and POME by microorganisms will produce hydrolytic enzyme which will degrade cellulose and hemicellulose during composting process. However, normal composting takes about four to six months to reach maturity. Hence, application of fungi into compost can shorten the period of composting. This study identifies the effect of xylanase and cellulase produced by Aspergillus niger and Trichoderma virens on composting process using EFB and POME. The degradation of EFB and POME indicates the lignocellulolytic capacity of Aspergillus niger and Trichoderma virens with more than 7% decrease in hemicellulose and more than 25% decrease in cellulose for both inoculated compost. Inoculation of Aspergillus niger and Trichoderma virens also increased the enzyme activities during the composting period compared to the control compost by 21% for both xylanase and cellulase. Rapid rise in the activities of cellulase and xylanase was observed by Aspergillus niger with the highest activities of 14.41 FPU/mg and 3.89 IU/mg, respectively. Increased activities of cellulase and xylanase also occurred in inoculation of Trichoderma virens with the highest activities obtained at 13.21 FPU/mg and 4.43 IU/mg, respectively. Therefore, it is evident that the inoculation of fungi can increase the enzyme activities hence effectively degrading the EFB and POME.

Ionic Liquid Pretreatment and Enzymatic Hydrolysis of Wood Biomass

Pretreatment of lignocellulosic biomass materials from poplar, acacia, oak, and fir with different ionic liquids (ILs) containing 1-alkyl-3-methyl-imidazolium cations and various anions has been carried out. The dissolved cellulose from biomass was precipitated by adding anti-solvents into the solution and vigorous stirring. Commercial cellulases Celluclast 1.5L and Accelerase 1000 have been used for hydrolysis of untreated and pretreated lignocellulosic biomass. Among the tested ILs, [Emim]COOCH3 showed the best efficiency, resulting in highest amount of liberated reducing sugars. Pretreatment of lignocellulosic biomass using glycerol-ionic liquids combined pretreatment and dilute acid-ionic liquids combined pretreatment were evaluated and compared with glycerol pretreatment, ionic liquids pretreatment and dilute acid pretreatment.

A Comparison of Marginal and Joint Generalized Quasi-likelihood Estimating Equations Based On the Com-Poisson GLM: Application to Car Breakdowns Data

In this paper, we apply and compare two generalized estimating equation approaches to the analysis of car breakdowns data in Mauritius. Number of breakdowns experienced by a machinery is a highly under-dispersed count random variable and its value can be attributed to the factors related to the mechanical input and output of that machinery. Analyzing such under-dispersed count observation as a function of the explanatory factors has been a challenging problem. In this paper, we aim at estimating the effects of various factors on the number of breakdowns experienced by a passenger car based on a study performed in Mauritius over a year. We remark that the number of passenger car breakdowns is highly under-dispersed. These data are therefore modelled and analyzed using Com-Poisson regression model. We use the two types of quasi-likelihood estimation approaches to estimate the parameters of the model: marginal and joint generalized quasi-likelihood estimating equation approaches. Under-dispersion parameter is estimated to be around 2.14 justifying the appropriateness of Com-Poisson distribution in modelling underdispersed count responses recorded in this study.

Rigorous Modeling of Fixed-Bed Reactors Containing Finite Hollow Cylindrical Catalyst with Michaelis-Menten Type of Kinetics

A large number of chemical, bio-chemical and pollution-control processes use heterogeneous fixed-bed reactors. The use of finite hollow cylindrical catalyst pellets can enhance conversion levels in such reactors. The absence of the pellet core can significantly lower the diffusional resistance associated with the solid phase. This leads to a better utilization of the catalytic material, which is reflected in the higher values for the effectiveness factor, leading ultimately to an enhanced conversion level in the reactor. It is however important to develop a rigorous heterogeneous model for the reactor incorporating the two-dimensional feature of the solid phase owing to the presence of the finite hollow cylindrical catalyst pellet. Presently, heterogeneous models reported in the literature invariably employ one-dimension solid phase models meant for spherical catalyst pellets. The objective of the paper is to present a rigorous model of the fixed-bed reactors containing finite hollow cylindrical catalyst pellets. The reaction kinetics considered here is the widely used Michaelis–Menten kinetics for the liquid-phase bio-chemical reactions. The reaction parameters used here are for the enzymatic degradation of urea. Results indicate that increasing the height to diameter ratio helps to improve the conversion level. On the other hand, decreasing the thickness is apparently not as effective. This could however be explained in terms of the higher void fraction of the bed that causes a smaller amount of the solid phase to be packed in the fixed-bed bio-chemical reactor.

Single-qubit Quantum Gates using Magneto-optic Kerr Effect

We propose the use of magneto-optic Kerr effect (MOKE) to realize single-qubit quantum gates. We consider longitudinal and polar MOKE in reflection geometry in which the magnetic field is parallel to both the plane of incidence and surface of the film. MOKE couples incident TE and TM polarized photons and the Hamiltonian that represents this interaction is isomorphic to that of a canonical two-level quantum system. By varying the phase and amplitude of the magnetic field, we can realize Hadamard, NOT, and arbitrary phase-shift single-qubit quantum gates. The principal advantage is operation with magnetically non-transparent materials.

IVE: Virtual Humans’ AI Prototyping Toolkit

IVE toolkit has been created for facilitating research,education and development in the field of virtual storytelling and computer games. Primarily, the toolkit is intended for modelling action selection mechanisms of virtual humans, investigating level-of-detail AI techniques for large virtual environments, and for exploring joint behaviour and role-passing technique (Sec. V). Additionally, the toolkit can be used as an AI middleware without any changes. The main facility of IVE is that it serves for prototyping both the AI and virtual worlds themselves. The purpose of this paper is to describe IVE's features in general and to present our current work - including an educational game - on this platform.

A Multi Objective Optimization Approach to Optimize Vehicle Ride and Handling Characteristics

Vehicle suspension design must fulfill some conflicting criteria. Among those is ride comfort which is attained by minimizing the acceleration transmitted to the sprung mass, via suspension spring and damper. Also good handling of a vehicle is a desirable property which requires stiff suspension and therefore is in contrast with a vehicle with good ride. Among the other desirable features of a suspension is the minimization of the maximum travel of suspension. This travel which is called suspension working space in vehicle dynamics literature is also a design constraint and it favors good ride. In this research a full car 8 degrees of freedom model has been developed and the three above mentioned criteria, namely: ride, handling and working space has been adopted as objective functions. The Multi Objective Programming (MOP) discipline has been used to find the Pareto Front and some reasoning used to chose a design point between these non dominated points of Pareto Front.

A New Approach for Prioritization of Failure Modes in Design FMEA using ANOVA

The traditional Failure Mode and Effects Analysis (FMEA) uses Risk Priority Number (RPN) to evaluate the risk level of a component or process. The RPN index is determined by calculating the product of severity, occurrence and detection indexes. The most critically debated disadvantage of this approach is that various sets of these three indexes may produce an identical value of RPN. This research paper seeks to address the drawbacks in traditional FMEA and to propose a new approach to overcome these shortcomings. The Risk Priority Code (RPC) is used to prioritize failure modes, when two or more failure modes have the same RPN. A new method is proposed to prioritize failure modes, when there is a disagreement in ranking scale for severity, occurrence and detection. An Analysis of Variance (ANOVA) is used to compare means of RPN values. SPSS (Statistical Package for the Social Sciences) statistical analysis package is used to analyze the data. The results presented are based on two case studies. It is found that the proposed new methodology/approach resolves the limitations of traditional FMEA approach.

Chaotic Response and Bifurcation Analysis of Gear-Bearing System with and without Porous Effect under Nonlinear Suspension

This study presents a systematic analysis of the dynamic behaviors of a gear-bearing system with porous squeeze film damper (PSFD) under nonlinear suspension, nonlinear oil-film force and nonlinear gear meshing force effect. It can be found that the system exhibits very rich forms of sub-harmonic and even the chaotic vibrations. The bifurcation diagrams also reveal that greater values of permeability may not only improve non-periodic motions effectively, but also suppress dynamic amplitudes of the system. Therefore, porous effect plays an important role to improve dynamic stability of gear-bearing systems or other mechanical systems. The results presented in this study provide some useful insights into the design and development of a gear-bearing system for rotating machinery that operates in highly rotational speed and highly nonlinear regimes.

Development of a Simple laser-based 2D Compensating System for the Contouring Accuracy of Machine Tools

The dynamical contouring error is a critical element for the accuracy of machine tools. The contouring error is defined as the difference between the processing actual path and commanded path, which is implemented by following the command curves from feeding driving system in machine tools. The contouring error is resulted from various factors, such as the external loads, friction, inertia moment, feed rate, speed control, servo control, and etc. Thus, the study proposes a 2D compensating system for the contouring accuracy of machine tools. Optical method is adopted by using stable frequency laser diode and the high precision position sensor detector (PSD) to performno-contact measurement. Results show the related accuracy of position sensor detector (PSD) of 2D contouring accuracy compensating system was ±1.5 μm for a calculated range of ±3 mm, and improvement accuracy is over 80% at high-speed feed rate.

Experiments and Modeling of Ion Exchange Resins for Nuclear Power Plants

Resins are used in nuclear power plants for water ultrapurification. Two approaches are considered in this work: column experiments and simulations. A software called OPTIPUR was developed, tested and used. The approach simulates the onedimensional reactive transport in porous medium with convectivedispersive transport between particles and diffusive transport within the boundary layer around the particles. The transfer limitation in the boundary layer is characterized by the mass transfer coefficient (MTC). The influences on MTC were measured experimentally. The variation of the inlet concentration does not influence the MTC; on the contrary of the Darcy velocity which influences. This is consistent with results obtained using the correlation of Dwivedi&Upadhyay. With the MTC, knowing the number of exchange site and the relative affinity, OPTIPUR can simulate the column outlet concentration versus time. Then, the duration of use of resins can be predicted in conditions of a binary exchange.

Multiple Moving Talker Tracking by Integration of Two Successive Algorithms

In this paper, an estimation accuracy of multiple moving talker tracking using a microphone array is improved. The tracking can be achieved by the adaptive method in which two algorithms are integrated, namely, the PAST (Projection Approximation Subspace Tracking) algorithm and the IPLS (Interior Point Least Square) algorithm. When either talker begins to speak again after a silent period, an appropriate feasible region for an evaluation function of the IPLS algorithm might not be set. Then, the tracking fails due to the incorrect updating. Therefore, if an increment of the number of active talkers is detected, the feasible region must be reset. Then, a low cost realization is required for the high speed tracking and a high accuracy realization is desired for the precise tracking. In this paper, the directions roughly estimated using the delayed-sum-array method are used for the resetting. Several results of experiments performed in an actual room environment show the effectiveness of the proposed method.

A Study of the Damages to Historical Monuments due to Climatic Factors and Air Pollution and Offering Solutions

Historical monuments as architectural heritage are, economically and culturally, considered one of the key aspects for modern communities. Cultural heritage represents a country-s national identity and pride and maintains and enriches that country-s culture. Therefore, conservation of the monuments remained from our ancestors requires everybody-s serious and unremitting effort. Conservation, renewal, restoration, and technical study of cultural and historical matters are issues which have a special status among various forms of art and science in the present century and this is due to two reasons: firstly, progress of humankind in this century has created a factor called environmental pollution which not only has caused new destructive processes of cultural/historical monuments but also has accelerated the previous destructive processes by several times, and secondly, the rapid advance of various sciences, especially chemistry, has lead to the contribution of new methods and materials to this significant issue.

Wave Vortex Parameters as an Indicator of Breaking Intensity

The study of the geometric shape of the plunging wave enclosed vortices as a possible indicator for the breaking intensity of ocean waves has been ongoing for almost 50 years with limited success. This paper investigates the validity of using the vortex ratio and vortex angle as methods of predicting breaking intensity. Previously published works on vortex parameters, based on regular wave flume results or solitary wave theory, present contradictory results and conclusions. Through the first complete analysis of field collected irregular wave breaking vortex parameters it is illustrated that the vortex ratio and vortex angle cannot be accurately predicted using standard breaking wave characteristics and hence are not suggested as a possible indicator for breaking intensity.

Organizational De-Evolution; the Small Group or Single Actor Terrorist

Traditionally, terror groups have been formed by ideologically aligned actors who perceive a lack of options for achieving political or social change. However, terrorist attacks have been increasingly carried out by small groups of actors or lone individuals who may be only ideologically affiliated with larger, formal terrorist organizations. The formation of these groups represents the inverse of traditional organizational growth, whereby structural de-evolution within issue-based organizations leads to the formation of small, independent terror cells. Ideological franchising – the bypassing of formal affiliation to the “parent" organization – represents the de-evolution of traditional concepts of organizational structure in favor of an organic, independent, and focused unit. Traditional definitions of dark networks that are issue-based include focus on an identified goal, commitment to achieving this goal through unrestrained actions, and selection of symbolic targets. The next step in the de-evolution of small dark networks is the miniorganization, consisting of only a handful of actors working toward a common, violent goal. Information-sharing through social media platforms, coupled with civil liberties of democratic nations, provide the communication systems, access to information, and freedom of movement necessary for small dark networks to flourish without the aid of a parent organization. As attacks such as the 7/7 bombings demonstrate the effectiveness of small dark networks, terrorist actors will feel increasingly comfortable aligning with an ideology only, without formally organizing. The natural result of this de-evolving organization is the single actor event, where an individual seems to subscribe to a larger organization-s violent ideology with little or no formal ties.

Software Reliability Prediction Model Analysis

Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.

Reconstitute Information about Discontinued Water Quality Variables in the Nile Delta Monitoring Network Using Two Record Extension Techniques

The world economic crises and budget constraints have caused authorities, especially those in developing countries, to rationalize water quality monitoring activities. Rationalization consists of reducing the number of monitoring sites, the number of samples, and/or the number of water quality variables measured. The reduction in water quality variables is usually based on correlation. If two variables exhibit high correlation, it is an indication that some of the information produced may be redundant. Consequently, one variable can be discontinued, and the other continues to be measured. Later, the ordinary least squares (OLS) regression technique is employed to reconstitute information about discontinued variable by using the continuously measured one as an explanatory variable. In this paper, two record extension techniques are employed to reconstitute information about discontinued water quality variables, the OLS and the Line of Organic Correlation (LOC). An empirical experiment is conducted using water quality records from the Nile Delta water quality monitoring network in Egypt. The record extension techniques are compared for their ability to predict different statistical parameters of the discontinued variables. Results show that the OLS is better at estimating individual water quality records. However, results indicate an underestimation of the variance in the extended records. The LOC technique is superior in preserving characteristics of the entire distribution and avoids underestimation of the variance. It is concluded from this study that the OLS can be used for the substitution of missing values, while LOC is preferable for inferring statements about the probability distribution.