A Renovated Cook's Distance Based On The Buckley-James Estimate In Censored Regression

There have been various methods created based on the regression ideas to resolve the problem of data set containing censored observations, i.e. the Buckley-James method, Miller-s method, Cox method, and Koul-Susarla-Van Ryzin estimators. Even though comparison studies show the Buckley-James method performs better than some other methods, it is still rarely used by researchers mainly because of the limited diagnostics analysis developed for the Buckley-James method thus far. Therefore, a diagnostic tool for the Buckley-James method is proposed in this paper. It is called the renovated Cook-s Distance, (RD* i ) and has been developed based on the Cook-s idea. The renovated Cook-s Distance (RD* i ) has advantages (depending on the analyst demand) over (i) the change in the fitted value for a single case, DFIT* i as it measures the influence of case i on all n fitted values Yˆ∗ (not just the fitted value for case i as DFIT* i) (ii) the change in the estimate of the coefficient when the ith case is deleted, DBETA* i since DBETA* i corresponds to the number of variables p so it is usually easier to look at a diagnostic measure such as RD* i since information from p variables can be considered simultaneously. Finally, an example using Stanford Heart Transplant data is provided to illustrate the proposed diagnostic tool.

Study on Specific Energy in Grinding of DRACs: A Response Surface Methodology Approach

In this study, the effects of machining parameters on specific energy during surface grinding of 6061Al-SiC35P composites are investigated. Vol% of SiC, feed and depth of cut were chosen as process variables. The power needed for the calculation of the specific energy is measured from the two watt meter method. Experiments are conducted using standard RSM design called Central composite design (CCD). A second order response surface model was developed for specific energy. The results identify the significant influence factors to minimize the specific energy. The confirmation results demonstrate the practicability and effectiveness of the proposed approach.

Research on Strategy for Automated Scaleless-Map Compilation

As a tool for human spatial cognition and thinking, the map has been playing an important role. Maps are perhaps as fundamental to society as language and the written word. Economic and social development requires extensive and in-depth understanding of their own living environment, from the scope of the overall global to urban housing. This has brought unprecedented opportunities and challenges for traditional cartography . This paper first proposed the concept of scaleless-map and its basic characteristics, through the analysis of the existing multi-scale representation techniques. Then some strategies are presented for automated mapping compilation. Taking into account the demand of automated map compilation, detailed proposed the software - WJ workstation must have four technical features, which are generalization operators, symbol primitives, dynamically annotation and mapping process template. This paper provides a more systematic new idea and solution to improve the intelligence and automation of the scaleless cartography.

Indonesian News Classification using Support Vector Machine

Digital news with a variety topics is abundant on the internet. The problem is to classify news based on its appropriate category to facilitate user to find relevant news rapidly. Classifier engine is used to split any news automatically into the respective category. This research employs Support Vector Machine (SVM) to classify Indonesian news. SVM is a robust method to classify binary classes. The core processing of SVM is in the formation of an optimum separating plane to separate the different classes. For multiclass problem, a mechanism called one against one is used to combine the binary classification result. Documents were taken from the Indonesian digital news site, www.kompas.com. The experiment showed a promising result with the accuracy rate of 85%. This system is feasible to be implemented on Indonesian news classification.

Conversion of Sugarcane Shoots to Reducing Sugars

Sugarcane Shoots is an abundantly available residual resources consisting of lignocelluloses which take it into the benefit. The present study was focused on utilizing of sugarcane shoot for reducing sugar production as a substrate in ethanol production. Physical and chemical pretreatments of sugarcane shoot were investigated. Results showed that the size of sugarcane shoot influenced the cellulose content. The maximum cellulose yield (60 %) can be obtained from alkaline pretreated sugarcane shoot with 1.0 M NaOH at 30 oC for 90 min. The cellulose yield reached up to 93.9% (w/w). Enzymatically hydrolyzed of cellulosic residual in 0.04 citrate buffer (pH 5) with celluclast 1.5L (0.7 FPU/ml) resulted in the highest amount of reducing sugar at a rate of 32.1 g/l after 4 h incubation at 50°C, and 100 oC for 5 min . Cellulose conversion was 55.5%.

Comparing Interval Estimators for Reliability in a Dependent Set-up

In this paper some procedures for building confidence intervals for the reliability in stress-strength models are discussed and empirically compared. The particular case of a bivariate normal setup is considered. The confidence intervals suggested are obtained employing approximations or asymptotic properties of maximum likelihood estimators. The coverage and the precision of these intervals are empirically checked through a simulation study. An application to real paired data is also provided.

A Critical Review of the Adequacy of EIA Reports-Evidence from Pakistan

The preparation of good-quality Environmental Impact Assessment (EIA) reports contribute to enhancing overall effectiveness of EIA. This component of the EIA process becomes more important in situation where public participation is weak and there is lack of expertise on the part of the competent authority. In Pakistan, EIA became mandatory for every project likely to cause adverse environmental impacts from July 1994. The competent authority also formulated guidelines for preparation and review of EIA reports in 1997. However, EIA is yet to prove as a successful decision support tool to help in environmental protection. One of the several reasons of this ineffectiveness is the generally poor quality of EIA reports. This paper critically reviews EIA reports of some randomly selected projects. Interviews of EIA consultants, project proponents and concerned government officials have also been conducted to underpin the root causes of poor quality of EIA reports. The analysis reveals several inadequacies particularly in areas relating to identification, evaluation and mitigation of key impacts and consideration of alternatives. The paper identifies some opportunities and suggests measures for improving the quality of EIA reports and hence making EIA an effective tool to help in environmental protection.

Viscous Potential Flow Analysis of Electrohydrodynamic Capillary Instability through Porous Media

The effect of porous medium on the capillary instability of a cylindrical interface in the presence of axial electric field has been investigated using viscous potential flow theory. In viscous potential flow, the viscous term in Navier-Stokes equation vanishes as vorticity is zero but viscosity is not zero. Viscosity enters through normal stress balance in the viscous potential flow theory and tangential stresses are not considered. A dispersion relation that accounts for the growth of axisymmetric waves is derived and stability is discussed theoretically as well as numerically. Stability criterion is given by critical value of applied electric field as well as critical wave number. Various graphs have been drawn to show the effect of various physical parameters such as electric field, viscosity ratio, permittivity ratio on the stability of the system. It has been observed that the axial electric field and porous medium both have stabilizing effect on the stability of the system.

Biometrical Comparison of Artemia urmiana Günther, 1899 (Crustacea: Anostraca) Cysts between Rainy and Drought Years (1994-2003/4) from Urmia Lake, Iran

Nowadays, biometrical characterizations of Artemia cysts are used as one of the most important factors in the study of Artemia populations and intraspecific particularity; meanwhile these characters can be used as economical indices. For example, typically high hatching efficiency is possible due to the small diameter of cysts (high number per gram); therefore small diameter of cysts show someway high quality of cysts. This study was performed during a ten year period, including two different ecological conditions: rainy and drought. It is important from two different aspects because it covers alteration of A. urmiana during ten years also its variation in the best and worst environmental situations in which salinity increased from 173.8 ppt in 1994 to 280.8 ppt in 2003/4. In this study the biometrical raw data of Artemia urmiana cysts at seven stations from the Urmia Lake in 1994 and their seven identical locations at 26 studied stations in 2003/4 were reanalyzed again and compared together. Biometrical comparison of untreated and decapsulated cysts in each of the seven similar stations showed a highly significant variation between 1994 and 2003/4. Based on this study, in whole stations the untreated and decapsulated cysts from 1994 were larger than cysts of 2003/4 without any exception. But there was no logical relationship between salinity and chorion thickness in the Urmia Lake. With regard to PCA analyses the stations of two different studied years certainly have been separated with factor 1 from each other. In conclusion, the interaction between genetic and environmental factors can determine and explain variation in the range of cysts diameter in Artemia.

Community Detection-based Analysis of the Human Interactome Network

The study of proteomics reached unexpected levels of interest, as a direct consequence of its discovered influence over some complex biological phenomena, such as problematic diseases like cancer. This paper presents a new technique that allows for an accurate analysis of the human interactome network. It is basically a two-step analysis process that involves, at first, the detection of each protein-s absolute importance through the betweenness centrality computation. Then, the second step determines the functionallyrelated communities of proteins. For this purpose, we use a community detection technique that is based on the edge betweenness calculation. The new technique was thoroughly tested on real biological data and the results prove some interesting properties of those proteins that are involved in the carcinogenesis process. Apart from its experimental usefulness, the novel technique is also computationally effective in terms of execution times. Based on the analysis- results, some topological features of cancer mutated proteins are presented and a possible optimization solution for cancer drugs design is suggested.

An Investigation on Thermo Chemical Conversions of Solid Waste for Energy Recovery

Solid waste can be considered as an urban burden or as a valuable resource depending on how it is managed. To meet the rising demand for energy and to address environmental concerns, a conversion from conventional energy systems to renewable resources is essential. For the sustainability of human civilization, an environmentally sound and techno-economically feasible waste treatment method is very important to treat recyclable waste. Several technologies are available for realizing the potential of solid waste as an energy source, ranging from very simple systems for disposing of dry waste to more complex technologies capable of dealing with large amounts of industrial waste. There are three main pathways for conversion of waste material to energy: thermo chemical, biochemical and physicochemical. This paper investigates the thermo chemical conversion of solid waste for energy recovery. The processes, advantages and dis-advantages of various thermo chemical conversion processes are discussed and compared. Special attention is given to Gasification process as it provides better solutions regarding public acceptance, feedstock flexibility, near-zero emissions, efficiency and security. Finally this paper presents comparative statements of thermo chemical processes and introduces an integrated waste management system.

The Creation of Sustainable Architecture by use of Transformable Intelligent Building Skins

Built environments have a large impact on environmental sustainability and if it is not considered properly can negatively affect our planet. The application of transformable intelligent building systems that automatically respond to environmental conditions is one of the best ways that can intelligently assist us to create sustainable environment. The significance of this issue is evident as energy crisis and environmental changes has made the sustainability the main concerns in many societies. The aim of this research is to review and evaluate the importance and influence of transformable intelligent structure on the creation of sustainable architecture. Intelligent systems in current buildings provide convenience through automatically responding to changes in environmental conditions, reducing energy dissipation and increase of the lifecycle of buildings. This paper by analyzing significant intelligent building systems will evaluate the potentials of transformable intelligent systems in the creation of sustainable architecture and environment.

A Comparative Study of Rigid and Modified Simplex Methods for Optimal Parameter Settings of ACO for Noisy Non-Linear Surfaces

There are two common types of operational research techniques, optimisation and metaheuristic methods. The latter may be defined as a sequential process that intelligently performs the exploration and exploitation adopted by natural intelligence and strong inspiration to form several iterative searches. An aim is to effectively determine near optimal solutions in a solution space. In this work, a type of metaheuristics called Ant Colonies Optimisation, ACO, inspired by a foraging behaviour of ants was adapted to find optimal solutions of eight non-linear continuous mathematical models. Under a consideration of a solution space in a specified region on each model, sub-solutions may contain global or multiple local optimum. Moreover, the algorithm has several common parameters; number of ants, moves, and iterations, which act as the algorithm-s driver. A series of computational experiments for initialising parameters were conducted through methods of Rigid Simplex, RS, and Modified Simplex, MSM. Experimental results were analysed in terms of the best so far solutions, mean and standard deviation. Finally, they stated a recommendation of proper level settings of ACO parameters for all eight functions. These parameter settings can be applied as a guideline for future uses of ACO. This is to promote an ease of use of ACO in real industrial processes. It was found that the results obtained from MSM were pretty similar to those gained from RS. However, if these results with noise standard deviations of 1 and 3 are compared, MSM will reach optimal solutions more efficiently than RS, in terms of speed of convergence.

Maintenance Management System for Upstream Operations in Oil and Gas Industry: Case Study

This paper explores the plant maintenance management system that has been used by giant oil and gas company in Malaysia. The system also called as PMMS used to manage the upstream operations for more than 100 plants of the case study company. Moreover, from the observations, focus group discussion with PMMS personnel and application through simulation (SAP R/3), the paper reviews the step-by-step approach and the elements that required for the PMMS. The findings show that the PMMS integrates the overall business strategy in upstream operations that consist of asset management, work management and performance management. In addition, PMMS roles are to help operations personnel organize and plan their daily activities, to improve productivity and reduce equipment downtime and to help operations management analyze the facilities and create performance, and to provide and maintain the operational effectiveness of the facilities.

Analysis of Endovascular Graft Features Affecting Endotension Following Endovascular Aneurysm Repair

Endovascular aneurysm repair is a new and minimally invasive repair for patients with abdominal aortic aneurysm (AAA). This method has potential advantages that are incomparable with other repair methods. However, the enlargement of aneurysm in the absence of endoleak, which is known as endotension, may occur as one of post-operative compliances of this method. Typically, endotension is mainly as a result of pressure transmitted to aneurysm sac by endovascular installed graft. After installation of graft the aneurysm sac reduces significantly but remains non-zero. There are some factors which affect this pressure transmitted. In this study, the geometry features of installed vascular graft have been considered. It is inferred that graft neck angle and iliac bifurcation angle are two factors which can affect the drag force on graft and consequently the pressure transmitted to aneurysm.

Intuition Operator: Providing Genomes with Reason

In this contribution, the use of a new genetic operator is proposed. The main advantage of using this operator is that it is able to assist the evolution procedure to converge faster towards the optimal solution of a problem. This new genetic operator is called ''intuition'' operator. Generally speaking, one can claim that this operator is a way to include any heuristic or any other local knowledge, concerning the problem, that cannot be embedded in the fitness function. Simulation results show that the use of this operator increases significantly the performance of the classic Genetic Algorithm by increasing the convergence speed of its population.

Trimmed Mean as an Adaptive Robust Estimator of a Location Parameter for Weibull Distribution

One of the purposes of the robust method of estimation is to reduce the influence of outliers in the data, on the estimates. The outliers arise from gross errors or contamination from distributions with long tails. The trimmed mean is a robust estimate. This means that it is not sensitive to violation of distributional assumptions of the data. It is called an adaptive estimate when the trimming proportion is determined from the data rather than being fixed a “priori-. The main objective of this study is to find out the robustness properties of the adaptive trimmed means in terms of efficiency, high breakdown point and influence function. Specifically, it seeks to find out the magnitude of the trimming proportion of the adaptive trimmed mean which will yield efficient and robust estimates of the parameter for data which follow a modified Weibull distribution with parameter λ = 1/2 , where the trimming proportion is determined by a ratio of two trimmed means defined as the tail length. Secondly, the asymptotic properties of the tail length and the trimmed means are also investigated. Finally, a comparison is made on the efficiency of the adaptive trimmed means in terms of the standard deviation for the trimming proportions and when these were fixed a “priori". The asymptotic tail lengths defined as the ratio of two trimmed means and the asymptotic variances were computed by using the formulas derived. While the values of the standard deviations for the derived tail lengths for data of size 40 simulated from a Weibull distribution were computed for 100 iterations using a computer program written in Pascal language. The findings of the study revealed that the tail lengths of the Weibull distribution increase in magnitudes as the trimming proportions increase, the measure of the tail length and the adaptive trimmed mean are asymptotically independent as the number of observations n becomes very large or approaching infinity, the tail length is asymptotically distributed as the ratio of two independent normal random variables, and the asymptotic variances decrease as the trimming proportions increase. The simulation study revealed empirically that the standard error of the adaptive trimmed mean using the ratio of tail lengths is relatively smaller for different values of trimming proportions than its counterpart when the trimming proportions were fixed a 'priori'.

Pattern Matching Based on Regular Tree Grammars

Pattern matching based on regular tree grammars have been widely used in many areas of computer science. In this paper, we propose a pattern matcher within the framework of code generation, based on a generic and a formalized approach. According to this approach, parsers for regular tree grammars are adapted to a general pattern matching solution, rather than adapting the pattern matching according to their parsing behavior. Hence, we first formalize the construction of the pattern matches respective to input trees drawn from a regular tree grammar in a form of the so-called match trees. Then, we adopt a recently developed generic parser and tightly couple its parsing behavior with such construction. In addition to its generality, the resulting pattern matcher is characterized by its soundness and efficient implementation. This is demonstrated by the proposed theory and by the derived algorithms for its implementation. A comparison with similar and well-known approaches, such as the ones based on tree automata and LR parsers, has shown that our pattern matcher can be applied to a broader class of grammars, and achieves better approximation of pattern matches in one pass. Furthermore, its use as a machine code selector is characterized by a minimized overhead, due to the balanced distribution of the cost computations into static ones, during parser generation time, and into dynamic ones, during parsing time.

Optimization for Reducing Handoff Latency and Utilization of Bandwidth in ATM Networks

To support mobility in ATM networks, a number of technical challenges need to be resolved. The impact of handoff schemes in terms of service disruption, handoff latency, cost implications and excess resources required during handoffs needs to be addressed. In this paper, a one phase handoff and route optimization solution using reserved PVCs between adjacent ATM switches to reroute connections during inter-switch handoff is studied. In the second phase, a distributed optimization process is initiated to optimally reroute handoff connections. The main objective is to find the optimal operating point at which to perform optimization subject to cost constraint with the purpose of reducing blocking probability of inter-switch handoff calls for delay tolerant traffic. We examine the relation between the required bandwidth resources and optimization rate. Also we calculate and study the handoff blocking probability due to lack of bandwidth for resources reserved to facilitate the rapid rerouting.

Access Policy Specification for SCADA Networks

Efforts to secure supervisory control and data acquisition (SCADA) systems must be supported under the guidance of sound security policies and mechanisms to enforce them. Critical elements of the policy must be systematically translated into a format that can be used by policy enforcement components. Ideally, the goal is to ensure that the enforced policy is a close reflection of the specified policy. However, security controls commonly used to enforce policies in the IT environment were not designed to satisfy the specific needs of the SCADA environment. This paper presents a language, based on the well-known XACML framework, for the expression of authorization policies for SCADA systems.