The Index of Sustainable Functionality: An Application for Measuring Sustainability

The index of sustainable functionality (ISF) is an adaptive, multi-criteria technique that is used to measure sustainability; it is a concept that can be transposed to many regions throughout the world. An ISF application of the Southern Regional Organisation of Councils (SouthROC) in South East Queensland (SEQ) – the fastest growing region in Australia – indicated over a 25 year period an increase of over 10% level of functionality from 58.0% to 68.3%. The ISF of SouthROC utilised methodologies that derived from an expert panel based approach. The overall results attained an intermediate level of functionality which amounted to related concerns of economic progress and lack of social awareness. Within the region, a solid basis for future testing by way of measured changes and developed trends can be established. In this regard as management tool, the ISF record offers support for regional sustainability practice and decision making alike. This research adaptively analyses sustainability – a concept that is lacking throughout much of the academic literature and any reciprocal experimentation. This lack of knowledge base has been the emphasis of where future sustainability research can grow from and prove useful in rapidly growing regions. It is the intentions of this research to help further develop the notions of index-based quantitative sustainability.

Combined Sewer Overflow forecasting with Feed-forward Back-propagation Artificial Neural Network

A feed-forward, back-propagation Artificial Neural Network (ANN) model has been used to forecast the occurrences of wastewater overflows in a combined sewerage reticulation system. This approach was tested to evaluate its applicability as a method alternative to the common practice of developing a complete conceptual, mathematical hydrological-hydraulic model for the sewerage system to enable such forecasts. The ANN approach obviates the need for a-priori understanding and representation of the underlying hydrological hydraulic phenomena in mathematical terms but enables learning the characteristics of a sewer overflow from the historical data. The performance of the standard feed-forward, back-propagation of error algorithm was enhanced by a modified data normalizing technique that enabled the ANN model to extrapolate into the territory that was unseen by the training data. The algorithm and the data normalizing method are presented along with the ANN model output results that indicate a good accuracy in the forecasted sewer overflow rates. However, it was revealed that the accurate forecasting of the overflow rates are heavily dependent on the availability of a real-time flow monitoring at the overflow structure to provide antecedent flow rate data. The ability of the ANN to forecast the overflow rates without the antecedent flow rates (as is the case with traditional conceptual reticulation models) was found to be quite poor.

An Impairment Sensitive and Reliable SR-ARQ Mechanism for Unreliable Feedback in GPRS

The advances in wireless communication have opened unlimited horizons but there are some challenges as well. The Nature derived air medium between MS (Mobile Station) and BS (Base Station) is beyond human control and produces channel impairment. The impact of the natural conditions at the air medium is the biggest issue in wireless communication. Natural conditions make reliability more cumbersome; here reliability refers to the efficient recovery of the lost or erroneous data. The SR-ARQ (Selective Repeat-Automatic Repeat Request) protocol is a de facto standard for any wireless technology at the air interface with its standard reliability features. Our focus in this research is on the reliability of the control or feedback signal of the SR-ARQ protocol. The proposed mechanism, RSR-ARQ (Reliable SR-ARQ) is an enhancement of the SR-ARQ protocol that has ensured the reliability of the control signals through channel impairment sensitive mechanism. We have modeled the system under two-state discrete time Markov Channel. The simulation results demonstrate the better recovery of the lost or erroneous data that will increase the overall system performance.

A Study on the Application of TRIZ to CAD/CAM System

This study created new graphical icons and operating functions in a CAD/CAM software system by analyzing icons in some of the popular systems, such as AutoCAD, AlphaCAM, Mastercam and the 1st edition of LiteCAM. These software systems all focused on geometric design and editing, thus how to transmit messages intuitively from icon itself to users is an important function of graphical icons. The primary purpose of this study is to design innovative icons and commands for new software. This study employed the TRIZ method, an innovative design method, to generate new concepts systematically. Through literature review, it then investigated and analyzed the relationship between TRIZ and idea development. Contradiction Matrix and 40 Principles were used to develop an assisting tool suitable for icon design in software development. We first gathered icon samples from the selected CAD/CAM systems. Then grouped these icons by meaningful functions, and compared useful and harmful properties. Finally, we developed new icons for new software systems in order to avoid intellectual property problem.

Arriving at an Optimum Value of Tolerance Factor for Compressing Medical Images

Medical imaging uses the advantage of digital technology in imaging and teleradiology. In teleradiology systems large amount of data is acquired, stored and transmitted. A major technology that may help to solve the problems associated with the massive data storage and data transfer capacity is data compression and decompression. There are many methods of image compression available. They are classified as lossless and lossy compression methods. In lossy compression method the decompressed image contains some distortion. Fractal image compression (FIC) is a lossy compression method. In fractal image compression an image is coded as a set of contractive transformations in a complete metric space. The set of contractive transformations is guaranteed to produce an approximation to the original image. In this paper FIC is achieved by PIFS using quadtree partitioning. PIFS is applied on different images like , Ultrasound, CT Scan, Angiogram, X-ray, Mammograms. In each modality approximately twenty images are considered and the average values of compression ratio and PSNR values are arrived. In this method of fractal encoding, the parameter, tolerance factor Tmax, is varied from 1 to 10, keeping the other standard parameters constant. For all modalities of images the compression ratio and Peak Signal to Noise Ratio (PSNR) are computed and studied. The quality of the decompressed image is arrived by PSNR values. From the results it is observed that the compression ratio increases with the tolerance factor and mammogram has the highest compression ratio. The quality of the image is not degraded upto an optimum value of tolerance factor, Tmax, equal to 8, because of the properties of fractal compression.

CFD Simulation and Validation of Flap Type Wave-Maker

A general purpose viscous flow solver Ansys CFX was used to solve the unsteady three-dimensional (3D) Reynolds Averaged Navier-Stokes Equation (RANSE) for simulating a 3D numerical viscous wave tank. A flap-type wave generator was incorporated in the computational domain to generate the desired incident waves. Authors have made effort to study the physical behaviors of Flap type wave maker with governing parameters. Dependency of the water fill depth, Time period of oscillations and amplitude of oscillations of flap were studied. Effort has been made to establish relations between parameters. A validation study was also carried out against CFD methodology with wave maker theory. It has been observed that CFD results are in good agreement with theoretical results. Beaches of different slopes were introduced to damp the wave, so that it should not cause any reflection from boundary. As a conclusion this methodology can simulate the experimental wave-maker for regular wave generation for different wave length and amplitudes.

Tool for Helping Rural Woman Giving Birth

Giving birth is a natural process and most women have to go through it. Gynecologist or Midwife usually uses the leg holder to position the cervix in the stitching process. In some part of rural areas in Indonesia, the labor process normally being done at homes by calling in a midwife or gynecologist. The facilities for this kind of labor process is not yet sufficient, as the use of leg holder supposedly on the obstetric bed. The reality is that it is impossible to bring in the obstetric bed to the patient-s house at the time they call for giving birth or the time when the stitching of the cervix need to be done. This research is redesigning the leg holder through Biomechanics and ergonomic approaches to obtain the optimal design which is suitable to the user of a developing country such as Indonesia.

Efficient Dimensionality Reduction of Directional Overcurrent Relays Optimal Coordination Problem

Directional over current relays (DOCR) are commonly used in power system protection as a primary protection in distribution and sub-transmission electrical systems and as a secondary protection in transmission systems. Coordination of protective relays is necessary to obtain selective tripping. In this paper, an approach for efficiency reduction of DOCRs nonlinear optimum coordination (OC) is proposed. This was achieved by modifying the objective function and relaxing several constraints depending on the four constraints classification, non-valid, redundant, pre-obtained and valid constraints. According to this classification, the far end fault effect on the objective function and constraints, and in consequently on relay operating time, was studied. The study was carried out, firstly by taking into account the near-end and far-end faults in DOCRs coordination problem formulation; and then faults very close to the primary relays (nearend faults). The optimal coordination (OC) was achieved by simultaneously optimizing all variables (TDS and Ip) in nonlinear environment by using of Genetic algorithm nonlinear programming techniques. The results application of the above two approaches on 6-bus and 26-bus system verify that the far-end faults consideration on OC problem formulation don-t lose the optimality.

Systematic Functional Analysis Methods for Design Retrieval and Documentation

Apart from geometry, functionality is one of the most significant hallmarks of a product. The functionality of a product can be considered as the fundamental justification for a product existence. Therefore a functional analysis including a complete and reliable descriptor has a high potential to improve product development process in various fields especially in knowledge-based design. One of the important applications of the functional analysis and indexing is in retrieval and design reuse concept. More than 75% of design activity for a new product development contains reusing earlier and existing design know-how. Thus, analysis and categorization of product functions concluded by functional indexing, influences directly in design optimization. This paper elucidates and evaluates major classes for functional analysis by discussing their major methods. Moreover it is finalized by presenting a noble hybrid approach for functional analysis.

[Ca(2,2'-bipyridine)3]2+ -Montmorillonite: A Potentiometric Sensor for Sulfide ion

Sulfide ion (S2-) is one of the most important ions to be monitored due to its high toxicity, especially for aquatic organisms. In this work, [Ca(2,2'-bipyridine)3]2+-intercalated montmorillonite was prepared and used as a sensor to construct a potentiometric electrode to measure sulfide ion in solution. The formation of [Ca(2,2'- bipyridine)3]2+ in montmorillonite was confirmed by Fourier Transform Infrared spectra. The electrode worked well at pH 4-12 and 4-10 in sulfide solution 10-2 M and 10-3 M, respectively, in terms of Nernstian slope. The sensor gave good precision and low cost.

Economic Evaluations Using Genetic Algorithms to Determine the Territorial Impact Caused by High Speed Railways

The evolution of technology and construction techniques has enabled the upgrading of transport networks. In particular, the high-speed rail networks allow convoys to peak at above 300 km/h. These structures, however, often significantly impact the surrounding environment. Among the effects of greater importance are the ones provoked by the soundwave connected to train transit. The wave propagation affects the quality of life in areas surrounding the tracks, often for several hundred metres. There are substantial damages to properties (buildings and land), in terms of market depreciation. The present study, integrating expertise in acoustics, computering and evaluation fields, outlines a useful model to select project paths so as to minimize the noise impact and reduce the causes of possible litigation. It also facilitates the rational selection of initiatives to contain the environmental damage to the already existing railway tracks. The research is developed with reference to the Italian regulatory framework (usually more stringent than European and international standards) and refers to a case study concerning the high speed network in Italy.

Some Relationships between Classes of Reverse Watson-Crick Finite Automata

A Watson-Crick automaton is recently introduced as a computational model of DNA computing framework. It works on tapes consisting of double stranded sequences of symbols. Symbols placed on the corresponding cells of the double-stranded sequences are related by a complimentary relation. In this paper, we investigate a variation of Watson-Crick automata in which both heads read the tape in reverse directions. They are called reverse Watson-Crick finite automata (RWKFA). We show that all of following four classes, i.e., simple, 1-limited, all-final, all-final and simple, are equal to non-restricted version of RWKFA.

Identification, Prediction and Detection of the Process Fault in a Cement Rotary Kiln by Locally Linear Neuro-Fuzzy Technique

In this paper, we use nonlinear system identification method to predict and detect process fault of a cement rotary kiln. After selecting proper inputs and output, an input-output model is identified for the plant. To identify the various operation points in the kiln, Locally Linear Neuro-Fuzzy (LLNF) model is used. This model is trained by LOLIMOT algorithm which is an incremental treestructure algorithm. Then, by using this method, we obtained 3 distinct models for the normal and faulty situations in the kiln. One of the models is for normal condition of the kiln with 15 minutes prediction horizon. The other two models are for the two faulty situations in the kiln with 7 minutes prediction horizon are presented. At the end, we detect these faults in validation data. The data collected from White Saveh Cement Company is used for in this study.

Photovoltaic Array Sizing for PV-Electrolyzer

Hydrogen that used as fuel in fuel cell vehicles can be produced from renewable sources such as wind, solar, and hydro technologies. PV-electrolyzer is one of the promising methods to produce hydrogen with zero pollution emission. Hydrogen production from a PV-electrolyzer system depends on the efficiency of the electrolyzer and photovoltaic array, and sun irradiance at that site. In this study, the amount of hydrogen is obtained using mathematical equations for difference driving distance and sun peak hours. The results show that the minimum of 99 PV modules are used to generate 1.75 kgH2 per day for two vehicles.

Framework for Delivery Reliability in European Machinery and Equipment Industry

Today-s manufacturing companies are facing multiple and dynamic customer-supplier-relationships embedded in nonhierarchical production networks. This complex environment leads to problems with delivery reliability and wasteful turbulences throughout the entire network. This paper describes an operational model based on a theoretical framework which improves delivery reliability of each individual customer-supplier-relationship within non-hierarchical production networks of the European machinery and equipment industry. By developing a non-centralized coordination mechanism based on determining the value of delivery reliability and derivation of an incentive system for suppliers the number of in time deliveries can be increased and thus the turbulences in the production network smoothened. Comparable to an electronic stock exchange the coordination mechanism will transform the manual and nontransparent process of determining penalties for delivery delays into an automated and transparent market mechanism creating delivery reliability.

Revisiting Distributed Protocols for Mobility at the Application Layer

During more than a decade, many proposals and standards have been designed to deal with the mobility issues; however, there are still some serious limitations in basing solutions on them. In this paper we discuss the possibility of handling mobility at the application layer. We do this while revisiting the conventional implementation of the Two Phase Commit (2PC) protocol which is a fundamental asset of transactional technology for ensuring the consistent commitment of distributed transactions. The solution is based on an execution framework providing an efficient extension that is aware of the mobility and preserves the 2PC principle.

Spatial Variability of Some Soil Properties in Mountain Rangelands of Northern Iran

In this paper spatial variability of some chemical and physical soil properties were investigated in mountain rangelands of Nesho, Mazandaran province, Iran. 110 soil samples from 0-30 cm depth were taken with systematic method on grid 30×30 m2 in regions with different vegetation cover and transported to laboratory. Then soil chemical and physical parameters including Acidity (pH), Electrical conductivity, Caco3, Bulk density, Particle density, total phosphorus, total Nitrogen, available potassium, Organic matter, Saturation moisture, Soil texture (percentage of sand, silt and clay), Sodium, Calcium, magnesium were measured in laboratory. Data normalization was performed then was done statistical analysis for description of soil properties and geostatistical analysis for indication spatial correlation between these properties and were perpetrated maps of spatial distribution of soil properties using Kriging method. Results indicated that in the study area Saturation moisture and percentage of Sand had highest and lowest spatial correlation respectively.

Inconsistency Discovery in Multiple State Diagrams

In this article, we introduce a new approach for analyzing UML designs to detect the inconsistencies between multiple state diagrams and sequence diagrams. The Super State Analysis (SSA) identifies the inconsistencies in super states, single step transitions, and sequences. Because SSA considers multiple UML state diagrams, it discovers inconsistencies that cannot be discovered when considering only a single UML state diagram. We have introduced a transition set that captures relationship information that is not specifiable in UML diagrams. The SSA model uses the transition set to link transitions of multiple state diagrams together. The analysis generates three different sets automatically. These sets are compared to the provided sets to detect the inconsistencies. SSA identifies five types of inconsistencies: impossible super states, unreachable super states, illegal transitions, missing transitions, and illegal sequences.

Actionable Rules: Issues and New Directions

Knowledge Discovery in Databases (KDD) is the process of extracting previously unknown, hidden and interesting patterns from a huge amount of data stored in databases. Data mining is a stage of the KDD process that aims at selecting and applying a particular data mining algorithm to extract an interesting and useful knowledge. It is highly expected that data mining methods will find interesting patterns according to some measures, from databases. It is of vital importance to define good measures of interestingness that would allow the system to discover only the useful patterns. Measures of interestingness are divided into objective and subjective measures. Objective measures are those that depend only on the structure of a pattern and which can be quantified by using statistical methods. While, subjective measures depend only on the subjectivity and understandability of the user who examine the patterns. These subjective measures are further divided into actionable, unexpected and novel. The key issues that faces data mining community is how to make actions on the basis of discovered knowledge. For a pattern to be actionable, the user subjectivity is captured by providing his/her background knowledge about domain. Here, we consider the actionability of the discovered knowledge as a measure of interestingness and raise important issues which need to be addressed to discover actionable knowledge.