Integrating Life Cycle Uncertainties for Evaluating a Building Overall Cost

Overall cost is a significant consideration in any decision-making process. Although many studies were carried out on overall cost in construction, little has treated the uncertainties of real life cycle development. On the basis of several case studies, a feedback process was performed on the historical data of studied buildings. This process enabled to identify some factors causing uncertainty during the operational period. As a result, the research proposes a new method for assessing the overall cost during a part of the building-s life cycle taking account of the building actual value, its end-of-life value and the influence of the identified life cycle uncertainty factors. The findings are a step towards a higher level of reliability in overall cost evaluation taking account of some usually unexpected uncertainty factors.

Research on Self-Perceptions of Pre-Service Turkish Language Teachers in Turkey with Regard to Problem Solving Skills

The aim of this research is to determine how preservice Turkish teachers perceive themselves in terms of problem solving skills. Students attending Department of Turkish Language Teaching of Gazi University Education Faculty in 2005-2006 academic year constitute the study group (n= 270) of this research in which survey model was utilized. Data were obtained by Problem Solving Inventory developed by Heppner & Peterson and Personal Information Form. Within the settings of this research, Cronbach Alpha reliability coefficient of the scale was found as .87. Besides, reliability coefficient obtained by split-half technique which splits odd and even numbered items of the scale was found as r=.81 (Split- Half Reliability). The findings of the research revealed that preservice Turkish teachers were sufficiently qualified on the subject of problem solving skills and statistical significance was found in favor of male candidates in terms of “gender" variable. According to the “grade" variable, statistical significance was found in favor of 4th graders.

Modeling Strategy and Numerical Validation of the Turbulent Flow over a two-Dimensional Flat Roof

The construction of a civil structure inside a urban area inevitably modifies the outdoor microclimate at the building site. Wind speed, wind direction, air pollution, driving rain, radiation and daylight are some of the main physical aspects that are subjected to the major changes. The quantitative amount of these modifications depends on the shape, size and orientation of the building and on its interaction with the surrounding environment.The flow field over a flat roof model building has been numerically investigated in order to determine two-dimensional CFD guidelines for the calculation of the turbulent flow over a structure immersed in an atmospheric boundary layer. To this purpose, a complete validation campaign has been performed through a systematic comparison of numerical simulations with wind tunnel experimental data.Several turbulence models and spatial node distributions have been tested for five different vertical positions, respectively from the upstream leading edge to the downstream bottom edge of the analyzed model. Flow field characteristics in the neighborhood of the building model have been numerically investigated, allowing a quantification of the capabilities of the CFD code to predict the flow separation and the extension of the recirculation regions.The proposed calculations have allowed the development of a preliminary procedure to be used as a guidance in selecting the appropriate grid configuration and corresponding turbulence model for the prediction of the flow field over a twodimensional roof architecture dominated by flow separation.

Study on Various Measures for Flood in Specific Region: A Case Study of the 2008 Lao Flood

In recent years, the number of natural disasters in Laos has a trend to increase, especially the disaster of flood. To make a flood plan risk management in the future, it is necessary to understand and analyze the characteristics of the rainfall and Mekong River level data. To reduce the damage, this paper presents the flood risk analysis in Luangprabang and Vientiane, the prefecture of Laos. In detail, the relationship between the rainfall and the Mekong River level has evaluated and appropriate countermeasure for flood was discussed.

Intrusion Detection based on Distance Combination

The intrusion detection problem has been frequently studied, but intrusion detection methods are often based on a single point of view, which always limits the results. In this paper, we introduce a new intrusion detection model based on the combination of different current methods. First we use a notion of distance to unify the different methods. Second we combine these methods using the Pearson correlation coefficients, which measure the relationship between two methods, and we obtain a combined distance. If the combined distance is greater than a predetermined threshold, an intrusion is detected. We have implemented and tested the combination model with two different public data sets: the data set of masquerade detection collected by Schonlau & al., and the data set of program behaviors from the University of New Mexico. The results of the experiments prove that the combination model has better performances.

Evaluation of Risk Attributes Driven by Periodically Changing System Functionality

Modeling of the distributed systems allows us to represent the whole its functionality. The working system instance rarely fulfils the whole functionality represented by model; usually some parts of this functionality should be accessible periodically. The reporting system based on the Data Warehouse concept seams to be an intuitive example of the system that some of its functionality is required only from time to time. Analyzing an enterprise risk associated with the periodical change of the system functionality, we should consider not only the inaccessibility of the components (object) but also their functions (methods), and the impact of such a situation on the system functionality from the business point of view. In the paper we suggest that the risk attributes should be estimated from risk attributes specified at the requirements level (Use Case in the UML model) on the base of the information about the structure of the model (presented at other levels of the UML model). We argue that it is desirable to consider the influence of periodical changes in requirements on the enterprise risk estimation. Finally, the proposition of such a solution basing on the UML system model is presented.

Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data

Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.

Optimizing of Fuzzy C-Means Clustering Algorithm Using GA

Fuzzy C-means Clustering algorithm (FCM) is a method that is frequently used in pattern recognition. It has the advantage of giving good modeling results in many cases, although, it is not capable of specifying the number of clusters by itself. In FCM algorithm most researchers fix weighting exponent (m) to a conventional value of 2 which might not be the appropriate for all applications. Consequently, the main objective of this paper is to use the subtractive clustering algorithm to provide the optimal number of clusters needed by FCM algorithm by optimizing the parameters of the subtractive clustering algorithm by an iterative search approach and then to find an optimal weighting exponent (m) for the FCM algorithm. In order to get an optimal number of clusters, the iterative search approach is used to find the optimal single-output Sugenotype Fuzzy Inference System (FIS) model by optimizing the parameters of the subtractive clustering algorithm that give minimum least square error between the actual data and the Sugeno fuzzy model. Once the number of clusters is optimized, then two approaches are proposed to optimize the weighting exponent (m) in the FCM algorithm, namely, the iterative search approach and the genetic algorithms. The above mentioned approach is tested on the generated data from the original function and optimal fuzzy models are obtained with minimum error between the real data and the obtained fuzzy models.

DIVAD: A Dynamic and Interactive Visual Analytical Dashboard for Exploring and Analyzing Transport Data

The advances in location-based data collection technologies such as GPS, RFID etc. and the rapid reduction of their costs provide us with a huge and continuously increasing amount of data about movement of vehicles, people and goods in an urban area. This explosive growth of geospatially-referenced data has far outpaced the planner-s ability to utilize and transform the data into insightful information thus creating an adverse impact on the return on the investment made to collect and manage this data. Addressing this pressing need, we designed and developed DIVAD, a dynamic and interactive visual analytics dashboard to allow city planners to explore and analyze city-s transportation data to gain valuable insights about city-s traffic flow and transportation requirements. We demonstrate the potential of DIVAD through the use of interactive choropleth and hexagon binning maps to explore and analyze large taxi-transportation data of Singapore for different geographic and time zones.

A New Biologically Inspired Pattern Recognition Spproach for Face Recognition

This paper reports a new pattern recognition approach for face recognition. The biological model of light receptors - cones and rods in human eyes and the way they are associated with pattern vision in human vision forms the basis of this approach. The functional model is simulated using CWD and WPD. The paper also discusses the experiments performed for face recognition using the features extracted from images in the AT & T face database. Artificial Neural Network and k- Nearest Neighbour classifier algorithms are employed for the recognition purpose. A feature vector is formed for each of the face images in the database and recognition accuracies are computed and compared using the classifiers. Simulation results show that the proposed method outperforms traditional way of feature extraction methods prevailing for pattern recognition in terms of recognition accuracy for face images with pose and illumination variations.

Ground Heat Exchanger Modeling Developed for Energy Flows of an Incompressible Fluid

Ground-source heat pumps achieve higher efficiencies than conventional air-source heat pumps because they exchange heat with the ground that is cooler in summer and hotter in winter than the air environment. Earth heat exchangers are essential parts of the ground-source heat pumps and the accurate prediction of their performance is of fundamental importance. This paper presents the development and validation of a numerical model through an incompressible fluid flow, for the simulation of energy and temperature changes in and around a U-tube borehole heat exchanger. The FlexPDE software is used to solve the resulting simultaneous equations that model the heat exchanger. The validated model (through a comparison with experimental data) is then used to extract conclusions on how various parameters like the U-tube diameter, the variation of the ground thermal conductivity and specific heat and the borehole filling material affect the temperature of the fluid.

A Simple Deterministic Model for the Spread of Leptospirosis in Thailand

In this work, we consider a deterministic model for the transmission of leptospirosis which is currently spreading in the Thai population. The SIR model which incorporates the features of this disease is applied to the epidemiological data in Thailand. It is seen that the numerical solutions of the SIR equations are in good agreement with real empirical data. Further improvements are discussed.

Predictive Clustering Hybrid Regression(pCHR) Approach and Its Application to Sucrose-Based Biohydrogen Production

A predictive clustering hybrid regression (pCHR) approach was developed and evaluated using dataset from H2- producing sucrose-based bioreactor operated for 15 months. The aim was to model and predict the H2-production rate using information available about envirome and metabolome of the bioprocess. Selforganizing maps (SOM) and Sammon map were used to visualize the dataset and to identify main metabolic patterns and clusters in bioprocess data. Three metabolic clusters: acetate coupled with other metabolites, butyrate only, and transition phases were detected. The developed pCHR model combines principles of k-means clustering, kNN classification and regression techniques. The model performed well in modeling and predicting the H2-production rate with mean square error values of 0.0014 and 0.0032, respectively.

Mining Network Data for Intrusion Detection through Naïve Bayesian with Clustering

Network security attacks are the violation of information security policy that received much attention to the computational intelligence society in the last decades. Data mining has become a very useful technique for detecting network intrusions by extracting useful knowledge from large number of network data or logs. Naïve Bayesian classifier is one of the most popular data mining algorithm for classification, which provides an optimal way to predict the class of an unknown example. It has been tested that one set of probability derived from data is not good enough to have good classification rate. In this paper, we proposed a new learning algorithm for mining network logs to detect network intrusions through naïve Bayesian classifier, which first clusters the network logs into several groups based on similarity of logs, and then calculates the prior and conditional probabilities for each group of logs. For classifying a new log, the algorithm checks in which cluster the log belongs and then use that cluster-s probability set to classify the new log. We tested the performance of our proposed algorithm by employing KDD99 benchmark network intrusion detection dataset, and the experimental results proved that it improves detection rates as well as reduces false positives for different types of network intrusions.

A Study on Flammability of Bio Oil Combustible Vapour Mixtures

Study of fire and explosion is very important mainly in oil and gas industries due to several accidents which have been reported in the past and present. In this work, we have investigated the flammability of bio oil vapour mixtures. This mixture may contribute to fire during the storage and transportation process. Bio oil sample derived from Palm Kernell shell was analysed using Gas Chromatography Mass Spectrometry (GC-MS) to examine the composition of the sample. Mole fractions of 12 selected components in the liquid phase were obtained from the GC-FID data and used to calculate mole fractions of components in the gas phase via modified Raoult-s law. Lower Flammability Limits (LFLs) and Upper Flammability Limits (UFLs) for individual components were obtained from published literature. However, stoichiometric concentration method was used to calculate the flammability limits of some components which their flammability limit values are not available in the literature. The LFL and UFL values for the mixture were calculated using the Le Chatelier equation. The LFLmix and UFLmix values were used to construct a flammability diagram and subsequently used to determine the flammability of the mixture. The findings of this study can be used to propose suitable inherently safer method to prevent the flammable mixture from occurring and to minimizing the loss of properties, business, and life due to fire accidents in bio oil productions.

Classifying Biomedical Text Abstracts based on Hierarchical 'Concept' Structure

Classifying biomedical literature is a difficult and challenging task, especially when a large number of biomedical articles should be organized into a hierarchical structure. In this paper, we present an approach for classifying a collection of biomedical text abstracts downloaded from Medline database with the help of ontology alignment. To accomplish our goal, we construct two types of hierarchies, the OHSUMED disease hierarchy and the Medline abstract disease hierarchies from the OHSUMED dataset and the Medline abstracts, respectively. Then, we enrich the OHSUMED disease hierarchy before adapting it to ontology alignment process for finding probable concepts or categories. Subsequently, we compute the cosine similarity between the vector in probable concepts (in the “enriched" OHSUMED disease hierarchy) and the vector in Medline abstract disease hierarchies. Finally, we assign category to the new Medline abstracts based on the similarity score. The results obtained from the experiments show the performance of our proposed approach for hierarchical classification is slightly better than the performance of the multi-class flat classification.

Thermodynamic Modeling of the High Temperature Shift Converter Reactor Using Minimization of Gibbs Free Energy

The equilibrium chemical reactions taken place in a converter reactor of the Khorasan Petrochemical Ammonia plant was studied using the minimization of Gibbs free energy method. In the minimization of the Gibbs free energy function the Davidon– Fletcher–Powell (DFP) optimization procedure using the penalty terms in the well-defined objective function was used. It should be noted that in the DFP procedure along with the corresponding penalty terms the Hessian matrices for the composition of constituents in the Converter reactor can be excluded. This, in fact, can be considered as the main advantage of the DFP optimization procedure. Also the effect of temperature and pressure on the equilibrium composition of the constituents was investigated. The results obtained in this work were compared with the data collected from the converter reactor of the Khorasan Petrochemical Ammonia plant. It was concluded that the results obtained from the method used in this work are in good agreement with the industrial data. Notably, the algorithm developed in this work, in spite of its simplicity, takes the advantage of short computation and convergence time.

An Efficient Data Mining Approach on Compressed Transactions

In an era of knowledge explosion, the growth of data increases rapidly day by day. Since data storage is a limited resource, how to reduce the data space in the process becomes a challenge issue. Data compression provides a good solution which can lower the required space. Data mining has many useful applications in recent years because it can help users discover interesting knowledge in large databases. However, existing compression algorithms are not appropriate for data mining. In [1, 2], two different approaches were proposed to compress databases and then perform the data mining process. However, they all lack the ability to decompress the data to their original state and improve the data mining performance. In this research a new approach called Mining Merged Transactions with the Quantification Table (M2TQT) was proposed to solve these problems. M2TQT uses the relationship of transactions to merge related transactions and builds a quantification table to prune the candidate itemsets which are impossible to become frequent in order to improve the performance of mining association rules. The experiments show that M2TQT performs better than existing approaches.

A Bionic Approach to Dynamic, Multimodal Scene Perception and Interpretation in Buildings

Today, building automation is advancing from simple monitoring and control tasks of lightning and heating towards more and more complex applications that require a dynamic perception and interpretation of different scenes occurring in a building. Current approaches cannot handle these newly upcoming demands. In this article, a bionically inspired approach for multimodal, dynamic scene perception and interpretation is presented, which is based on neuroscientific and neuro-psychological research findings about the perceptual system of the human brain. This approach bases on data from diverse sensory modalities being processed in a so-called neuro-symbolic network. With its parallel structure and with its basic elements being information processing and storing units at the same time, a very efficient method for scene perception is provided overcoming the problems and bottlenecks of classical dynamic scene interpretation systems.

Segmenting Ultrasound B-Mode Images Using RiIG Distributions and Stochastic Optimization

In this paper, we propose a novel algorithm for delineating the endocardial wall from a human heart ultrasound scan. We assume that the gray levels in the ultrasound images are independent and identically distributed random variables with different Rician Inverse Gaussian (RiIG) distributions. Both synthetic and real clinical data will be used for testing the algorithm. Algorithm performance will be evaluated using the expert radiologist evaluation of a soft copy of an ultrasound scan during the scanning process and secondly, doctor’s conclusion after going through a printed copy of the same scan. Successful implementation of this algorithm should make it possible to differentiate normal from abnormal soft tissue and help disease identification, what stage the disease is in and how best to treat the patient. We hope that an automated system that uses this algorithm will be useful in public hospitals especially in Third World countries where problems such as shortage of skilled radiologists and shortage of ultrasound machines are common. These public hospitals are usually the first and last stop for most patients in these countries.