Data Envelopment Analysis under Uncertainty and Risk

Data Envelopment Analysis (DEA) is one of the most widely used technique for evaluating the relative efficiency of a set of homogeneous decision making units. Traditionally, it assumes that input and output variables are known in advance, ignoring the critical issue of data uncertainty. In this paper, we deal with the problem of efficiency evaluation under uncertain conditions by adopting the general framework of the stochastic programming. We assume that output parameters are represented by discretely distributed random variables and we propose two different models defined according to a neutral and risk-averse perspective. The models have been validated by considering a real case study concerning the evaluation of the technical efficiency of a sample of individual firms operating in the Italian leather manufacturing industry. Our findings show the validity of the proposed approach as ex-ante evaluation technique by providing the decision maker with useful insights depending on his risk aversion degree.

A Video-based Algorithm for Moving Objects Detection at Signalized Intersection

Mixed-traffic (e.g., pedestrians, bicycles, and vehicles) data at an intersection is one of the essential factors for intersection design and traffic control. However, some data such as pedestrian volume cannot be directly collected by common detectors (e.g. inductive loop, sonar and microwave sensors). In this paper, a video based detection algorithm is proposed for mixed-traffic data collection at intersections using surveillance cameras. The algorithm is derived from Gaussian Mixture Model (GMM), and uses a mergence time adjustment scheme to improve the traditional algorithm. Real-world video data were selected to test the algorithm. The results show that the proposed algorithm has the faster processing speed and more accuracy than the traditional algorithm. This indicates that the improved algorithm can be applied to detect mixed-traffic at signalized intersection, even when conflicts occur.

Use of Time-Depend Effects for Mixing and Separation of the Two-Phase Flows

The paper shows some ability to manage two-phase flows arising from the use of unsteady effects. In one case, we consider the condition of fragmentation of the interface between the two components leads to the intensification of mixing. The problem is solved when the temporal and linear scale are small for the appearance of the developed mixing layer. Showing that exist such conditions for unsteady flow velocity at the surface of the channel, which will lead to the creation and fragmentation of vortices at Re numbers of order unity. Also showing that the Re is not a criterion of similarity for this type of flows, but we can introduce a criterion that depends on both the Re, and the frequency splitting of the vortices. It turned out that feature of this situation is that streamlines behave stable, and if we analyze the behavior of the interface between the components it satisfies all the properties of unstable flows. The other problem we consider the behavior of solid impurities in the extensive system of channels. Simulated unsteady periodic flow modeled breaths. Consider the behavior of the particles along the trajectories. It is shown that, depending on the mass and diameter of the particles, they can be collected in a caustic on the channel walls, stop in a certain place or fly back. Of interest is the distribution of particle velocity in frequency. It turned out that by choosing a behavior of the velocity field of the carrier gas can affect the trajectory of individual particles including force them to fly back.

e-Service Innovation within Open Innovation Networks

Service innovations are central concerns in fast changing environment. Due to the fitness in customer demands and advances in information technologies (IT) in service management, an expanded conceptualization of e-service innovation is required. Specially, innovation practices have become increasingly more challenging, driving managers to employ a different open innovation model to maintain competitive advantages. At the same time, firms need to interact with external and internal customers in innovative environments, like the open innovation networks, to co-create values. Based on these issues, an important conceptual framework of e-service innovation is developed. This paper aims to examine the contributing factors on e-service innovation and firm performance, including financial and non-financial aspects. The study concludes by showing how e-service innovation will play a significant role in growing the overall values of the firm. The discussion and conclusion will lead to a stronger understanding of e-service innovation and co-creating values with customers within open innovation networks.

Context for Simplicity: A Basis for Context-aware Systems Based on the 3GPP Generic User Profile

The paper focuses on the area of context modeling with respect to the specification of context-aware systems supporting ubiquitous applications. The proposed approach, followed within the SIMPLICITY IST project, uses a high-level system ontology to derive context models for system components which consequently are mapped to the system's physical entities. For the definition of user and device-related context models in particular, the paper suggests a standard-based process consisting of an analysis phase using the Common Information Model (CIM) methodology followed by an implementation phase that defines 3GPP based components. The benefits of this approach are further depicted by preliminary examples of XML grammars defining profiles and components, component instances, coupled with descriptions of respective ubiquitous applications.

Estimating Shortest Circuit Path Length Complexity

When binary decision diagrams are formed from uniformly distributed Monte Carlo data for a large number of variables, the complexity of the decision diagrams exhibits a predictable relationship to the number of variables and minterms. In the present work, a neural network model has been used to analyze the pattern of shortest path length for larger number of Monte Carlo data points. The neural model shows a strong descriptive power for the ISCAS benchmark data with an RMS error of 0.102 for the shortest path length complexity. Therefore, the model can be considered as a method of predicting path length complexities; this is expected to lead to minimum time complexity of very large-scale integrated circuitries and related computer-aided design tools that use binary decision diagrams.

Simulating Pathogen Transport with in a Naturally Ventilated Hospital Ward

Understanding how airborne pathogens are transported through hospital wards is essential for determining the infection risk to patients and healthcare workers. This study utilizes Computational Fluid Dynamics (CFD) simulations to explore possible pathogen transport within a six-bed partitioned Nightingalestyle hospital ward. Grid independence of a ward model was addressed using the Grid Convergence Index (GCI) from solutions obtained using three fullystructured grids. Pathogens were simulated using source terms in conjunction with a scalar transport equation and a RANS turbulence model. Errors were found to be less than 4% in the calculation of air velocities but an average of 13% was seen in the scalar field. A parametric study of variations in the pathogen release point illustrated that its distribution is strongly influenced by the local velocity field and the degree of air mixing present.

Indonesian Store Loyalty Factors for Modern Retailing Market

Modern retailers such as hypermarket/supermarket need to be more customer-oriented in order to survive in today-s competitive business world. As a result, the investigation of determinant factors of store loyalty becomes important issue for modern retailing players. This study suggests that consumers- store loyalty in the modern retailing market (hypermarkets and supermarkets) is influenced by environmental factors (such as store image, store personnel). Using a model of stimulus-organismresponse (S-O-R), this research examines S-R relationship of store loyalty. S-O-R framework is derived from the existence literature and tested empirically based on Indonesian consumers- experience. The stimuli for this study are store image, store personnel, satisfaction and culture factors. Affect, or the consumers- liking to modern retailing stores, mediates the chosen environmental factors on consumer-s store loyalty. The findings showed that store image, store satisfaction and culture have significant positive relationship to store loyalty via affect.

The Self-Propelled Model of a Boat, Based on the Wave Thrust

We attempted investigate a boat model, based on the conversion of energy of surface wave into a sequence of unidirectional pulses of jet spurts, in other words - model of the boat, which is thrusting by the waves field on water surface. These pulses are forming some average reactive stream from the output nozzle on the stern of boat. The suggested model provides the conversion of its oscillatory motions (both pitching and rolling) into a jet flow. This becomes possible due to special construction of the boat and due to several details, sensitive to the local wave field. The boat model presents the uniflow jet engine without slow conversions of mechanical energy into intermediate forms and without any external sources of energy (besides surface waves). Motion of boat is characterized by fast jerks and average onward velocity, which exceeds the velocities of liquid particles in the wave.

Generating Qualitative Causal Graph using Modeling Constructs of Qualitative Process Theory for Explaining Organic Chemistry Reactions

This paper discusses the causal explanation capability of QRIOM, a tool aimed at supporting learning of organic chemistry reactions. The development of the tool is based on the hybrid use of Qualitative Reasoning (QR) technique and Qualitative Process Theory (QPT) ontology. Our simulation combines symbolic, qualitative description of relations with quantity analysis to generate causal graphs. The pedagogy embedded in the simulator is to both simulate and explain organic reactions. Qualitative reasoning through a causal chain will be presented to explain the overall changes made on the substrate; from initial substrate until the production of final outputs. Several uses of the QPT modeling constructs in supporting behavioral and causal explanation during run-time will also be demonstrated. Explaining organic reactions through causal graph trace can help improve the reasoning ability of learners in that their conceptual understanding of the subject is nurtured.

A Maximum Parsimony Model to Reconstruct Phylogenetic Network in Honey Bee Evolution

Phylogenies ; The evolutionary histories of groups of species are one of the most widely used tools throughout the life sciences, as well as objects of research with in systematic, evolutionary biology. In every phylogenetic analysis reconstruction produces trees. These trees represent the evolutionary histories of many groups of organisms, bacteria due to horizontal gene transfer and plants due to process of hybridization. The process of gene transfer in bacteria and hybridization in plants lead to reticulate networks, therefore, the methods of constructing trees fail in constructing reticulate networks. In this paper a model has been employed to reconstruct phylogenetic network in honey bee. This network represents reticulate evolution in honey bee. The maximum parsimony approach has been used to obtain this reticulate network.

Solving Partially Monotone Problems with Neural Networks

In many applications, it is a priori known that the target function should satisfy certain constraints imposed by, for example, economic theory or a human-decision maker. Here we consider partially monotone problems, where the target variable depends monotonically on some of the predictor variables but not all. We propose an approach to build partially monotone models based on the convolution of monotone neural networks and kernel functions. The results from simulations and a real case study on house pricing show that our approach has significantly better performance than partially monotone linear models. Furthermore, the incorporation of partial monotonicity constraints not only leads to models that are in accordance with the decision maker's expertise, but also reduces considerably the model variance in comparison to standard neural networks with weight decay.

Software Effort Estimation Using Soft Computing Techniques

Various models have been derived by studying large number of completed software projects from various organizations and applications to explore how project sizes mapped into project effort. But, still there is a need to prediction accuracy of the models. As Neuro-fuzzy based system is able to approximate the non-linear function with more precision. So, Neuro-Fuzzy system is used as a soft computing approach to generate model by formulating the relationship based on its training. In this paper, Neuro-Fuzzy technique is used for software estimation modeling of on NASA software project data and performance of the developed models are compared with the Halstead, Walston-Felix, Bailey-Basili and Doty Models mentioned in the literature.

Modelling Indoor Air Carbon Dioxide (CO2)Concentration using Neural Network

The use of neural networks is popular in various building applications such as prediction of heating load, ventilation rate and indoor temperature. Significant is, that only few papers deal with indoor carbon dioxide (CO2) prediction which is a very good indicator of indoor air quality (IAQ). In this study, a data-driven modelling method based on multilayer perceptron network for indoor air carbon dioxide in an apartment building is developed. Temperature and humidity measurements are used as input variables to the network. Motivation for this study derives from the following issues. First, measuring carbon dioxide is expensive and sensors power consumptions is high and secondly, this leads to short operating times of battery-powered sensors. The results show that predicting CO2 concentration based on relative humidity and temperature measurements, is difficult. Therefore, more additional information is needed.

A Self-Consistent Scheme for Elastic-Plastic Asperity Contact

In this paper, a generalized self-consistent scheme, or “three phase model", is used to set up a micro-mechanics model for rough surface contact with randomly distributed asperities. The dimensionless average real pressure p is obtained as function of the ratio of the real contact area to the apparent contact area, 0 A / A r . Both elastic and plastic materials are considered, and the influence of the plasticity of material on p is discussed. Both two-dimensional and three-dimensional rough surface contact problems are considered.

Compressive Properties of a Synthetic Bone Substitute for Vertebral Cancellous Bone

Transpedicular screw fixation in spinal fractures, degenerative changes, or deformities is a well-established procedure. However, important rate of fixation failure due to screw bending, loosening, or pullout are still reported particularly in weak bone stock in osteoporosis. To overcome the problem, mechanism of failure has to be fully investigated in vitro. Post-mortem human subjects are less accessible and animal cadavers comprise limitations due to different geometry and mechanical properties. Therefore, the development of a synthetic model mimicking the realistic human vertebra is highly demanded. A bone surrogate, composed of Polyurethane (PU) foam analogous to cancellous bone porous structure, was tested for 3 different densities in this study. The mechanical properties were investigated under uniaxial compression test by minimizing the end artifacts on specimens. The results indicated that PU foam of 0.32 g.cm-3 density has comparable mechanical properties to human cancellous bone in terms of young-s modulus and yield strength. Therefore, the obtained information can be considered as primary step for developing a realistic cancellous bone of human vertebral body. Further evaluations are also recommended for other density groups.

Application of Vortex Tubes for Extracting Sediments Using SHARC Software - A Case Study of the Western Canal in the Dez Diversion Weir

Sediment loads transfer in hydraulic installations and their consequences for the O&M of modern canal systems is emerging as one of the most important considerations in hydraulic engineering projects apriticularly those which are inteded to feed the irrigation and draiange schemes of large command areas such as the Dez and Mogahn in Iran.. The aim of this paper is to investigate the applicability of the vortex tube as a viable means of extracting sediment loads entering the canal systems in general and the water inatke structures in particulars. The Western conveyance canal of the Dez Diversion weir which feeds the Karkheh Flood Plain in Sothwestern Dezful has been used as the case study using the data from the Dastmashan Hydrometric Station. The SHARC software has been used as an analytical framework to interprete the data. Results show that given the grain size D50 and the canal turbulence the adaption length from the beginning of the canal and after the diversion dam is estimated at 477 m, a point which is suitable for laying the vortex tube.

Face Recognition Using Double Dimension Reduction

In this paper a new approach to face recognition is presented that achieves double dimension reduction making the system computationally efficient with better recognition results. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results improve with increase in face image resolution and levels off when arriving at a certain resolution level. In the proposed model of face recognition, first image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to better computational speed and feature extraction potential of Discrete Cosine Transform (DCT) it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A trade of between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL database, Yale database and a color database. The proposed technique has performed much better compared to other techniques. The significance of the model is two fold: (1) dimension reduction up to an effective and suitable face image resolution (2) appropriate DCT coefficients are retained to achieve best recognition results with varying image poses, intensity and illumination level.

Electrical and Magnetic Modelling of a Power Transformer: A Bond Graph Approach

Bond graph models of an electrical transformer including the nonlinear saturation are presented. The transformer using electrical and magnetic circuits are modelled. These models determine the relation between self and mutual inductances, and the leakage and magnetizing inductances of power transformers with two windings using the properties of a bond graph. The equivalence between electrical and magnetic variables is given. The modelling and analysis using this methodology to three phase power transformers can be extended.