Analyzing Methods of the Relation between Concepts based on a Concept Hierarchy

Data objects are usually organized hierarchically, and the relations between them are analyzed based on a corresponding concept hierarchy. The relation between data objects, for example how similar they are, are usually analyzed based on the conceptual distance in the hierarchy. If a node is an ancestor of another node, it is enough to analyze how close they are by calculating the distance vertically. However, if there is not such relation between two nodes, the vertical distance cannot express their relation explicitly. This paper tries to fill this gap by improving the analysis method for data objects based on hierarchy. The contributions of this paper include: (1) proposing an improved method to evaluate the vertical distance between concepts; (2) defining the concept horizontal distance and a method to calculate the horizontal distance; and (3) discussing the methods to confine a range by the horizontal distance and the vertical distance, and evaluating the relation between concepts.

Targeting the Pulmonary Delivery via Optimizing Physicochemical Characteristics of Instilled Liquid and Exploring Distribution of Produced Liquids by Bench-Top Models and Scintigraphy of Rabbits- Lungs

We aimed to investigate how can target and optimize pulmonary delivery distribution by changing physicochemical characteristics of instilled liquid.Therefore, we created a new liquids group: a. eligible for desired distribution within lung because of assorted physicochemical characteristics b. capable of being augmented with a broad range of chemicals inertly c. no interference on respiratory function d. compatible with airway surface liquid We developed forty types of new liquid,were composed of Carboxymethylcellulose sodium,Glycerin and different types of Polysorbates.Viscosity was measured using a Programmable Rheometer and surface tension by KRUSS Tensiometer.We subsequently examined the liquids and delivery protocols by simple and branched glass capillary tube models of airways.Eventually,we explored pulmonary distribution of liquids being augmented with technetium-99m in mechanically ventilated rabbits.We used a single head large field of view gamma camera.Kinematic viscosity between 0.265Stokes and 0.289Stokes,density between 1g/cm3 and 1.5g/cm3 and surface tension between 25dyn/cm and 35dyn/cm were the most acceptable.

High Impedance Faults Detection Technique Based on Wavelet Transform

The purpose of this paper is to solve the problem of protecting aerial lines from high impedance faults (HIFs) in distribution systems. This investigation successfully applies 3I0 zero sequence current to solve HIF problems. The feature extraction system based on discrete wavelet transform (DWT) and the feature identification technique found on statistical confidence are then applied to discriminate effectively between the HIFs and the switch operations. Based on continuous wavelet transform (CWT) pattern recognition of HIFs is proposed, also. Staged fault testing results demonstrate that the proposed wavelet based algorithm is feasible performance well.

Atomic Force Microscopy (AFM)Topographical Surface Characterization of Multilayer-Coated and Uncoated Carbide Inserts

In recent years, scanning probe atomic force microscopy SPM AFM has gained acceptance over a wide spectrum of research and science applications. Most fields focuses on physical, chemical, biological while less attention is devoted to manufacturing and machining aspects. The purpose of the current study is to assess the possible implementation of the SPM AFM features and its NanoScope software in general machining applications with special attention to the tribological aspects of cutting tool. The surface morphology of coated and uncoated as-received carbide inserts is examined, analyzed, and characterized through the determination of the appropriate scanning setting, the suitable data type imaging techniques and the most representative data analysis parameters using the MultiMode SPM AFM in contact mode. The NanoScope operating software is used to capture realtime three data types images: “Height", “Deflection" and “Friction". Three scan sizes are independently performed: 2, 6, and 12 μm with a 2.5 μm vertical range (Z). Offline mode analysis includes the determination of three functional topographical parameters: surface “Roughness", power spectral density “PSD" and “Section". The 12 μm scan size in association with “Height" imaging is found efficient to capture every tiny features and tribological aspects of the examined surface. Also, “Friction" analysis is found to produce a comprehensive explanation about the lateral characteristics of the scanned surface. Configuration of many surface defects and drawbacks has been precisely detected and analyzed.

Behavioral Analysis of Team Members in Virtual Organization based on Trust Dimension and Learning

Trust management and Reputation models are becoming integral part of Internet based applications such as CSCW, E-commerce and Grid Computing. Also the trust dimension is a significant social structure and key to social relations within a collaborative community. Collaborative Decision Making (CDM) is a difficult task in the context of distributed environment (information across different geographical locations) and multidisciplinary decisions are involved such as Virtual Organization (VO). To aid team decision making in VO, Decision Support System and social network analysis approaches are integrated. In such situations social learning helps an organization in terms of relationship, team formation, partner selection etc. In this paper we focus on trust learning. Trust learning is an important activity in terms of information exchange, negotiation, collaboration and trust assessment for cooperation among virtual team members. In this paper we have proposed a reinforcement learning which enhances the trust decision making capability of interacting agents during collaboration in problem solving activity. Trust computational model with learning that we present is adapted for best alternate selection of new project in the organization. We verify our model in a multi-agent simulation where the agents in the community learn to identify trustworthy members, inconsistent behavior and conflicting behavior of agents.

A Multilingual Virtual Simulated Patient Framework for Training Primary Health Care Students

This paper describes the Multilingual Virtual Simulated Patient framework. It has been created to train the social skills and testing the knowledge of primary health care medical students. The framework generates conversational agents which perform in serveral languages as virtual simulated patients that help to improve the communication and diagnosis skills of the students complementing their training process.

Designing Ontology-Based Knowledge Integration for Preprocessing of Medical Data in Enhancing a Machine Learning System for Coding Assignment of a Multi-Label Medical Text

This paper discusses the designing of knowledge integration of clinical information extracted from distributed medical ontologies in order to ameliorate a machine learning-based multilabel coding assignment system. The proposed approach is implemented using a decision tree technique of the machine learning on the university hospital data for patients with Coronary Heart Disease (CHD). The preliminary results obtained show a satisfactory finding that the use of medical ontologies improves the overall system performance.

The Contribution of Growth Rate to the Pathogenicity of Candida spp.

Fungal infections are becoming more common and the range of susceptible individuals has expanded. While Candida albicans remains the most common infective species, other Candida spp. are becoming increasingly significant. In a range of large-scale studies of candidaemia between 1999 and 2006, about 52% of 9717 cases involved C. albicans, about 30% involved either C. glabrata or C. parapsilosis and less than 15% involved C. tropicalis, C. krusei or C. guilliermondii. However, the probability of mortality within 30 days of infection with a particular species was at least 40% for C. tropicalis, C. albicans, C. glabrata and C. krusei and only 22% for C. parapsilopsis. Clinical isolates of Candida spp. grew at rates ranging from 1.65 h-1 to 4.9 h-1. Three species (C. krusei, C. albicans and C. glabrata) had relatively high growth rates (μm > 4 h-1), C. tropicalis and C. dubliniensis grew moderately quickly (Ôëê 3 h-1) and C. parapsilosis and C. guilliermondii grew slowly (< 2 h-1). Based on these data, the log of the odds of mortality within 30 days of diagnosis was linearly related to μm. From this the underlying probability of mortality is 0.13 (95% CI: 0.10-0.17) and it increases by about 0.09 ± 0.02 for each unit increase in μm. Given that the overall crude mortality is about 0.36, the growth of Candida spp. approximately doubles the rate, consistent with the results of larger case-matched studies of candidaemia.

Orders Preparation and Control on the Productive Process Efficiency Preparation

The main objective of this paper is to analyse the influence of preparation and control of orders on performance. The focused activities explored in this research are: procurement, production and distribution. These changes in performance were obtained through improvement of the supply chain. It is proved using all the company activities that it is possible to increase de efficiency and do services in an adequate way, placing the products in the market efficiently. For that, it was explored the importance of the supply chain, with privilege to the practical environment and the quantification of the obtained results.

The Role of Contextual Ontologies in Enterprise Modeling

Information sharing and exchange, rather than information processing, is what characterizes information technology in the 21st century. Ontologies, as shared common understanding, gain increasing attention, as they appear as the most promising solution to enable information sharing both at a semantic level and in a machine-processable way. Domain Ontology-based modeling has been exploited to provide shareability and information exchange among diversified, heterogeneous applications of enterprises. Contextual ontologies are “an explicit specification of contextual conceptualization". That is: ontology is characterized by concepts that have multiple representations and they may exist in several contexts. Hence, contextual ontologies are a set of concepts and relationships, which are seen from different perspectives. Contextualization is to allow for ontologies to be partitioned according to their contexts. The need for contextual ontologies in enterprise modeling has become crucial due to the nature of today's competitive market. Information resources in enterprise is distributed and diversified and is in need to be shared and communicated locally through the intranet and globally though the internet. This paper discusses the roles that ontologies play in an enterprise modeling, and how ontologies assist in building a conceptual model in order to provide communicative and interoperable information systems. The issue of enterprise modeling based on contextual domain ontology is also investigated, and a framework is proposed for an enterprise model that consists of various applications.

Microstructure Changes of Machined Surfaceson Austenitic 304 Stainless Steel

This paper presents a experiment to estimate the influences of cutting conditions in microstructure changes of machining austenitic 304 stainless steel, especially for wear insert. The wear insert were prefabricated with a width of 0.5 mm. And the forces, temperature distribution, RS, and microstructure changes were measured by force dynamometer, infrared thermal camera, X-ray diffraction, XRD, SEM, respectively. The results told that the different combinations of machining condition have a significant influence on machined surface microstructure changes. In addition to that, the ANOVA and AOMwere used to tell the different influences of cutting speed, feed rate, and wear insert.

Momentum Accounting in Public Management: A Case Study in a Brazilian Navy-s Services Provider Military Organization

This study examines the possibility to apply the theory of multidimensional accounting (momentum accounting) in a Brazilian Navy-s Services Provider Military Organization (Organização Militar Prestadora de Serviços - OMPS). In general, the core of the said theory is the fact that Accounting does not recognize the inertia of transactions occurring in an entity, and that occur repeatedly in some cases, regardless of the implementation of new actions by its managers. The study evaluates the possibility of greater use of information recorded in the financial statements of the unit of analysis, within the strategic decisions of the organization. As a research strategy, we adopted the case study. The results infer that it is possible to use the theory in the context of a multidimensional OMPS, promoting useful information for decision-making and thereby contributing to the strengthening of the necessary alignment of its administration with the current desires of the Brazilian society.

A Virtual Learning Environment for Deaf Children: Design and Evaluation

The object of this research is the design and evaluation of an immersive Virtual Learning Environment (VLE) for deaf children. Recently we have developed a prototype immersive VR game to teach sign language mathematics to deaf students age K- 4 [1] [2]. In this paper we describe a significant extension of the prototype application. The extension includes: (1) user-centered design and implementation of two additional interactive environments (a clock store and a bakery), and (2) user-centered evaluation including development of user tasks, expert panel-based evaluation, and formative evaluation. This paper is one of the few to focus on the importance of user-centered, iterative design in VR application development, and to describe a structured evaluation method.

Distribution Centers Reliability Cost in Capacitated Facility Location Problem

Recently studies in area of supply chain network (SCN) have focused on the disruption issues in distribution systems. Also this paper extends the previous literature by providing a new biobjective model for cost minimization of designing a three echelon SCN across normal and failure scenarios with considering multi capacity option for manufacturers and distribution centers. Moreover, in order to solve the problem by means of LINGO software, novel model will be reformulated through a branch of LP-Metric method called Min-Max approach.

An Optimal Feature Subset Selection for Leaf Analysis

This paper describes an optimal approach for feature subset selection to classify the leaves based on Genetic Algorithm (GA) and Kernel Based Principle Component Analysis (KPCA). Due to high complexity in the selection of the optimal features, the classification has become a critical task to analyse the leaf image data. Initially the shape, texture and colour features are extracted from the leaf images. These extracted features are optimized through the separate functioning of GA and KPCA. This approach performs an intersection operation over the subsets obtained from the optimization process. Finally, the most common matching subset is forwarded to train the Support Vector Machine (SVM). Our experimental results successfully prove that the application of GA and KPCA for feature subset selection using SVM as a classifier is computationally effective and improves the accuracy of the classifier.

A Fault Tolerant Token-based Algorithm for Group Mutual Exclusion in Distributed Systems

The group mutual exclusion (GME) problem is a variant of the mutual exclusion problem. In the present paper a token-based group mutual exclusion algorithm, capable of handling transient faults, is proposed. The algorithm uses the concept of dynamic request sets. A time out mechanism is used to detect the token loss; also, a distributed scheme is used to regenerate the token. The worst case message complexity of the algorithm is n+1. The maximum concurrency and forum switch complexity of the algorithm are n and min (n, m) respectively, where n is the number of processes and m is the number of groups. The algorithm also satisfies another desirable property called smooth admission. The scheme can also be adapted to handle the extended group mutual exclusion problem.

The Effect of IT Service Quality Attributes on Supply Chain Management and Performance

Nowadays, where most of the leading economies are service oriented and e-business is being widely used for their management, supply chain management has become one of the most studied and practiced fields. Quality has an important role on today-s business processes, so it is important to understand the impact of IT service quality on the performance of supply chains. This paper will start by analyzing the Supply Chain Operations Reference (SCOR) model and each of its five activities: Plan, Source, Make, Delivery, and Return. This article proposes a framework for analyzing Effect of IT Service Quality on Supply Chain Performance. Using the proposed framework, hypotheses are framed for the direct effect of IT service quality on Supply Chain Performance and its indirect effect through effective Supply Chain Management. The framework will be validated empirically based on the surveys of executives of various organizations and statistical analyses of the data collected.

Practical Guidelines and Examples for the Users of the TMS320C6713 DSK

This paper describes how the correct endian mode of the TMS320C6713 DSK board can be identified. It also explains how the TMS320C6713 DSK board can be used in the little endian and in the big endian modes for assembly language programming in particular and for signal processing in general. Similarly, it discusses how crucially important it is for a user of the TMS320C6713 DSK board to identify the mode of operation and then use it correctly during the development stages of the assembly language programming; otherwise, it will cause unnecessary confusion and erroneous results as far as storing data into the memory and loading data from the memory is concerned. Furthermore, it highlights and strongly recommends to the users of the TMS320C6713 DSK board to be aware of the availability and importance of various display options in the Code Composer Studio (CCS) for correctly interpreting and displaying the desired data in the memory. The information presented in this paper will be of great importance and interest to those practitioners and developers who wants to use the TMS320C6713 DSK board for assembly language programming as well as input-output signal processing manipulations. Finally, examples that clearly illustrate the concept are presented.

A Matlab / Simulink Based Tool for Power Electronic Circuits

Transient simulation of power electronic circuits is of considerable interest to the designer. The switching nature of the devices used permits development of specialized algorithms which allow a considerable reduction in simulation time compared to general purpose simulation algorithms. This paper describes a method used to simulate a power electronic circuits using the SIMULINK toolbox within MATLAB software. Theoretical results are presented provides the basis of transient analysis of a power electronic circuits.

Electromagnetic Imaging of Inhomogeneous Dielectric Cylinders Buried in a Slab Mediumby TE Wave Illumination

The electromagnetic imaging of inhomogeneous dielectric cylinders buried in a slab medium by transverse electric (TE) wave illumination is investigated. Dielectric cylinders of unknown permittivities are buried in second space and scattered a group of unrelated waves incident from first space where the scattered field is recorded. By proper arrangement of the various unrelated incident fields, the difficulties of ill-posedness and nonlinearity are circumvented, and the permittivity distribution can be reconstructed through simple matrix operations. The algorithm is based on the moment method and the unrelated illumination method. Numerical results are given to demonstrate the capability of the inverse algorithm. Good reconstruction is obtained even in the presence of additive Gaussian random noise in measured data. In addition, the effect of noise on the reconstruction result is also investigated.