Synthesis and Foam Power of New Biodegradable Surfactant

This work deals with the synthesis and the determination of some surface properties of a new anionic surfactant belonging to sulfonamide derivatives. The interest in this new surfactant is that its behavior in aqueous solution is interesting both from a fundamental and a practice point of view. Indeed, it is well known that this kind of surfactant leads to the formation of bilayer structures, and the microstructures obtained have applications in various fields, ranging from cosmetics to detergents, to biological systems such as cell membranes and bioreactors. The surfactant synthesized from pure n-alkane by photosulfochlorination and derivatized using N-ethanol amine is a mixture of position isomers. These compounds have been analyzed by Gas Chromatography coupled to Mass Spectrometry by Electron Impact mode (GC -MS/IE), and IR. The surface tension measurements were carried out, leading to the determination of the critical micelle concentration (CMC), surface excess and the area occupied per molecule at the interface. The foaming power has also been determined by Bartsch method, and the results have been compared to those of commercial surfactants. The stability of the foam formed has also been evaluated. These compounds show good foaming power characterized in most cases by dry foam.

Automatic Tuning for a Systemic Model of Banking Originated Losses (SYMBOL) Tool on Multicore

Nowadays, the mathematical/statistical applications are developed with more complexity and accuracy. However, these precisions and complexities have brought as result that applications need more computational power in order to be executed faster. In this sense, the multicore environments are playing an important role to improve and to optimize the execution time of these applications. These environments allow us the inclusion of more parallelism inside the node. However, to take advantage of this parallelism is not an easy task, because we have to deal with some problems such as: cores communications, data locality, memory sizes (cache and RAM), synchronizations, data dependencies on the model, etc. These issues are becoming more important when we wish to improve the application’s performance and scalability. Hence, this paper describes an optimization method developed for Systemic Model of Banking Originated Losses (SYMBOL) tool developed by the European Commission, which is based on analyzing the application's weakness in order to exploit the advantages of the multicore. All these improvements are done in an automatic and transparent manner with the aim of improving the performance metrics of our tool. Finally, experimental evaluations show the effectiveness of our new optimized version, in which we have achieved a considerable improvement on the execution time. The time has been reduced around 96% for the best case tested, between the original serial version and the automatic parallel version.

Adaptive Dynamic Time Warping for Variable Structure Pattern Recognition

Pattern discovery from time series is of fundamental importance. Particularly, when information about the structure of a pattern is not complete, an algorithm to discover specific patterns or shapes automatically from the time series data is necessary. The dynamic time warping is a technique that allows local flexibility in aligning time series. Because of this, it is widely used in many fields such as science, medicine, industry, finance and others. However, a major problem of the dynamic time warping is that it is not able to work with structural changes of a pattern. This problem arises when the structure is influenced by noise, which is a common thing in practice for almost every application. This paper addresses this problem by means of developing a novel technique called adaptive dynamic time warping.

Experimental Study on the Creep Characteristics of FRC Base for Composite Pavement System

The composite pavement system considered in this paper is composed of a functional surface layer, a fiber reinforced asphalt middle layer and a fiber reinforced lean concrete base layer. The mix design of the fiber reinforced lean concrete corresponds to the mix composition of conventional lean concrete but reinforced by fibers. The quasi-absence of research on the durability or long-term performances (fatigue, creep, etc.) of such mix design stresses the necessity to evaluate experimentally the long-term characteristics of this layer composition. This study tests the creep characteristics as one of the long-term characteristics of the fiber reinforced lean concrete layer for composite pavement using a new creep device. The test results reveal that the lean concrete mixed with fiber reinforcement and fly ash develops smaller creep than the conventional lean concrete. The results of the application of the CEB-FIP prediction equation indicate that a modified creep prediction equation should be developed to fit with the new mix design of the layer.

Control Signal from EOG Analysis and Its Application

A game using electro-oculography (EOG) as control signal was introduced in this study. Various EOG signals are generated by eye movements. Even though EOG is a quite complex type of signal, distinct and separable EOG signals could be classified from horizontal and vertical, left and right eye movements. Proper signal processing was incorporated since EOG signal has very small amplitude in the order of micro volts and contains noises influenced by external conditions. Locations of the electrodes were set to be above and below as well as left and right positions of the eyes. Four control signals of up, down, left and right were generated. A microcontroller processed signals in order to simulate a DDR game. A LCD display showed arrows falling down with four different head directions. This game may be used as eye exercise for visual concentration and acuity. Our proposed EOG control signal can be utilized in many other applications of human machine interfaces such as wheelchair, computer keyboard and home automation.

A General Mandatory Access Control Framework in Distributed Environments

In this paper, we propose a general mandatory access framework for distributed systems. The framework can be applied into multiple operating systems and can handle multiple stakeholders. Despite considerable advancements in the area of mandatory access control, a certain approach to enforcing mandatory access control can only be applied in a specific operating system. Other than PC market in which windows captures the overwhelming shares, there are a number of popular operating systems in the emerging smart phone environment, i.e. Android, Windows mobile, Symbian, RIM. It should be noted that more and more stakeholders are involved in smartphone software, such as devices owners, service providers and application providers. Our framework includes three parts—local decision layer, the middle layer and the remote decision layer. The middle layer takes charge of managing security contexts, OS API, operations and policy combination. The design of the remote decision layer doesn’t depend on certain operating systems because of the middle layer’s existence. We implement the framework in windows, linux and other popular embedded systems.

Logistic Changeability - Application of a Methodological Framework for Designing Logistic Changeability

In the past decades, the environment of production companies showed a permanent increase in dynamic and volatility in the form of demand fluctuations, new technologies or global crises. As a reaction to these new requirements, changeability of production systems came into attention. A changeable production system can adapt to these changes quickly and with little effort. Even though demand for changeable production exists for some time, the practical application is still insufficient. To overcome this deficit, a three year research project at the Department of Production Systems and Logistics at the Leibniz University of Hanover/ Germany was initiated. As a result of this project, different concepts have been developed to design production changeable. An excerpt of the results will be presented in this paper. An eight step procedure will be presented to design the changeability of production logistics. This procedure has been applied at a German manufacturer of high demanding weighing machines. The developed procedure, their application in industry, as well as the major results of the application will be presented.

Application of Scientific Metrics to Evaluate Academic Reputation in Different Research Areas

In this paper, we address the problem of identifying academic reputation of researchers using scientific metrics in different research areas. Due to the characteristics of each area, researchers can present different behaviors. In previous work, we define Rep-Index that makes use of a profile template to individually identify the reputation of researchers. The Rep-Index is comprehensive and adaptive because involves hole trajectory of the researcher built throughout his career and can be used in different areas and in different contexts. Now, we compare our metric (Rep-Index) with the h-index and the g-index through experiments with researchers in the fields of Economics, Dentistry and Computer Science. We analyze the trajectory of 830 Brazilian researchers from the National Council of Technological and Scientific Development (CNPq), which receive grants research productivity. The grants are aimed at productivity researchers that stand out among their peers, enhancing their scientific normative criteria established by CNPq. Of the 830 researchers, 210 are in the area of Economics, 216 of Dentistry e 404 of Computer Science. The experiments show that our metric is strongly correlated with h-index, g-index and CNPq ranking. We also show good results for our hypothesis that our metric can be used to evaluate research in several areas. We apply our metric (Rep-Index) to compare the behavior of researchers in relation to their h-index and g-index through extensive experiments. The experiments showed that our metric is strongly correlated with h-index, g-index and CNPq ranking.

Bio-Surfactant Production and Its Application in Microbial EOR

There are various sources of energies available worldwide and among them, crude oil plays a vital role. Oil recovery is achieved using conventional primary and secondary recovery methods. In-order to recover the remaining residual oil, technologies like Enhanced Oil Recovery (EOR) are utilized which is also known as tertiary recovery. Among EOR, Microbial enhanced oil recovery (MEOR) is a technique which enables the improvement of oil recovery by injection of bio-surfactant produced by microorganisms. Bio-surfactant can retrieve unrecoverable oil from the cap rock which is held by high capillary force. Bio-surfactant is a surface active agent which can reduce the interfacial tension and reduce viscosity of oil and thereby oil can be recovered to the surface as the mobility of the oil is increased. Research in this area has shown promising results besides the method is echo-friendly and cost effective compared with other EOR techniques. In our research, on laboratory scale we produced bio-surfactant using the strain Pseudomonas putida (MTCC 2467) and injected into designed simple sand packed column which resembles actual petroleum reservoir. The experiment was conducted in order to determine the efficiency of produced bio-surfactant in oil recovery. The column was made of plastic material with 10 cm in length. The diameter was 2.5 cm. The column was packed with fine sand material. Sand was saturated with brine initially followed by oil saturation. Water flooding followed by bio-surfactant injection was done to determine the amount of oil recovered. Further, the injection of bio-surfactant volume was varied and checked how effectively oil recovery can be achieved. A comparative study was also done by injecting Triton X 100 which is one of the chemical surfactant. Since, bio-surfactant reduced surface and interfacial tension oil can be easily recovered from the porous sand packed column.

Effects of Mobile Design Quality and Innovation Characteristics on Intention to Use Mobile Tourism Guide

This study investigates theoretical model of tourist intention in the context of mobile tourism guide. The research model consists of three constructs: mobile design quality, innovation characteristics, and intention to use mobile tourism guide. In order to investigate the effects of determinants and examine the relationships, partial least squares is employed for data analysis and research model development. The results show that mobile design quality and innovation quality significantly impact on tourists’ intention to use mobile tourism guide. Furthermore, mobile design quality has a strong influence on innovation characteristics, and cannot be the moderator on the relationship between innovation characteristics and tourists’ intention to use mobile tourism guide. Our findings propose theoretical model for mobile research and provide an important guideline for developing mobile application.

Benefits from a SMED Application in a Punching Machine

This paper presents an application of the Single-Minute Exchange of Die (SMED) methodology to a turret punching machine in an elevators company, in Portugal. The work was developed during five months, in the ambit of a master thesis in Industrial Engineering and Management. The Lean Production tool SMED was applied to reduce setup times in order to improve the production flexibility of the machine. The main results obtained were a reduction of 64% in setup time (from 15.1 to 5.4min), 50% in work-in-process amount (from 12.8 to 6.4 days) and 99% in the distance traveled by the operator during the internal period (from 136.7 to 1.7m). These improvements correspond to gains of about €7,315.38 per year.

Optical Flow Based Moving Object Detection and Tracking for Traffic Surveillance

Automated motion detection and tracking is a challenging task in traffic surveillance. In this paper, a system is developed to gather useful information from stationary cameras for detecting moving objects in digital videos. The moving detection and tracking system is developed based on optical flow estimation together with application and combination of various relevant computer vision and image processing techniques to enhance the process. To remove noises, median filter is used and the unwanted objects are removed by applying thresholding algorithms in morphological operations. Also the object type restrictions are set using blob analysis. The results show that the proposed system successfully detects and tracks moving objects in urban videos.

Role of Process Parameters on Pocket Milling with Abrasive Water Jet Machining Technique

Abrasive Water Jet Machining is an unconventional machining process well known for machining hard to cut materials. The primary research focus on the process was for through cutting and a very limited literature is available on pocket milling using AWJM. The present work is an attempt to use this process for milling applications considering a set of various process parameters. Four different input parameters, which were considered by researchers for part separation, are selected for the above application, i.e., abrasive size, flow rate, standoff distance and traverse speed. Pockets of definite size are machined to investigate surface roughness, material removal rate and pocket depth. Based on the data available through experiments on SS304 material, it is observed that higher traverse speeds gives a better finish because of reduction in the particle energy density and lower depth is also observed. Increase in the standoff distance and abrasive flow rate reduces the rate of material removal as the jet loses its focus and occurrence of collisions within the particles. ANOVA for individual output parameter has been studied to know the significant process parameters.

Early-Warning Lights Classification Management System for Industrial Parks in Taiwan

This paper presents the early-warning lights classification management system for industrial parks promoted by the Taiwan Environmental Protection Administration (EPA) since 2011, including the definition of each early-warning light, objectives, action program and accomplishments. All of the 151 industrial parks in Taiwan were classified into four early-warning lights, including red, orange, yellow and green, for carrying out respective pollution management according to the monitoring data of soil and groundwater quality, regulatory compliance, and regulatory listing of control site or remediation site. The Taiwan EPA set up a priority list for high potential polluted industrial parks and investigated their soil and groundwater qualities based on the results of the light classification and pollution potential assessment. In 2011-2013, there were 44 industrial parks selected and carried out different investigation, such as the early warning groundwater well networks establishment and pollution investigation/verification for the red and orange-light industrial parks and the environmental background survey for the yellow-light industrial parks. Among them, 22 industrial parks were newly or continuously confirmed that the concentrations of pollutants exceeded those in soil or groundwater pollution control standards. Thus, the further investigation, groundwater use restriction, listing of pollution control site or remediation site, and pollutant isolation measures were implemented by the local environmental protection and industry competent authorities; the early warning lights of those industrial parks were proposed to adjust up to orange or red-light. Up to the present, the preliminary positive effect of the soil and groundwater quality management system for industrial parks has been noticed in several aspects, such as environmental background information collection, early warning of pollution risk, pollution investigation and control, information integration and application, and inter-agency collaboration. Finally, the work and goal of self-initiated quality management of industrial parks will be carried out on the basis of the inter-agency collaboration by the classified lights system of early warning and management as well as the regular announcement of the status of each industrial park.

A Distance Function for Data with Missing Values and Its Application

Missing values in data are common in real world applications. Since the performance of many data mining algorithms depend critically on it being given a good metric over the input space, we decided in this paper to define a distance function for unlabeled datasets with missing values. We use the Bhattacharyya distance, which measures the similarity of two probability distributions, to define our new distance function. According to this distance, the distance between two points without missing attributes values is simply the Mahalanobis distance. When on the other hand there is a missing value of one of the coordinates, the distance is computed according to the distribution of the missing coordinate. Our distance is general and can be used as part of any algorithm that computes the distance between data points. Because its performance depends strongly on the chosen distance measure, we opted for the k nearest neighbor classifier to evaluate its ability to accurately reflect object similarity. We experimented on standard numerical datasets from the UCI repository from different fields. On these datasets we simulated missing values and compared the performance of the kNN classifier using our distance to other three basic methods. Our  experiments show that kNN using our distance function outperforms the kNN using other methods. Moreover, the runtime performance of our method is only slightly higher than the other methods.

Probabilistic Bhattacharya Based Active Contour Model in Structure Tensor Space

Object identification and segmentation application requires extraction of object in foreground from the background. In this paper the Bhattacharya distance based probabilistic approach is utilized with an active contour model (ACM) to segment an object from the background. In the proposed approach, the Bhattacharya histogram is calculated on non-linear structure tensor space. Based on the histogram, new formulation of active contour model is proposed to segment images. The results are tested on both color and gray images from the Berkeley image database. The experimental results show that the proposed model is applicable to both color and gray images as well as both texture images and natural images. Again in comparing to the Bhattacharya based ACM in ICA space, the proposed model is able to segment multiple object too.

Beneficiation of Pyrolitic Carbon Black

This research investigated treatment of crude carbon black produced from pyrolysis of waste tyres in order to evaluate its quality and possible industrial applications. A representative sample of crude carbon black was dry screened to determine the initial particle size distribution. This was followed by pulverizing the crude carbon black and leaching in hot concentrated sulphuric acid for the removal of heavy metals and other contaminants. Analysis of the refined carbon black showed a significant improvement of the product quality compared to crude carbon black. It was discovered that refined carbon black can be further classified into multiple high value products for various industrial applications such as filler, paint pigment, activated carbon and fuel briquettes.

Augmented Reality on Android

Augmented Reality is an application which combines a live view of real-world environment and computer-generated images. This paper studies and demonstrates an efficient Augmented Reality development in the mobile Android environment with the native Java language and Android SDK. Major components include Barcode Reader, File Loader, Marker Detector, Transform Matrix Generator, and a cloud database.

A Study of Priority Evaluation and Resource Allocation for Revitalization of Cultural Heritages in the Urban Development

Proper maintenance and preservation of significant cultural heritages or historic buildings is necessary. It can not only enhance environmental benefits and a sense of community, but also preserve a city's history and people’s memory. It allows the next generation to be able to get a glimpse of our past, and achieve the goal of sustainable preserved cultural assets. However, the management of maintenance work has not been appropriate for many designated heritages or historic buildings so far. The planning and implementation of the reuse has yet to have a breakthrough specification. It leads the heritages to a mere formality of being “reserved”, instead of the real meaning of “conservation”. For the restoration and preservation of cultural heritages study issues, it is very important due to the consideration of historical significance, symbolism, and economic benefits effects. However, the decision makers such as the officials from public sector they often encounter which heritage should be prioritized to be restored first under the available limited budgets. Only very few techniques are available today to determine the appropriately restoration priorities for the diverse historical heritages, perhaps because of a lack of systematized decision-making aids been proposed before. In the past, the discussions of management and maintenance towards cultural assets were limited to the selection of reuse alternatives instead of the allocation of resources. In view of this, this research will adopt some integrated research methods to solve the existing problems that decision-makers might encounter when allocating resources in the management and maintenance of heritages and historic buildings. The purpose of this study is to develop a sustainable decision making model for local governments to resolve these problems. We propose an alternative decision support model to prioritize restoration needs within the limited budgets. The model is constructed based on fuzzy Delphi, fuzzy analysis network process (FANP) and goal programming (GP) methods. In order to avoid misallocate resources; this research proposes a precise procedure that can take multi-stakeholders views, limited costs and resources into consideration. Also, the combination of many factors and goals has been taken into account to find the highest priority and feasible solution results. To illustrate the approach we propose in this research, seven cultural heritages in Taipei city as one example has been used as an empirical study, and the results are in depth analyzed to explain the application of our proposed approach.

The Implementation of the Multi-Agent Classification System (MACS) in Compliance with FIPA Specifications

The paper discusses the implementation of the MultiAgent classification System (MACS) and utilizing it to provide an automated and accurate classification of end users developing applications in the spreadsheet domain. However, different technologies have been brought together to build MACS. The strength of the system is the integration of the agent technology with the FIPA specifications together with other technologies, which are the .NET widows service based agents, the Windows Communication Foundation (WCF) services, the Service Oriented Architecture (SOA), and Oracle Data Mining (ODM). The Microsoft's .NET widows service based agents were utilized to develop the monitoring agents of MACS, the .NET WCF services together with SOA approach allowed the distribution and communication between agents over the WWW. The Monitoring Agents (MAs) were configured to execute automatically to monitor excel spreadsheets development activities by content. Data gathered by the Monitoring Agents from various resources over a period of time was collected and filtered by a Database Updater Agent (DUA) residing in the .NET client application of the system. This agent then transfers and stores the data in Oracle server database via Oracle stored procedures for further processing that leads to the classification of the end user developers.