Abstract: In this paper the intelligent control of full automatic car wash using a programmable logic controller (PLC) has been investigated and designed to do all steps of carwashing. The Intelligent control of full automatic carwash has the ability to identify and profile the geometrical dimensions of the vehicle chassis. Vehicle dimension identification is an important point in this control system to adjust the washing brushes position and time duration. The study also tries to design a control set for simulating and building the automatic carwash. The main purpose of the simulation is to develop criteria for designing and building this type of carwash in actual size to overcome challenges of automation. The results of this research indicate that the proposed method in process control not only increases productivity, speed, accuracy and safety but also reduce the time and cost of washing based on dynamic model of the vehicle. A laboratory prototype based on an advanced intelligent control has been built to study the validity of the design and simulation which it’s appropriate performance confirms the validity of this study.
Abstract: Various studies have showed that about 90% of single line to ground faults occurred on High voltage transmission lines have transient nature. This type of faults is cleared by temporary outage (by the single phase auto-reclosure). The interval between opening and reclosing of the faulted phase circuit breakers is named “Dead Time” that is varying about several hundred milliseconds. For adjustment of traditional single phase auto-reclosures that usually are not intelligent, it is necessary to calculate the dead time in the off-line condition precisely. If the dead time used in adjustment of single phase auto-reclosure is less than the real dead time, the reclosing of circuit breakers threats the power systems seriously. So in this paper a novel approach for precise calculation of dead time in power transmission lines based on the network equivalencing in time domain is presented. This approach has extremely higher precision in comparison with the traditional method based on Thevenin equivalent circuit. For comparison between the proposed approach in this paper and the traditional method, a comprehensive simulation by EMTP-ATP is performed on an extensive power network.
Abstract: Telecenter is a place where communities can access computers, the Internet, and other digital technologies to enable them to gather information, create, learn, and communicate with others. However, previous studies found that sustainability issues related to economic, political and institutional, social and technology is one of the major problem faced by the telecenter. Based on that problem this research is planning to design a possible solution on rural telecenters sustainability with the support of business intelligence (BI). The empirical study will be conducted through qualitative and quantitative method including interviews and observations with a range of stakeholders including ministry officers, telecenters managers and operators. Result from the data collection will be analyzed using causal modeling approach of SEM SmartPLS for the validity. The expected finding from this research is the Business Intelligent Requirement Model as a guild for sustainability of the rural telecenters.
Abstract: Manufacturers have been exploring innovative strategies to achieve and sustain competitive advantages as they face a new era of intensive global competition. Such strategy is known as Supply Chain Management (SCM), which has gained a tremendous amount of attention from both researchers and practitioners over the last decade. Supply chain management (SCM) is considered as the most popular operating strategy for improving organizational competitiveness in the twenty-first century. It has attracted a lot of attention recently due to its role involving all of the activities in industrial organizations, ranging from raw material procurement to final product delivery to customers. Well-designed supply chain systems can substantially improve efficiency and product quality, and eventually enhance customer satisfaction and profitability. In this paper, a manufacturing engineering perspective on supply chain integration is presented. Research issues discussed include the product and process design for the supply chain, design evaluation of manufacturing in the supply chain, agent-based techniques for supply chain integration, intelligent information for sharing across the supply chain, and a development of standards for product, process, and production data exchange to facilitate electronic commerce. The objective is to provide guidelines and references for manufacturing engineers and researchers interested in supply chain integration.
Abstract: An innovative design for intelligent Chemical Mechanical Polishing (CMP) system is proposed and verified by experiments in this report. On-line measurement and real-time feedback are integrated to eliminate the shortcomings of traditional approaches, e.g., the batch-to-batch discrepancy of required polishing time, over consumption of chemical slurry, and non-uniformity across the wafer. The major advantage of the proposed method is that the finish of local surface roughness can be consistent, no matter where the inner-ring region or outer-ring region is concerned. Secondly, it is able to eliminate the Edge effect. Conventionally, the interfacial induced stress near the wafer edge is generally much higher than that near the wafer center. At last, by using the proposed intelligent chemical mechanical polishing strategy, the cost of the entire machining cycle can be much reduced while the quality of the finished goods certainly upgraded.
Abstract: In many applications retransmissions of lost packets are not permitted. OFDM is a multi-carrier modulation scheme having excellent performance which allows overlapping in frequency domain. With OFDM there is a simple way of dealing with multipath relatively simple DSP algorithms.
In this paper, an image frame is compressed using DWT, and the compressed data is arranged in data vectors, each with equal number of coefficients. These vectors are quantized and binary coded to get the bit steams, which are then packetized and intelligently mapped to the OFDM system. Based on one-bit channel state information at the transmitter, the descriptions in order of descending priority are assigned to the currently good channels such that poorer sub-channels can only affect the lesser important data vectors. We consider only one-bit channel state information available at the transmitter, informing only about the sub-channels to be good or bad. For a good sub-channel, instantaneous received power should be greater than a threshold Pth. Otherwise, the sub-channel is in fading state and considered bad for that batch of coefficients. In order to reduce the system power consumption, the mapped descriptions onto the bad sub channels are dropped at the transmitter. The binary channel state information gives an opportunity to map the bit streams intelligently and to save a reasonable amount of power. By using MAT LAB simulation we can analysis the performance of our proposed scheme, in terms of system energy saving without compromising the received quality in terms of peak signal-noise ratio.
Abstract: The Tropical Data Hub (TDH) is a virtual research environment that provides researchers with an e-research infrastructure to congregate significant tropical data sets for data reuse, integration, searching, and correlation. However, researchers often require data and metadata synthesis across disciplines for cross-domain analyses and knowledge discovery. A triplestore offers a semantic layer to achieve a more intelligent method of search to support the synthesis requirements by automating latent linkages in the data and metadata. Presently, the benchmarks to aid the decision of which triplestore is best suited for use in an application environment like the TDH are limited to performance. This paper describes a new evaluation tool developed to analyze both features and performance. The tool comprises a weighted decision matrix to evaluate the interoperability, functionality, performance, and support availability of a range of integrated and native triplestores to rank them according to requirements of the TDH.
Abstract: Conventional controller’s usually required a prior knowledge of mathematical modelling of the process. The inaccuracy of mathematical modelling degrades the performance of the process, especially for non-linear and complex control problem. The process used is Water-Bath system, which is most widely used and nonlinear to some extent. For Water-Bath system, it is necessary to attain desired temperature within a specified period of time to avoid the overshoot and absolute error, with better temperature tracking capability, else the process is disturbed.
To overcome above difficulties intelligent controllers, Fuzzy Logic (FL) and Adaptive Neuro-Fuzzy Inference System (ANFIS), are proposed in this paper. The Fuzzy controller is designed to work with knowledge in the form of linguistic control rules. But the translation of these linguistic rules into the framework of fuzzy set theory depends on the choice of certain parameters, for which no formal method is known. To design ANFIS, Fuzzy-Inference-System is combined with learning capability of Neural-Network.
It is analyzed that ANFIS is best suitable for adaptive temperature control of above system. As compared to PID and FLC, ANFIS produces a stable control signal. It has much better temperature tracking capability with almost zero overshoot and minimum absolute error.
Abstract: Breast cancer is a major health burden worldwide being a major cause of death amongst women. In this paper, Fuzzy Inference Systems (FIS) are developed for the evaluation of breast cancer risk using Mamdani-type and Sugeno-type models. The paper outlines the basic difference between Mamdani-type FIS and Sugeno-type FIS. The results demonstrated the performance comparison of the two systems and the advantages of using Sugeno- type over Mamdani-type.
Abstract: Automated intelligent, clinical decision support systems generally promote to help or to assist physicians and patients regarding to prevention of diseases or treatment of illnesses using computer represented knowledge and information. In this paper, assessment factors affecting the proper design of clinical decision support system were investigated. The required procedure steps for gathering the data from clinical trial and extracting the information from large volume of healthcare repositories were listed, which are necessary for validation and verification of evidence-based implementation of clinical decision support system. The goal of this paper is to extract useful evaluation factors affecting the quality of the clinical decision support system in the design, development, and implementation of a computer-based decision support system.
Abstract: Educational data mining is a specific data mining field applied to data originating from educational environments, it relies on different approaches to discover hidden knowledge from the available data. Among these approaches are machine learning techniques which are used to build a system that acquires learning from previous data. Machine learning can be applied to solve different regression, classification, clustering and optimization problems.
In our research, we propose a “Student Advisory Framework” that utilizes classification and clustering to build an intelligent system. This system can be used to provide pieces of consultations to a first year university student to pursue a certain education track where he/she will likely succeed in, aiming to decrease the high rate of academic failure among these students. A real case study in Cairo Higher Institute for Engineering, Computer Science and Management is presented using real dataset collected from 2000−2012.The dataset has two main components: pre-higher education dataset and first year courses results dataset. Results have proved the efficiency of the suggested framework.
Abstract: In this paper, a simple moving human detection method is proposed for video surveillance system or access monitoring system. The frame difference and noise threshold are used for initial detection of a moving human-object, and simple labeling method is applied for final human-object segmentation. The simulated results show that the applied algorithm is fast to detect the moving human-objects by performing 95% of correct detection rate. The proposed algorithm has confirmed that can be used as an intelligent video access monitoring system.
Abstract: The computer aided for design, analysis, control, visualization and simulation of robotized workcells is very interesting in this time. Computer Aided Robot Control (CARC) is a subsystem of the system CIM including the computer aided systems of all activities connected with visualization and working of robotized workcells. There are three basic ideas: current CAD/CAM/CAE systems for design and 3D visualization, special PC based control and simulation systems and Augmented Reality Aided Manufacturing (ARAM) systems. This paper describes example of Open Source software application that can to be utilized at planning of the robotized workcells, visualization and off-line programming the automated processes realized by authors.
Abstract: Localization and Recognition of License registration characters from the moving vehicle is a computationally complex task in the field of machine vision and is of substantial interest because of its diverse applications such as cross border security, law enforcement and various other intelligent transportation applications. Previous research used the plate specific details such as aspect ratio, character style, color or dimensions of the plate in the complex task of plate localization. In this paper, license registration character is localized by Enhanced Weight based density map (EWBDM) method, which is independent of such constraints. In connection with our previous method, this paper proposes a method that relaxes constraints in lighting conditions, different fonts of character occurred in the plate and plates with hand-drawn characters in various aspect quotients. The robustness of this method is well suited for applications where the appearance of plates seems to be varied widely. Experimental results show that this approach is suited for recognizing license plates in different external environments.
Abstract: Improvements in the data fusion and data analysis phase of research are imperative due to the exponential growth of sensed data. Currently, there are developments in the Semantic Sensor Web community to explore efficient methods for reuse, correlation and integration of web-based data sets and live data streams. This paper describes the integration of remotely sensed data with web-available static data for use in observational hypothesis testing and the analysis phase of research. The Semantic Reef system combines semantic technologies (e.g., well-defined ontologies and logic systems) with scientific workflows to enable hypothesis-based research. A framework is presented for how the data fusion concepts from the Semantic Reef architecture map to the Smart Environment Monitoring and Analysis Technologies (SEMAT) intelligent sensor network initiative. The data collected via SEMAT and the inferred knowledge from the Semantic Reef system are ingested to the Tropical Data Hub for data discovery, reuse, curation and publication.
Abstract: The planning in manufacturing system is becoming complicated day by day due to the expanding networks and shortage of skilled people to manage change. Consequently, faster lead time and rising demands for eco-efficient evaluation of manufacturing products and processes need exploitation of new and intelligent knowledge management concepts for manufacturing planning. This paper highlights motivation for incorporation of new features in the manufacturing planning system. Furthermore, it elaborates requirements for the development of intelligent knowledge management concept to support planning related decisions. Afterwards, the derived concept is presented in this paper considering two case studies. The first case study is concerned with the automotive ramp-up planning. The second case study specifies requirements for knowledge management system to support decisions in eco-efficient evaluation of manufacturing products and processes
Abstract: The performance of state of the art worldwide telecommunication networks strongly depends on the efficiency of the applied routing mechanism. Game theoretical approaches to this problem offer new solutions. In this paper a new continuous network routing model is defined to describe data transfer in fixed telecommunication networks of multiple hosts. The nodes of the network correspond to routers whose latency is assumed to be traffic dependent. We propose that the whole traffic of the network can be decomposed to a finite number of tasks, which belong to various hosts. To describe the different latency-sensitivity, utility functions are defined for each task. The model is used to compare router and host intelligent types of routing methods, corresponding to various data transfer protocols. We analyze host intelligent routing as a transferable utility cooperative game with externalities. The main aim of the paper is to provide a framework in which the efficiency of various routing algorithms can be compared and the transferable utility game arising in the cooperative case can be analyzed.
Abstract: Short-Term Load Forecasting (STLF) plays an important role for the economic and secure operation of power systems. In this paper, Continuous Genetic Algorithm (CGA) is employed to evolve the optimum large neural networks structure and connecting weights for one-day ahead electric load forecasting problem. This study describes the process of developing three layer feed-forward large neural networks for load forecasting and then presents a heuristic search algorithm for performing an important task of this process, i.e. optimal networks structure design. The proposed method is applied to STLF of the local utility. Data are clustered due to the differences in their characteristics. Special days are extracted from the normal training sets and handled separately. In this way, a solution is provided for all load types, including working days and weekends and special days. We find good performance for the large neural networks. The proposed methodology gives lower percent errors all the time. Thus, it can be applied to automatically design an optimal load forecaster based on historical data.
Abstract: Nowadays, more engineering systems are using some
kind of Artificial Intelligence (AI) for the development of their
processes. Some well-known AI techniques include artificial neural
nets, fuzzy inference systems, and neuro-fuzzy inference systems
among others. Furthermore, many decision-making applications base
their intelligent processes on Fuzzy Logic; due to the Fuzzy
Inference Systems (FIS) capability to deal with problems that are
based on user knowledge and experience. Also, knowing that users
have a wide variety of distinctiveness, and generally, provide
uncertain data, this information can be used and properly processed
by a FIS. To properly consider uncertainty and inexact system input
values, FIS normally use Membership Functions (MF) that represent
a degree of user satisfaction on certain conditions and/or constraints.
In order to define the parameters of the MFs, the knowledge from
experts in the field is very important. This knowledge defines the MF
shape to process the user inputs and through fuzzy reasoning and
inference mechanisms, the FIS can provide an “appropriate" output.
However an important issue immediately arises: How can it be
assured that the obtained output is the optimum solution? How can it
be guaranteed that each MF has an optimum shape? A viable solution
to these questions is through the MFs parameter optimization. In this
Paper a novel parameter optimization process is presented. The
process for FIS parameter optimization consists of the five simple
steps that can be easily realized off-line. Here the proposed process
of FIS parameter optimization it is demonstrated by its
implementation on an Intelligent Interface section dealing with the
on-line customization / personalization of internet portals applied to
E-commerce.
Abstract: A challenged control problem is when the
performance is pushed to the limit. The state-derivative feedback
control strategy directly uses acceleration information for feedback
and state estimation. The derivative part is concerned with the rateof-
change of the error with time. If the measured variable approaches
the set point rapidly, then the actuator is backed off early to allow it
to coast to the required level. Derivative action makes a control
system behave much more intelligently. A sensor measures the
variable to be controlled and the measured in formation is fed back to
the controller to influence the controlled variable. A high gain
problem can be also formulated for proportional plus derivative
feedback transformation. Using MATLAB Simulink dynamic
simulation tool this paper examines a system with a proportional plus
derivative feedback and presents an automatic implementation of
finding an acceptable controlled system. Using feedback
transformations the system is transformed into another system.