Persistence of Termination for Term Rewriting Systems with Ordered Sorts

A property is persistent if for any many-sorted term rewriting system , has the property if and only if term rewriting system , which results from by omitting its sort information, has the property. Zantema showed that termination is persistent for term rewriting systems without collapsing or duplicating rules. In this paper, we show that the Zantema's result can be extended to term rewriting systems on ordered sorts, i.e., termination is persistent for term rewriting systems on ordered sorts without collapsing, decreasing or duplicating rules. Furthermore we give the example as application of this result. Also we obtain that completeness is persistent for this class of term rewriting systems.

Solving Fully Fuzzy Linear Systems by use of a Certain Decomposition of the Coefficient Matrix

In this paper, we give a certain decomposition of the coefficient matrix of the fully fuzzy linear system (FFLS) to obtain a simple algorithm for solving these systems. The new algorithm can solve FFLS in a smaller computing process. We will illustrate our method by solving some examples.

Self-evolving Artificial Immune System via Developing T and B Cell for Permutation Flow-shop Scheduling Problems

Artificial Immune System is applied as a Heuristic Algorithm for decades. Nevertheless, many of these applications took advantage of the benefit of this algorithm but seldom proposed approaches for enhancing the efficiency. In this paper, a Self-evolving Artificial Immune System is proposed via developing the T and B cell in Immune System and built a self-evolving mechanism for the complexities of different problems. In this research, it focuses on enhancing the efficiency of Clonal selection which is responsible for producing Affinities to resist the invading of Antigens. T and B cell are the main mechanisms for Clonal Selection to produce different combinations of Antibodies. Therefore, the development of T and B cell will influence the efficiency of Clonal Selection for searching better solution. Furthermore, for better cooperation of the two cells, a co-evolutional strategy is applied to coordinate for more effective productions of Antibodies. This work finally adopts Flow-shop scheduling instances in OR-library to validate the proposed algorithm.

Ec-A: A Task Allocation Algorithm for Energy Minimization in Multiprocessor Systems

With the necessity of increased processing capacity with less energy consumption; power aware multiprocessor system has gained more attention in the recent future. One of the additional challenges that is to be solved in a multi-processor system when compared to uni-processor system is job allocation. This paper presents a novel task dependent job allocation algorithm: Energy centric- Allocation (Ec-A) and Rate Monotonic (RM) scheduling to minimize energy consumption in a multiprocessor system. A simulation analysis is carried out to verify the performance increase with reduction in energy consumption and required number of processors in the system.

Mobile Velocity Based Bidirectional Call Overflow Scheme in Hierarchical Cellular System

In the age of global communications, heterogeneous networks are seen to be the best choice of strategy to ensure continuous and uninterruptible services. This will allow mobile terminal to stay in connection even they are migrating into different segment coverage through the handoff process. With the increase of teletraffic demands in mobile cellular system, hierarchical cellular systems have been adopted extensively for more efficient channel utilization and better QoS (Quality of Service). This paper presents a bidirectional call overflow scheme between two layers of microcells and macrocells, where handoffs are decided by the velocity of mobile making the call. To ensure that handoff calls are given higher priorities, it is assumed that guard channels are assigned in both macrocells and microcells. A hysteresis value introduced in mobile velocity is used to allow mobile roam in the same cell if its velocity changes back within the set threshold values. By doing this the number of handoffs is reduced thereby reducing the processing overhead and enhancing the quality of service to the end user.

Experimental Study of Light Crude Oil-Water Emulsions

This paper made an attempt to investigate the problem associated with enhancement of emulsions of light crude oil-water recovery in an oil field of Algerian Sahara. Measurements were taken through experiments using RheoStress (RS600). Factors such as shear rate, temperature and light oil concentration on the viscosity behavior were considered. Experimental measurements were performed in terms of shear stress–shear rate, yield stress and flow index on mixture of light crude oil–water. The rheological behavior of emulsion showed Non-Newtonian shear thinning behavior (Herschel-Bulkley). The experiments done in the laboratory showed the stability of some water in light crude oil emulsions form during consolidate oil recovery process. To break the emulsion using additives may involve higher cost and could be very expensive. Therefore, further research should be directed to find solution of these problems that have been encountered.

On the Mathematical Structure and Algorithmic Implementation of Biochemical Network Models

Modeling and simulation of biochemical reactions is of great interest in the context of system biology. The central dogma of this re-emerging area states that it is system dynamics and organizing principles of complex biological phenomena that give rise to functioning and function of cells. Cell functions, such as growth, division, differentiation and apoptosis are temporal processes, that can be understood if they are treated as dynamic systems. System biology focuses on an understanding of functional activity from a system-wide perspective and, consequently, it is defined by two hey questions: (i) how do the components within a cell interact, so as to bring about its structure and functioning? (ii) How do cells interact, so as to develop and maintain higher levels of organization and functions? In recent years, wet-lab biologists embraced mathematical modeling and simulation as two essential means toward answering the above questions. The credo of dynamics system theory is that the behavior of a biological system is given by the temporal evolution of its state. Our understanding of the time behavior of a biological system can be measured by the extent to which a simulation mimics the real behavior of that system. Deviations of a simulation indicate either limitations or errors in our knowledge. The aim of this paper is to summarize and review the main conceptual frameworks in which models of biochemical networks can be developed. In particular, we review the stochastic molecular modelling approaches, by reporting the principal conceptualizations suggested by A. A. Markov, P. Langevin, A. Fokker, M. Planck, D. T. Gillespie, N. G. van Kampfen, and recently by D. Wilkinson, O. Wolkenhauer, P. S. Jöberg and by the author.

Changes in Selected Fuel Properties of Sewage Sludge as a Result of its Storage

The article presents test results on the changes occurring in sewage sludge during the process of its storage. Tests were conducted on mechanically dehydrated sewage sludge derived from large municipal sewage treatment plants equipped with biological sewage treatment systems. In testing presented in the paper the focus was on the basic fuel properties of sewage sludge: moisture content, heat of combustion, carbon share. In the first part of the article the overview of the issues concerning the sewage sludge management is presented and the genesis of tests is explained. Further in the paper, selected results of conducted tests are discussed. Changes in tested parameters were determined in the period of a 10- month sewage storage.

An Efficient Data Mining Approach on Compressed Transactions

In an era of knowledge explosion, the growth of data increases rapidly day by day. Since data storage is a limited resource, how to reduce the data space in the process becomes a challenge issue. Data compression provides a good solution which can lower the required space. Data mining has many useful applications in recent years because it can help users discover interesting knowledge in large databases. However, existing compression algorithms are not appropriate for data mining. In [1, 2], two different approaches were proposed to compress databases and then perform the data mining process. However, they all lack the ability to decompress the data to their original state and improve the data mining performance. In this research a new approach called Mining Merged Transactions with the Quantification Table (M2TQT) was proposed to solve these problems. M2TQT uses the relationship of transactions to merge related transactions and builds a quantification table to prune the candidate itemsets which are impossible to become frequent in order to improve the performance of mining association rules. The experiments show that M2TQT performs better than existing approaches.

A Bionic Approach to Dynamic, Multimodal Scene Perception and Interpretation in Buildings

Today, building automation is advancing from simple monitoring and control tasks of lightning and heating towards more and more complex applications that require a dynamic perception and interpretation of different scenes occurring in a building. Current approaches cannot handle these newly upcoming demands. In this article, a bionically inspired approach for multimodal, dynamic scene perception and interpretation is presented, which is based on neuroscientific and neuro-psychological research findings about the perceptual system of the human brain. This approach bases on data from diverse sensory modalities being processed in a so-called neuro-symbolic network. With its parallel structure and with its basic elements being information processing and storing units at the same time, a very efficient method for scene perception is provided overcoming the problems and bottlenecks of classical dynamic scene interpretation systems.

Double Reduction of Ada-ECATNet Representation using Rewriting Logic

One major difficulty that faces developers of concurrent and distributed software is analysis for concurrency based faults like deadlocks. Petri nets are used extensively in the verification of correctness of concurrent programs. ECATNets [2] are a category of algebraic Petri nets based on a sound combination of algebraic abstract types and high-level Petri nets. ECATNets have 'sound' and 'complete' semantics because of their integration in rewriting logic [12] and its programming language Maude [13]. Rewriting logic is considered as one of very powerful logics in terms of description, verification and programming of concurrent systems. We proposed in [4] a method for translating Ada-95 tasking programs to ECATNets formalism (Ada-ECATNet). In this paper, we show that ECATNets formalism provides a more compact translation for Ada programs compared to the other approaches based on simple Petri nets or Colored Petri nets (CPNs). Such translation doesn-t reduce only the size of program, but reduces also the number of program states. We show also, how this compact Ada-ECATNet may be reduced again by applying reduction rules on it. This double reduction of Ada-ECATNet permits a considerable minimization of the memory space and run time of corresponding Maude program.

Segmenting Ultrasound B-Mode Images Using RiIG Distributions and Stochastic Optimization

In this paper, we propose a novel algorithm for delineating the endocardial wall from a human heart ultrasound scan. We assume that the gray levels in the ultrasound images are independent and identically distributed random variables with different Rician Inverse Gaussian (RiIG) distributions. Both synthetic and real clinical data will be used for testing the algorithm. Algorithm performance will be evaluated using the expert radiologist evaluation of a soft copy of an ultrasound scan during the scanning process and secondly, doctor’s conclusion after going through a printed copy of the same scan. Successful implementation of this algorithm should make it possible to differentiate normal from abnormal soft tissue and help disease identification, what stage the disease is in and how best to treat the patient. We hope that an automated system that uses this algorithm will be useful in public hospitals especially in Third World countries where problems such as shortage of skilled radiologists and shortage of ultrasound machines are common. These public hospitals are usually the first and last stop for most patients in these countries.

Statistical Analysis of Different Configurations of Hybrid Doped Fiber Amplifiers

Wavelength multiplexing (WDM) technology along with optical amplifiers is used for optical communication systems in S-band, C-band and L-band. To improve the overall system performance Hybrid amplifiers consisting of cascaded TDFA and EDFA with different gain bandwidths are preferred for long haul wavelength multiplexed optical communication systems. This paper deals with statistical analysis of different configuration of hybrid amplifier i.e. analysis of TDFA-EDFA configuration and EDFA – TDFA configuration. In this paper One-Way ANOVA method is used for statistical analysis.

A K-Means Based Clustering Approach for Finding Faulty Modules in Open Source Software Systems

Prediction of fault-prone modules provides one way to support software quality engineering. Clustering is used to determine the intrinsic grouping in a set of unlabeled data. Among various clustering techniques available in literature K-Means clustering approach is most widely being used. This paper introduces K-Means based Clustering approach for software finding the fault proneness of the Object-Oriented systems. The contribution of this paper is that it has used Metric values of JEdit open source software for generation of the rules for the categorization of software modules in the categories of Faulty and non faulty modules and thereafter empirically validation is performed. The results are measured in terms of accuracy of prediction, probability of Detection and Probability of False Alarms.

Modeling and Optimization of Abrasive Waterjet Parameters using Regression Analysis

Abrasive waterjet is a novel machining process capable of processing wide range of hard-to-machine materials. This research addresses modeling and optimization of the process parameters for this machining technique. To model the process a set of experimental data has been used to evaluate the effects of various parameter settings in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. Depth of cut, as one of the most important output characteristics, has been evaluated based on different parameter settings. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. The pairwise effects of process parameters settings on process response outputs are also shown graphically. The proposed model is then embedded into a Simulated Annealing algorithm to optimize the process parameters. The optimization is carried out for any desired values of depth of cut. The objective is to determine proper levels of process parameters in order to obtain a certain level of depth of cut. Computational results demonstrate that the proposed solution procedure is quite effective in solving such multi-variable problems.

The Determination of Rating Points of Objects with Qualitative Characteristics and their Usagein Decision Making Problems

The paper presents the method developed to assess rating points of objects with qualitative indexes. The novelty of the method lies in the fact that the authors use linguistic scales that allow to formalize the values of the indexes with the help of fuzzy sets. As a result it is possible to operate correctly with dissimilar indexes on the unified basis and to get stable final results. The obtained rating points are used in decision making based on fuzzy expert opinions.

Eye Gesture Analysis with Head Movement for Advanced Driver Assistance Systems

Road traffic accidents are a major cause of death worldwide. In an attempt to reduce accidents, some research efforts have focused on creating Advanced Driver Assistance Systems (ADAS) able to detect vehicle, driver and environmental conditions and to use this information to identify cues for potential accidents. This paper presents continued work on a novel Non-intrusive Intelligent Driver Assistance and Safety System (Ni-DASS) for assessing driver point of regard within vehicles. It uses an on-board CCD camera to observe the driver-s face. A template matching approach is used to compare the driver-s eye-gaze pattern with a set of eye-gesture templates of the driver looking at different focal points within the vehicle. The windscreen is divided into cells and comparison of the driver-s eye-gaze pattern with templates of a driver-s eyes looking at each cell is used to determine the driver-s point of regard on the windscreen. Results indicate that the proposed technique could be useful in situations where low resolution estimates of driver point of regard are adequate. For instance, To allow ADAS systems to alert the driver if he/she has positively failed to observe a hazard.

A New Heuristic Approach to Solving U-shape Assembly Line Balancing Problems Type-1

Assembly line balancing is a very important issue in mass production systems due to production cost. Although many studies have been done on this topic, but because assembly line balancing problems are so complex they are categorized as NP-hard problems and researchers strongly recommend using heuristic methods. This paper presents a new heuristic approach called the critical task method (CTM) for solving U-shape assembly line balancing problems. The performance of the proposed heuristic method is tested by solving a number of test problems and comparing them with 12 other heuristics available in the literature to confirm the superior performance of the proposed heuristic. Furthermore, to prove the efficiency of the proposed CTM, the objectives are increased to minimize the number of workstation (or equivalently maximize line efficiency), and minimizing the smoothness index. Finally, it is proven that the proposed heuristic is more efficient than the others to solve the U-shape assembly line balancing problem.

LAYMOD; A Layered and Modular Platform for CAx Collaboration Management and Supporting Product data Integration based on STEP Standard

Nowadays companies strive to survive in a competitive global environment. To speed up product development/modifications, it is suggested to adopt a collaborative product development approach. However, despite the advantages of new IT improvements still many CAx systems work separately and locally. Collaborative design and manufacture requires a product information model that supports related CAx product data models. To solve this problem many solutions are proposed, which the most successful one is adopting the STEP standard as a product data model to develop a collaborative CAx platform. However, the improvement of the STEP-s Application Protocols (APs) over the time, huge number of STEP AP-s and cc-s, the high costs of implementation, costly process for conversion of older CAx software files to the STEP neutral file format; and lack of STEP knowledge, that usually slows down the implementation of the STEP standard in collaborative data exchange, management and integration should be considered. In this paper the requirements for a successful collaborative CAx system is discussed. The STEP standard capability for product data integration and its shortcomings as well as the dominant platforms for supporting CAx collaboration management and product data integration are reviewed. Finally a platform named LAYMOD to fulfil the requirements of CAx collaborative environment and integrating the product data is proposed. The platform is a layered platform to enable global collaboration among different CAx software packages/developers. It also adopts the STEP modular architecture and the XML data structures to enable collaboration between CAx software packages as well as overcoming the STEP standard limitations. The architecture and procedures of LAYMOD platform to manage collaboration and avoid contradicts in product data integration are introduced.

Artificial Visual Percepts for Image Understanding

Visual inputs are one of the key sources from which humans perceive the environment and 'understand' what is happening. Artificial systems perceive the visual inputs as digital images. The images need to be processed and analysed. Within the human brain, processing of visual inputs and subsequent development of perception is one of its major functionalities. In this paper we present part of our research project, which aims at the development of an artificial model for visual perception (or 'understanding') based on the human perceptive and cognitive systems. We propose a new model for perception from visual inputs and a way of understaning or interpreting images using the model. We demonstrate the implementation and use of the model with a real image data set.