Intelligent System for Breast Cancer Prognosis using Multiwavelet Packets and Neural Network

This paper presents an approach for early breast cancer diagnostic by employing combination of artificial neural networks (ANN) and multiwaveletpacket based subband image decomposition. The microcalcifications correspond to high-frequency components of the image spectrum, detection of microcalcifications is achieved by decomposing the mammograms into different frequency subbands,, reconstructing the mammograms from the subbands containing only high frequencies. For this approach we employed different types of multiwaveletpacket. We used the result as an input of neural network for classification. The proposed methodology is tested using the Nijmegen and the Mammographic Image Analysis Society (MIAS) mammographic databases and images collected from local hospitals. Results are presented as the receiver operating characteristic (ROC) performance and are quantified by the area under the ROC curve.

Low-MAC FEC Controller for JPEG2000 Image Transmission Over IEEE 802.15.4

In this paper, we propose the low-MAC FEC controller for practical implementation of JPEG2000 image transmission using IEEE 802.15.4. The proposed low-MAC FEC controller has very small HW size and spends little computation to estimate channel state. Because of this advantage, it is acceptable to apply IEEE 802.15.4 which has to operate more than 1 year with battery. For the image transmission, we integrate the low-MAC FEC controller and RCPC coder in sensor node of LR-WPAN. The modified sensor node has increase of 3% hardware size than conventional zigbee sensor node.

New Class of Chaotic Mappings in Symbol Space

Symbolic dynamics studies dynamical systems on the basis of the symbol sequences obtained for a suitable partition of the state space. This approach exploits the property that system dynamics reduce to a shift operation in symbol space. This shift operator is a chaotic mapping. In this article we show that in the symbol space exist other chaotic mappings.

Low Power and Less Area Architecture for Integer Motion Estimation

Full search block matching algorithm is widely used for hardware implementation of motion estimators in video compression algorithms. In this paper we are proposing a new architecture, which consists of a 2D parallel processing unit and a 1D unit both working in parallel. The proposed architecture reduces both data access power and computational power which are the main causes of power consumption in integer motion estimation. It also completes the operations with nearly the same number of clock cycles as compared to a 2D systolic array architecture. In this work sum of absolute difference (SAD)-the most repeated operation in block matching, is calculated in two steps. The first step is to calculate the SAD for alternate rows by a 2D parallel unit. If the SAD calculated by the parallel unit is less than the stored minimum SAD, the SAD of the remaining rows is calculated by the 1D unit. Early termination, which stops avoidable computations has been achieved with the help of alternate rows method proposed in this paper and by finding a low initial SAD value based on motion vector prediction. Data reuse has been applied to the reference blocks in the same search area which significantly reduced the memory access.

Metaheuristic Algorithms for Decoding Binary Linear Codes

This paper introduces two decoders for binary linear codes based on Metaheuristics. The first one uses a genetic algorithm and the second is based on a combination genetic algorithm with a feed forward neural network. The decoder based on the genetic algorithms (DAG) applied to BCH and convolutional codes give good performances compared to Chase-2 and Viterbi algorithm respectively and reach the performances of the OSD-3 for some Residue Quadratic (RQ) codes. This algorithm is less complex for linear block codes of large block length; furthermore their performances can be improved by tuning the decoder-s parameters, in particular the number of individuals by population and the number of generations. In the second algorithm, the search space, in contrast to DAG which was limited to the code word space, now covers the whole binary vector space. It tries to elude a great number of coding operations by using a neural network. This reduces greatly the complexity of the decoder while maintaining comparable performances.

Methods for Better Assessment of Fatigue and Deterioration in Bridges and Other Steel or Concrete Constructions

Large metal and concrete structures suffer by various kinds of deterioration, and accurate prediction of the remaining life is important. This paper informs about two methods for its assessment. One method, suitable for steel bridges and other constructions exposed to fatigue, monitors the loads and damage accumulation using information systems for the operation and the finite element model of the construction. In addition to the operation load, the dead weight of the construction and thermal stresses can be included into the model. The second method is suitable for concrete bridges and other structures, which suffer by carbonatation and other degradation processes, driven by diffusion. The diffusion constant, important for the prediction of future development, can be determined from the depth-profile of pH, obtained by pH measurement at various depths. Comparison with measurements on real objects illustrates the suitability of both methods.

Optimum Operating Conditions for Direct Oxidation of H2S in a Fluidized Bed Reactor

In this research a mathematical model for direct oxidization of hydrogen sulfide into elemental sulfur in a fluidized bed reactor with external circulation was developed. As the catalyst is deactivated in the fluidized bed, it might be placed in a reduction tank in order to remove sulfur through heating above its dew point. The reactor model demonstrated via MATLAB software. It was shown that variations of H2S conversion as well as; products formed were reasonable in comparison with corresponding results of a fixed bed reactor. Through analyzing results of this model, it became possible to propose the main optimized operating conditions for the process considered. These conditions included; the temperature range of 100-130ºC and utilizing the catalyst as much as possible providing the highest bed density respect to dimensions of bed, economical aspects that the bed ever remained in fluidized mode. A high active and stable catalyst under the optimum conditions exhibited 100% conversion in a fluidized bed reactor.

Integrating Life Cycle Uncertainties for Evaluating a Building Overall Cost

Overall cost is a significant consideration in any decision-making process. Although many studies were carried out on overall cost in construction, little has treated the uncertainties of real life cycle development. On the basis of several case studies, a feedback process was performed on the historical data of studied buildings. This process enabled to identify some factors causing uncertainty during the operational period. As a result, the research proposes a new method for assessing the overall cost during a part of the building-s life cycle taking account of the building actual value, its end-of-life value and the influence of the identified life cycle uncertainty factors. The findings are a step towards a higher level of reliability in overall cost evaluation taking account of some usually unexpected uncertainty factors.

Agent-based Framework for Energy Efficiency in Wireless Sensor Networks

Wireless sensor networks are consisted of hundreds or thousands of small sensors that have limited resources. Energy-efficient techniques are the main issue of wireless sensor networks. This paper proposes an energy efficient agent-based framework in wireless sensor networks. We adopt biologically inspired approaches for wireless sensor networks. Agent operates automatically with their behavior policies as a gene. Agent aggregates other agents to reduce communication and gives high priority to nodes that have enough energy to communicate. Agent behavior policies are optimized by genetic operation at the base station. Simulation results show that our proposed framework increases the lifetime of each node. Each agent selects a next-hop node with neighbor information and behavior policies. Our proposed framework provides self-healing, self-configuration, self-optimization properties to sensor nodes.

Data Transformation Services (DTS): Creating Data Mart by Consolidating Multi-Source Enterprise Operational Data

Trends in business intelligence, e-commerce and remote access make it necessary and practical to store data in different ways on multiple systems with different operating systems. As business evolve and grow, they require efficient computerized solution to perform data update and to access data from diverse enterprise business applications. The objective of this paper is to demonstrate the capability of DTS [1] as a database solution for automatic data transfer and update in solving business problem. This DTS package is developed for the sales of variety of plants and eventually expanded into commercial supply and landscaping business. Dimension data modeling is used in DTS package to extract, transform and load data from heterogeneous database systems such as MySQL, Microsoft Access and Oracle that consolidates into a Data Mart residing in SQL Server. Hence, the data transfer from various databases is scheduled to run automatically every quarter of the year to review the efficient sales analysis. Therefore, DTS is absolutely an attractive solution for automatic data transfer and update which meeting today-s business needs.

Carotenoids and Colour Before and After Storage of Organically and Conventionally Cultivated Potato Genotypes in Latvia

Potatoes are a good source of carotenoids, which are lipophilic compounds synthesized in plastids from isoprenoids. The aim of this research was to determine the content of carotenoids in relationship with the colour of organically and conventionally cultivated potato genotypes before and after period of storage. In cooperation with the State Priekuli Plant Breeding Institute (Latvia), six potato genotypes were studied: 'Agrie dzeltenie', 'Prelma', 'Imanta', 'S-03135-10', 'S-99108-8' and 'S-01063-5'. All the genotypes were cultivated under three different conditions: organically and conventionally (two conditions). The content of carotenoids was determined by using spectrophotometer and the colour – L*a*b* system. The results of current research show that after the period of storage, carotenoid amount has increased and in conventionally cultivated potatoes it varies from 228.514 to 552.434 μg 100 g-1 while in organically cultivated potato genotypes – from 45.485 to 662.699 μg 100 g-1 FW. Colour of potato flesh was changing during storage.

Chips of Ti-6Al-2Sn-4Zr-6Mo Alloy – A Detailed Geometry Study

Titanium alloys like Ti-6Al-2Sn-4Zr-6Mo (Ti- 6246) are widely used in aerospace applications. Component manufacturing, however, is difficult and expensive as their machinability is extremely poor. A thorough understanding of the chip formation process is needed to improve related metal cutting operations.In the current study, orthogonal cutting experiments have been performed and theresulting chips were analyzed by optical microscopy and scanning electron microscopy.Chips from aTi- 6246ingot were produced at different cutting speeds and cutting depths. During the experiments, depending of the cutting conditions, continuous or segmented chips were formed. Narrow, highly deformed and grain oriented zones, the so-called shear zone, separated individual segments. Different material properties have been measured in the shear zones and the segments.

Self-evolving Artificial Immune System via Developing T and B Cell for Permutation Flow-shop Scheduling Problems

Artificial Immune System is applied as a Heuristic Algorithm for decades. Nevertheless, many of these applications took advantage of the benefit of this algorithm but seldom proposed approaches for enhancing the efficiency. In this paper, a Self-evolving Artificial Immune System is proposed via developing the T and B cell in Immune System and built a self-evolving mechanism for the complexities of different problems. In this research, it focuses on enhancing the efficiency of Clonal selection which is responsible for producing Affinities to resist the invading of Antigens. T and B cell are the main mechanisms for Clonal Selection to produce different combinations of Antibodies. Therefore, the development of T and B cell will influence the efficiency of Clonal Selection for searching better solution. Furthermore, for better cooperation of the two cells, a co-evolutional strategy is applied to coordinate for more effective productions of Antibodies. This work finally adopts Flow-shop scheduling instances in OR-library to validate the proposed algorithm.

Predictive Clustering Hybrid Regression(pCHR) Approach and Its Application to Sucrose-Based Biohydrogen Production

A predictive clustering hybrid regression (pCHR) approach was developed and evaluated using dataset from H2- producing sucrose-based bioreactor operated for 15 months. The aim was to model and predict the H2-production rate using information available about envirome and metabolome of the bioprocess. Selforganizing maps (SOM) and Sammon map were used to visualize the dataset and to identify main metabolic patterns and clusters in bioprocess data. Three metabolic clusters: acetate coupled with other metabolites, butyrate only, and transition phases were detected. The developed pCHR model combines principles of k-means clustering, kNN classification and regression techniques. The model performed well in modeling and predicting the H2-production rate with mean square error values of 0.0014 and 0.0032, respectively.

Development of Autonomous Cable Inspection Robot for Nuclear Power Plant

The cables in a nuclear power plant are designed to be used for about 40 years in safe operation environment. However, the heat and radiation in the nuclear power plant causes the rapid performance deterioration of cables in nuclear vessels and heat exchangers, which requires cable lifetime estimation. The most accurate method of estimating the cable lifetime is to evaluate the cables in a laboratory. However, removing cables while the plant is operating is not allowed because of its safety and cost. In this paper, a robot system to estimate the cable lifetime in nuclear power plants is developed and tested. The developed robot system can calculate a modulus value to estimate the cable lifetime even when the nuclear power plant is in operation.

MJPEG Real-Time Transmission in Industrial Environments Using a CBR Channel

Currently, there are many local area industrial networks that can give guaranteed bandwidth to synchronous traffic, particularly providing CBR channels (Constant Bit Rate), which allow improved bandwidth management. Some of such networks operate over Ethernet, delivering channels with enough capacity, specially with compressors, to integrate multimedia traffic in industrial monitoring and image processing applications with many sources. In these industrial environments where a low latency is an essential requirement, JPEG is an adequate compressing technique but it generates VBR traffic (Variable Bit Rate). Transmitting VBR traffic in CBR channels is inefficient and current solutions to this problem significantly increase the latency or further degrade the quality. In this paper an R(q) model is used which allows on-line calculation of the JPEG quantification factor. We obtained increased quality, a lower requirement for the CBR channel with reduced number of discarded frames along with better use of the channel bandwidth.

Scheduling a Flexible Flow Shops Problem using DEA

This paper considers a scheduling problem in flexible flow shops environment with the aim of minimizing two important criteria including makespan and cumulative tardiness of jobs. Since the proposed problem is known as an Np-hard problem in literature, we have to develop a meta-heuristic to solve it. We considered general structure of Genetic Algorithm (GA) and developed a new version of that based on Data Envelopment Analysis (DEA). Two objective functions assumed as two different inputs for each Decision Making Unit (DMU). In this paper we focused on efficiency score of DMUs and efficient frontier concept in DEA technique. After introducing the method we defined two different scenarios with considering two types of mutation operator. Also we provided an experimental design with some computational results to show the performance of algorithm. The results show that the algorithm implements in a reasonable time.

Public Transport: Punctuality Index for Bus Operation

Public bus service plays a significant role in our society as people movers and to facilitate travels within towns and districts. The quality of service of public bus is always being regarded as poor, or rather, underestimated as second class means of transportation. Reliability of service, or the ability to deliver service as planned, is one key element in perceiving the quality of bus service and the punctuality index is one of the performance parameters in determining the service reliability. This study concentrates on evaluating the reliability performance of bus operation using punctuality index assessment. A week data for each of six city bus routes is recorded using the on-board methodology to calculate the punctuality index for city bus service in Kota Bharu. The results revealed that the punctuality index for the whole city bus network is 94.25% (LOS B).

Integrate Communication Modeling into the Design Modeling at Early Stages of the Design Flow Case Study: Unmanned Aerial Vehicle (UAV)

This paper shows how we can integrate communication modeling into the design modeling at early stages of the design flow. We consider effect of incorporating noise such as impulsive noise on system stability. We show that with change of the system model and investigate the system performance under the different communication effects. We modeled a unmanned aerial vehicle (UAV) as a demonstration using SystemC methodology. Moreover the system is modeled by joining the capabilities of UML and SystemC to operate at system level.

The Determination of Rating Points of Objects with Qualitative Characteristics and their Usagein Decision Making Problems

The paper presents the method developed to assess rating points of objects with qualitative indexes. The novelty of the method lies in the fact that the authors use linguistic scales that allow to formalize the values of the indexes with the help of fuzzy sets. As a result it is possible to operate correctly with dissimilar indexes on the unified basis and to get stable final results. The obtained rating points are used in decision making based on fuzzy expert opinions.