A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation

Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.

Design for Manufacturability and Concurrent Engineering for Product Development

In the 1980s, companies began to feel the effect of three major influences on their product development: newer and innovative technologies, increasing product complexity and larger organizations. And therefore companies were forced to look for new product development methods. This paper tries to focus on the two of new product development methods (DFM and CE). The aim of this paper is to see and analyze different product development methods specifically on Design for Manufacturability and Concurrent Engineering. Companies can achieve and be benefited by minimizing product life cycle, cost and meeting delivery schedule. This paper also presents simplified models that can be modified and used by different companies based on the companies- objective and requirements. Methodologies that are followed to do this research are case studies. Two companies were taken and analysed on the product development process. Historical data, interview were conducted on these companies in addition to that, Survey of literatures and previous research works on similar topics has been done during this research. This paper also tries to show the implementation cost benefit analysis and tries to calculate the implementation time. From this research, it has been found that the two companies did not achieve the delivery time to the customer. Some of most frequently coming products are analyzed and 50% to 80 % of their products are not delivered on time to the customers. The companies are following the traditional way of product development that is sequentially design and production method, which highly affect time to market. In the case study it is found that by implementing these new methods and by forming multi disciplinary team in designing and quality inspection; the company can reduce the workflow steps from 40 to 30.

Attribute Selection Methods Comparison for Classification of Diffuse Large B-Cell Lymphoma

The most important subtype of non-Hodgkin-s lymphoma is the Diffuse Large B-Cell Lymphoma. Approximately 40% of the patients suffering from it respond well to therapy, whereas the remainder needs a more aggressive treatment, in order to better their chances of survival. Data Mining techniques have helped to identify the class of the lymphoma in an efficient manner. Despite that, thousands of genes should be processed to obtain the results. This paper presents a comparison of the use of various attribute selection methods aiming to reduce the number of genes to be searched, looking for a more effective procedure as a whole.

Improving Location Management in Mobile IPv4 Networks

The Mobile IP Standard has been developed to support mobility over the Internet. This standard contains several drawbacks as in the cases where packets are routed via sub-optimal paths and significant amount of signaling messages is generated due to the home registration procedure which keeps the network aware of the current location of the mobile nodes. Recently, a dynamic hierarchical mobility management strategy for mobile IP networks (DHMIP) has been proposed to reduce home registrations costs. However, this strategy induces a packet delivery delay and increases the risk of packet loss. In this paper, we propose an enhanced version of the dynamic hierarchical strategy that reduces the packet delivery delay and minimizes the risk of packet loss. Preliminary results obtained from simulations are promising. They show that the enhanced version outperforms the original dynamic hierarchical mobility management strategy version.

A Face-to-Face Education Support System Capable of Lecture Adaptation and Q&A Assistance Based On Probabilistic Inference

Keys to high-quality face-to-face education are ensuring flexibility in the way lectures are given, and providing care and responsiveness to learners. This paper describes a face-to-face education support system that is designed to raise the satisfaction of learners and reduce the workload on instructors. This system consists of a lecture adaptation assistance part, which assists instructors in adapting teaching content and strategy, and a Q&A assistance part, which provides learners with answers to their questions. The core component of the former part is a “learning achievement map", which is composed of a Bayesian network (BN). From learners- performance in exercises on relevant past lectures, the lecture adaptation assistance part obtains information required to adapt appropriately the presentation of the next lecture. The core component of the Q&A assistance part is a case base, which accumulates cases consisting of questions expected from learners and answers to them. The Q&A assistance part is a case-based search system equipped with a search index which performs probabilistic inference. A prototype face-to-face education support system has been built, which is intended for the teaching of Java programming, and this approach was evaluated using this system. The expected degree of understanding of each learner for a future lecture was derived from his or her performance in exercises on past lectures, and this expected degree of understanding was used to select one of three adaptation levels. A model for determining the adaptation level most suitable for the individual learner has been identified. An experimental case base was built to examine the search performance of the Q&A assistance part, and it was found that the rate of successfully finding an appropriate case was 56%.

Sustainable Development in Disaster Affected Rural Areas: The Case of Dinar Villages

Post-disaster reconstruction projects offer opportunities to facilitate physical, social and economic development and to reduce future hazard vulnerability long after the disasters. Sustainability of post-disaster reconstruction project conducted in the villages of Dinar following the 1995 earthquake was investigated in this paper. Officials of the Government who were involved in the project were interviewed. Besides, two field surveys were done in 12 villages of Dinar in winter months of 2008. Beneficiaries were interviewed and physical, socio-cultural and economic impacts of the reconstruction were examined. The research revealed that the postdisaster reconstruction project has negative aspects from the point view of sustainability. The physical, socio-cultural and economic factors were not considered during decision making process of the project.

Time-Cost-Quality Trade-off Software by using Simplified Genetic Algorithm for Typical Repetitive Construction Projects

Time-Cost Optimization "TCO" is one of the greatest challenges in construction project planning and control, since the optimization of either time or cost, would usually be at the expense of the other. Since there is a hidden trade-off relationship between project and cost, it might be difficult to predict whether the total cost would increase or decrease as a result of the schedule compression. Recently third dimension in trade-off analysis is taken into consideration that is quality of the projects. Few of the existing algorithms are applied in a case of construction project with threedimensional trade-off analysis, Time-Cost-Quality relationships. The objective of this paper is to presents the development of a practical software system; that named Automatic Multi-objective Typical Construction Resource Optimization System "AMTCROS". This system incorporates the basic concepts of Line Of Balance "LOB" and Critical Path Method "CPM" in a multi-objective Genetic Algorithms "GAs" model. The main objective of this system is to provide a practical support for typical construction planners who need to optimize resource utilization in order to minimize project cost and duration while maximizing its quality simultaneously. The application of these research developments in planning the typical construction projects holds a strong promise to: 1) Increase the efficiency of resource use in typical construction projects; 2) Reduce construction duration period; 3) Minimize construction cost (direct cost plus indirect cost); and 4) Improve the quality of newly construction projects. A general description of the proposed software for the Time-Cost-Quality Trade-Off "TCQTO" is presented. The main inputs and outputs of the proposed software are outlined. The main subroutines and the inference engine of this software are detailed. The complexity analysis of the software is discussed. In addition, the verification, and complexity of the proposed software are proved and tested using a real case study.

Water Security in Rural Areas through Solar Energy in Baja California Sur, Mexico

This study aims to assess the potential of solar energy technology for improving access to water and hence the livelihood strategies of rural communities in Baja California Sur, Mexico. It focuses on livestock ranches and photovoltaic water-pumptechnology as well as other water extraction methods. The methodology used are the Sustainable Livelihoods and the Appropriate Technology approaches. A household survey was applied in June of 2006 to 32 ranches in the municipality, of which 22 used PV pumps; and semi-structured interviews were conducted. Findings indicate that solar pumps have in fact helped people improve their quality of life by allowing them to pursue a different livelihood strategy and that improved access to water -not necessarily as more water but as less effort to extract and collect it- does not automatically imply overexploitation of the resource; consumption is based on basic needs as well as on storage and pumping capacity. Justification for such systems lies in the avoidance of logistical problems associated to fossil fuels, PV pumps proved to be the most beneficial when substituting gasoline or diesel equipment but of dubious advantage if intended to replace wind or gravity systems. Solar water pumping technology-s main obstacle to dissemination are high investment and repairs costs and it is therefore not suitable for all cases even when insolation rates and water availability are adequate. In cases where affordability is not an obstacle it has become an important asset that contributes –by means of reduced expenses, less effort and saved time- to the improvement of livestock, the main livelihood provider for these ranches.

The Anti-Noise and Anti-Wear Systems for Railways

In recent years there has been a continuous increase of axle loads, tonnage, train speed and train length which has increased both the productivity in the rail sector and the risk of rail breaks and derailments. On the other hand, the environmental requirements (e.g. noise reduction) for railway operations will become tighter in the future. In our research we developed a new composite material which does not change braking properties, is capable of taking extremely high pressure loads, reduces noise and is environmentally friendly. Part of our research was also the development of technology which will be able to apply this material to the rail. The result of our research was the system which reduces the wear out significantly and almost completely eliminates the squealing noise at the same time, and by using only one special material.

Optimization Based Tuning of Autopilot Gains for a Fixed Wing UAV

Unmanned Aerial Vehicles (UAVs) have gained tremendous importance, in both Military and Civil, during first decade of this century. In a UAV, onboard computer (autopilot) autonomously controls the flight and navigation of the aircraft. Based on the aircraft role and flight envelope, basic to complex and sophisticated controllers are used to stabilize the aircraft flight parameters. These controllers constitute the autopilot system for UAVs. The autopilot systems, most commonly, provide lateral and longitudinal control through Proportional-Integral-Derivative (PID) controllers or Phase-lead or Lag Compensators. Various techniques are commonly used to ‘tune’ gains of these controllers. Some techniques used are, in-flight step-by-step tuning, software-in-loop or hardware-in-loop tuning methods. Subsequently, numerous in-flight tests are required to actually ‘fine-tune’ these gains. However, an optimization-based tuning of these PID controllers or compensators, as presented in this paper, can greatly minimize the requirement of in-flight ‘tuning’ and substantially reduce the risks and cost involved in flight-testing.

Effect of Coolant on Cutting Forces and Surface Roughness in Grinding of CSM GFRP

This paper presents a comparative study on dry and wet grinding through experimental investigation in the grinding of CSM glass fibre reinforced polymer laminates using a pink aluminium oxide wheel. Different sets of experiments were performed to study the effects of the independent grinding parameters such as grinding wheel speed, feed and depth of cut on dependent performance criteria such as cutting forces and surface finish. Experimental conditions were laid out using design of experiment central composite design. An effective coolant was sought in this study to minimise cutting forces and surface roughness for GFRP laminates grinding. Test results showed that the use of coolants reduces surface roughness, although not necessarily the cutting forces. These research findings provide useful economic machining solution in terms of optimized grinding conditions for grinding CSM GFRP.

A Prototype of Augmented Reality for Visualising Large Sensors’ Datasets

In this paper we discuss the development of an Augmented Reality (AR) - based scientific visualization system prototype that supports identification, localisation, and 3D visualisation of oil leakages sensors datasets. Sensors generates significant amount of multivariate datasets during normal and leak situations. Therefore we have developed a data model to effectively manage such data and enhance the computational support needed for the effective data explorations. A challenge of this approach is to reduce the data inefficiency powered by the disparate, repeated, inconsistent and missing attributes of most available sensors datasets. To handle this challenge, this paper aim to develop an AR-based scientific visualization interface which automatically identifies, localise and visualizes all necessary data relevant to a particularly selected region of interest (ROI) along the virtual pipeline network. Necessary system architectural supports needed as well as the interface requirements for such visualizations are also discussed in this paper.

Enhance Performance of Secure Image Using Wavelet Compression

The increase popularity of multimedia application especially in image processing places a great demand on efficient data storage and transmission techniques. Network communication such as wireless network can easily be intercepted and cause of confidential information leaked. Unfortunately, conventional compression and encryption methods are too slow; it is impossible to carry out real time secure image processing. In this research, Embedded Zerotree Wavelet (EZW) encoder which specially designs for wavelet compression is examined. With this algorithm, three methods are proposed to reduce the processing time, space and security protection that will be secured enough to protect the data.

Energy Efficient Clustering Algorithm with Global and Local Re-clustering for Wireless Sensor Networks

Wireless Sensor Networks consist of inexpensive, low power sensor nodes deployed to monitor the environment and collect data. Gathering information in an energy efficient manner is a critical aspect to prolong the network lifetime. Clustering  algorithms have an advantage of enhancing the network lifetime. Current clustering algorithms usually focus on global re-clustering and local re-clustering separately. This paper, proposed a combination of those two reclustering methods to reduce the energy consumption of the network. Furthermore, the proposed algorithm can apply to homogeneous as well as heterogeneous wireless sensor networks. In addition, the cluster head rotation happens, only when its energy drops below a dynamic threshold value computed by the algorithm. The simulation result shows that the proposed algorithm prolong the network lifetime compared to existing algorithms.

Monte Carlo Simulation of Copolymer Heterogeneity in Atom Transfer Radical Copolymerization of Styrene and N-Butyl Acrylate

A high-performance Monte Carlo simulation, which simultaneously takes diffusion-controlled and chain-length-dependent bimolecular termination reactions into account, is developed to simulate atom transfer radical copolymerization of styrene and nbutyl acrylate. As expected, increasing initial feed fraction of styrene raises the fraction of styrene-styrene dyads (fAA) and reduces that of n-butyl acrylate dyads (fBB). The trend of variation in randomness parameter (fAB) during the copolymerization also varies significantly. Also, there is a drift in copolymer heterogeneity and the highest drift occurs in the initial feeds containing lower percentages of styrene, i.e. 20% and 5%.

High Speed Video Transmission for Telemedicine using ATM Technology

In this paper, we study statistical multiplexing of VBR video in ATM networks. ATM promises to provide high speed realtime multi-point to central video transmission for telemedicine applications in rural hospitals and in emergency medical services. Video coders are known to produce variable bit rate (VBR) signals and the effects of aggregating these VBR signals need to be determined in order to design a telemedicine network infrastructure capable of carrying these signals. We first model the VBR video signal and simulate it using a generic continuous-data autoregressive (AR) scheme. We carry out the queueing analysis by the Fluid Approximation Model (FAM) and the Markov Modulated Poisson Process (MMPP). The study has shown a trade off: multiplexing VBR signals reduces burstiness and improves resource utilization, however, the buffer size needs to be increased with an associated economic cost. We also show that the MMPP model and the Fluid Approximation model fit best, respectively, the cell region and the burst region. Therefore, a hybrid MMPP and FAM completely characterizes the overall performance of the ATM statistical multiplexer. The ramifications of this technology are clear: speed, reliability (lower loss rate and jitter), and increased capacity in video transmission for telemedicine. With migration to full IP-based networks still a long way to achieving both high speed and high quality of service, the proposed ATM architecture will remain of significant use for telemedicine.

Broadband PowerLine Communications: Performance Analysis

Power line channel is proposed as an alternative for broadband data transmission especially in developing countries like Tanzania [1]. However the channel is affected by stochastic attenuation and deep notches which can lead to the limitation of channel capacity and achievable data rate. Various studies have characterized the channel without giving exactly the maximum performance and limitation in data transfer rate may be this is due to complexity of channel modeling being used. In this paper the channel performance of medium voltage, low voltage and indoor power line channel is presented. In the investigations orthogonal frequency division multiplexing (OFDM) with phase shift keying (PSK) as carrier modulation schemes is considered, for indoor, medium and low voltage channels with typical ten branches and also Golay coding is applied for medium voltage channel. From channels, frequency response deep notches are observed in various frequencies which can lead to reduce the achievable data rate. However, is observed that data rate up to 240Mbps is realized for a signal to noise ratio of about 50dB for indoor and low voltage channels, however for medium voltage a typical link with ten branches is affected by strong multipath and coding is required for feasible broadband data transfer.

Trapping Efficiency of Diesel Particles Through a Square Duct

Diesel Engines emit complex mixtures of inorganic and organic compounds in the form of both solid and vapour phase particles. Most of the particulates released are ultrafine nanoparticles which are detrimental to human health and can easily enter the body by respiration. The emissions standards on particulate matter release from diesel engines are constantly upgraded within the European Union and with future regulations based on the particles numbers released instead of merely mass, the need for effective aftertreatment devices will increase. Standard particulate filters in the form of wall flow filters can have problems with high soot accumulation, producing a large exhaust backpressure. A potential solution would be to combine the standard filter with a flow through filter to reduce the load on the wall flow filter. In this paper soot particle trapping has been simulated in different continuous flow filters of monolithic structure including the use of promoters, at laminar flow conditions. An Euler Lagrange model, the discrete phase model in Ansys used with user defined functions for forces acting on particles. A method to quickly screen trapping of 5 nm and 10 nm particles in different catalysts designs with tracers was also developed. Simulations of square duct monoliths with promoters show that the strength of the vortices produced are not enough to give a high amount of particle deposition on the catalyst walls. The smallest particles in the simulations, 5 and 10 nm particles were trapped to a higher extent, than larger particles up to 1000 nm, in all studied geometries with the predominant deposition mechanism being Brownian diffusion. The comparison of the different filters designed with a wall flow filter does show that the options for altering a design of a flow through filter, without imposing a too large pressure drop penalty are good.

A Study of the Hand-Hold Impact on the EM Interaction of a Cellular Handset and a Human

This paper investigates the impact of the hand-hold positions on both antenna performance and the specific absorption rate (SAR) induced in the user-s head. A cellular handset with external antenna operating at GSM-900 frequency is modeled and simulated using a finite difference time-domain (FDTD)-based platform SEMCAD-X. A specific anthropomorphic mannequin (SAM) is adopted to simulate the user-s head, whereas a semirealistic CAD-model of three-tissues is designed to simulate the user-s hand. The results show that in case of the handset in hand close to head at different positions; the antenna total efficiency gets reduced to (14.5% - 5.9%) at cheek-position and to (27.5% to 11.8%) at tilt-position. The peak averaged SAR1g values in head close to handset without hand, are 4.67 W/Kg and 2.66 W/Kg at cheek and tilt-position, respectively. Due to the presence of hand, the SAR1g in head gets reduced to (3.67-3.31 W/Kg) at cheek-position and to (1.84-1.64 W/Kg) at tilt-position, depending on the hand-hold position.

An Experimental Study of a Self-Supervised Classifier Ensemble

Learning using labeled and unlabelled data has received considerable amount of attention in the machine learning community due its potential in reducing the need for expensive labeled data. In this work we present a new method for combining labeled and unlabeled data based on classifier ensembles. The model we propose assumes each classifier in the ensemble observes the input using different set of features. Classifiers are initially trained using some labeled samples. The trained classifiers learn further through labeling the unknown patterns using a teaching signals that is generated using the decision of the classifier ensemble, i.e. the classifiers self-supervise each other. Experiments on a set of object images are presented. Our experiments investigate different classifier models, different fusing techniques, different training sizes and different input features. Experimental results reveal that the proposed self-supervised ensemble learning approach reduces classification error over the single classifier and the traditional ensemble classifier approachs.