Diagnosing the Cause and its Timing of Changes in Multivariate Process Mean Vector from Quality Control Charts using Artificial Neural Network

Quality control charts are very effective in detecting out of control signals but when a control chart signals an out of control condition of the process mean, searching for a special cause in the vicinity of the signal time would not always lead to prompt identification of the source(s) of the out of control condition as the change point in the process parameter(s) is usually different from the signal time. It is very important to manufacturer to determine at what point and which parameters in the past caused the signal. Early warning of process change would expedite the search for the special causes and enhance quality at lower cost. In this paper the quality variables under investigation are assumed to follow a multivariate normal distribution with known means and variance-covariance matrix and the process means after one step change remain at the new level until the special cause is being identified and removed, also it is supposed that only one variable could be changed at the same time. This research applies artificial neural network (ANN) to identify the time the change occurred and the parameter which caused the change or shift. The performance of the approach was assessed through a computer simulation experiment. The results show that neural network performs effectively and equally well for the whole shift magnitude which has been considered.

Genetic Algorithm Parameters Optimization for Bi-Criteria Multiprocessor Task Scheduling Using Design of Experiments

Multiprocessor task scheduling is a NP-hard problem and Genetic Algorithm (GA) has been revealed as an excellent technique for finding an optimal solution. In the past, several methods have been considered for the solution of this problem based on GAs. But, all these methods consider single criteria and in the present work, minimization of the bi-criteria multiprocessor task scheduling problem has been considered which includes weighted sum of makespan & total completion time. Efficiency and effectiveness of genetic algorithm can be achieved by optimization of its different parameters such as crossover, mutation, crossover probability, selection function etc. The effects of GA parameters on minimization of bi-criteria fitness function and subsequent setting of parameters have been accomplished by central composite design (CCD) approach of response surface methodology (RSM) of Design of Experiments. The experiments have been performed with different levels of GA parameters and analysis of variance has been performed for significant parameters for minimisation of makespan and total completion time simultaneously.

CFD Modeling of a Radiator Axial Fan for Air Flow Distribution

The fluid mechanics principle is used extensively in designing axial flow fans and their associated equipment. This paper presents a computational fluid dynamics (CFD) modeling of air flow distribution from a radiator axial flow fan used in an acid pump truck Tier4 (APT T4) Repower. This axial flow fan augments the transfer of heat from the engine mounted on the APT T4. CFD analysis was performed for an area weighted average static pressure difference at the inlet and outlet of the fan. Pressure contours, velocity vectors, and path lines were plotted for detailing the flow characteristics for different orientations of the fan blade. The results were then compared and verified against known theoretical observations and actual experimental data. This study shows that a CFD simulation can be very useful for predicting and understanding the flow distribution from a radiator fan for further research work.

A New Biometric Human Identification Based On Fusion Fingerprints and Finger Veins Using monoLBP Descriptor

Single biometric modality recognition is not able to meet the high performance supplies in most cases with its application become more and more broadly. Multimodal biometrics identification represents an emerging trend recently. This paper investigates a novel algorithm based on fusion of both fingerprint and fingervein biometrics. For both biometric recognition, we employ the Monogenic Local Binary Pattern (MonoLBP). This operator integrate the orginal LBP (Local Binary Pattern ) with both other rotation invariant measures: local phase and local surface type. Experimental results confirm that a weighted sum based proposed fusion achieves excellent identification performances opposite unimodal biometric systems. The AUC of proposed approach based on combining the two modalities has very close to unity (0.93).

Physical Properties and Stability of Emulsions as Affected by Native and Modified Yam Starches

This study was conducted in order to determine the physical properties and stability of mayonnaise-like emulsions as affected by modified yam starches. Native yam starch was modified via pre-gelatinization and cross-linking phosphorylation procedures. The emulsions (50% oil dispersed phase) were prepared with 0.3% native potato, native yam, pre-gelatinized yam and cross-linking phosphorylation yam starches. The droplet size of surface weighted mean diameter was found to be significantly (p < 0.05) lower in the sample with cross-linking phosphorylation yam starch as compared to other samples. Moreover, the viscosity of the sample with pregelatinized yam starch was observed to be higher than that of other samples. The phase separation stability was low in the freshly prepared and stored (45 days, 5°C) emulsions containing native yam starch. This study thus generally suggested that modified yam starches were more suitable (i.e. better physical properties and stability) to be used as stabilizers in a similar system i.e. light mayonnaises, rather than a native yam starch.

A Survey on Metric of Software Cognitive Complexity for OO design

In modern era, the biggest challenge facing the software industry is the upcoming of new technologies. So, the software engineers are gearing up themselves to meet and manage change in large software system. Also they find it difficult to deal with software cognitive complexities. In the last few years many metrics were proposed to measure the cognitive complexity of software. This paper aims at a comprehensive survey of the metric of software cognitive complexity. Some classic and efficient software cognitive complexity metrics, such as Class Complexity (CC), Weighted Class Complexity (WCC), Extended Weighted Class Complexity (EWCC), Class Complexity due to Inheritance (CCI) and Average Complexity of a program due to Inheritance (ACI), are discussed and analyzed. The comparison and the relationship of these metrics of software complexity are also presented.

The Household-Based Socio-Economic Index for Every District in Peninsular Malaysia

Deprivation indices are widely used in public health study. These indices are also referred as the index of inequalities or disadvantage. Even though, there are many indices that have been built before, it is believed to be less appropriate to use the existing indices to be applied in other countries or areas which had different socio-economic conditions and different geographical characteristics. The objective of this study is to construct the index based on the geographical and socio-economic factors in Peninsular Malaysia which is defined as the weighted household-based deprivation index. This study has employed the variables based on household items, household facilities, school attendance and education level obtained from Malaysia 2000 census report. The factor analysis is used to extract the latent variables from indicators, or reducing the observable variable into smaller amount of components or factor. Based on the factor analysis, two extracted factors were selected, known as Basic Household Amenities and Middle-Class Household Item factor. It is observed that the district with a lower index values are located in the less developed states like Kelantan, Terengganu and Kedah. Meanwhile, the areas with high index values are located in developed states such as Pulau Pinang, W.P. Kuala Lumpur and Selangor.

Health Hazards Related to Computer Use: Experience of the National Institute for Medical Research in Tanzania

This paper is based on a study conducted in 2006 to assess the impact of computer usage on health of National Institute for Medical Research (NIMR) staff. NIMR being a research Institute, most of its staff spend substantial part of their working time on computers. There was notion among NIMR staff on possible prolonged computer usage health hazards. Hence, a study was conducted to establish facts and possible mitigation measures. A total of 144 NIMR staff were involved in the study of whom 63.2% were males and 36.8% females aged between 20 and 59 years. All staff cadres were included in the sample. The functions performed by Institute staff using computers includes; data management, proposal development and report writing, research activities, secretarial duties, accounting and administrative duties, on-line information retrieval and online communication through e-mail services. The interviewed staff had been using computers for 1-8 hours a day and for a period ranging from 1 to 20 years. The study has indicated ergonomic hazards for a significant proportion of interviewees (63%) of various kinds ranging from backache to eyesight related problems. The authors highlighted major issues which are substantially applicable in preventing occurrences of computer related problems and they urged NIMR Management and/or the government of Tanzania opts to adapt their practicability.

Attribute Weighted Class Complexity: A New Metric for Measuring Cognitive Complexity of OO Systems

In general, class complexity is measured based on any one of these factors such as Line of Codes (LOC), Functional points (FP), Number of Methods (NOM), Number of Attributes (NOA) and so on. There are several new techniques, methods and metrics with the different factors that are to be developed by the researchers for calculating the complexity of the class in Object Oriented (OO) software. Earlier, Arockiam et.al has proposed a new complexity measure namely Extended Weighted Class Complexity (EWCC) which is an extension of Weighted Class Complexity which is proposed by Mishra et.al. EWCC is the sum of cognitive weights of attributes and methods of the class and that of the classes derived. In EWCC, a cognitive weight of each attribute is considered to be 1. The main problem in EWCC metric is that, every attribute holds the same value but in general, cognitive load in understanding the different types of attributes cannot be the same. So here, we are proposing a new metric namely Attribute Weighted Class Complexity (AWCC). In AWCC, the cognitive weights have to be assigned for the attributes which are derived from the effort needed to understand their data types. The proposed metric has been proved to be a better measure of complexity of class with attributes through the case studies and experiments

Mathematical Rescheduling Models for Railway Services

This paper presents the review of past studies concerning mathematical models for rescheduling passenger railway services, as part of delay management in the occurrence of railway disruption. Many past mathematical models highlighted were aimed at minimizing the service delays experienced by passengers during service disruptions. Integer programming (IP) and mixed-integer programming (MIP) models are critically discussed, focusing on the model approach, decision variables, sets and parameters. Some of them have been tested on real-life data of railway companies worldwide, while a few have been validated on fictive data. Based on selected literatures on train rescheduling, this paper is able to assist researchers in the model formulation by providing comprehensive analyses towards the model building. These analyses would be able to help in the development of new approaches in rescheduling strategies or perhaps to enhance the existing rescheduling models and make them more powerful or more applicable with shorter computing time.

The Frequency Graph for the Traveling Salesman Problem

Traveling salesman problem (TSP) is hard to resolve when the number of cities and routes become large. The frequency graph is constructed to tackle the problem. A frequency graph maintains the topological relationships of the original weighted graph. The numbers on the edges are the frequencies of the edges emulated from the local optimal Hamiltonian paths. The simplest kind of local optimal Hamiltonian paths are computed based on the four vertices and three lines inequality. The search algorithm is given to find the optimal Hamiltonian circuit based on the frequency graph. The experiments show that the method can find the optimal Hamiltonian circuit within several trials.

A Software Framework for Predicting Oil-Palm Yield from Climate Data

Intelligent systems based on machine learning techniques, such as classification, clustering, are gaining wide spread popularity in real world applications. This paper presents work on developing a software system for predicting crop yield, for example oil-palm yield, from climate and plantation data. At the core of our system is a method for unsupervised partitioning of data for finding spatio-temporal patterns in climate data using kernel methods which offer strength to deal with complex data. This work gets inspiration from the notion that a non-linear data transformation into some high dimensional feature space increases the possibility of linear separability of the patterns in the transformed space. Therefore, it simplifies exploration of the associated structure in the data. Kernel methods implicitly perform a non-linear mapping of the input data into a high dimensional feature space by replacing the inner products with an appropriate positive definite function. In this paper we present a robust weighted kernel k-means algorithm incorporating spatial constraints for clustering the data. The proposed algorithm can effectively handle noise, outliers and auto-correlation in the spatial data, for effective and efficient data analysis by exploring patterns and structures in the data, and thus can be used for predicting oil-palm yield by analyzing various factors affecting the yield.

Dextran/Poly(L-histidine) Graft Copolymer for pH-Responsive Drug Delivery

pH-sensitive drug targeting using nanoparticles for cancer chemotherapy have been spotlighted in recent decades. Graft copolymer composed of poly (L-histidine) (PHS) and dextran (DexPHS) was synthesized and pH-sensitive nanoparticles were fabricated for pH-responsive drug delivery of doxorubicin (DOX). Nanoparticles of DexPHS showed pH-sensitive changes in particle sizes and drug release behavior, i.e. particle sizes and drug release rate were increased at acidic pH, indicating that DexPHS nanoparticles have pH-sensitive drug delivery potentials. Antitumor activity of DOX-incorporated DexPHS nanoparticles were studied using CT26 colorectal carcinoma cells. Results indicated that fluorescence intensity was higher at acidic pH than basic pH. These results indicated that DexPHS nanoparticles have pH-responsive drug targeting.

Designing a Rescue System for Earthquake-Stricken Area with the Aim of Facilitation and Accelerating Accessibilities (Case Study: City of Tehran)

Natural disasters, including earthquake, kill many people around the world every year. Society rescue actions, which start after the earthquake and are called LAST in abbreviation, include locating, access, stabilization and transportation. In the present article, we have studied the process of local accessibility to the injured and transporting them to health care centers. With regard the heavy traffic load due to earthquake, the destruction of connecting roads and bridges and the heavy debris in alleys and street, which put the lives of the injured and the people buried under the debris in danger, accelerating the rescue actions and facilitating the accessibilities are of great importance, obviously. Tehran, the capital of Iran, is among the crowded cities in the world and is the center of extensive economic, political, cultural and social activities. Tehran has a population of about 9.5 millions and because of the immigration of people from the surrounding cities. Furthermore, considering the fact that Tehran is located on two important and large faults, a 6 Richter magnitude earthquake in this city could lead to the greatest catastrophe during the entire human history. The present study is a kind of review and a major part of the required information for it, has been obtained from libraries all of the rescue vehicles around the world, including rescue helicopters, ambulances, fire fighting vehicles and rescue boats, and their applied technology, and also the robots specifically designed for the rescue system and the advantages and disadvantages of them, have been investigated. The studies show that there is a significant relationship between the rescue team-s arrival time at the incident zone and the number of saved people; so that, if the duration of burial under debris 30 minutes, the probability of survival is %99.3, after a day is %81, after 2days is %19 and after 5days is %7.4. The exiting transport systems all have some defects. If these defects are removed, more people could be saved each hour and the preparedness against natural disasters is increased. In this study, transport system has been designed for the rescue team and the injured; which could carry the rescue team to the incident zone and the injured to the health care centers. In addition, this system is able to fly in the air and move on the earth as well; so that the destruction of roads and the heavy traffic load could not prevent the rescue team from arriving early at the incident zone. The system also has the equipment required firebird for debris removing, optimum transport of the injured and first aid.

A Weighted Sum Technique for the Joint Optimization of Performance and Power Consumption in Data Centers

With data centers, end-users can realize the pervasiveness of services that will be one day the cornerstone of our lives. However, data centers are often classified as computing systems that consume the most amounts of power. To circumvent such a problem, we propose a self-adaptive weighted sum methodology that jointly optimizes the performance and power consumption of any given data center. Compared to traditional methodologies for multi-objective optimization problems, the proposed self-adaptive weighted sum technique does not rely on a systematical change of weights during the optimization procedure. The proposed technique is compared with the greedy and LR heuristics for large-scale problems, and the optimal solution for small-scale problems implemented in LINDO. the experimental results revealed that the proposed selfadaptive weighted sum technique outperforms both of the heuristics and projects a competitive performance compared to the optimal solution.

Weighted Harmonic Arnoldi Method for Large Interior Eigenproblems

The harmonic Arnoldi method can be used to find interior eigenpairs of large matrices. However, it has been shown that this method may converge erratically and even may fail to do so. In this paper, we present a new method for computing interior eigenpairs of large nonsymmetric matrices, which is called weighted harmonic Arnoldi method. The implementation of the method has been tested by numerical examples, the results show that the method converges fast and works with high accuracy.

Local Image Descriptor using VQ-SIFT for Image Retrieval

In this paper, we present local image descriptor using VQ-SIFT for more effective and efficient image retrieval. Instead of SIFT's weighted orientation histograms, we apply vector quantization (VQ) histogram as an alternate representation for SIFT features. Experimental results show that SIFT features using VQ-based local descriptors can achieve better image retrieval accuracy than the conventional algorithm while the computational cost is significantly reduced.

The Effect of Contrived Success in Calculation Tasks on the Self-efficacy of Junior High School Students

This study examines whether contrived success on a task closely related to school subjects would promote students- self-efficacy. In our previous study, junior high school students who experienced contrived success on anagram tasks raised their sense of self-efficacy and kept it high for a year.We tried to replicate that study, substituting calculation tasks for the anagrams. One hundred eighteen junior high school students participated in this study, 18 of whom were surreptitiously given easier tasks than their classmates. Those students with easier tasks outperformed their peers and thereby raised their sense of self-efficacy. However, elevated self-efficacy did not persist, falling to the starting level after only three months.

Experimental and Numerical Study of the Effect of Lateral Wind on the Feeder Airship

Feeder is one of the airships of the Multibody Advanced Airship for Transport (MAAT) system, under development within the EU FP7 project. MAAT is based on a modular concept composed of two different parts that have the possibility to join; respectively they are the so-called Cruiser and Feeder, designed on the lighter than air principle. Feeder, also named ATEN (Airship Transport Elevator Network), is the smaller one which joins the bigger one, Cruiser, also named PTAH (Photovoltaic modular Transport Airship for High altitude),envisaged to happen at 15km altitude. During the MAAT design phase, the aerodynamic studies of the both airships and their interactions are analyzed. The objective of these studies is to understand the aerodynamic behavior of all the preselected configurations, as an important element in the overall MAAT system design. The most of these configurations are only simulated by CFD, while the most feasible one is experimentally analyzed in order to validate and thrust the CFD predictions. This paper presents the numerical and experimental investigation of the Feeder “conical like" shape configuration. The experiments are focused on the aerodynamic force coefficients and the pressure distribution over the Feeder outer surface, while the numerical simulation cover also the analysis of the velocity and pressure distribution. Finally, the wind tunnel experiment is compared with its CFD model in order to validate such specific simulations with respective experiments and to better understand the difference between the wind tunnel and in-flight circumstances.

Phytoremediation of Wastewater Using Some of Aquatic Macrophytes as Biological Purifiers for Irrigation Purposes

An attempt was made for availability of wastewater reuse/reclamation for irrigation purposes using phytoremediation “the low cost and less technology", using six local aquatic macrophytes “e.g. T. angustifolia, B. maritimus, Ph. australis, A. donax, A. plantago-aquatica and M. longifolia (Linn)" as biological waste purifiers. Outdoor experiments/designs were conducted from May 03, 2007 till October 15, 2008, close to one of the main sewage channels of Sulaimani City/Iraq*. All processes were mainly based on conventional wastewater treatment processes, besides two further modifications were tested, the first was sand filtration pots, implanted by individual species of experimental macrophytes and the second was constructed wetlands implanted by experimental macrophytes all together. Untreated and treated wastewater samples were analyzed for their key physico-chemical properties (only heavy metals Fe, Mn, Zn and Cu with particular reference to removal efficiency by experimental macrophytes are highlighted in this paper). On the other hand, vertical contents of heavy metals were also evaluated from both pots and the cells of constructed wetland. After 135 days, macrophytes were harvested and heavy metals were analyzed in their biomass (roots/shoots) for removal efficiency assessment (i.e. uptake/ bioaccumulation rate). Results showed that; removal efficiency of all studied heavy metals was much higher in T. angustifolia followed by Ph. Australis, B. maritimus and A. donax in triple experiment sand pots. Constructed wetland experiments have revealed that; the more replicated constructed wetland cells the highest heavy metal removal efficiency was indicated.