The Investigation of Motor Cooling Performance

This study experimentally and numerically investigates motor cooling performance. The motor consists of a centrifugal fan, two axial fans, a shaft, a stator, a rotor and a heat exchanger with 637 cooling tubes. The pressure rise-flow rate (P-Q) performance curves of the cooling fans at 1800 rpm are tested using a test apparatus complying with the Chinese National Standard (CNS) 2726. Compared with the experimental measurements, the numerical analysis results show that the P-Q performance curves of the axial fan and centrifugal fan can be estimated within about 2% and 6%, respectively. By using the simplified model, setting up the heat exchanger and stator as porous media, the flow field in the motor is calculated. By using the results of the flow field near the rotor and stator, and subjecting the heat generation rate as a boundary condition, the temperature distributions of the stator and rotor are also calculated. The simulation results show that the calculated temperature of the stator winding near the axial fans is lower by about 5% than the measured value, and the calculated temperature of the stator core located at the center of the stator is about 1% higher than the measured value. Besides, discussion is made to improve the motor cooling performance.

A Parameter-Tuning Framework for Metaheuristics Based on Design of Experiments and Artificial Neural Networks

In this paper, a framework for the simplification and standardization of metaheuristic related parameter-tuning by applying a four phase methodology, utilizing Design of Experiments and Artificial Neural Networks, is presented. Metaheuristics are multipurpose problem solvers that are utilized on computational optimization problems for which no efficient problem specific algorithm exist. Their successful application to concrete problems requires the finding of a good initial parameter setting, which is a tedious and time consuming task. Recent research reveals the lack of approach when it comes to this so called parameter-tuning process. In the majority of publications, researchers do have a weak motivation for their respective choices, if any. Because initial parameter settings have a significant impact on the solutions quality, this course of action could lead to suboptimal experimental results, and thereby a fraudulent basis for the drawing of conclusions.

Chemotherapy Safety Protocol for Oncology Nurses: It's Effect on Their Protective Measures Practices

Background: Widespread use of chemotherapeutic drugs in the treatment of cancer has lead to higher health hazards among employee who handle and administer such drugs, so nurses should know how to protect themselves, their patients and their work environment against toxic effects of chemotherapy. Aim of this study was carried out to examine the effect of chemotherapy safety protocol for oncology nurses on their protective measure practices. Design: A quasi experimental research design was utilized. Setting: The study was carried out in oncology department of Menoufia university hospital and Tanta oncology treatment center. Sample: A convenience sample of forty five nurses in Tanta oncology treatment center and eighteen nurses in Menoufiya oncology department. Tools: 1. an interviewing questionnaire that covering sociodemographic data, assessment of unit and nurses' knowledge about chemotherapy. II: Obeservational check list to assess nurses' actual practices of handling and adminestration of chemotherapy. A base line data were assessed before implementing Chemotherapy Safety protocol, then Chemotherapy Safety protocol was implemented, and after 2 monthes they were assessed again. Results: reveled that 88.9% of study group I and 55.6% of study group II improved to good total knowledge scores after educating on the safety protocol, also 95.6% of study group I and 88.9% of study group II had good total practice score after educating on the safety protocol. Moreover less than half of group I (44.4%) reported that heavy workload is the most barriers for them, while the majority of group II (94.4%) had many barriers for adhering to the safety protocol such as they didn’t know the protocol, the heavy work load and inadequate equipment. Conclusions: Safety protocol for Oncology Nurses seemed to have positive effect on improving nurses' knowledge and practice. Recommendation: chemotherapy safety protocol should be instituted for all oncology nurses who are working in any oncology unit and/ or center to enhance compliance, and this protocol should be done at frequent intervals.

Optimum Time Coordination of Overcurrent Relays using Two Phase Simplex Method

Overcurrent (OC) relays are the major protection devices in a distribution system. The operating time of the OC relays are to be coordinated properly to avoid the mal-operation of the backup relays. The OC relay time coordination in ring fed distribution networks is a highly constrained optimization problem which can be stated as a linear programming problem (LPP). The purpose is to find an optimum relay setting to minimize the time of operation of relays and at the same time, to keep the relays properly coordinated to avoid the mal-operation of relays. This paper presents two phase simplex method for optimum time coordination of OC relays. The method is based on the simplex algorithm which is used to find optimum solution of LPP. The method introduces artificial variables to get an initial basic feasible solution (IBFS). Artificial variables are removed using iterative process of first phase which minimizes the auxiliary objective function. The second phase minimizes the original objective function and gives the optimum time coordination of OC relays.

Genetic Algorithm Parameters Optimization for Bi-Criteria Multiprocessor Task Scheduling Using Design of Experiments

Multiprocessor task scheduling is a NP-hard problem and Genetic Algorithm (GA) has been revealed as an excellent technique for finding an optimal solution. In the past, several methods have been considered for the solution of this problem based on GAs. But, all these methods consider single criteria and in the present work, minimization of the bi-criteria multiprocessor task scheduling problem has been considered which includes weighted sum of makespan & total completion time. Efficiency and effectiveness of genetic algorithm can be achieved by optimization of its different parameters such as crossover, mutation, crossover probability, selection function etc. The effects of GA parameters on minimization of bi-criteria fitness function and subsequent setting of parameters have been accomplished by central composite design (CCD) approach of response surface methodology (RSM) of Design of Experiments. The experiments have been performed with different levels of GA parameters and analysis of variance has been performed for significant parameters for minimisation of makespan and total completion time simultaneously.

Influence of Deep Cold Rolling and Low Plasticity Burnishing on Surface Hardness and Surface Roughness of AISI 4140 Steel

Deep cold rolling (DCR) and low plasticity burnishing (LPB) process are cold working processes, which easily produce a smooth and work-hardened surface by plastic deformation of surface irregularities. The present study focuses on the surface roughness and surface hardness aspects of AISI 4140 work material, using fractional factorial design of experiments. The assessment of the surface integrity aspects on work material was done, in order to identify the predominant factors amongst the selected parameters. They were then categorized in order of significance followed by setting the levels of the factors for minimizing surface roughness and/or maximizing surface hardness. In the present work, the influence of main process parameters (force, feed rate, number of tool passes/overruns, initial roughness of the work piece, ball material, ball diameter and lubricant used) on the surface roughness and the hardness of AISI 4140 steel were studied for both LPB and DCR process and the results are compared. It was observed that by using LPB process surface hardness has been improved by 167% and in DCR process surface hardness has been improved by 442%. It was also found that the force, ball diameter, number of tool passes and initial roughness of the workpiece are the most pronounced parameters, which has a significant effect on the work piece-s surface during deep cold rolling and low plasticity burnishing process.

Optimization of Methods for Development of Fermented-Distillate of Passion Fruit Beverage

Fermented beverages have high expression in the market for beverages in general, is increasingly valued in situations where the characteristic aroma and flavor of the material that gave rise to them are kept after processing. This study aimed to develop a distilled beverage from passion fruit, and assess, by sensory tests and chromatographic profile, the influence of different treatments (FM1- spirit with pulp addiction and FM2 – spirit with bigger ratio of pulp in must) in the setting of volatiles in the fruit drink, and performing chemical characterization taking into account the main parameters of quality established by the legislation. The chromatograms and the first sensorial tests had indicated that sample FM1 possess better characteristics of aroma, as much of how much quantitative the qualitative point of view. However, it analyzes it sensorial end (preference test) disclosed the biggest preference of the cloth provers for sample FM2-2 (note 7.93), being the attributes of decisive color and flavor in this reply, confirmed for the observed values lowest of fixed and total acidity in the samples of treatment FM2.

Capacitor Placement in Distribution Systems Using Simulating Annealing (SA)

This paper undertakes the problem of optimal capacitor placement in a distribution system. The problem is how to optimally determine the locations to install capacitors, the types and sizes of capacitors to he installed and, during each load level,the control settings of these capacitors in order that a desired objective function is minimized while the load constraints,network constraints and operational constraints (e.g. voltage profile) at different load levels are satisfied. The problem is formulated as a combinatorial optimization problem with a nondifferentiable objective function. Four solution mythologies based on algorithms (GA),tabu search (TS), and hybrid GA-SA algorithms are presented.The solution methodologies are preceded by a sensitivity analysis to select the candidate capacitor installation locations.

Solution of Optimal Reactive Power Flow using Biogeography-Based Optimization

Optimal reactive power flow is an optimization problem with one or more objective of minimizing the active power losses for fixed generation schedule. The control variables are generator bus voltages, transformer tap settings and reactive power output of the compensating devices placed on different bus bars. Biogeography- Based Optimization (BBO) technique has been applied to solve different kinds of optimal reactive power flow problems subject to operational constraints like power balance constraint, line flow and bus voltages limits etc. BBO searches for the global optimum mainly through two steps: Migration and Mutation. In the present work, BBO has been applied to solve the optimal reactive power flow problems on IEEE 30-bus and standard IEEE 57-bus power systems for minimization of active power loss. The superiority of the proposed method has been demonstrated. Considering the quality of the solution obtained, the proposed method seems to be a promising one for solving these problems.

Toward a Model for Knowledge Development in Virtual Environments: Strategies for Student Ownership

This article discusses the concept of student ownership of knowledge and seeks to determine how to move students from knowledge acquisition to knowledge application and ultimately to knowledge generation in a virtual setting. Instructional strategies for fostering student engagement in a virtual environment are critical to the learner-s strategic ownership of the knowledge. A number of relevant theories that focus on learning, affect, needs and adult concerns are presented to provide a basis for exploring the transfer of knowledge from teacher to learner. A model under development is presented that combines the dimensions of knowledge approach, the teacher-student relationship with regards to knowledge authority and teaching approach to demonstrate the recursive and scaffolded design for creation of virtual learning environments.

Downlink Scheduling and Radio Resource Allocation in Adaptive OFDMA Wireless Communication Systems for User-Individual QoS

In this paper, we address the problem of adaptive radio resource allocation (RRA) and packet scheduling in the downlink of a cellular OFDMA system, and propose a downlink multi-carrier proportional fair (MPF) scheduler and its joint with adaptive RRA algorithm to distribute radio resources among multiple users according to their individual QoS requirements. The allocation and scheduling objective is to maximize the total throughput, while at the same time maintaining the fairness among users. The simulation results demonstrate that the methods presented provide for user more explicit fairness relative to RRA algorithm, but the joint scheme achieves the higher sum-rate capacity with flexible parameters setting compared with MPF scheduler.

The Regional Concept, Public Policy and Policy Spaces: The ARC and TVA

This paper examines two policy spaces–the ARC and TVA–and their spatialized politics. The research observes that the regional concept informs public policy and can contribute to the formation of stable policy initiatives. Using the subsystem framework to understand the political viability of policy regimes, the authors conclude policy geographies that appeal to traditional definitions of regions are more stable over time. In contrast, geographies that fail to reflect pre-existing representations of space are engaged in more competitive subsystem politics. The paper demonstrates that the spatial practices of policy regions and their directional politics influence the political viability of programs. The paper concludes that policy spaces should institutionalize pre-existing geographies–not manufacture new ones.

Initialization Method of Reference Vectors for Improvement of Recognition Accuracy in LVQ

Initial values of reference vectors have significant influence on recognition accuracy in LVQ. There are several existing techniques, such as SOM and k-means, for setting initial values of reference vectors, each of which has provided some positive results. However, those results are not sufficient for the improvement of recognition accuracy. This study proposes an ACO-used method for initializing reference vectors with an aim to achieve recognition accuracy higher than those obtained through conventional methods. Moreover, we will demonstrate the effectiveness of the proposed method by applying it to the wine data and English vowel data and comparing its results with those of conventional methods.

Fracture Location Characterizations of Dissimilar Friction Stir Welds

This paper reports the tensile fracture location characterizations of dissimilar friction stir welds between 5754 aluminium alloy and C11000 copper. The welds were produced using three shoulder diameter tools; namely, 15, 18 and 25 mm by varying the process parameters. The rotational speeds considered were 600, 950 and 1200 rpm while the feed rates employed were 50, 150 and 300 mm/min to represent the low, medium and high settings respectively. The tensile fracture locations were evaluated using the optical microscope to identify the fracture locations and were characterized. It was observed that 70% of the tensile samples failed in the Thermo Mechanically Affected Zone (TMAZ) of copper at the weld joints. Further evaluation of the fracture surfaces of the pulled tensile samples revealed that welds with low Ultimate Tensile Strength either have defects or intermetallics present at their joint interfaces.

Application of GM (1, 1) Model Group Based on Recursive Solution in China's Energy Demand Forecasting

To learn about China-s future energy demand, this paper first proposed GM(1,1) model group based on recursive solutions of parameters estimation, setting up a general solving-algorithm of the model group. This method avoided the problems occurred on the past researches that remodeling, loss of information and large amount of calculation. This paper established respectively all-data-GM(1,1), metabolic GM(1,1) and new information GM (1,1)model according to the historical data of energy consumption in China in the year 2005-2010 and the added data of 2011, then modeling, simulating and comparison of accuracies we got the optimal models and to predict. Results showed that the total energy demand of China will be 37.2221 billion tons of equivalent coal in 2012 and 39.7973 billion tons of equivalent coal in 2013, which are as the same as the overall planning of energy demand in The 12th Five-Year Plan.

Atomic Force Microscopy (AFM)Topographical Surface Characterization of Multilayer-Coated and Uncoated Carbide Inserts

In recent years, scanning probe atomic force microscopy SPM AFM has gained acceptance over a wide spectrum of research and science applications. Most fields focuses on physical, chemical, biological while less attention is devoted to manufacturing and machining aspects. The purpose of the current study is to assess the possible implementation of the SPM AFM features and its NanoScope software in general machining applications with special attention to the tribological aspects of cutting tool. The surface morphology of coated and uncoated as-received carbide inserts is examined, analyzed, and characterized through the determination of the appropriate scanning setting, the suitable data type imaging techniques and the most representative data analysis parameters using the MultiMode SPM AFM in contact mode. The NanoScope operating software is used to capture realtime three data types images: “Height", “Deflection" and “Friction". Three scan sizes are independently performed: 2, 6, and 12 μm with a 2.5 μm vertical range (Z). Offline mode analysis includes the determination of three functional topographical parameters: surface “Roughness", power spectral density “PSD" and “Section". The 12 μm scan size in association with “Height" imaging is found efficient to capture every tiny features and tribological aspects of the examined surface. Also, “Friction" analysis is found to produce a comprehensive explanation about the lateral characteristics of the scanned surface. Configuration of many surface defects and drawbacks has been precisely detected and analyzed.

Color Lighting Efficiency of Light Emitting Diode Tube to Lure the Adult Coconut Hispine Beetle

The objective of this research was to investigate the efficiency of the light emitting diode (LED) tube in various color lights used to lure the adult coconut hispine beetle. The research was conducted by setting the forward bias on LED tubes, and the next step was to test luminous efficacy and quantity of electricity used to power each LED tube in different color lights. Finally, the researcher examined the efficiency of each color-light LED tube to lure the adult coconut hispine beetle. The results showed that the ultraviolet LED tubes had the most capacity to allure the adult coconut hispine beetles with the percentage of 82.92, followed by the blue LED tubes with the percentage of 59.76. Whereas the yellow, pink, red and warm white LED tubes had no influence to the adult coconut hispine beetles.

New Curriculum Approach in Teaching Network Security Subjects for ICT Courses in Malaysia

This paper discusses a curriculum approach that will give emphasis on practical portions of teaching network security subjects in information and communication technology courses. As we are well aware, the need to use a practice and application oriented approach in education is paramount. Research on active learning and cooperative groups have shown that students grasps more and have more tendency towards obtaining and realizing soft skills like leadership, communication and team work as opposed to the more traditional theory and exam based teaching and learning. While this teaching and learning paradigm is relatively new in Malaysia, it has been practiced widely in the West. This paper examines a certain approach whereby students learning wireless security are divided into and work in small and manageable groups where there will be 2 teams which consist of black hat and white hat teams. The former will try to find and expose vulnerabilities in a wireless network while the latter will try their best to prevent such attacks on their wireless networks using hardware, software, design and enforcement of security policy and etc. This paper will try to show that the approach taken plus the use of relevant and up to date software and hardware and with suitable environment setting will hopefully expose students to a more fruitful outcome in terms of understanding of concepts, theories and their motivation to learn.

Developing New Processes and Optimizing Performance Using Response Surface Methodology

Response surface methodology (RSM) is a very efficient tool to provide a good practical insight into developing new process and optimizing them. This methodology could help engineers to raise a mathematical model to represent the behavior of system as a convincing function of process parameters. Through this paper the sequential nature of the RSM surveyed for process engineers and its relationship to design of experiments (DOE), regression analysis and robust design reviewed. The proposed four-step procedure in two different phases could help system analyst to resolve the parameter design problem involving responses. In order to check accuracy of the designed model, residual analysis and prediction error sum of squares (PRESS) described. It is believed that the proposed procedure in this study can resolve a complex parameter design problem with one or more responses. It can be applied to those areas where there are large data sets and a number of responses are to be optimized simultaneously. In addition, the proposed procedure is relatively simple and can be implemented easily by using ready-made standard statistical packages.

EEG Spikes Detection, Sorting, and Localization

This study introduces a new method for detecting, sorting, and localizing spikes from multiunit EEG recordings. The method combines the wavelet transform, which localizes distinctive spike features, with Super-Paramagnetic Clustering (SPC) algorithm, which allows automatic classification of the data without assumptions such as low variance or Gaussian distributions. Moreover, the method is capable of setting amplitude thresholds for spike detection. The method makes use of several real EEG data sets, and accordingly the spikes are detected, clustered and their times were detected.