Application of Nano Cutting Fluid under Minimum Quantity Lubrication (MQL) Technique to Improve Grinding of Ti – 6Al – 4V Alloy

Minimum Quantity Lubrication (MQL) technique obtained a significant attention in machining processes to reduce environmental loads caused by usage of conventional cutting fluids. Recently nanofluids are finding an extensive application in the field of mechanical engineering because of their superior lubrication and heat dissipation characteristics. This paper investigates the use of a nanofluid under MQL mode to improve grinding characteristics of Ti-6Al-4V alloy. Taguchi-s experimental design technique has been used in the present investigation and a second order model has been established to predict grinding forces and surface roughness. Different concentrations of water based Al2O3 nanofluids were applied in the grinding operation through MQL setup developed in house and the results have been compared with those of conventional coolant and pure water. Experimental results showed that grinding forces reduced significantly when nano cutting fluid was used even at low concentration of the nano particles and surface finish has been found to improve with higher concentration of the nano particles.

Mathematical Modeling of Non-Isothermal Multi-Component Fluid Flow in Pipes Applying to Rapid Gas Decompression in Rich and Base Gases

The paper presents a one-dimensional transient mathematical model of compressible non-isothermal multicomponent fluid mixture flow in a pipe. The set of the mass, momentum and enthalpy conservation equations for gas phase is solved in the model. Thermo-physical properties of multi-component gas mixture are calculated by solving the Equation of State (EOS) model. The Soave-Redlich-Kwong (SRK-EOS) model is chosen. Gas mixture viscosity is calculated on the basis of the Lee-Gonzales- Eakin (LGE) correlation. Numerical analysis of rapid gas decompression process in rich and base natural gases is made on the basis of the proposed mathematical model. The model is successfully validated on the experimental data [1]. The proposed mathematical model shows a very good agreement with the experimental data [1] in a wide range of pressure values and predicts the decompression in rich and base gas mixtures much better than analytical and mathematical models, which are available from the open source literature.

Wiener Filter as an Optimal MMSE Interpolator

The ideal sinc filter, ignoring the noise statistics, is often applied for generating an arbitrary sample of a bandlimited signal by using the uniformly sampled data. In this article, an optimal interpolator is proposed; it reaches a minimum mean square error (MMSE) at its output in the presence of noise. The resulting interpolator is thus a Wiener filter, and both the optimal infinite impulse response (IIR) and finite impulse response (FIR) filters are presented. The mean square errors (MSE-s) for the interpolator of different length impulse responses are obtained by computer simulations; it shows that the MSE-s of the proposed interpolators with a reasonable length are improved about 0.4 dB under flat power spectra in noisy environment with signal-to-noise power ratio (SNR) equal 10 dB. As expected, the results also demonstrate the improvements for the MSE-s with various fractional delays of the optimal interpolator against the ideal sinc filter under a fixed length impulse response.

Leaf Chlorophyll of Corn, Sweet basil and Borage under Intercropping System in Weed Interference

Intercropping is one of the sustainable agricultural factors. The SPAD meter can be used to predict nitrogen index reliably, it may also be a useful tool for assessing the relative impact of weeds on crops. In order to study the effect of weeds on SPAD in corn (Zea mays L.), sweet basil (Ocimum basilicum L.) and borage (Borago officinalis L.) in intercropping system, a factorial experiment was conducted in three replications in 2011. Experimental factors were included intercropping of corn with sweet basil and borage in different ratios (100:0, 75:25, 50:50, 25:75 and 0:100 corn: borage or sweet basil) and weed infestation (weed control and weed interference). The results showed that intercropping of corn with sweet basil and borage increased the SPAD value of corn compare to monoculture in weed interference condition. Sweet basil SPAD value in weed control treatments (43.66) was more than weed interference treatments (40.17). Corn could increase the borage SPAD value compare to monoculture in weed interference treatments.

Federal Open Agent System Platform

Open Agent System platform based on High Level Architecture is firstly proposed to support the application involving heterogeneous agents. The basic idea is to develop different wrappers for different agent systems, which are wrapped as federates to join a federation. The platform is based on High Level Architecture and the advantages for this open standard are naturally inherited, such as system interoperability and reuse. Especially, the federal architecture allows different federates to be heterogeneous so as to support the integration of different agent systems. Furthermore, both implicit communication and explicit communication between agents can be supported. Then, as the wrapper RTI_JADE an example, the components are discussed. Finally, the performance of RTI_JADE is analyzed. The results show that RTI_JADE works very efficiently.

Blind Channel Estimation Based on URV Decomposition Technique for Uplink of MC-CDMA

In this paper, we investigate a blind channel estimation method for Multi-carrier CDMA systems that use a subspace decomposition technique. This technique exploits the orthogonality property between the noise subspace and the received user codes to obtain channel of each user. In the past we used Singular Value Decomposition (SVD) technique but SVD have most computational complexity so in this paper use a new algorithm called URV Decomposition, which serve as an intermediary between the QR decomposition and SVD, replaced in SVD technique to track the noise space of the received data. Because of the URV decomposition has almost the same estimation performance as the SVD, but has less computational complexity.

Directors- Islamic Code of Ethics

This paper discusses a new model of Islamic code of ethics for directors. Several corporate scandals and local (example Transmile and Megan Media) and overseas corporate (example Parmalat and Enron) collapses show that the current corporate governance and regulatory reform are unable to prevent these events from recurring. Arguably, the code of ethics for directors is under research and the current code of ethics only concentrates on binding the work of the employee of the organization as a whole, without specifically putting direct attention to the directors, the group of people responsible for the performance of the company. This study used a semi-structured interview survey of well-known Islamic scholars such as the Mufti to develop the model. It is expected that the outcome of the research is a comprehensive model of code of ethics based on the Islamic principles that can be applied and used by the company to construct a code of ethics for their directors.

A Generalized Coordination Setting Method for Distribution Systems with Closed-loop

The protection issues in distribution systems with open and closed-loop are studied, and a generalized protection setting scheme based on the traditional over current protection theories is proposed to meet the new requirements. The setting method is expected to be easier realized using computer program, so that the on-line adaptive setting for coordination in distribution system can be implemented. An automatic setting program is created and several cases are taken into practice. The setting results are verified by the coordination curves of the protective devices which are plotted using MATLAB.

Investigation on Adjustable Mirror Bender Using Light Beam Size

In this research, the use of light beam size to design the adjustable mirror bender is presented. The focused beam line characterized by its size towards the synchrotron light beam line is investigated. The COSMOSWorks is used in all simulation components of curvature adjustment system to analyze in finite element method. The results based on simulation covers the use of applied forces during adjustment of the mirror radius are presented.

A High Performance Technique in Harmonic Omitting Based on Predictive Current Control of a Shunt Active Power Filter

The perfect operation of common Active Filters is depended on accuracy of identification system distortion. Also, using a suitable method in current injection and reactive power compensation, leads to increased filter performance. Due to this fact, this paper presents a method based on predictive current control theory in shunt active filter applications. The harmonics of the load current is identified by using o–d–q reference frame on load current and eliminating the DC part of d–q components. Then, the rest of these components deliver to predictive current controller as a Threephase reference current by using Park inverse transformation. System is modeled in discreet time domain. The proposed method has been tested using MATLAB model for a nonlinear load (with Total Harmonic Distortion=20%). The simulation results indicate that the proposed filter leads to flowing a sinusoidal current (THD=0.15%) through the source. In addition, the results show that the filter tracks the reference current accurately.

Probabilistic Model Development for Project Performance Forecasting

In this paper, based on the past project cost and time performance, a model for forecasting project cost performance is developed. This study presents a probabilistic project control concept to assure an acceptable forecast of project cost performance. In this concept project activities are classified into sub-groups entitled control accounts. Then obtain the Stochastic S-Curve (SS-Curve), for each sub-group and the project SS-Curve is obtained by summing sub-groups- SS-Curves. In this model, project cost uncertainties are considered through Beta distribution functions of the project activities costs required to complete the project at every selected time sections through project accomplishment, which are extracted from a variety of sources. Based on this model, after a percentage of the project progress, the project performance is measured via Earned Value Management to adjust the primary cost probability distribution functions. Then, accordingly the future project cost performance is predicted by using the Monte-Carlo simulation method.

Non-negative Principal Component Analysis for Face Recognition

Principle component analysis is often combined with the state-of-art classification algorithms to recognize human faces. However, principle component analysis can only capture these features contributing to the global characteristics of data because it is a global feature selection algorithm. It misses those features contributing to the local characteristics of data because each principal component only contains some levels of global characteristics of data. In this study, we present a novel face recognition approach using non-negative principal component analysis which is added with the constraint of non-negative to improve data locality and contribute to elucidating latent data structures. Experiments are performed on the Cambridge ORL face database. We demonstrate the strong performances of the algorithm in recognizing human faces in comparison with PCA and NREMF approaches.

Designing and Implementing an Innovative Course about World Wide Web, Based on the Conceptual Representations of Students

Internet is nowadays included to all National Curriculums of the elementary school. A comparative study of their goals leads to the conclusion that a complete curriculum should aim to student-s acquisition of the abilities to navigate and search for information and additionally to emphasize on the evaluation of the information provided by the World Wide Web. In a constructivistic knowledge framework the design of a course has to take under consideration the conceptual representations of students. The following paper presents the conceptual representation of students of eleven years old, attending the Sixth Grade of Greek Elementary School about World Wide Web and their use in the design and implementation of an innovative course.

Solving Part Type Selection and Loading Problem in Flexible Manufacturing System Using Real Coded Genetic Algorithms – Part I: Modeling

This paper and its companion (Part 2) deal with modeling and optimization of two NP-hard problems in production planning of flexible manufacturing system (FMS), part type selection problem and loading problem. The part type selection problem and the loading problem are strongly related and heavily influence the system-s efficiency and productivity. The complexity of the problems is harder when flexibilities of operations such as the possibility of operation processed on alternative machines with alternative tools are considered. These problems have been modeled and solved simultaneously by using real coded genetic algorithms (RCGA) which uses an array of real numbers as chromosome representation. These real numbers can be converted into part type sequence and machines that are used to process the part types. This first part of the papers focuses on the modeling of the problems and discussing how the novel chromosome representation can be applied to solve the problems. The second part will discuss the effectiveness of the RCGA to solve various test bed problems.

Identifying Significant Factors of Brick Laying Process through Design of Experiment and Computer Simulation: A Case Study

Improving performance measures in the construction processes has been a major concern for managers and decision makers in the industry. They seek for ways to recognize the key factors which have the largest effect on the process. Identifying such factors can guide them to focus on the right parts of the process in order to gain the best possible result. In the present study design of experiment (DOE) has been applied to a computer simulation model of brick laying process to determine significant factors while productivity has been chosen as the response of the experiment. To this end, four controllable factors and their interaction have been experimented and the best factor level has been calculated for each one. The results indicate that three factors, namely, labor of brick, labor of mortar and inter arrival time of mortar along with interaction of labor of brick and labor of mortar are significant.

The Residual Effects of Different Doses of Atrazine+Alachlor and Foramsulfuron on the Growth and Physiology of Rapeseed (Brassica napus L.)

A pot experiment was carried out under controlled conditions to evaluate the residual effects of different doses of atrazine+alachlor and foramsulfuron used in corn fields on the growth and physiology of rapeseed (Brassica napus L.). A split-plot experiment in CRD with 4 replications was used. The main plots consisted of herbicide type (atrazine+alachlor mixture and foramsulfuron) and the sub-plots were different residual doses of the herbicides (0, 1%, 5%, 10%, 20%, 40%, 50% and 100%). 7 cm diameter pots were filled with a virgin soil and seeds of rapeseed cv. Hayola were planted in them. The pots were kept under controlled conditions for 8 weeks after germination. At harvest, the growth parameters and the chlorophyll contents of the leaves were determined. The results showed that the growth of rapeseed plants was completely prevented at the highest residual doses of the herbicides (50 and 100 %). The growth parameters of rapeseed plants were affected by all doses of both types of the herbicide as compared to the controls. The residual effects of atrazine+alachlor mixture in reducing the growth parameters of rapeseed were more pronounced as compared to the residual effects of foramsulfuron alone.

Investigating Simple Multipath Compensation for Frequency Modulated Signals at Lower Frequencies

Radio propagation from point-to-point is affected by the physical channel in many ways. A signal arriving at a destination travels through a number of different paths which are referred to as multi-paths. Research in this area of wireless communications has progressed well over the years with the research taking different angles of focus. By this is meant that some researchers focus on ways of reducing or eluding Multipath effects whilst others focus on ways of mitigating the effects of Multipath through compensation schemes. Baseband processing is seen as one field of signal processing that is cardinal to the advancement of software defined radio technology. This has led to wide research into the carrying out certain algorithms at baseband. This paper considers compensating for Multipath for Frequency Modulated signals. The compensation process is carried out at Radio frequency (RF) and at Quadrature baseband (QBB) and the results are compared. Simulations are carried out using MatLab so as to show the benefits of working at lower QBB frequencies than at RF.

Hydrogen Integration in Petrochemical Complexes, Using Modified Automated Targeting Method

Owing to extensive use of hydrogen in refining or petrochemical units, it is essential to manage hydrogen network in order to make the most efficient utilization of hydrogen. On the other hand, hydrogen is an important byproduct not properly used through petrochemical complexes and mostly sent to the fuel system. A few works have been reported in literature to improve hydrogen network for petrochemical complexes. In this study a comprehensive analysis is carried out on petrochemical units using a modified automated targeting technique which is applied to determine the minimum hydrogen consumption. Having applied the modified targeting method in two petrochemical cases, the results showed a significant reduction in required fresh hydrogen.

An Adaptive Model for Blind Image Restoration using Bayesian Approach

Image restoration involves elimination of noise. Filtering techniques were adopted so far to restore images since last five decades. In this paper, we consider the problem of image restoration degraded by a blur function and corrupted by random noise. A method for reducing additive noise in images by explicit analysis of local image statistics is introduced and compared to other noise reduction methods. The proposed method, which makes use of an a priori noise model, has been evaluated on various types of images. Bayesian based algorithms and technique of image processing have been described and substantiated with experimentation using MATLAB.

Characterization and Development of Anthropomorphic Phantoms Liver for Use in Nuclear Medicine

The objective this study was to characterize and develop anthropomorphic liver phantoms in tomography hepatic procedures for quality control and improvement professionals in nuclear medicine. For the conformation of the anthropomorphic phantom was used in plaster and acrylic. We constructed three phantoms representing processes with liver cirrhosis. The phantoms were filled with 99mTc diluted with water to obtain the scintigraphic images. Tomography images were analyzed anterior and posterior phantom representing a body with a greater degree cirrhotic. It was noted that the phantoms allow the acquisition of images similar to real liver with cirrhosis. Simulations of hemangiomas may contribute to continued professional education of nuclear medicine, on the question of image acquisition, allowing of the study parameters such of the matrix, energy window and count statistics.