Compromise Ratio Method for Decision Making under Fuzzy Environment using Fuzzy Distance Measure

The aim of this paper is to adopt a compromise ratio (CR) methodology for fuzzy multi-attribute single-expert decision making proble. In this paper, the rating of each alternative has been described by linguistic terms, which can be expressed as triangular fuzzy numbers. The compromise ratio method for fuzzy multi-attribute single expert decision making has been considered here by taking the ranking index based on the concept that the chosen alternative should be as close as possible to the ideal solution and as far away as possible from the negative-ideal solution simultaneously. From logical point of view, the distance between two triangular fuzzy numbers also is a fuzzy number, not a crisp value. Therefore a fuzzy distance measure, which is itself a fuzzy number, has been used here to calculate the difference between two triangular fuzzy numbers. Now in this paper, with the help of this fuzzy distance measure, it has been shown that the compromise ratio is a fuzzy number and this eases the problem of the decision maker to take the decision. The computation principle and the procedure of the compromise ratio method have been described in detail in this paper. A comparative analysis of the compromise ratio method previously proposed [1] and the newly adopted method have been illustrated with two numerical examples.

The Effect of Harmonic Power Fluctuation for Estimating Flicker

Voltage flicker problems have long existed in several of the distribution areas served by the Taiwan Power Company. In the past, those research results indicating that the estimated ΔV10 value based on the conventional method is significantly smaller than the survey value. This paper is used to study the relationship between the voltage flicker problems and harmonic power variation for the power system with electric arc furnaces. This investigation discussed thought the effect of harmonic power fluctuation with flicker estimate value. The method of field measurement, statistics and simulation is used. The survey results demonstrate that 10 ΔV estimate must account for the effect of harmonic power variation.

Machining of FRP Composites by Abrasive Jet Machining Optimization Using Taguchi

Abrasive Jet Machining is an Unconventional machining process in which the metal is removed from brittle and hard material in the form of micro-chips. With increase in need of materials like ceramics, composites, in manufacturing of various Mechanical & Electronic components, AJM has become a useful technique for micro machining. The present study highlights the influence of different parameters like Pressure, SOD, Time, Abrasive grain size, nozzle diameter on the Metal removal of FRP (Fiber Reinforced Polymer) composite by Abrasive jet machining. The results of the Experiments conducted were analyzed and optimized with TAGUCHI method of Optimization and ANOVA for Optimal Value.

The Relations between the Fractal Properties of the River Networks and the River Flow Time Series

All the geophysical phenomena including river networks and flow time series are fractal events inherently and fractal patterns can be investigated through their behaviors. A non-linear system like a river basin can well be analyzed by a non-linear measure such as the fractal analysis. A bilateral study is held on the fractal properties of the river network and the river flow time series. A moving window technique is utilized to scan the fractal properties of them. Results depict both events follow the same strategy regarding to the fractal properties. Both the river network and the time series fractal dimension tend to saturate in a distinct value.

Prediction of Optimum Cutting Parameters to obtain Desired Surface in Finish Pass end Milling of Aluminium Alloy with Carbide Tool using Artificial Neural Network

End milling process is one of the common metal cutting operations used for machining parts in manufacturing industry. It is usually performed at the final stage in manufacturing a product and surface roughness of the produced job plays an important role. In general, the surface roughness affects wear resistance, ductility, tensile, fatigue strength, etc., for machined parts and cannot be neglected in design. In the present work an experimental investigation of end milling of aluminium alloy with carbide tool is carried out and the effect of different cutting parameters on the response are studied with three-dimensional surface plots. An artificial neural network (ANN) is used to establish the relationship between the surface roughness and the input cutting parameters (i.e., spindle speed, feed, and depth of cut). The Matlab ANN toolbox works on feed forward back propagation algorithm is used for modeling purpose. 3-12-1 network structure having minimum average prediction error found as best network architecture for predicting surface roughness value. The network predicts surface roughness for unseen data and found that the result/prediction is better. For desired surface finish of the component to be produced there are many different combination of cutting parameters are available. The optimum cutting parameter for obtaining desired surface finish, to maximize tool life is predicted. The methodology is demonstrated, number of problems are solved and algorithm is coded in Matlab®.

Hybrid Neuro Fuzzy Approach for Automatic Generation Control of Two -Area Interconnected Power System

The main objective of Automatic Generation Control (AGC) is to balance the total system generation against system load losses so that the desired frequency and power interchange with neighboring systems is maintained. Any mismatch between generation and demand causes the system frequency to deviate from its nominal value. Thus high frequency deviation may lead to system collapse. This necessitates a very fast and accurate controller to maintain the nominal system frequency. This paper deals with a novel approach of artificial intelligence (AI) technique called Hybrid Neuro-Fuzzy (HNF) approach for an (AGC). The advantage of this controller is that it can handle the non-linearities at the same time it is faster than other conventional controllers. The effectiveness of the proposed controller in increasing the damping of local and inter area modes of oscillation is demonstrated in a two area interconnected power system. The result shows that intelligent controller is having improved dynamic response and at the same time faster than conventional controller.

SWNT Sensors for Monitoring the Oxidation of Edible Oils

There are several means to measure the oxidation of edible oils, such as the acid value, the peroxide value, and the anisidine value. However, these means require large quantities of reagents and are time-consuming tasks. Therefore, a more convenient and time-saving way to measure the oxidation of edible oils is required. In this report, an edible oil condition sensor was fabricated by using single-walled nanotubes (SWNT). In order to test the sensor, oxidized edible oils, each one at a different acid value, were prepared. The SWNT sensors were immersed into these oxidized oils and the resistance changes in the sensors were measured. It was found that the conductivity of the sensors decreased as the oxidation level of oil increased. This result suggests that a change of the oil components induced by the oxidation process in edible oils is related to the conductivity change in the SWNT sensor.

EMD-Based Signal Noise Reduction

This paper introduces a new signal denoising based on the Empirical mode decomposition (EMD) framework. The method is a fully data driven approach. Noisy signal is decomposed adaptively into oscillatory components called Intrinsic mode functions (IMFs) by means of a process called sifting. The EMD denoising involves filtering or thresholding each IMF and reconstructs the estimated signal using the processed IMFs. The EMD can be combined with a filtering approach or with nonlinear transformation. In this work the Savitzky-Golay filter and shoftthresholding are investigated. For thresholding, IMF samples are shrinked or scaled below a threshold value. The standard deviation of the noise is estimated for every IMF. The threshold is derived for the Gaussian white noise. The method is tested on simulated and real data and compared with averaging, median and wavelet approaches.

Theoretical Investigation of the Instantaneous Folding Force during the First Fold Creation in a Square Column

In this paper, a theoretical formula is presented to predict the instantaneous folding force of the first fold creation in a square column under axial loading. Calculations are based on analysis of “Basic Folding Mechanism" introduced by Wierzbicki and Abramowicz. For this purpose, the sum of dissipated energy rate under bending around horizontal and inclined hinge lines and dissipated energy rate under extensional deformations are equated to the work rate of the external force on the structure. Final formula obtained in this research, reasonably predicts the instantaneous folding force of the first fold creation versus folding distance and folding angle and also predicts the instantaneous folding force instead of the average value. Finally, according to the calculated theoretical relation, instantaneous folding force of the first fold creation in a square column was sketched versus folding distance and was compared to the experimental results which showed a good correlation.

Managing Iterations in Product Design and Development

The inherent iterative nature of product design and development poses significant challenge to reduce the product design and development time (PD). In order to shorten the time to market, organizations have adopted concurrent development where multiple specialized tasks and design activities are carried out in parallel. Iterative nature of work coupled with the overlap of activities can result in unpredictable time to completion and significant rework. Many of the products have missed the time to market window due to unanticipated or rather unplanned iteration and rework. The iterative and often overlapped processes introduce greater amounts of ambiguity in design and development, where the traditional methods and tools of project management provide less value. In this context, identifying critical metrics to understand the iteration probability is an open research area where significant contribution can be made given that iteration has been the key driver of cost and schedule risk in PD projects. Two important questions that the proposed study attempts to address are: Can we predict and identify the number of iterations in a product development flow? Can we provide managerial insights for a better control over iteration? The proposal introduces the concept of decision points and using this concept intends to develop metrics that can provide managerial insights into iteration predictability. By characterizing the product development flow as a network of decision points, the proposed research intends to delve further into iteration probability and attempts to provide more clarity.

Two Stage Control Method Using a Disturbance Observer and a Kalman Filter

This paper describes the two stage control using a disturbance observer and a Kalman filter. The system feedback uses the estimated state when it controls the speed. After the change-over point, its feedback uses the controlled plant output when it controls the position. To change the system continually, a change-over point has to be determined pertinently, and the controlled plant input has to be adjusted by the addition of the appropriate value. The proposed method has noise-reduction effect. It changes the system continually, even if the controlled plant identification has the error. Although the conventional method needs a speed sensor, the proposed method does not need it. The proposed method has a superior robustness compared with the conventional two stage control.

eLearning Tools Evaluation based on Quality Concept Distance Computing. A Case Study

Despite the extensive use of eLearning systems, there is no consensus on a standard framework for evaluating this kind of quality system. Hence, there is only a minimum set of tools that can supervise this judgment and gives information about the course content value. This paper presents two kinds of quality set evaluation indicators for eLearning courses based on the computational process of three known metrics, the Euclidian, Hamming and Levenshtein distances. The “distance" calculus is applied to standard evaluation templates (i.e. the European Commission Programme procedures vs. the AFNOR Z 76-001 Standard), determining a reference point in the evaluation of the e-learning course quality vs. the optimal concept(s). The case study, based on the results of project(s) developed in the framework of the European Programme “Leonardo da Vinci", with Romanian contractors, try to put into evidence the benefits of such a method.

Performance of Histogram-Based Skin Colour Segmentation for Arms Detection in Human Motion Analysis Application

Arms detection is one of the fundamental problems in human motion analysis application. The arms are considered as the most challenging body part to be detected since its pose and speed varies in image sequences. Moreover, the arms are usually occluded with other body parts such as the head and torso. In this paper, histogram-based skin colour segmentation is proposed to detect the arms in image sequences. Six different colour spaces namely RGB, rgb, HSI, TSL, SCT and CIELAB are evaluated to determine the best colour space for this segmentation procedure. The evaluation is divided into three categories, which are single colour component, colour without luminance and colour with luminance. The performance is measured using True Positive (TP) and True Negative (TN) on 250 images with manual ground truth. The best colour is selected based on the highest TN value followed by the highest TP value.

A Complexity-Based Approach in Image Compression using Neural Networks

In this paper we present an adaptive method for image compression that is based on complexity level of the image. The basic compressor/de-compressor structure of this method is a multilayer perceptron artificial neural network. In adaptive approach different Back-Propagation artificial neural networks are used as compressor and de-compressor and this is done by dividing the image into blocks, computing the complexity of each block and then selecting one network for each block according to its complexity value. Three complexity measure methods, called Entropy, Activity and Pattern-based are used to determine the level of complexity in image blocks and their ability in complexity estimation are evaluated and compared. In training and evaluation, each image block is assigned to a network based on its complexity value. Best-SNR is another alternative in selecting compressor network for image blocks in evolution phase which chooses one of the trained networks such that results best SNR in compressing the input image block. In our evaluations, best results are obtained when overlapping the blocks is allowed and choosing the networks in compressor is based on the Best-SNR. In this case, the results demonstrate superiority of this method comparing with previous similar works and JPEG standard coding.

Design of Thermal Control Subsystem for TUSAT Telecommunication Satellite

TUSAT is a prospective Turkish Communication Satellite designed for providing mainly data communication and broadcasting services through Ku-Band and C-Band channels. Thermal control is a vital issue in satellite design process. Therefore, all satellite subsystems and equipments should be maintained in the desired temperature range from launch to end of maneuvering life. The main function of the thermal control is to keep the equipments and the satellite structures in a given temperature range for various phases and operating modes of spacecraft during its lifetime. This paper describes the thermal control design which uses passive and active thermal control concepts. The active thermal control is based on heaters regulated by software via thermistors. Alternatively passive thermal control composes of heat pipes, multilayer insulation (MLI) blankets, radiators, paints and surface finishes maintaining temperature level of the overall carrier components within an acceptable value. Thermal control design is supported by thermal analysis using thermal mathematical models (TMM).

A Study on the Heading of Spur Gears: Numerical Analysis and Experiments

In this study, the precision heading process of spur gears has been investigated by means of numerical analysis. The effect of some parameters such as teeth number and module on the forming force and material flow were presented. The simulation works were performed rigid-plastic finite element method using DEFORM 3D software. In order to validate the estimated numerical results, they were compared with those obtained experimentally during heading of spur gear using lead as a model material. Results showed that the optimum number of gear teeth is between 10 to 20, that is because of being the specific pressure in its minimum value.

Equal Sharing Solutions for Bicooperative Games

In this paper, we discuss the egalitarianism solution (ES) and center-of-gravity of the imputation-set value (CIV) for bicooperative games, which can be seen as the extensions of the solutions for traditional games given by Dutta and Ray [1] and Driessen and Funaki [2]. Furthermore, axiomatic systems for the given values are proposed. Finally, a numerical example is offered to illustrate the player ES and CTV.

High Performance In0.42Ga0.58As/In0.26Ga0.74As Vertical Cavity Surface Emitting Quantum Well Laser on In0.31Ga0.69As Ternary Substrate

This paper reports on the theoretical performance analysis of the 1.3 μm In0.42Ga0.58As /In0.26Ga0.74As multiple quantum well (MQW) vertical cavity surface emitting laser (VCSEL) on the ternary In0.31Ga0.69As substrate. The output power of 2.2 mW has been obtained at room temperature for 7.5 mA injection current. The material gain has been estimated to be ~3156 cm-1 at room temperature with the injection carrier concentration of 2×1017 cm-3. The modulation bandwidth of this laser is measured to be 9.34 GHz at room temperature for the biasing current of 2 mA above the threshold value. The outcomes reveal that the proposed InGaAsbased MQW laser is the promising one for optical communication system.

The Impact of Semantic Web on E-Commerce

Semantic Web Technologies enable machines to interpret data published in a machine-interpretable form on the web. At the present time, only human beings are able to understand the product information published online. The emerging semantic Web technologies have the potential to deeply influence the further development of the Internet Economy. In this paper we propose a scenario based research approach to predict the effects of these new technologies on electronic markets and business models of traders and intermediaries and customers. Over 300 million searches are conducted everyday on the Internet by people trying to find what they need. A majority of these searches are in the domain of consumer ecommerce, where a web user is looking for something to buy. This represents a huge cost in terms of people hours and an enormous drain of resources. Agent enabled semantic search will have a dramatic impact on the precision of these searches. It will reduce and possibly eliminate information asymmetry where a better informed buyer gets the best value. By impacting this key determinant of market prices semantic web will foster the evolution of different business and economic models. We submit that there is a need for developing these futuristic models based on our current understanding of e-commerce models and nascent semantic web technologies. We believe these business models will encourage mainstream web developers and businesses to join the “semantic web revolution."

The Relationship between Business-model Innovation and Firm Value: A Dynamic Perspective

When consistently innovative business-models can give companies a competitive advantage, longitudinal empirical research, which can reflect dynamic business-model changes, has yet to prove a definitive connection. This study consequently employs a dynamic perspective in conjunction with innovation theory to examine the relationship between the types of business-model innovation and firm value. This study tries to examine various types of business-model innovation in high-end and low-end technology industries such as HTC and the 7-Eleven chain stores with research periods of 14 years and 32 years, respectively. The empirical results suggest that adopting radical business-model innovation in addition to expanding new target markets can successfully lead to a competitive advantage. Sustained advanced technological competences and service/product innovation are the key successful factors in high-end and low-end technology industry business-models respectively. In sum up, the business-model innovation can yield a higher market value and financial value in high-end technology industries than low-end ones.