Improving Digital Image Edge Detection by Fuzzy Systems

Image Edge Detection is one of the most important parts of image processing. In this paper, by fuzzy technique, a new method is used to improve digital image edge detection. In this method, a 3x3 mask is employed to process each pixel by means of vicinity. Each pixel is considered a fuzzy input and by examining fuzzy rules in its vicinity, the edge pixel is specified and by utilizing calculation algorithms in image processing, edges are displayed more clearly. This method shows significant improvement compared to different edge detection methods (e.g. Sobel, Canny).

To Join or Not to Join: The Effects of Healthcare Networks

This study uses a simulation to establish a realistic environment for laboratory research on Accountable Care Organizations. We study network attributes in order to gain insights regarding healthcare providers- conduct and performance. Our findings indicate how network structure creates significant differences in organizational performance. We demonstrate how healthcare providers positioning themselves at the central, pivotal point of the network while maintaining their alliances with their partners produce better outcomes.

A Comparative Cross-sectional Study of Religious Behavior in High School and University Students

The purpose of this study was to investigate the religious behavior of students in high school and universality in Lamerd , a town in the south of Iran, with respect to increase in their level of education and age. The participants were 450 high school and university students in all levels from first year of junior high school to the senior university students who were chosen through multistage cluster sampling method and their religious behavior was studied. Through the revised questionnaire by Nezar Alany from the University of Bahrain (r = 0/797), the religious behavior of the subjects were analyzed. Results showed that students in high school in religious behavior were superior to the students of university (003/0>p) and there was a decline of religious behavior in junior high school third year students to second students of the same school (042/0>p). More important is that the decrease in religious behavior was associated with increase in educational levels (017/0>p) and age (043/0>p).

An UML Statechart Diagram-Based MM-Path Generation Approach for Object-Oriented Integration Testing

MM-Path, an acronym for Method/Message Path, describes the dynamic interactions between methods in object-oriented systems. This paper discusses the classifications of MM-Path, based on the characteristics of object-oriented software. We categorize it according to the generation reasons, the effect scope and the composition of MM-Path. A formalized representation of MM-Path is also proposed, which has considered the influence of state on response method sequences of messages. .Moreover, an automatic MM-Path generation approach based on UML Statechart diagram has been presented, and the difficulties in identifying and generating MM-Path can be solved. . As a result, it provides a solid foundation for further research on test cases generation based on MM-Path.

Efficient Time Synchronization in Wireless Sensor Networks

Energy efficiency is the key requirement in wireless sensor network as sensors are small, cheap and are deployed in very large number in a large geographical area, so there is no question of replacing the batteries of the sensors once deployed. Different ways can be used for efficient energy transmission including Multi-Hop algorithms, collaborative communication, cooperativecommunication, Beam- forming, routing algorithm, phase, frequency and time synchronization. The paper reviews the need for time synchronization and proposed a BFS based synchronization algorithm to achieve energy efficiency. The efficiency of our protocol has been tested and verified by simulation

Location of Vortex Formation Threshold at Suction Inlets near Ground Planes – Ascending and Descending Conditions

Vortices can develop in intakes of turbojet and turbo fan aero engines during high power operation in the vicinity of solid surfaces. These vortices can cause catastrophic damage to the engine. The factors determining the formation of the vortex include both geometric dimensions as well as flow parameters. It was shown that the threshold at which the vortex forms or disappears is also dependent on the initial flow condition (i.e. whether a vortex forms after stabilised non vortex flow or vice-versa). A computational fluid dynamics study was conducted to determine the difference in thresholds between the two conditions. This is the first reported numerical investigation of the “memory effect". The numerical results reproduce the phenomenon reported in previous experimental studies and additional factors, which had not been previously studied, were investigated. They are the rate at which ambient velocity changes and the initial value of ambient velocity. The former was found to cause a shift in the threshold but not the later. It was also found that the varying condition thresholds are not symmetrical about the neutral threshold. The vortex to no vortex threshold lie slightly further away from the neutral threshold compared to the no vortex to vortex threshold. The results suggests that experimental investigation of vortex formation threshold performed either in vortex to no vortex conditions, or vice versa, solely may introduce mis-predictions greater than 10%.

A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation

Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.

Dynamic Threshold Adjustment Approach For Neural Networks

The use of neural networks for recognition application is generally constrained by their inherent parameters inflexibility after the training phase. This means no adaptation is accommodated for input variations that have any influence on the network parameters. Attempts were made in this work to design a neural network that includes an additional mechanism that adjusts the threshold values according to the input pattern variations. The new approach is based on splitting the whole network into two subnets; main traditional net and a supportive net. The first deals with the required output of trained patterns with predefined settings, while the second tolerates output generation dynamically with tuning capability for any newly applied input. This tuning comes in the form of an adjustment to the threshold values. Two levels of supportive net were studied; one implements an extended additional layer with adjustable neuronal threshold setting mechanism, while the second implements an auxiliary net with traditional architecture performs dynamic adjustment to the threshold value of the main net that is constructed in dual-layer architecture. Experiment results and analysis of the proposed designs have given quite satisfactory conducts. The supportive layer approach achieved over 90% recognition rate, while the multiple network technique shows more effective and acceptable level of recognition. However, this is achieved at the price of network complexity and computation time. Recognition generalization may be also improved by accommodating capabilities involving all the innate structures in conjugation with Intelligence abilities with the needs of further advanced learning phases.

An Improved Genetic Algorithm to Solve the Traveling Salesman Problem

The Genetic Algorithm (GA) is one of the most important methods used to solve many combinatorial optimization problems. Therefore, many researchers have tried to improve the GA by using different methods and operations in order to find the optimal solution within reasonable time. This paper proposes an improved GA (IGA), where the new crossover operation, population reformulates operation, multi mutation operation, partial local optimal mutation operation, and rearrangement operation are used to solve the Traveling Salesman Problem. The proposed IGA was then compared with three GAs, which use different crossover operations and mutations. The results of this comparison show that the IGA can achieve better results for the solutions in a faster time.

Design for Manufacturability and Concurrent Engineering for Product Development

In the 1980s, companies began to feel the effect of three major influences on their product development: newer and innovative technologies, increasing product complexity and larger organizations. And therefore companies were forced to look for new product development methods. This paper tries to focus on the two of new product development methods (DFM and CE). The aim of this paper is to see and analyze different product development methods specifically on Design for Manufacturability and Concurrent Engineering. Companies can achieve and be benefited by minimizing product life cycle, cost and meeting delivery schedule. This paper also presents simplified models that can be modified and used by different companies based on the companies- objective and requirements. Methodologies that are followed to do this research are case studies. Two companies were taken and analysed on the product development process. Historical data, interview were conducted on these companies in addition to that, Survey of literatures and previous research works on similar topics has been done during this research. This paper also tries to show the implementation cost benefit analysis and tries to calculate the implementation time. From this research, it has been found that the two companies did not achieve the delivery time to the customer. Some of most frequently coming products are analyzed and 50% to 80 % of their products are not delivered on time to the customers. The companies are following the traditional way of product development that is sequentially design and production method, which highly affect time to market. In the case study it is found that by implementing these new methods and by forming multi disciplinary team in designing and quality inspection; the company can reduce the workflow steps from 40 to 30.

Contact Stress on the Surface of Gear Teeth with Different Profile

Contact stress is an important problem in industry. This is a problem that in the first attention may be don-t appears, but disregard of these stresses cause a lot of damages in machines. These stresses occur at locations such as gear teeth, bearings, cams and between a locomotive wheel and the railroad rail. These stresses cause failure by excessive elastic deformation, yielding and fracture. In this paper we intend show the effective parameters in contact stress and ponder effect of curvature. In this paper we study contact stresses on the surface of gear teeth and compare these stresses for four popular profiles of gear teeth (involute, cycloid, epicycloids, and hypocycloid). We study this problem with mathematical and finite element methods and compare these two methods on different profile surfaces.

Feature Based Unsupervised Intrusion Detection

The goal of a network-based intrusion detection system is to classify activities of network traffics into two major categories: normal and attack (intrusive) activities. Nowadays, data mining and machine learning plays an important role in many sciences; including intrusion detection system (IDS) using both supervised and unsupervised techniques. However, one of the essential steps of data mining is feature selection that helps in improving the efficiency, performance and prediction rate of proposed approach. This paper applies unsupervised K-means clustering algorithm with information gain (IG) for feature selection and reduction to build a network intrusion detection system. For our experimental analysis, we have used the new NSL-KDD dataset, which is a modified dataset for KDDCup 1999 intrusion detection benchmark dataset. With a split of 60.0% for the training set and the remainder for the testing set, a 2 class classifications have been implemented (Normal, Attack). Weka framework which is a java based open source software consists of a collection of machine learning algorithms for data mining tasks has been used in the testing process. The experimental results show that the proposed approach is very accurate with low false positive rate and high true positive rate and it takes less learning time in comparison with using the full features of the dataset with the same algorithm.

A Critical Social Research Perspective on Self-Directed Learning and Information Technology Practitioners

Information systems practitioners are frequently required to master new technology, often without the aid of formal training. They require the skill to manage their own learning and, when this skill is developed in their formal training, their adaptability to new technology may be improved. Self- directed learning is the ability of the learner to manage his or her own learning experience with some guidance from a facilitator. Self-directed learning skills are best improved when practiced. This paper reflects on a critical social research project to improve the self-directed learning skills of fourth year Information Systems students. Critical social research differs from other research paradigms in that the researcher is viewed as the agent of change to achieve the desired outcome in the problem situation.

Speech Data Compression using Vector Quantization

Mostly transforms are used for speech data compressions which are lossy algorithms. Such algorithms are tolerable for speech data compression since the loss in quality is not perceived by the human ear. However the vector quantization (VQ) has a potential to give more data compression maintaining the same quality. In this paper we propose speech data compression algorithm using vector quantization technique. We have used VQ algorithms LBG, KPE and FCG. The results table shows computational complexity of these three algorithms. Here we have introduced a new performance parameter Average Fractional Change in Speech Sample (AFCSS). Our FCG algorithm gives far better performance considering mean absolute error, AFCSS and complexity as compared to others.

An Efficient Technique for Extracting Fuzzy Rulesfrom Neural Networks

Artificial neural networks (ANN) have the ability to model input-output relationships from processing raw data. This characteristic makes them invaluable in industry domains where such knowledge is scarce at best. In the recent decades, in order to overcome the black-box characteristic of ANNs, researchers have attempted to extract the knowledge embedded within ANNs in the form of rules that can be used in inference systems. This paper presents a new technique that is able to extract a small set of rules from a two-layer ANN. The extracted rules yield high classification accuracy when implemented within a fuzzy inference system. The technique targets industry domains that possess less complex problems for which no expert knowledge exists and for which a simpler solution is preferred to a complex one. The proposed technique is more efficient, simple, and applicable than most of the previously proposed techniques.

Attribute Selection Methods Comparison for Classification of Diffuse Large B-Cell Lymphoma

The most important subtype of non-Hodgkin-s lymphoma is the Diffuse Large B-Cell Lymphoma. Approximately 40% of the patients suffering from it respond well to therapy, whereas the remainder needs a more aggressive treatment, in order to better their chances of survival. Data Mining techniques have helped to identify the class of the lymphoma in an efficient manner. Despite that, thousands of genes should be processed to obtain the results. This paper presents a comparison of the use of various attribute selection methods aiming to reduce the number of genes to be searched, looking for a more effective procedure as a whole.

Customer Value Creation by CRM System in Electronic Device Companies

The service industry accounts for about 70% of GDP of Japan, and the importance of the service innovation is pointed out. The importance of the system use and the support service increases in the information system that is one of the service industries. However, because the system is not used enough, the purpose for which it was originally intended cannot often be achieved in the CRM system. To promote the use of the system, the effective service method is needed. It is thought that the service model's making and the clarification of the success factors are necessary to improve the operation service of the CRM system. In this research the model of the operation service in the CRM system is made.

Improved Automated Classification of Alcoholics and Non-alcoholics

In this paper, several improvements are proposed to previous work of automated classification of alcoholics and nonalcoholics. In the previous paper, multiplayer-perceptron neural network classifying energy of gamma band Visual Evoked Potential (VEP) signals gave the best classification performance using 800 VEP signals from 10 alcoholics and 10 non-alcoholics. Here, the dataset is extended to include 3560 VEP signals from 102 subjects: 62 alcoholics and 40 non-alcoholics. Three modifications are introduced to improve the classification performance: i) increasing the gamma band spectral range by increasing the pass-band width of the used filter ii) the use of Multiple Signal Classification algorithm to obtain the power of the dominant frequency in gamma band VEP signals as features and iii) the use of the simple but effective knearest neighbour classifier. To validate that these two modifications do give improved performance, a 10-fold cross validation classification (CVC) scheme is used. Repeat experiments of the previously used methodology for the extended dataset are performed here and improvement from 94.49% to 98.71% in maximum averaged CVC accuracy is obtained using the modifications. This latest results show that VEP based classification of alcoholics is worth exploring further for system development.

Active Suspension - Case Study on Robust Control

Automotive suspension system is important part of car comfort and safety. In this article automotive active suspension with linear motor as actuator is designed using H-infinity control. This paper is focused on comparison of different controller designed for quart, half or full-car model (and always used for “full" car). Special attention is placed on energy demand of the whole system. Each controller configuration is simulated and then verified on the hydraulic quarter car test bed.

NOHIS-Tree: High-Dimensional Index Structure for Similarity Search

In Content-Based Image Retrieval systems it is important to use an efficient indexing technique in order to perform and accelerate the search in huge databases. The used indexing technique should also support the high dimensions of image features. In this paper we present the hierarchical index NOHIS-tree (Non Overlapping Hierarchical Index Structure) when we scale up to very large databases. We also present a study of the influence of clustering on search time. The performance test results show that NOHIS-tree performs better than SR-tree. Tests also show that NOHIS-tree keeps its performances in high dimensional spaces. We include the performance test that try to determine the number of clusters in NOHIS-tree to have the best search time.