Computing the Loop Bound in Iterative Data Flow Graphs Using Natural Token Flow

Signal processing applications which are iterative in nature are best represented by data flow graphs (DFG). In these applications, the maximum sampling frequency is dependent on the topology of the DFG, the cyclic dependencies in particular. The determination of the iteration bound, which is the reciprocal of the maximum sampling frequency, is critical in the process of hardware implementation of signal processing applications. In this paper, a novel technique to compute the iteration bound is proposed. This technique is different from all previously proposed techniques, in the sense that it is based on the natural flow of tokens into the DFG rather than the topology of the graph. The proposed algorithm has lower run-time complexity than all known algorithms. The performance of the proposed algorithm is illustrated through analytical analysis of the time complexity, as well as through simulation of some benchmark problems.

An Enterprise Intelligent System Development and Solution Framework

The recent trend has been using hybrid approach rather than using a single intelligent technique to solve the problems. In this paper, we describe and discuss a framework to develop enterprise solutions that are backed by intelligent techniques. The framework not only uses intelligent techniques themselves but it is a complete environment that includes various interfaces and components to develop the intelligent solutions. The framework is completely Web-based and uses XML extensively. It can work like shared plat-form to be accessed by multiple developers, users and decision makers.

Groundwater Level Prediction at a Pilot Area in Southeastern Part of the UAE using Shallow Seismic Method

The groundwater is one of the main sources for sustainability in the United Arab Emirates (UAE). Intensive developments in Al-Ain area lead to increase water demand, which consequently reduced the overall groundwater quantity in major aquifers. However, in certain residential areas within Al-Ain, it has been noticed that the groundwater level is rising, for example in Sha-ab Al Askher area. The reasons for the groundwater rising phenomenon are yet to be investigated. In this work, twenty four seismic refraction profiles have been carried out along the study pilot area; as well as field measurement of the groundwater level in a number of available water wells in the area. The processed seismic data indicated the deepest and shallowest groundwater levels are 15m and 2.3 meters respectively. This result is greatly consistent with the proper field measurement of the groundwater level. The minimum detected value may be referred to perched subsurface water which may be associated to the infiltration from the surrounding water bodies such as lakes, and elevated farms. The maximum values indicate the accurate groundwater level within the study area. The findings of this work may be considered as a preliminary help to the decision makers.

Extraction Condition of Phaseolus vulgaris

Theoptimal extraction condition of dried Phaseolus vulgaris powderwas studied. The three independent variables are raw material concentration, shaking and centrifugaltime. The dependent variables are both yield percentage of crude extract and alphaamylase enzyme inhibition activity. The experimental design was based on box-behnkendesign. Highest yield percentage of crude extract could get from extraction condition at concentration of 1, 0,1, concentration of 0.15 M ,extraction time for 2hour, and separationtime for60 min. Moreover, the crude extract with highest alpha-amylase enzyme inhibition activityoccurred by extraction condition at concentration of 0.10 M, extraction time for 2 min, and separation time for 45 min

Unified, Low-Cost Analysis Framework for the Cycling Situation in Cities

We propose a low-cost uniform analysis framework allowing comparison of the strengths and weaknesses of the bicycling experience within and between cities. A primary component is an expedient, one-page mobility survey from which mode share is calculated. The bicycle mode share of many cities remains unknown, creating a serious barrier for both scientists and policy makers aiming to understand and increase rates of bicycling. Because of its low cost and expedience, this framework could be replicated widely, uniformly filling the data gap. The framework has been applied to 13 Central European cities with success. Data is collected on multiple modes with specific questions regarding both behavior and quality of travel experience. Individual preferences are also collected, examining the conditions under which respondents would change behavior to adopt more sustainable modes (bicycling or public transportation). A broad analysis opportunity results, intended to inform policy choices.

Automated Textile Defect Recognition System Using Computer Vision and Artificial Neural Networks

Least Development Countries (LDC) like Bangladesh, whose 25% revenue earning is achieved from Textile export, requires producing less defective textile for minimizing production cost and time. Inspection processes done on these industries are mostly manual and time consuming. To reduce error on identifying fabric defects requires more automotive and accurate inspection process. Considering this lacking, this research implements a Textile Defect Recognizer which uses computer vision methodology with the combination of multi-layer neural networks to identify four classifications of textile defects. The recognizer, suitable for LDC countries, identifies the fabric defects within economical cost and produces less error prone inspection system in real time. In order to generate input set for the neural network, primarily the recognizer captures digital fabric images by image acquisition device and converts the RGB images into binary images by restoration process and local threshold techniques. Later, the output of the processed image, the area of the faulty portion, the number of objects of the image and the sharp factor of the image, are feed backed as an input layer to the neural network which uses back propagation algorithm to compute the weighted factors and generates the desired classifications of defects as an output.

Simultaneous Optimization of Machining Parameters and Tool Geometry Specifications in Turning Operation of AISI1045 Steel

Machining is an important manufacturing process used to produce a wide variety of metallic parts. Among various machining processes, turning is one of the most important one which is employed to shape cylindrical parts. In turning, the quality of finished product is measured in terms of surface roughness. In turn, surface quality is determined by machining parameters and tool geometry specifications. The main objective of this study is to simultaneously model and optimize machining parameters and tool geometry in order to improve the surface roughness for AISI1045 steel. Several levels of machining parameters and tool geometry specifications are considered as input parameters. The surface roughness is selected as process output measure of performance. A Taguchi approach is employed to gather experimental data. Then, based on signal-to-noise (S/N) ratio, the best sets of cutting parameters and tool geometry specifications have been determined. Using these parameters values, the surface roughness of AISI1045 steel parts may be minimized. Experimental results are provided to illustrate the effectiveness of the proposed approach.

Blind Source Separation Using Modified Gaussian FastICA

This paper addresses the problem of source separation in images. We propose a FastICA algorithm employing a modified Gaussian contrast function for the Blind Source Separation. Experimental result shows that the proposed Modified Gaussian FastICA is effectively used for Blind Source Separation to obtain better quality images. In this paper, a comparative study has been made with other popular existing algorithms. The peak signal to noise ratio (PSNR) and improved signal to noise ratio (ISNR) are used as metrics for evaluating the quality of images. The ICA metric Amari error is also used to measure the quality of separation.

The Necessity of Biomass Application for Developing Combined Heat and Power (CHP)with Biogas Fuel: Case Study

The daily increase of organic waste materials resulting from different activities in the country is one of the main factors for the pollution of environment. Today, with regard to the low level of the output of using traditional methods, the high cost of disposal waste materials and environmental pollutions, the use of modern methods such as anaerobic digestion for the production of biogas has been prevailing. The collected biogas from the process of anaerobic digestion, as a renewable energy source similar to natural gas but with a less methane and heating value is usable. Today, with the help of technologies of filtration and proper preparation, access to biogas with features fully similar to natural gas has become possible. At present biogas is one of the main sources of supplying electrical and thermal energy and also an appropriate option to be used in four stroke engine, diesel engine, sterling engine, gas turbine, gas micro turbine and fuel cell to produce electricity. The use of biogas for different reasons which returns to socio-economic and environmental advantages has been noticed in CHP for the production of energy in the world. The production of biogas from the technology of anaerobic digestion and its application in CHP power plants in Iran can not only supply part of the energy demands in the country, but it can materialize moving in line with the sustainable development. In this article, the necessity of the development of CHP plants with biogas fuels in the country will be dealt based on studies performed from the economic, environmental and social aspects. Also to prove the importance of the establishment of these kinds of power plants from the economic point of view, necessary calculations has been done as a case study for a CHP power plant with a biogas fuel.

Initializing K-Means using Genetic Algorithms

K-Means (KM) is considered one of the major algorithms widely used in clustering. However, it still has some problems, and one of them is in its initialization step where it is normally done randomly. Another problem for KM is that it converges to local minima. Genetic algorithms are one of the evolutionary algorithms inspired from nature and utilized in the field of clustering. In this paper, we propose two algorithms to solve the initialization problem, Genetic Algorithm Initializes KM (GAIK) and KM Initializes Genetic Algorithm (KIGA). To show the effectiveness and efficiency of our algorithms, a comparative study was done among GAIK, KIGA, Genetic-based Clustering Algorithm (GCA), and FCM [19].

Particle Swarm Optimization Approach on Flexible Structure at Wiper Blade System

Application of flexible structures has been significantly, increased in industry and aerospace missions due to their contributions and unique advantages over the rigid counterparts. In this paper, vibration analysis of a flexible structure i.e., automobile wiper blade is investigated and controlled. The wiper generates unwanted noise and vibration during the wiping the rain and other particles on windshield which may cause annoying noise in different ranges of frequency. A two dimensional analytical modeled wiper blade whose model accuracy is verified by numerical studies in literature is considered in this study. Particle swarm optimization (PSO) is employed in alliance with input shaping (IS) technique in order to control or to attenuate the amplitude level of unwanted noise/vibration of the wiper blade.

A Quality Optimization Approach: An Application on Next Generation Networks

The next generation wireless systems, especially the cognitive radio networks aim at utilizing network resources more efficiently. They share a wide range of available spectrum in an opportunistic manner. In this paper, we propose a quality management model for short-term sub-lease of unutilized spectrum bands to different service providers. We built our model on competitive secondary market architecture. To establish the necessary conditions for convergent behavior, we utilize techniques from game theory. Our proposed model is based on potential game approach that is suitable for systems with dynamic decision making. The Nash equilibrium point tells the spectrum holders the ideal price values where profit is maximized at the highest level of customer satisfaction. Our numerical results show that the price decisions of the network providers depend on the price and QoS of their own bands as well as the prices and QoS levels of their opponents- bands.

Comparison of Artificial Neural Network and Multivariate Regression Methods in Prediction of Soil Cation Exchange Capacity

Investigation of soil properties like Cation Exchange Capacity (CEC) plays important roles in study of environmental reaserches as the spatial and temporal variability of this property have been led to development of indirect methods in estimation of this soil characteristic. Pedotransfer functions (PTFs) provide an alternative by estimating soil parameters from more readily available soil data. 70 soil samples were collected from different horizons of 15 soil profiles located in the Ziaran region, Qazvin province, Iran. Then, multivariate regression and neural network model (feedforward back propagation network) were employed to develop a pedotransfer function for predicting soil parameter using easily measurable characteristics of clay and organic carbon. The performance of the multivariate regression and neural network model was evaluated using a test data set. In order to evaluate the models, root mean square error (RMSE) was used. The value of RMSE and R2 derived by ANN model for CEC were 0.47 and 0.94 respectively, while these parameters for multivariate regression model were 0.65 and 0.88 respectively. Results showed that artificial neural network with seven neurons in hidden layer had better performance in predicting soil cation exchange capacity than multivariate regression.

Effect of Information System Strategies on Supply Chain Strategies and Supply Chain Performance

In order to achieve competitive advantage and better performance of firm, supply chain management (SCM) strategy should support and drive forward business strategy. It means that supply chain should be aligned with business strategy, at the same time supply chain (SC) managers need to use appropriate information system (IS) solution to support their strategy, which would lead to stay competitive. There are different kinds of IS strategies which enable managers to meet the SC requirement by selecting the best IS strategy. Therefore, it is important to align IS strategies and practices with SC strategies and practices, which could help us to plan for an IS application that supports and enhances a SCMS. In this study, aligning IS with SC in strategy level is considered. The main aim of this paper is to align the various IS strategies with SCM strategies and demonstrate their impact on SC and firm performance.

Noise Estimation for Speech Enhancement in Non-Stationary Environments-A New Method

This paper presents a new method for estimating the nonstationary noise power spectral density given a noisy signal. The method is based on averaging the noisy speech power spectrum using time and frequency dependent smoothing factors. These factors are adjusted based on signal-presence probability in individual frequency bins. Signal presence is determined by computing the ratio of the noisy speech power spectrum to its local minimum, which is updated continuously by averaging past values of the noisy speech power spectra with a look-ahead factor. This method adapts very quickly to highly non-stationary noise environments. The proposed method achieves significant improvements over a system that uses voice activity detector (VAD) in noise estimation.

Performance Analysis of Evolutionary ANN for Output Prediction of a Grid-Connected Photovoltaic System

This paper presents performance analysis of the Evolutionary Programming-Artificial Neural Network (EPANN) based technique to optimize the architecture and training parameters of a one-hidden layer feedforward ANN model for the prediction of energy output from a grid connected photovoltaic system. The ANN utilizes solar radiation and ambient temperature as its inputs while the output is the total watt-hour energy produced from the grid-connected PV system. EP is used to optimize the regression performance of the ANN model by determining the optimum values for the number of nodes in the hidden layer as well as the optimal momentum rate and learning rate for the training. The EPANN model is tested using two types of transfer function for the hidden layer, namely the tangent sigmoid and logarithmic sigmoid. The best transfer function, neural topology and learning parameters were selected based on the highest regression performance obtained during the ANN training and testing process. It is observed that the best transfer function configuration for the prediction model is [logarithmic sigmoid, purely linear].

Electronic Markets has Weakened the “Tradeoff between Reach and Richness“ in the Internet

This paper has two main ideas. Firstly, it describes Evans and Wurster-s concepts “the trade-off between reach and richness", and relates them to the impact of technology on the virtual markets. Authors Evans and Wurster see the transfer of information as a 'trade'off between richness and reach-. Reach refers to the number of people who share particular information, with Richness ['Rich'] being a more complex concept combining: bandwidth, customization, interactivity, reliability, security and currency. Traditional shopping limits the number of shops the shopper is able to visit due to time and other cost constraints; the time spent traveling consequently leaves the shopper with less time to evaluate the product. The paper concludes that although the Web provides Reach, offering Richness and the sense of community required for creating and sustaining relationships with potential clients could be difficult.

Feature-Based Machining using Macro

This paper presents an on-going research work on the implementation of feature-based machining via macro programming. Repetitive machining features such as holes, slots, pockets etc can readily be encapsulated in macros. Each macro consists of methods on how to machine the shape as defined by the feature. The macro programming technique comprises of a main program and subprograms. The main program allows user to select several subprograms that contain features and define their important parameters. With macros, complex machining routines can be implemented easily and no post processor is required. A case study on machining of a part that comprised of planar face, hole and pocket features using the macro programming technique was carried out. It is envisaged that the macro programming technique can be extended to other feature-based machining fields such as the newly developed STEP-NC domain.

K-Means for Spherical Clusters with Large Variance in Sizes

Data clustering is an important data exploration technique with many applications in data mining. The k-means algorithm is well known for its efficiency in clustering large data sets. However, this algorithm is suitable for spherical shaped clusters of similar sizes and densities. The quality of the resulting clusters decreases when the data set contains spherical shaped with large variance in sizes. In this paper, we introduce a competent procedure to overcome this problem. The proposed method is based on shifting the center of the large cluster toward the small cluster, and recomputing the membership of small cluster points, the experimental results reveal that the proposed algorithm produces satisfactory results.

Technology Readiness Index (TRI) among USM Distance Education Students According to Age

This paper reports the findings of a research conducted to evaluate the ownership and usage of technology devices within Distance Education students- according to their age. This research involved 45 Distance Education students from USM Universiti Sains Malaysia (DEUSM) as its respondents. Data was collected through questionnaire that had been developed by the researchers based on some literature review. The data was analyzed to find out the frequencies of respondents agreements towards ownership of technology devices and the use of technology devices. The findings shows that all respondents own mobile phone and majority of them reveal that they use mobile on regular basis. The student in the age 30-39 has the heist ownership of the technology devices.