A Functional Interpretation of Quantum Theory

In this paper a functional interpretation of quantum theory (QT) with emphasis on quantum field theory (QFT) is proposed. Besides the usual statements on relations between a functions initial state and final state, a functional interpretation also contains a description of the dynamic evolution of the function. That is, it describes how things function. The proposed functional interpretation of QT/QFT has been developed in the context of the author-s work towards a computer model of QT with the goal of supporting the largest possible scope of QT concepts. In the course of this work, the author encountered a number of problems inherent in the translation of quantum physics into a computer program. He came to the conclusion that the goal of supporting the major QT concepts can only be satisfied, if the present model of QT is supplemented by a "functional interpretation" of QT/QFT. The paper describes a proposal for that

Mimicking Morphogenesis for Robust Behaviour of Cellular Architectures

Morphogenesis is the process that underpins the selforganised development and regeneration of biological systems. The ability to mimick morphogenesis in artificial systems has great potential for many engineering applications, including production of biological tissue, design of robust electronic systems and the co-ordination of parallel computing. Previous attempts to mimick these complex dynamics within artificial systems have relied upon the use of evolutionary algorithms that have limited their size and complexity. This paper will present some insight into the underlying dynamics of morphogenesis, then show how to, without the assistance of evolutionary algorithms, design cellular architectures that converge to complex patterns.

Medical Image Segmentation Based On Vigorous Smoothing and Edge Detection Ideology

Medical image segmentation based on image smoothing followed by edge detection assumes a great degree of importance in the field of Image Processing. In this regard, this paper proposes a novel algorithm for medical image segmentation based on vigorous smoothening by identifying the type of noise and edge diction ideology which seems to be a boom in medical image diagnosis. The main objective of this algorithm is to consider a particular medical image as input and make the preprocessing to remove the noise content by employing suitable filter after identifying the type of noise and finally carrying out edge detection for image segmentation. The algorithm consists of three parts. First, identifying the type of noise present in the medical image as additive, multiplicative or impulsive by analysis of local histograms and denoising it by employing Median, Gaussian or Frost filter. Second, edge detection of the filtered medical image is carried out using Canny edge detection technique. And third part is about the segmentation of edge detected medical image by the method of Normalized Cut Eigen Vectors. The method is validated through experiments on real images. The proposed algorithm has been simulated on MATLAB platform. The results obtained by the simulation shows that the proposed algorithm is very effective which can deal with low quality or marginal vague images which has high spatial redundancy, low contrast and biggish noise, and has a potential of certain practical use of medical image diagnosis.

Hospitality Program Postgraduate Theses: What Hinders Their Accomplishment?

Postgraduate education is generally aimed at providing in-depth knowledge and understanding that include general philosophy in the world sciences, management, technologies, applications and other elements closely related to specific areas. In most universities, besides core and non-core subjects, a thesis is one of the requirements for the postgraduate student to accomplish before graduating. This paper reports on the empirical investigation into attributes that are associated with the obstacles to thesis accomplishment among postgraduate students. Using the quantitative approach the experiences of postgraduate students were tapped. Findings clearly revealed that information seeking, writing skills and other factors which refer to supervisor and time management, in particular, are recognized as contributory factors which positively or negatively influence postgraduates’ thesis accomplishment. Among these, writing skills dimensions were found to be the most difficult process in thesis accomplishment compared to information seeking and other factors. This pessimistic indication has provided some implications not only for the students but supervisors and institutions as a whole.

A Strategy to Optimize the SPC Scheme for Mass Production of HDD Arm with ClusteringTechnique and Three-Way Control Chart

Consider a mass production of HDD arms where hundreds of CNC machines are used to manufacturer the HDD arms. According to an overwhelming number of machines and models of arm, construction of separate control chart for monitoring each HDD arm model by each machine is not feasible. This research proposed a strategy to optimize the SPC management on shop floor. The procedure started from identifying the clusters of the machine with similar manufacturing performance using clustering technique. The three way control chart ( I - MR - R ) is then applied to each clustered group of machine. This proposed research has advantageous to the manufacturer in terms of not only better performance of the SPC but also the quality management paradigm.

Development of a Catchment Water Quality Model for Continuous Simulations of Pollutants Build-up and Wash-off

Estimation of runoff water quality parameters is required to determine appropriate water quality management options. Various models are used to estimate runoff water quality parameters. However, most models provide event-based estimates of water quality parameters for specific sites. The work presented in this paper describes the development of a model that continuously simulates the accumulation and wash-off of water quality pollutants in a catchment. The model allows estimation of pollutants build-up during dry periods and pollutants wash-off during storm events. The model was developed by integrating two individual models; rainfall-runoff model, and catchment water quality model. The rainfall-runoff model is based on the time-area runoff estimation method. The model allows users to estimate the time of concentration using a range of established methods. The model also allows estimation of the continuing runoff losses using any of the available estimation methods (i.e., constant, linearly varying or exponentially varying). Pollutants build-up in a catchment was represented by one of three pre-defined functions; power, exponential, or saturation. Similarly, pollutants wash-off was represented by one of three different functions; power, rating-curve, or exponential. The developed runoff water quality model was set-up to simulate the build-up and wash-off of total suspended solids (TSS), total phosphorus (TP) and total nitrogen (TN). The application of the model was demonstrated using available runoff and TSS field data from road and roof surfaces in the Gold Coast, Australia. The model provided excellent representation of the field data demonstrating the simplicity yet effectiveness of the proposed model.

An Improved Algorithm for Calculation of the Third-order Orthogonal Tensor Product Expansion by Using Singular Value Decomposition

As a method of expanding a higher-order tensor data to tensor products of vectors we have proposed the Third-order Orthogonal Tensor Product Expansion (3OTPE) that did similar expansion as Higher-Order Singular Value Decomposition (HOSVD). In this paper we provide a computation algorithm to improve our previous method, in which SVD is applied to the matrix that constituted by the contraction of original tensor data and one of the expansion vector obtained. The residual of the improved method is smaller than the previous method, truncating the expanding tensor products to the same number of terms. Moreover, the residual is smaller than HOSVD when applying to color image data. It is able to be confirmed that the computing time of improved method is the same as the previous method and considerably better than HOSVD.

Optimization of a Triangular Fin with Variable Fin Base Thickness

A triangular fin with variable fin base thickness is analyzed and optimized using a two-dimensional analytical method. The influence of fin base height and fin base thickness on the temperature in the fin is listed. For the fixed fin volumes, the maximum heat loss, the corresponding optimum fin effectiveness, fin base height and fin tip length as a function of the fin base thickness, convection characteristic number and dimensionless fin volume are represented. One of the results shows that the optimum heat loss increases whereas the corresponding optimum fin effectiveness decreases with the increase of fin volume.

The Influence of Voltage Flicker for the Wind Generator upon Distribution System

One of the most important power quality issues is voltage flicker. Nowadays this issue also impacts the power system all over the world. The fact of the matter is that the more and the larger capacity of wind generator has been installed. Under unstable wind power situation, the variation of output current and voltage have caused trouble to voltage flicker. Hence, the major purpose of this study is to analyze the impact of wind generator on voltage flicker of power system. First of all, digital simulation and analysis are carried out based on wind generator operating under various system short circuit capacity, impedance angle, loading, and power factor of load. The simulation results have been confirmed by field measurements.

A New Technique for Solar Activity Forecasting Using Recurrent Elman Networks

In this paper we present an efficient approach for the prediction of two sunspot-related time series, namely the Yearly Sunspot Number and the IR5 Index, that are commonly used for monitoring solar activity. The method is based on exploiting partially recurrent Elman networks and it can be divided into three main steps: the first one consists in a “de-rectification" of the time series under study in order to obtain a new time series whose appearance, similar to a sum of sinusoids, can be modelled by our neural networks much better than the original dataset. After that, we normalize the derectified data so that they have zero mean and unity standard deviation and, finally, train an Elman network with only one input, a recurrent hidden layer and one output using a back-propagation algorithm with variable learning rate and momentum. The achieved results have shown the efficiency of this approach that, although very simple, can perform better than most of the existing solar activity forecasting methods.

Temperature Field Study of Brake Disc in a Belt Conveyor Brake

To reveal the temperature field distribution of disc brake in downward belt conveyor, mathematical models of heat transfer for disc brake were established combined with heat transfer theory. Then, the simulation process was stated in detail and the temperature field of disc brake under conditions of dynamic speed and dynamic braking torque was numerically simulated by using ANSYS software. Finally the distribution and variation laws of temperature field in the braking process were analyzed. Results indicate that the maximum surface temperature occurs at a time before the brake end and there exist large temperature gradients in both radial and axial directions, while it is relatively small in the circumferential direction.

Cellular Phone Users- Willingness to Shop Online

This study aims to identify cellular phone users- shopping motivating factors towards online shopping. 100 university students located in Klang Valley, Malaysia were involved as the respondents. They were required to complete a set of questionnaire and had to own a cellular phone in order to be selected as sample in this study. Three from five proposed hypotheses were supported: purchasing information, shopping utilities and service quality. As a result, marketers and retailers should concentrate more on the less important factors in order to encourage and create willingness of the consumers to purchase online. Recommendation for future research is also presented.

Odor Discrimination Using Neural Decoding of Olfactory Bulbs in Rats

This paper presents a novel method for inferring the odor based on neural activities observed from rats- main olfactory bulbs. Multi-channel extra-cellular single unit recordings were done by micro-wire electrodes (tungsten, 50μm, 32 channels) implanted in the mitral/tufted cell layers of the main olfactory bulb of anesthetized rats to obtain neural responses to various odors. Neural response as a key feature was measured by substraction of neural firing rate before stimulus from after. For odor inference, we have developed a decoding method based on the maximum likelihood (ML) estimation. The results have shown that the average decoding accuracy is about 100.0%, 96.0%, 84.0%, and 100.0% with four rats, respectively. This work has profound implications for a novel brain-machine interface system for odor inference.

Comparative Study of Three DGS Unit Shapes and Compact Microstrip Low-Pass and Band-Pass Filters Designs

In this paper, three types of defected ground structure (DGS) units which are triangular-head (TH), rectangular-head (RH) and U-shape (US) are investigated. They are further used to low-pass and band-pass filters designs (LPF and BPF) and the obtained performances are examined. The LPF employing RH-DGS geometry presents the advantages of compact size, low-insertion loss and wide stopband compared to the other filters. It provides cutoff frequency of 2.5 GHz, largest rejection band width of 20 dB from 2.98 to 8.76 GHz, smallest transition region and smallest sharpness of the cutoff frequency. The BPF based on RH-DGS has the highest bandwidth (BW) of about 0.74 GHz and the lowest center frequency of 3.24 GHz, whereas the other BPFs have BWs less than 0.7 GHz.

Inventory Control for a Joint Replenishment Problem with Stochastic Demand

Most papers model Joint Replenishment Problem (JRP) as a (kT,S) where kT is a multiple value for a common review period T,and S is a predefined order up to level. In general the (T,S) policy is characterized by a long out of control period which requires a large amount of safety stock compared to the (R,Q) policy. In this paper a probabilistic model is built where an item, call it item(i), with the shortest order time between interval (T)is modeled under (R,Q) policy and its inventory is continuously reviewed, while the rest of items (j) are periodically reviewed at a definite time corresponding to item

Dynamic Time Warping in Gait Classificationof Motion Capture Data

The method of gait identification based on the nearest neighbor classification technique with motion similarity assessment by the dynamic time warping is proposed. The model based kinematic motion data, represented by the joints rotations coded by Euler angles and unit quaternions is used. The different pose distance functions in Euler angles and quaternion spaces are considered. To evaluate individual features of the subsequent joints movements during gait cycle, joint selection is carried out. To examine proposed approach database containing 353 gaits of 25 humans collected in motion capture laboratory is used. The obtained results are promising. The classifications, which takes into consideration all joints has accuracy over 91%. Only analysis of movements of hip joints allows to correctly identify gaits with almost 80% precision.

Presenting a Combinatorial Feature to Estimate Depth of Anesthesia

Determining depth of anesthesia is a challenging problem in the context of biomedical signal processing. Various methods have been suggested to determine a quantitative index as depth of anesthesia, but most of these methods suffer from high sensitivity during the surgery. A novel method based on energy scattering of samples in the wavelet domain is suggested to represent the basic content of electroencephalogram (EEG) signal. In this method, first EEG signal is decomposed into different sub-bands, then samples are squared and energy of samples sequence is constructed through each scale and time, which is normalized and finally entropy of the resulted sequences is suggested as a reliable index. Empirical Results showed that applying the proposed method to the EEG signals can classify the awake, moderate and deep anesthesia states similar to BIS.

Adaptive Motion Estimator Based on Variable Block Size Scheme

This paper presents an adaptive motion estimator that can be dynamically reconfigured by the best algorithm depending on the variation of the video nature during the lifetime of an application under running. The 4 Step Search (4SS) and the Gradient Search (GS) algorithms are integrated in the estimator in order to be used in the case of rapid and slow video sequences respectively. The Full Search Block Matching (FSBM) algorithm has been also integrated in order to be used in the case of the video sequences which are not real time oriented. In order to efficiently reduce the computational cost while achieving better visual quality with low cost power, the proposed motion estimator is based on a Variable Block Size (VBS) scheme that uses only the 16x16, 16x8, 8x16 and 8x8 modes. Experimental results show that the adaptive motion estimator allows better results in term of Peak Signal to Noise Ratio (PSNR), computational cost, FPGA occupied area, and dissipated power relatively to the most popular variable block size schemes presented in the literature.

A Generalized Approach for State Analysis and Parameter Estimation of Bilinear Systems using Haar Connection Coefficients

Three novel and significant contributions are made in this paper Firstly, non-recursive formulation of Haar connection coefficients, pioneered by the present authors is presented, which can be computed very efficiently and avoid stack and memory overflows. Secondly, the generalized approach for state analysis of singular bilinear time-invariant (TI) and time-varying (TV) systems is presented; vis-˜a-vis diversified and complex works reported by different authors. Thirdly, a generalized approach for parameter estimation of bilinear TI and TV systems is also proposed. The unified framework of the proposed method is very significant in that the digital hardware once-designed can be used to perform the complex tasks of state analysis and parameter estimation of different types of bilinear systems single-handedly. The simplicity, effectiveness and generalized nature of the proposed method is established by applying it to different types of bilinear systems for the two tasks.

Effect of Flowrate and Coolant Temperature on the Efficiency of Progressive Freeze Concentration on Simulated Wastewater

Freeze concentration freezes or crystallises the water molecules out as ice crystals and leaves behind a highly concentrated solution. In conventional suspension freeze concentration where ice crystals formed as a suspension in the mother liquor, separation of ice is difficult. The size of the ice crystals is still very limited which will require usage of scraped surface heat exchangers, which is very expensive and accounted for approximately 30% of the capital cost. This research is conducted using a newer method of freeze concentration, which is progressive freeze concentration. Ice crystals were formed as a layer on the designed heat exchanger surface. In this particular research, a helical structured copper crystallisation chamber was designed and fabricated. The effect of two operating conditions on the performance of the newly designed crystallisation chamber was investigated, which are circulation flowrate and coolant temperature. The performance of the design was evaluated by the effective partition constant, K, calculated from the volume and concentration of the solid and liquid phase. The system was also monitored by a data acquisition tool in order to see the temperature profile throughout the process. On completing the experimental work, it was found that higher flowrate resulted in a lower K, which translated into high efficiency. The efficiency is the highest at 1000 ml/min. It was also found that the process gives the highest efficiency at a coolant temperature of -6 °C.