Segmentation of Breast Lesions in Ultrasound Images Using Spatial Fuzzy Clustering and Structure Tensors

Segmentation in ultrasound images is challenging due to the interference from speckle noise and fuzziness of boundaries. In this paper, a segmentation scheme using fuzzy c-means (FCM) clustering incorporating both intensity and texture information of images is proposed to extract breast lesions in ultrasound images. Firstly, the nonlinear structure tensor, which can facilitate to refine the edges detected by intensity, is used to extract speckle texture. And then, a spatial FCM clustering is applied on the image feature space for segmentation. In the experiments with simulated and clinical ultrasound images, the spatial FCM clustering with both intensity and texture information gets more accurate results than the conventional FCM or spatial FCM without texture information.

Takagi-Sugeno Fuzzy Control of Induction Motor

This paper deals with the synthesis of fuzzy state feedback controller of induction motor with optimal performance. First, the Takagi-Sugeno (T-S) fuzzy model is employed to approximate a non linear system in the synchronous d-q frame rotating with electromagnetic field-oriented. Next, a fuzzy controller is designed to stabilise the induction motor and guaranteed a minimum disturbance attenuation level for the closed-loop system. The gains of fuzzy control are obtained by solving a set of Linear Matrix Inequality (LMI). Finally, simulation results are given to demonstrate the controller-s effectiveness.

Optimization of SAD Algorithm on VLIW DSP

SAD (Sum of Absolute Difference) algorithm is heavily used in motion estimation which is computationally highly demanding process in motion picture encoding. To enhance the performance of motion picture encoding on a VLIW processor, an efficient implementation of SAD algorithm on the VLIW processor is essential. SAD algorithm is programmed as a nested loop with a conditional branch. In VLIW processors, loop is usually optimized by software pipelining, but researches on optimal scheduling of software pipelining for nested loops, especially nested loops with conditional branches are rare. In this paper, we propose an optimal scheduling and implementation of SAD algorithm with conditional branch on a VLIW DSP processor. The proposed optimal scheduling first transforms the nested loop with conditional branch into a single loop with conditional branch with consideration of full utilization of ILP capability of the VLIW processor and realization of earlier escape from the loop. Next, the proposed optimal scheduling applies a modulo scheduling technique developed for single loop. Based on this optimal scheduling strategy, optimal implementation of SAD algorithm on TMS320C67x, a VLIW DSP is presented. Through experiments on TMS320C6713 DSK, it is shown that H.263 encoder with the proposed SAD implementation performs better than other H.263 encoder with other SAD implementations, and that the code size of the optimal SAD implementation is small enough to be appropriate for embedded environments.

The Robust Clustering with Reduction Dimension

A clustering is process to identify a homogeneous groups of object called as cluster. Clustering is one interesting topic on data mining. A group or class behaves similarly characteristics. This paper discusses a robust clustering process for data images with two reduction dimension approaches; i.e. the two dimensional principal component analysis (2DPCA) and principal component analysis (PCA). A standard approach to overcome this problem is dimension reduction, which transforms a high-dimensional data into a lower-dimensional space with limited loss of information. One of the most common forms of dimensionality reduction is the principal components analysis (PCA). The 2DPCA is often called a variant of principal component (PCA), the image matrices were directly treated as 2D matrices; they do not need to be transformed into a vector so that the covariance matrix of image can be constructed directly using the original image matrices. The decomposed classical covariance matrix is very sensitive to outlying observations. The objective of paper is to compare the performance of robust minimizing vector variance (MVV) in the two dimensional projection PCA (2DPCA) and the PCA for clustering on an arbitrary data image when outliers are hiden in the data set. The simulation aspects of robustness and the illustration of clustering images are discussed in the end of paper

Smart Surveillance using PDA

The aim of this research is to develop a fast and reliable surveillance system based on a personal digital assistant (PDA) device. This is to extend the capability of the device to detect moving objects which is already available in personal computers. Secondly, to compare the performance between Background subtraction (BS) and Temporal Frame Differencing (TFD) techniques for PDA platform as to which is more suitable. In order to reduce noise and to prepare frames for the moving object detection part, each frame is first converted to a gray-scale representation and then smoothed using a Gaussian low pass filter. Two moving object detection schemes i.e., BS and TFD have been analyzed. The background frame is updated by using Infinite Impulse Response (IIR) filter so that the background frame is adapted to the varying illuminate conditions and geometry settings. In order to reduce the effect of noise pixels resulting from frame differencing morphological filters erosion and dilation are applied. In this research, it has been found that TFD technique is more suitable for motion detection purpose than the BS in term of speed. On average TFD is approximately 170 ms faster than the BS technique

Automatic Camera Calibration for Images of Soccer Match

Camera calibration plays an important role in the domain of the analysis of sports video. Considering soccer video, in most cases, the cross-points can be used for calibration at the center of the soccer field are not sufficient, so this paper introduces a new automatic camera calibration algorithm focus on solving this problem by using the properties of images of the center circle, halfway line and a touch line. After the theoretical analysis, a practicable automatic algorithm is proposed. Very little information used though, results of experiments with both synthetic data and real data show that the algorithm is applicable.

Effects of Dry Period Length on, Milk Production and Composition, Blood Metabolites and Complete Blood Count in Subsequent Lactation of Holstein Dairy Cows

Twenty - nine Holstein cows were used to evaluate the effects of different dry period (DP) lengths on milk yield and composition, some blood metabolites, and complete blood count (CBC). Cows were assigned to one of 2 treatments: 1) 60-d dry period, 2) 35-d DP. Milk yield, from calving to 60 days, was not different for cows on the treatments (p =0.130). Cows in the 35-d DP produced more milk protein and SNF compare with cows in treatment 1 (p ≤ 0.05). Serum glucose, non-esterified fatty acids (NEFA), beta hydroxyl butyrate acid (BHBA), blood urea nitrogen (BUN), urea, and glutamic oxaloacetic transaminase (GOT) were all similar among the treatments. Body condition score (BCS), body weight (BW), complete blood count (CBC) and health problems were similar between the treatments. The results of this study demonstrated we can reduce the dry period length to 35 days with no problems.

Functional Near Infrared Spectroscope for Cognition Brain Tasks by Wavelets Analysis and Neural Networks

Brain Computer Interface (BCI) has been recently increased in research. Functional Near Infrared Spectroscope (fNIRs) is one the latest technologies which utilize light in the near-infrared range to determine brain activities. Because near infrared technology allows design of safe, portable, wearable, non-invasive and wireless qualities monitoring systems, fNIRs monitoring of brain hemodynamics can be value in helping to understand brain tasks. In this paper, we present results of fNIRs signal analysis indicating that there exist distinct patterns of hemodynamic responses which recognize brain tasks toward developing a BCI. We applied two different mathematics tools separately, Wavelets analysis for preprocessing as signal filters and feature extractions and Neural networks for cognition brain tasks as a classification module. We also discuss and compare with other methods while our proposals perform better with an average accuracy of 99.9% for classification.

Gluten-Free Cookies Enriched with Blueberry Pomace: Optimization of Baking Process

With the aim of improving nutritional profile and antioxidant capacity of gluten-free cookies, blueberry pomace, by-product of juice production, was processed into a new food ingredient by drying and grinding and used for a gluten-free cookie formulation. Since the quality of a baked product is highly influenced by the baking conditions, the objective of this work was to optimize the baking time and thickness of dough pieces, by applying Response Surface Methodology (RSM) in order to obtain the best technological quality of the cookies. The experiments were carried out according to a Central Composite Design (CCD) by selecting the dough thickness and baking time as independent variables, while hardness, color parameters (L*, a* and b* values), water activity, diameter and short/long ratio were response variables. According to the results of RSM analysis, the baking time of 13.74min and dough thickness of 4.08mm was found to be the optimal for the baking temperature of 170°C. As similar optimal parameters were obtained by previously conducted experiment based on sensory analysis, response surface methodology (RSM) can be considered as a suitable approach to optimize the baking process.

Multiple Sequence Alignment Using Optimization Algorithms

Proteins or genes that have similar sequences are likely to perform the same function. One of the most widely used techniques for sequence comparison is sequence alignment. Sequence alignment allows mismatches and insertion/deletion, which represents biological mutations. Sequence alignment is usually performed only on two sequences. Multiple sequence alignment, is a natural extension of two-sequence alignment. In multiple sequence alignment, the emphasis is to find optimal alignment for a group of sequences. Several applicable techniques were observed in this research, from traditional method such as dynamic programming to the extend of widely used stochastic optimization method such as Genetic Algorithms (GAs) and Simulated Annealing. A framework with combination of Genetic Algorithm and Simulated Annealing is presented to solve Multiple Sequence Alignment problem. The Genetic Algorithm phase will try to find new region of solution while Simulated Annealing can be considered as an alignment improver for any near optimal solution produced by GAs.

Knowledge Continuity as a Part of Business Continuity Management

Today the intangible assets are the capital of knowledge and are the most important and the most valuable resource for organizations. All employees have knowledge independently of the kind of jobs they do. Knowledge is thus an asset, which influences business operations. The objective of this article is to identify knowledge continuity as an objective of business continuity management. The article has been prepared based on the analysis of secondary sources and the evaluation of primary sources of data by means of a quantitative survey conducted in the Czech Republic. The conclusion of the article is that organizations that apply business continuity management do not focus on the preservation of the knowledge of key employees. Organizations ensure knowledge continuity only intuitively, on a random basis, non-systematically and discontinuously. The non-ensuring of knowledge continuity represents a threat of loss of key knowledge for organizations and can also negatively affect business continuity.

Study of Remote Sensing and Satellite Images Ability in Preparing Agricultural Land Use Map (ALUM)

In this research the Preparation of Land use map of scanner LISS III satellite data, belonging to the IRS in the Aghche region in Isfahan province, is studied carefully. For this purpose, the IRS satellite images of August 2008 and various land preparation uses in region including rangelands, irrigation farming, dry farming, gardens and urban areas were separated and identified. Therefore, the GPS and Erdas Imaging software were used and three methods of Maximum Likelihood, Mahalanobis Distance and Minimum Distance were analyzed. In each of these methods, matrix error and Kappa index were calculated and accuracy of each method, based on percentages: 53.13, 56.64 and 48.44, were obtained respectively. Considering the low accuracy of these methods in separation of land preparation use, the visual interpretation of the map was used. Finally, regional visits of 150 points were noted at random and no error was observed. It shows that the map prepared by visual interpretation is in high accuracy. Although the probable errors due to visual interpretation and geometric correction might happen but the desired accuracy of the map which is more than 85 percent is reliable.

Increase of Heat Index over Bangladesh: Impact of Climate Change

Heat Index describes the combined effect of temperature and humidity on human body. This combined effect is causing a serious threat to the health of people because of the changing climate. With climate change, climate variability and thus the occurrence of heat waves is likely to increase. Evidence is emerging from the analysis of long-term climate records of an increase in the frequency and duration of extreme temperature events in all over Bangladesh particularly during summer. Summer season has prolonged while winters have become short in Bangladesh. Summers have become hotter and thus affecting the lives of the people engaged in outdoor activities during scorching sun hours. In 2003 around 62 people died due to heat wave across the country. In this paper Bangladesh is divided in four regions and heat index has been calculated from 1960 to 2010 in these regions of the country. The aim of this paper is to identify the spots most vulnerable to heat strokes and heat waves due to high heat index. The results show upward trend of heat index in almost all the regions of Bangladesh. The highest increase in heat index value has been observed in areas of South-west region and North-west Region. The highest change in average heat index has been found in Jessore by almost 5.50C.

Operational Risk – Scenario Analysis

This paper focuses on operational risk measurement techniques and on economic capital estimation methods. A data sample of operational losses provided by an anonymous Central European bank is analyzed using several approaches. Loss Distribution Approach and scenario analysis method are considered. Custom plausible loss events defined in a particular scenario are merged with the original data sample and their impact on capital estimates and on the financial institution is evaluated. Two main questions are assessed – What is the most appropriate statistical method to measure and model operational loss data distribution? and What is the impact of hypothetical plausible events on the financial institution? The g&h distribution was evaluated to be the most suitable one for operational risk modeling. The method based on the combination of historical loss events modeling and scenario analysis provides reasonable capital estimates and allows for the measurement of the impact of extreme events on banking operations.

Noise Performance Optimization of a Fast Wavelength Calibration Algorithm for OSAs

A new fast correlation algorithm for calibrating the wavelength of Optical Spectrum Analyzers (OSAs) was introduced in [1]. The minima of acetylene gas spectra were measured and correlated with saved theoretical data [2]. So it is possible to find the correct wavelength calibration data using a noisy reference spectrum. First tests showed good algorithmic performance for gas line spectra with high noise. In this article extensive performance tests were made to validate the noise resistance of this algorithm. The filter and correlation parameters of the algorithm were optimized for improved noise performance. With these parameters the performance of this wavelength calibration was simulated to predict the resulting wavelength error in real OSA systems. Long term simulations were made to evaluate the performance of the algorithm over the lifetime of a real OSA.

High Performance Liquid Chromatographic Method for Determination of Colistin Sulfate and its Application in Medicated Premixand Animal Feed

The aim of the present study was to develop and validate an inexpensive and simple high performance liquid chromatographic (HPLC) method for the determination of colistin sulfate. Separation of colistin sulfate was achieved on a ZORBAX Eclipse XDB-C18 column using UV detection at λ=215 nm. The mobile phase was 30 mM sulfate buffer (pH 2.5):acetonitrile(76:24). An excellent linearity (r2=0.998) was found in the concentration range of 25 - 400 μg/mL. Intra- day and inter-day precisions of method (%RSD, n=3) were less than 7.9%.The developed and validated method was applied to determination of the content of colistin sulfate in medicated premix and animal feed sample.The recovery of colistin from animal feed was satisfactorily ranged from 90.92 to 93.77%. The results demonstrated that the HPLC method developed in this work is appropriate for direct determination of colistin sulfate in commercial medicated premixes and animal feed.

Advanced Stochastic Models for Partially Developed Speckle

Speckled images arise when coherent microwave, optical, and acoustic imaging techniques are used to image an object, surface or scene. Examples of coherent imaging systems include synthetic aperture radar, laser imaging systems, imaging sonar systems, and medical ultrasound systems. Speckle noise is a form of object or target induced noise that results when the surface of the object is Rayleigh rough compared to the wavelength of the illuminating radiation. Detection and estimation in images corrupted by speckle noise is complicated by the nature of the noise and is not as straightforward as detection and estimation in additive noise. In this work, we derive stochastic models for speckle noise, with an emphasis on speckle as it arises in medical ultrasound images. The motivation for this work is the problem of segmentation and tissue classification using ultrasound imaging. Modeling of speckle in this context involves partially developed speckle model where an underlying Poisson point process modulates a Gram-Charlier series of Laguerre weighted exponential functions, resulting in a doubly stochastic filtered Poisson point process. The statistical distribution of partially developed speckle is derived in a closed canonical form. It is observed that as the mean number of scatterers in a resolution cell is increased, the probability density function approaches an exponential distribution. This is consistent with fully developed speckle noise as demonstrated by the Central Limit theorem.

An Approach for Reducing the End-to-end Delay and Increasing Network Lifetime in Mobile Adhoc Networks

Mobile adhoc network (MANET) is a collection of mobile devices which form a communication network with no preexisting wiring or infrastructure. Multiple routing protocols have been developed for MANETs. As MANETs gain popularity, their need to support real time applications is growing as well. Such applications have stringent quality of service (QoS) requirements such as throughput, end-to-end delay, and energy. Due to dynamic topology and bandwidth constraint supporting QoS is a challenging task. QoS aware routing is an important building block for QoS support. The primary goal of the QoS aware protocol is to determine the path from source to destination that satisfies the QoS requirements. This paper proposes a new energy and delay aware protocol called energy and delay aware TORA (EDTORA) based on extension of Temporally Ordered Routing Protocol (TORA).Energy and delay verifications of query packet have been done in each node. Simulation results show that the proposed protocol has a higher performance than TORA in terms of network lifetime, packet delivery ratio and end-to-end delay.

Difference in Psychological Well-Being Based On Comparison of Religions: A Case Study in Pekan District, Pahang, Malaysia

The psychological well-being of a family is a subjective matter for evaluation, all the more when it involves the element of religions, whether Islam, Christianity, Buddhism or Hinduism. Each of these religions emphasises similar values and morals on family psychological well-being. This comparative study is specifically to determine the role of religion on family psychological well-being in Pekan district, Pahang, Malaysia. The study adopts a quantitative and qualitative mixed method design and considers a total of 412 samples of parents and children for the quantitative study, and 21 samples for the qualitative study. The quantitative study uses simple random sampling, whereas the qualitative sampling is purposive. The instrument for quantitative study is Ryff’s Psychological Well-being Scale and the qualitative study involves the construction of a guidelines protocol for in-depth interviews of respondents. The quantitative study uses the SPSS version .19 with One Way Anova, and the qualitative analysis is manual based on transcripts with specific codes and themes. The results show nonsignificance, that is, no significant difference among religions in all family psychological well-being constructs in the comparison of Islam, Christianity, Buddhism and Hinduism, thereby accepting a null hypothesis and rejecting an alternative hypothesis. The qualitative study supports the quantitative study, that is, all 21 respondents explain that no difference exists in psychological wellbeing in the comparison of teachings in all the religious mentioned. These implications may be used as guidelines for government and non-government bodies in considering religion as an important element in family psychological well-being in the long run.