Moment Invariants in Image Analysis

This paper aims to present a survey of object recognition/classification methods based on image moments. We review various types of moments (geometric moments, complex moments) and moment-based invariants with respect to various image degradations and distortions (rotation, scaling, affine transform, image blurring, etc.) which can be used as shape descriptors for classification. We explain a general theory how to construct these invariants and show also a few of them in explicit forms. We review efficient numerical algorithms that can be used for moment computation and demonstrate practical examples of using moment invariants in real applications.

Fractal Analysis on Human Colonic Pressure Activities based on the Box-counting Method

The colonic tissue is a complicated dynamic system and the colonic activities it generates are composed of irregular segmental waves, which are referred to as erratic fluctuations or spikes. They are also highly irregular with subunit fractal structure. The traditional time-frequency domain statistics like the averaged amplitude, the motility index and the power spectrum, etc. are insufficient to describe such fluctuations. Thus the fractal box-counting dimension is proposed and the fractal scaling behaviors of the human colonic pressure activities under the physiological conditions are studied. It is shown that the dimension of the resting activity is smaller than that of the normal one, whereas the clipped version, which corresponds to the activity of the constipation patient, shows with higher fractal dimension. It may indicate a practical application to assess the colonic motility, which is often indicated by the colonic pressure activity.

A Balanced Cost Cluster-Heads Selection Algorithm for Wireless Sensor Networks

This paper focuses on reducing the power consumption of wireless sensor networks. Therefore, a communication protocol named LEACH (Low-Energy Adaptive Clustering Hierarchy) is modified. We extend LEACHs stochastic cluster-head selection algorithm by a modifying the probability of each node to become cluster-head based on its required energy to transmit to the sink. We present an efficient energy aware routing algorithm for the wireless sensor networks. Our contribution consists in rotation selection of clusterheads considering the remoteness of the nodes to the sink, and then, the network nodes residual energy. This choice allows a best distribution of the transmission energy in the network. The cluster-heads selection algorithm is completely decentralized. Simulation results show that the energy is significantly reduced compared with the previous clustering based routing algorithm for the sensor networks.

An Efficient Classification Method for Inverse Synthetic Aperture Radar Images

This paper proposes an efficient method to classify inverse synthetic aperture (ISAR) images. Because ISAR images can be translated and rotated in the 2-dimensional image place, invariance to the two factors is indispensable for successful classification. The proposed method achieves invariance to translation and rotation of ISAR images using a combination of two-dimensional Fourier transform, polar mapping and correlation-based alignment of the image. Classification is conducted using a simple matching score classifier. In simulations using the real ISAR images of five scaled models measured in a compact range, the proposed method yields classification ratios higher than 97 %.

Controller Design for Euler-Bernoulli Smart Structures Using Robust Decentralized POF via Reduced Order Modeling

This paper features the proposed modeling and design of a Robust Decentralized Periodic Output Feedback (RDPOF) control technique for the active vibration control of smart flexible multimodel Euler-Bernoulli cantilever beams for a multivariable (MIMO) case by retaining the first 6 vibratory modes. The beam structure is modeled in state space form using the concept of piezoelectric theory, the Euler-Bernoulli beam theory and the Finite Element Method (FEM) technique by dividing the beam into 4 finite elements and placing the piezoelectric sensor / actuator at two finite element locations (positions 2 and 4) as collocated pairs, i.e., as surface mounted sensor / actuator, thus giving rise to a multivariable model of the smart structure plant with two inputs and two outputs. Five such multivariable models are obtained by varying the dimensions (aspect ratios) of the aluminum beam, thus giving rise to a multimodel of the smart structure system. Using model order reduction technique, the reduced order model of the higher order system is obtained based on dominant eigen value retention and the method of Davison. RDPOF controllers are designed for the above 5 multivariable-multimodel plant. The closed loop responses with the RDPOF feedback gain and the magnitudes of the control input are observed and the performance of the proposed multimodel smart structure system with the controller is evaluated for vibration control.

Methods for Case Maintenance in Case-Based Reasoning

Case-Based Reasoning (CBR) is one of machine learning algorithms for problem solving and learning that caught a lot of attention over the last few years. In general, CBR is composed of four main phases: retrieve the most similar case or cases, reuse the case to solve the problem, revise or adapt the proposed solution, and retain the learned cases before returning them to the case base for learning purpose. Unfortunately, in many cases, this retain process causes the uncontrolled case base growth. The problem affects competence and performance of CBR systems. This paper proposes competence-based maintenance method based on deletion policy strategy for CBR. There are three main steps in this method. Step 1, formulate problems. Step 2, determine coverage and reachability set based on coverage value. Step 3, reduce case base size. The results obtained show that this proposed method performs better than the existing methods currently discussed in literature.

ZBTB17 Gene rs10927875 Polymorphism in Slovak Patients with Dilated Cardiomyopathy

Dilated cardiomyopathy (DCM) is a severe cardiovascular disorder characterized by progressive systolic dysfunction due to cardiac chamber dilatation and inefficient myocardial contractility often leading to chronic heart failure. Recently, a genome-wide association studies (GWASs) on DCM indicate that the ZBTB17 gene rs10927875 single nucleotide polymorphism is associated with DCM. The aim of the study was to identify the distribution of ZBTB17 gene rs10927875 polymorphism in 50 Slovak patients with DCM and 80 healthy control subjects using the Custom Taqman®SNP Genotyping assays. Risk factors detected at baseline in each group included age, sex, body mass index, smoking status, diabetes and blood pressure. The mean age of patients with DCM was 52.9±6.3 years; the mean age of individuals in control group was 50.3±8.9 years. The distribution of investigated genotypes of rs10927875 polymorphism within ZBTB17 gene in the cohort of Slovak patients with DCM was as follows: CC (38.8%), CT (55.1%), TT (6.1%), in controls: CC (43.8%), CT (51.2%), TT (5.0%). The risk allele T was more common among the patients with dilated cardiomyopathy than in normal controls (33.7% versus 30.6%). The differences in genotype or allele frequencies of ZBTB17 gene rs10927875 polymorphism were not statistically significant (p=0.6908; p=0.6098). The results of this study suggest that ZBTB17 gene rs10927875 polymorphism may be a risk factor for susceptibility to DCM in Slovak patients with DCM. Studies of numerous files and additional functional investigations are needed to fully understand the roles of genetic associations.

Image Segmentation Using Suprathreshold Stochastic Resonance

In this paper a new concept of partial complement of a graph G is introduced and using the same a new graph parameter, called completion number of a graph G, denoted by c(G) is defined. Some basic properties of graph parameter, completion number, are studied and upperbounds for completion number of classes of graphs are obtained , the paper includes the characterization also.

A Comparative Study of Fine Grained Security Techniques Based on Data Accessibility and Inference

This paper analyzes different techniques of the fine grained security of relational databases for the two variables-data accessibility and inference. Data accessibility measures the amount of data available to the users after applying a security technique on a table. Inference is the proportion of information leakage after suppressing a cell containing secret data. A row containing a secret cell which is suppressed can become a security threat if an intruder generates useful information from the related visible information of the same row. This paper measures data accessibility and inference associated with row, cell, and column level security techniques. Cell level security offers greatest data accessibility as it suppresses secret data only. But on the other hand, there is a high probability of inference in cell level security. Row and column level security techniques have least data accessibility and inference. This paper introduces cell plus innocent security technique that utilizes the cell level security method but suppresses some innocent data to dodge an intruder that a suppressed cell may not necessarily contain secret data. Four variations of the technique namely cell plus innocent 1/4, cell plus innocent 2/4, cell plus innocent 3/4, and cell plus innocent 4/4 respectively have been introduced to suppress innocent data equal to 1/4, 2/4, 3/4, and 4/4 percent of the true secret data inside the database. Results show that the new technique offers better control over data accessibility and inference as compared to the state-of-theart security techniques. This paper further discusses the combination of techniques together to be used. The paper shows that cell plus innocent 1/4, 2/4, and 3/4 techniques can be used as a replacement for the cell level security.

Real-time ROI Acquisition for Unsupervised and Touch-less Palmprint

In this paper we proposed a novel method to acquire the ROI (Region of interest) of unsupervised and touch-less palmprint captured from a web camera in real-time. We use Viola-Jones approach and skin model to get the target area in real time. Then an innovative course-to-fine approach to detect the key points on the hand is described. A new algorithm is used to find the candidate key points coarsely and quickly. In finely stage, we verify the hand key points with the shape context descriptor. To make the user much comfortable, it can process the hand image with different poses, even the hand is closed. Experiments show promising result by using the proposed method in various conditions.

Elimination Noise by Adaptive Wavelet Threshold

Due to some reasons, observed images are degraded which are mainly caused by noise. Recently image denoising using the wavelet transform has been attracting much attention. Waveletbased approach provides a particularly useful method for image denoising when the preservation of edges in the scene is of importance because the local adaptivity is based explicitly on the values of the wavelet detail coefficients. In this paper, we propose several methods of noise removal from degraded images with Gaussian noise by using adaptive wavelet threshold (Bayes Shrink, Modified Bayes Shrink and Normal Shrink). The proposed thresholds are simple and adaptive to each subband because the parameters required for estimating the threshold depend on subband data. Experimental results show that the proposed thresholds remove noise significantly and preserve the edges in the scene.

Current Distribution and Cathode Flooding Prediction in a PEM Fuel Cell

Non-uniform current distribution in polymer electrolyte membrane fuel cells results in local over-heating, accelerated ageing, and lower power output than expected. This issue is very critical when fuel cell experiences water flooding. In this work, the performance of a PEM fuel cell is investigated under cathode flooding conditions. Two-dimensional partially flooded GDL models based on the conservation laws and electrochemical relations are proposed to study local current density distributions along flow fields over a wide range of cell operating conditions. The model results show a direct association between cathode inlet humidity increases and that of average current density but the system becomes more sensitive to flooding. The anode inlet relative humidity shows a similar effect. Operating the cell at higher temperatures would lead to higher average current densities and the chance of system being flooded is reduced. In addition, higher cathode stoichiometries prevent system flooding but the average current density remains almost constant. The higher anode stoichiometry leads to higher average current density and higher sensitivity to cathode flooding.

Multi-Objective Planning and Operation of Water Supply Systems Subject to Climate Change

Many water supply systems in Australia are currently undergoing significant reconfiguration due to reductions in long term average rainfall and resulting low inflows to water supply reservoirs since the second half of the 20th century. When water supply systems undergo change, it is necessary to develop new operating rules, which should consider climate, because the climate change is likely to further reduce inflows. In addition, water resource systems are increasingly intended to be operated to meet complex and multiple objectives representing social, economic, environmental and sustainability criteria. This is further complicated by conflicting preferences on these objectives from diverse stakeholders. This paper describes a methodology to develop optimum operating rules for complex multi-reservoir systems undergoing significant change, considering all of the above issues. The methodology is demonstrated using the Grampians water supply system in northwest Victoria, Australia. Initial work conducted on the project is also presented in this paper.

A Contribution to the Application of the Structural Analysis Method in Entrepreneurial Practice

Quantitative methods of economic decision-making as the methodological base of the so called operational research represent an important set of tools for managing complex economic systems,both at the microeconomic level and on the macroeconomic scale. Mathematical models of controlled and controlling processes allow, by means of artificial experiments, obtaining information foroptimalor optimum approaching managerial decision-making.The quantitative methods of economic decision-making usually include a methodology known as structural analysis -an analysisof interdisciplinary production-consumption relations.

Influence of Flood Detention Capability in Flood Prevention for Flood Disaster of Depression Area

Rainfall records of rainfall station including the rainfall potential per hour and rainfall mass of five heavy storms are explored, respectively from 2001 to 2010. The rationalization formula is to investigate the capability of flood peak duration of flood detention pond in different rainfall conditions. The stable flood detention model is also proposed by using system dynamic control theory to get the message of flood detention pond in this research. When rainfall frequency of one hour rainfall duration is more than 100-year frequency which exceeds the flood detention standard of 20-year frequency for the flood detention pond, the flood peak duration of flood detention pond is 1.7 hours at most even though the flood detention pond with maximum drainage potential about 15.0 m3/s of pumping system is constructed. If the rainfall peak current is more than maximum drainage potential, the flood peak duration of flood detention pond is about 1.9 hours at most. The flood detention pond is the key factor of stable drainage control and flood prevention. The critical factors of flood disaster is not only rainfall mass, but also rainfall frequency of heavy storm in different rainfall duration and flood detention frequency of flood detention system.

Bridging the Gap Between CBR and VBR for H264 Standard

This paper provides a flexible way of controlling Variable-Bit-Rate (VBR) of compressed digital video, applicable to the new H264 video compression standard. The entire video sequence is assessed in advance and the quantisation level is then set such that bit rate (and thus the frame rate) remains within predetermined limits compatible with the bandwidth of the transmission system and the capabilities of the remote end, while at the same time providing constant quality similar to VBR encoding. A process for avoiding buffer starvation by selectively eliminating frames from the encoded output at times when the frame rate is slow (large number of bits per frame) will be also described. Finally, the problem of buffer overflow will be solved by selectively eliminating frames from the received input to the decoder. The decoder detects the omission of the frames and resynchronizes the transmission by monitoring time stamps and repeating frames if necessary.

The Open Knowledge Kernel

Web services are pieces of software that can be invoked via a standardized protocol. They can be combined via formalized taskflow languages. The Open Knowledge system is a fully distributed system using P2P technology, that allows users to publish the setaskflows, and programmers to register their web services or publish implementations of them, for the roles described in these workflows.Besides this, the system offers the functionality to select a peer that could coordinate such an interaction model and inform web services when it is their 'turn'. In this paper we describe the architecture and implementation of the Open Knowledge Kernel which provides the core functionality of the Open Knowledge system.

RAPD Analysis of Genetic Diversity of Castor Bean

The aim of this work was to detect genetic variability among the set of 40 castor genotypes using 8 RAPD markers. Amplification of genomic DNA of 40 genotypes, using RAPD analysis, yielded in 66 fragments, with an average of 8.25 polymorphic fragments per primer. Number of amplified fragments ranged from 3 to 13, with the size of amplicons ranging from 100 to 1200 bp. Values of the polymorphic information content (PIC) value ranged from 0.556 to 0.895 with an average of 0.784 and diversity index (DI) value ranged from 0.621 to 0.896 with an average of 0.798. The dendrogram based on hierarchical cluster analysis using UPGMA algorithm was prepared and analyzed genotypes were grouped into two main clusters and only two genotypes could not be distinguished. Knowledge on the genetic diversity of castor can be used for future breeding programs for increased oil production for industrial uses.

Comparison of MFCC and Cepstral Coefficients as a Feature Set for PCG Biometric Systems

Heart sound is an acoustic signal and many techniques used nowadays for human recognition tasks borrow speech recognition techniques. One popular choice for feature extraction of accoustic signals is the Mel Frequency Cepstral Coefficients (MFCC) which maps the signal onto a non-linear Mel-Scale that mimics the human hearing. However the Mel-Scale is almost linear in the frequency region of heart sounds and thus should produce similar results with the standard cepstral coefficients (CC). In this paper, MFCC is investigated to see if it produces superior results for PCG based human identification system compared to CC. Results show that the MFCC system is still superior to CC despite linear filter-banks in the lower frequency range, giving up to 95% correct recognition rate for MFCC and 90% for CC. Further experiments show that the high recognition rate is due to the implementation of filter-banks and not from Mel-Scaling.

Design of Integration Security System using XML Security

In this paper, we design an integration security system that provides authentication service, authorization service, and management service of security data and a unified interface for the management service. The interface is originated from XKMS protocol and is used to manage security data such as XACML policies, SAML assertions and other authentication security data including public keys. The system includes security services such as authentication, authorization and delegation of authentication by employing SAML and XACML based on security data such as authentication data, attributes information, assertions and polices managed with the interface in the system. It also has SAML producer that issues assertions related on the result of the authentication and the authorization services.