Dempster-Shafer Information Filtering in Multi-Modality Wireless Sensor Networks

A framework to estimate the state of dynamically varying environment where data are generated from heterogeneous sources possessing partial knowledge about the environment is presented. This is entirely derived within Dempster-Shafer and Evidence Filtering frameworks. The belief about the current state is expressed as belief and plausibility functions. An addition to Single Input Single Output Evidence Filter, Multiple Input Single Output Evidence Filtering approach is introduced. Variety of applications such as situational estimation of an emergency environment can be developed within the framework successfully. Fire propagation scenario is used to justify the proposed framework, simulation results are presented.

A Practical Approach for Testing the Process Quality

Process capability index Cpk is the most widely used index in making managerial decisions since it provides bounds on the process yield for normally distributed processes. However, existent methods for assessing process performance which constructed by statistical inference may unfortunately lead to fine results, because uncertainties exist in most real-world applications. Thus, this study adopts fuzzy inference to deal with testing of Cpk . A brief score is obtained for assessing a supplier’s process instead of a severe evaluation.

Boundary Segmentation of Microcalcification using Parametric Active Contours

A mammography image is composed of low contrast area where the breast tissues and the breast abnormalities such as microcalcification can hardly be differentiated by the medical practitioner. This paper presents the application of active contour models (Snakes) for the segmentation of microcalcification in mammography images. Comparison on the microcalcifiation areas segmented by the Balloon Snake, Gradient Vector Flow (GVF) Snake, and Distance Snake is done against the true value of the microcalcification area. The true area value is the average microcalcification area in the original mammography image traced by the expert radiologists. From fifty images tested, the result obtained shows that the accuracy of the Balloon Snake, GVF Snake, and Distance Snake in segmenting boundaries of microcalcification are 96.01%, 95.74%, and 95.70% accuracy respectively. This implies that the Balloon Snake is a better segmentation method to locate the exact boundary of a microcalcification region.

Carbon Disulfide Production via Hydrogen Sulfide Methane Reformation

Carbon disulfide is widely used for the production of viscose rayon, rubber, and other organic materials and it is a feedstock for the synthesis of sulfuric acid. The objective of this paper is to analyze possibilities for efficient production of CS2 from sour natural gas reformation (H2SMR) (2H2S+CH4 =CS2 +4H2) . Also, the effect of H2S to CH4 feed ratio and reaction temperature on carbon disulfide production is investigated numerically in a reforming reactor. The chemical reaction model is based on an assumed Probability Density Function (PDF) parameterized by the mean and variance of mixture fraction and β-PDF shape. The results show that the major factors influencing CS2 production are reactor temperature. The yield of carbon disulfide increases with increasing H2S to CH4 feed gas ratio (H2S/CH4≤4). Also the yield of C(s) increases with increasing temperature until the temperature reaches to 1000°K, and then due to increase of CS2 production and consumption of C(s), yield of C(s) drops with further increase in the temperature. The predicted CH4 and H2S conversion and yield of carbon disulfide are in good agreement with result of Huang and TRaissi.

Low Energy Method for Data Delivery in Ubiquitous Network

Recent advances in wireless sensor networks have led to many routing methods designed for energy-efficiency in wireless sensor networks. Despite that many routing methods have been proposed in USN, a single routing method cannot be energy-efficient if the environment of the ubiquitous sensor network varies. We present the controlling network access to various hosts and the services they offer, rather than on securing them one by one with a network security model. When ubiquitous sensor networks are deployed in hostile environments, an adversary may compromise some sensor nodes and use them to inject false sensing reports. False reports can lead to not only false alarms but also the depletion of limited energy resource in battery powered networks. The interleaved hop-by-hop authentication scheme detects such false reports through interleaved authentication. This paper presents a LMDD (Low energy method for data delivery) algorithm that provides energy-efficiency by dynamically changing protocols installed at the sensor nodes. The algorithm changes protocols based on the output of the fuzzy logic which is the fitness level of the protocols for the environment.

The Effect of Geometry Dimensions on the Earthquake Response of the Finite Element Method

In this paper, the effect of width and height of the model on the earthquake response in the finite element method is discussed. For this purpose an earth dam as a soil structure under earthquake has been considered. Various dam-foundation models are analyzed by Plaxis, a finite element package for solving geotechnical problems. The results indicate considerable differences in the seismic responses.

Numerical Investigation of Flow Patterns and Thermal Comfort in Air-Conditioned Lecture Rooms

The present paper was concerned primarily with the analysis, simulation of the air flow and thermal patterns in a lecture room. The paper is devoted to numerically investigate the influence of location and number of ventilation and air conditioning supply and extracts openings on air flow properties in a lecture room. The work focuses on air flow patterns, thermal behaviour in lecture room where large number of students. The effectiveness of an air flow system is commonly assessed by the successful removal of sensible and latent loads from occupants with additional of attaining air pollutant at a prescribed level to attain the human thermal comfort conditions and to improve the indoor air quality; this is the main target during the present paper. The study is carried out using computational fluid dynamics (CFD) simulation techniques as embedded in the commercially available CFD code (FLUENT 6.2). The CFD modelling techniques solved the continuity, momentum and energy conservation equations in addition to standard k – ε model equations for turbulence closure. Throughout the investigations, numerical validation is carried out by way of comparisons of numerical and experimental results. Good agreement is found among both predictions.

A CFD Study of Turbulent Convective Heat Transfer Enhancement in Circular Pipeflow

Addition of milli or micro sized particles to the heat transfer fluid is one of the many techniques employed for improving heat transfer rate. Though this looks simple, this method has practical problems such as high pressure loss, clogging and erosion of the material of construction. These problems can be overcome by using nanofluids, which is a dispersion of nanosized particles in a base fluid. Nanoparticles increase the thermal conductivity of the base fluid manifold which in turn increases the heat transfer rate. Nanoparticles also increase the viscosity of the basefluid resulting in higher pressure drop for the nanofluid compared to the base fluid. So it is imperative that the Reynolds number (Re) and the volume fraction have to be optimum for better thermal hydraulic effectiveness. In this work, the heat transfer enhancement using aluminium oxide nanofluid using low and high volume fraction nanofluids in turbulent pipe flow with constant wall temperature has been studied by computational fluid dynamic modeling of the nanofluid flow adopting the single phase approach. Nanofluid, up till a volume fraction of 1% is found to be an effective heat transfer enhancement technique. The Nusselt number (Nu) and friction factor predictions for the low volume fractions (i.e. 0.02%, 0.1 and 0.5%) agree very well with the experimental values of Sundar and Sharma (2010). While, predictions for the high volume fraction nanofluids (i.e. 1%, 4% and 6%) are found to have reasonable agreement with both experimental and numerical results available in the literature. So the computationally inexpensive single phase approach can be used for heat transfer and pressure drop prediction of new nanofluids.

Mirror Neuron System Study on Elderly Using Dynamic Causal Modeling fMRI Analysis

Dynamic Causal Modeling (DCM) functional Magnetic Resonance Imaging (fMRI) is a promising technique to study the connectivity among brain regions and effects of stimuli through modeling neuronal interactions from time-series neuroimaging. The aim of this study is to study characteristics of a mirror neuron system (MNS) in elderly group (age: 60-70 years old). Twenty volunteers were MRI scanned with visual stimuli to study a functional brain network. DCM was employed to determine the mechanism of mirror neuron effects. The results revealed major activated areas including precentral gyrus, inferior parietal lobule, inferior occipital gyrus, and supplementary motor area. When visual stimuli were presented, the feed-forward connectivity from visual area to conjunction area was increased and forwarded to motor area. Moreover, the connectivity from the conjunction areas to premotor area was also increased. Such findings can be useful for future diagnostic process for elderly with diseases such as Parkinson-s and Alzheimer-s.

A Robust LS-SVM Regression

In comparison to the original SVM, which involves a quadratic programming task; LS–SVM simplifies the required computation, but unfortunately the sparseness of standard SVM is lost. Another problem is that LS-SVM is only optimal if the training samples are corrupted by Gaussian noise. In Least Squares SVM (LS–SVM), the nonlinear solution is obtained, by first mapping the input vector to a high dimensional kernel space in a nonlinear fashion, where the solution is calculated from a linear equation set. In this paper a geometric view of the kernel space is introduced, which enables us to develop a new formulation to achieve a sparse and robust estimate.

Gas Flow Rate Identification in Biomass Power Plants by Response Surface Method

The utilize of renewable energy sources becomes more crucial and fascinatingly, wider application of renewable energy devices at domestic, commercial and industrial levels is not only affect to stronger awareness but also significantly installed capacities. Moreover, biomass principally is in form of woods and converts to be energy for using by humans for a long time. Gasification is a process of conversion of solid carbonaceous fuel into combustible gas by partial combustion. Many gasified models have various operating conditions because the parameters kept in each model are differentiated. This study applied the experimental data including three inputs variables including biomass consumption; temperature at combustion zone and ash discharge rate and gas flow rate as only one output variable. In this paper, response surface methods were applied for identification of the gasified system equation suitable for experimental data. The result showed that linear model gave superlative results.

Investigations of Free-to-Roll Motions and its Active Control under Pitch-up Maneuvers

Experiments have been carried out at sub-critical Reynolds number to investigate free-to-roll motions induced by forebody and/or wings complex flow on a 30° swept back nonslender wings-slender body-model for static and dynamic (pitch-up) cases. For the dynamic (pitch-up) case it has been observed that roll amplitude decreases and lag increases with increase in pitching speed. Decrease in roll amplitude with increase in pitch rate is attributed to low disturbing rolling moment due to weaker interaction between forebody and wing flow components. Asymmetric forebody vortices dominate and control the roll motion of the model in dynamic case when non-dimensional pitch rate ≥ 1x10-2. Effectiveness of the active control scheme utilizing rotating nose with artificial tip perturbation is observed to be low in the angle of attack region where the complex flow over the wings has contributions from both forebody and wings.

The Effects of Processing and Preservation on the Sensory Qualities of Prickly Pear Juice

Prickly pear juice has received renewed attention with regard to the effects of processing and preservation on its sensory qualities (colour, taste, flavour, aroma, astringency, visual browning and overall acceptability). Juice was prepared by homogenizing fruit and treating the pulp with pectinase (Aspergillus niger). Juice treatments applied were sugar addition, acidification, heat-treatment, refrigeration, and freezing and thawing. Prickly pear pulp and juice had unique properties (low pH 3.88, soluble solids 3.68 oBrix and high titratable acidity 0.47). Sensory profiling and descriptive analyses revealed that non-treated juice had a bitter taste with high astringency whereas treated prickly pear was significantly sweeter. All treated juices had a good sensory acceptance with values approximating or exceeding 7. Regression analysis of the consumer sensory attributes for non-treated prickly pear juice indicated an overwhelming rejection, while treated prickly pear juice received overall acceptability. Thus, educed favourable sensory responses and may have positive implications for consumer acceptability.

CScheme in Traditional Concurrency Problems

CScheme, a concurrent programming paradigm based on scheme concept enables concurrency schemes to be constructed from smaller synchronization units through a GUI based composer and latter be reused on other concurrency problems of a similar nature. This paradigm is particularly important in the multi-core environment prevalent nowadays. In this paper, we demonstrate techniques to separate concurrency from functional code using the CScheme paradigm. Then we illustrate how the CScheme methodology can be used to solve some of the traditional concurrency problems – critical section problem, and readers-writers problem - using synchronization schemes such as Single Threaded Execution Scheme, and Readers Writers Scheme.

Social Relation between the Malays and Chinese Communities from a Civilizational Perspectives

Towards the end of 19th century, the discovery of tin and the growing importance of rubber, had led Malaya to once again become the centre of attraction to western colonization, which later on caused the region to be influxed by cheap labour from China and India. One of the factors which attracted the alien communities was the characteristics of social relation offered by the Malays. If one analyzes the history of social relation of the Malays either among themselves or their relation with alien communities, it is apparent that the community places high regards to values such as tolerant, cooperative, respectful and helpful with each other. In fact, all these values are deeply rooted in the value of 'budi'. With the arrival of Islam, the value of 'budi' had been well assimilated with Islamic values thus giving birth to the value of 'budi-Islam'. Through 'budi- Islam', the Malay conducted their dealings with British as well the other communities during the time of peace or conflict. This value is well nurtured due to the geographical circumstances like the fertile, naturally rich land and bountiful marine life. Besides, a set of Malay customs known as 'adat' custom contributed in enhancing the values of budi.

String Matching using Inverted Lists

This paper proposes a new solution to string matching problem. This solution constructs an inverted list representing a  string pattern to be searched for. It then uses a new algorithm to process an input string in a single pass. The preprocessing phase  takes 1) time complexity O(m) 2) space complexity O(1) where m is  the length of pattern. The searching phase time complexity takes 1)  O(m+α ) in average case 2) O(n/m) in the best case and 3) O(n) in  the worst case, where α is the number of comparing leading to  mismatch and n is the length of input text.

Modeling the Fischer-Tropsch Reaction In a Slurry Bubble Column Reactor

Fischer-Tropsch synthesis is one of the most important catalytic reactions that convert the synthetic gas to light and heavy hydrocarbons. One of the main issues is selecting the type of reactor. The slurry bubble reactor is suitable choice for Fischer- Tropsch synthesis because of its good qualification to transfer heat and mass, high durability of catalyst, low cost maintenance and repair. The more common catalysts for Fischer-Tropsch synthesis are Iron-based and Cobalt-based catalysts, the advantage of these catalysts on each other depends on which type of hydrocarbons we desire to produce. In this study, Fischer-Tropsch synthesis is modeled with Iron and Cobalt catalysts in a slurry bubble reactor considering mass and momentum balance and the hydrodynamic relations effect on the reactor behavior. Profiles of reactant conversion and reactant concentration in gas and liquid phases were determined as the functions of residence time in the reactor. The effects of temperature, pressure, liquid velocity, reactor diameter, catalyst diameter, gasliquid and liquid-solid mass transfer coefficients and kinetic coefficients on the reactant conversion have been studied. With 5% increase of liquid velocity (with Iron catalyst), H2 conversions increase about 6% and CO conversion increase about 4%, With 8% increase of liquid velocity (with Cobalt catalyst), H2 conversions increase about 26% and CO conversion increase about 4%. With 20% increase of gas-liquid mass transfer coefficient (with Iron catalyst), H2 conversions increase about 12% and CO conversion increase about 10% and with Cobalt catalyst H2 conversions increase about 10% and CO conversion increase about 6%. Results show that the process is sensitive to gas-liquid mass transfer coefficient and optimum condition operation occurs in maximum possible liquid velocity. This velocity must be more than minimum fluidization velocity and less than terminal velocity in such a way that avoid catalysts particles from leaving the fluidized bed.

Gene Expression Data Classification Using Discriminatively Regularized Sparse Subspace Learning

Sparse representation which can represent high dimensional data effectively has been successfully used in computer vision and pattern recognition problems. However, it doesn-t consider the label information of data samples. To overcome this limitation, we develop a novel dimensionality reduction algorithm namely dscriminatively regularized sparse subspace learning(DR-SSL) in this paper. The proposed DR-SSL algorithm can not only make use of the sparse representation to model the data, but also can effective employ the label information to guide the procedure of dimensionality reduction. In addition,the presented algorithm can effectively deal with the out-of-sample problem.The experiments on gene-expression data sets show that the proposed algorithm is an effective tool for dimensionality reduction and gene-expression data classification.

Intellectual Capital and Competitive Advantage: An Analysis of the Biotechnology Industry

Intellectual capital measurement is a central aspect of knowledge management. The measurement and the evaluation of intangible assets play a key role in allowing an effective management of these assets as sources of competitiveness. For these reasons, managers and practitioners need conceptual and analytical tools taking into account the unique characteristics and economic significance of Intellectual Capital. Following this lead, we propose an efficiency and productivity analysis of Intellectual Capital, as a determinant factor of the company competitive advantage. The analysis is carried out by means of Data Envelopment Analysis (DEA) and Malmquist Productivity Index (MPI). These techniques identify Bests Practice companies that have accomplished competitive advantage implementing successful strategies of Intellectual Capital management, and offer to inefficient companies development paths by means of benchmarking. The proposed methodology is employed on the Biotechnology industry in the period 2007-2010.

Synchronization of Oestrus in Goats with Progestogen Sponges and Short Term Combined FGA, PGF2α Protocols

The study aimed to evaluated the reproductive performance response to short term oestrus synchronization during the transition period. One hundred and sixty-five indigenous multiparous non-lactating goats were subdivided into the following six treatment groups for oestrus synchronization: NT control Group (N= 30), Fe-21d, FGA vaginal sponge for 21days+eCG at 19thd; FPe- 11d, FGA 11d + PGF2α and eCG at 9th d; FPe-10d, FGA 10d+ PGF2α and eCG at 8th d; FPe-9d, FGA 9d +PGF2α and eCG at 7thd; PFe-5d, PGF2α at d0 + FGA 5d + eCG at 5thd. The goats were natural mated (1 male/6 females). Fecundity rates (n. births /n. females treated x 100) were statistically higher (P < 0.05) in short term FPe-9d (157.9%), FPe- 11d (115.4%), FPe-10d (111.1%) and PFe-5d (107.7%) groups compared to the NT control Group (66.7%).