A Study of Gaps in CBMIR Using Different Methods and Prospective

In recent years, rapid advances in software and hardware in the field of information technology along with a digital imaging revolution in the medical domain facilitate the generation and storage of large collections of images by hospitals and clinics. To search these large image collections effectively and efficiently poses significant technical challenges, and it raises the necessity of constructing intelligent retrieval systems. Content-based Image Retrieval (CBIR) consists of retrieving the most visually similar images to a given query image from a database of images[5]. Medical CBIR (content-based image retrieval) applications pose unique challenges but at the same time offer many new opportunities. On one hand, while one can easily understand news or sports videos, a medical image is often completely incomprehensible to untrained eyes.

An Efficient Technique for EMI Mitigation in Fluorescent Lamps using Frequency Modulation and Evolutionary Programming

Electromagnetic interference (EMI) is one of the serious problems in most electrical and electronic appliances including fluorescent lamps. The electronic ballast used to regulate the power flow through the lamp is the major cause for EMI. The interference is because of the high frequency switching operation of the ballast. Formerly, some EMI mitigation techniques were in practice, but they were not satisfactory because of the hardware complexity in the circuit design, increased parasitic components and power consumption and so on. The majority of the researchers have their spotlight only on EMI mitigation without considering the other constraints such as cost, effective operation of the equipment etc. In this paper, we propose a technique for EMI mitigation in fluorescent lamps by integrating Frequency Modulation and Evolutionary Programming. By the Frequency Modulation technique, the switching at a single central frequency is extended to a range of frequencies, and so, the power is distributed throughout the range of frequencies leading to EMI mitigation. But in order to meet the operating frequency of the ballast and the operating power of the fluorescent lamps, an optimal modulation index is necessary for Frequency Modulation. The optimal modulation index is determined using Evolutionary Programming. Thereby, the proposed technique mitigates the EMI to a satisfactory level without disturbing the operation of the fluorescent lamp.

Tool Failure Detection Based on Statistical Analysis of Metal Cutting Acoustic Emission Signals

The analysis of Acoustic Emission (AE) signal generated from metal cutting processes has often approached statistically. This is due to the stochastic nature of the emission signal as a result of factors effecting the signal from its generation through transmission and sensing. Different techniques are applied in this manner, each of which is suitable for certain processes. In metal cutting where the emission generated by the deformation process is rather continuous, an appropriate method for analysing the AE signal based on the root mean square (RMS) of the signal is often used and is suitable for use with the conventional signal processing systems. The aim of this paper is to set a strategy in tool failure detection in turning processes via the statistic analysis of the AE generated from the cutting zone. The strategy is based on the investigation of the distribution moments of the AE signal at predetermined sampling. The skews and kurtosis of these distributions are the key elements in the detection. A normal (Gaussian) distribution has first been suggested then this was eliminated due to insufficiency. The so called Beta distribution was then considered, this has been used with an assumed β density function and has given promising results with regard to chipping and tool breakage detection.

Compression and Filtering of Random Signals under Constraint of Variable Memory

We study a new technique for optimal data compression subject to conditions of causality and different types of memory. The technique is based on the assumption that some information about compressed data can be obtained from a solution of the associated problem without constraints of causality and memory. This allows us to consider two separate problem related to compression and decompression subject to those constraints. Their solutions are given and the analysis of the associated errors is provided.

Hardiness vs Alienation Personality Construct Essentially Explains Burnout Proclivity and Erroneous Computer Entry Problems in Rural Hellenic Hospital Labs

Erroneous computer entry problems [here: 'e'errors] in hospital labs threaten the patients-–health carers- relationship, undermining the health system credibility. Are e-errors random, and do lab professionals make them accidentally, or may they be traced through meaningful determinants? Theories on internal causality of mistakes compel to seek specific causal ascriptions of hospital lab eerrors instead of accepting some inescapability. Undeniably, 'To Err is Human'. But in view of rapid global health organizational changes, e-errors are too expensive to lack in-depth considerations. Yet, that efunction might supposedly be entrenched in the health carers- job description remains under dispute – at least for Hellenic labs, where e-use falls behind generalized(able) appreciation and application. In this study: i) an empirical basis of a truly high annual cost of e-errors at about €498,000.00 per rural Hellenic hospital was established, hence interest in exploring the issue was sufficiently substantiated; ii) a sample of 270 lab-expert nurses, technicians and doctors were assessed on several personality, burnout and e-error measures, and iii) the hypothesis that the Hardiness vs Alienation personality construct disposition explains resistance vs proclivity to e-errors was tested and verified: Hardiness operates as a resilience source in the encounter of high pressures experienced in the hospital lab, whereas its 'opposite', i.e., Alienation, functions as a predictor, not only of making e-errors, but also of leading to burn-out. Implications for apt interventions are discussed.

Adsorption of Textile Reactive Dye by Palm Shell Activated Carbon: Response Surface Methodology

The adsorption of simulated aqueous solution containing textile remazol reactive dye, namely Red 3BS by palm shell activated carbon (PSAC) as adsorbent was carried out using Response Surface Methodology (RSM). A Box-Behnken design in three most important operating variables; initial dye concentration, dosage of adsorbent and speed of impeller was employed for experimental design and optimization of results. The significance of independent variables and their interactions were tested by means of the analysis of variance (ANOVA) with 95% confidence limits. Model indicated that with the increasing of dosage and speed give the result of removal up to 90% with the capacity uptake more than 7 mg/g. High regression coefficient between the variables and the response (R-Sq = 93.9%) showed of good evaluation of experimental data by polynomial regression model.

Gait Recognition System: Bundle Rectangle Approach

Biometrics methods include recognition techniques such as fingerprint, iris, hand geometry, voice, face, ears and gait. The gait recognition approach has some advantages, for example it does not need the prior concern of the observed subject and it can record many biometric features in order to make deeper analysis, but most of the research proposals use high computational cost. This paper shows a gait recognition system with feature subtraction on a bundle rectangle drawn over the observed person. Statistical results within a database of 500 videos are shown.

Split-Pipe Design of Water Distribution Network Using Simulated Annealing

In this paper a procedure for the split-pipe design of looped water distribution network based on the use of simulated annealing is proposed. Simulated annealing is a heuristic-based search algorithm, motivated by an analogy of physical annealing in solids. It is capable for solving the combinatorial optimization problem. In contrast to the split-pipe design that is derived from a continuous diameter design that has been implemented in conventional optimization techniques, the split-pipe design proposed in this paper is derived from a discrete diameter design where a set of pipe diameters is chosen directly from a specified set of commercial pipes. The optimality and feasibility of the solutions are found to be guaranteed by using the proposed method. The performance of the proposed procedure is demonstrated through solving the three well-known problems of water distribution network taken from the literature. Simulated annealing provides very promising solutions and the lowest-cost solutions are found for all of these test problems. The results obtained from these applications show that simulated annealing is able to handle a combinatorial optimization problem of the least cost design of water distribution network. The technique can be considered as an alternative tool for similar areas of research. Further applications and improvements of the technique are expected as well.

State Dependent Riccati Equation Based Roll Autopilot for 122mm Artillery Rocket

State-dependent Riccati equation based controllers are becoming increasingly popular because of having attractive properties like optimality, stability and robustness. This paper focuses on the design of a roll autopilot for a fin stabilized and canard controlled 122mm artillery rocket using state-dependent Riccati equation technique. Initial spin is imparted to rocket during launch and it quickly decays due to straight tail fins. After the spin phase, the roll orientation of rocket is brought to zero with the canard deflection commands generated by the roll autopilot. Roll autopilot has been developed by considering uncoupled roll, pitch and yaw channels. The canard actuator is modeled as a second-order nonlinear system. Elements of the state weighing matrix for Riccati equation have been chosen to be state dependent to exploit the design flexibility offered by the Riccati equation technique. Simulation results under varying conditions of flight demonstrate the wide operating range of the proposed autopilot.

Signalling Cost Analysis of PDE-NEMO

A Personal Distributed Environment (PDE) is an example of an IP-based system architecture designed for future mobile communications. In a single PDE, there exist several Subnetworks hosting devices located across the infrastructure, which will inter-work with one another through the coordination of a Device Management Entity (DME). Some of these Sub-networks are fixed and some are mobile. In order to support Mobile Sub-networks mobility in the PDE, the PDE-NEMO protocol was proposed. This paper discussed the signalling cost analysis of PDE-NEMO by use of a detailed simulation model. The paper started with the introduction of the protocol, followed by the experiments and results and then followed by discussions.

A ±0.5V BiCMOS Class-A Current Conveyor

In this paper, a new BiCMOS CCII and CCCII, capable of operate at ±0.5V and having wide dynamic range with achieved bandwidth of 480MHz and 430MHz respectively have been proposed. The structures have been found to be insensitive to the threshold voltage variations. The proposed circuits are suitable for implementation using 0.25μm BiCMOS technology. Pspice simulations confirm the performance of the proposed structures.

A Quantitative Approach to Strategic Design of Component-Based Business Process Models

A new paradigm for software design and development models software by its business process, translates the model into a process execution language, and has it run by a supporting execution engine. This process-oriented paradigm promotes modeling of software by less technical users or business analysts as well as rapid development. Since business process models may be shared by different organizations and sometimes even by different business domains, it is interesting to apply a technique used in traditional software component technology to design reusable business processes. This paper discusses an approach to apply a technique for software component fabrication to the design of process-oriented software units, called process components. These process components result from decomposing a business process of a particular application domain into subprocesses with an aim that the process components can be reusable in different process-based software models. The approach is quantitative because the quality of process component design is measured from technical features of the process components. The approach is also strategic because the measured quality is determined against business-oriented component management goals. A software tool has been developed to measure how good a process component design is, according to the required managerial goals and comparing to other designs. We also discuss how we benefit from reusable process components.

The Effect of Chemical Treatment on TL Glow Curves of CdS/ZnS Thin Films Deposited by Vacuum Deposition Method

The effect of chemical treatment in CdCl2 and thermal annealing in 400°C, on the defect structures of potentially useful ZnS\CdS solar cell thin films deposited onto quartz substrate and prepared by vacuum deposition method was studied using the Thermoluminesence (TL) techniques. A series of electron and hole traps are found in the various deposited samples studied. After annealing, however, it was observed that the intensity and activation energy of TL signal increases with loss of the low temperature electron traps.

A Software Framework for Predicting Oil-Palm Yield from Climate Data

Intelligent systems based on machine learning techniques, such as classification, clustering, are gaining wide spread popularity in real world applications. This paper presents work on developing a software system for predicting crop yield, for example oil-palm yield, from climate and plantation data. At the core of our system is a method for unsupervised partitioning of data for finding spatio-temporal patterns in climate data using kernel methods which offer strength to deal with complex data. This work gets inspiration from the notion that a non-linear data transformation into some high dimensional feature space increases the possibility of linear separability of the patterns in the transformed space. Therefore, it simplifies exploration of the associated structure in the data. Kernel methods implicitly perform a non-linear mapping of the input data into a high dimensional feature space by replacing the inner products with an appropriate positive definite function. In this paper we present a robust weighted kernel k-means algorithm incorporating spatial constraints for clustering the data. The proposed algorithm can effectively handle noise, outliers and auto-correlation in the spatial data, for effective and efficient data analysis by exploring patterns and structures in the data, and thus can be used for predicting oil-palm yield by analyzing various factors affecting the yield.

A Framework for Identifying the Critical Factors Affecting the Decision to Adopt and Use Inter-Organizational Information Systems

The importance of inter-organizational system (IOS) has been increasingly recognized by organizations. However, IOS adoption has proved to be difficult and, at this stage, why this is so is not fully uncovered. In practice, benefits have often remained concentrated, primarily accruing to the dominant party, resulting in low rates of adoption and usage, and often culminating in the failure of the IOS. The main research question is why organizations initiate or join IOS and what factors influence their adoption and use levels. This paper reviews the literature on IOS adoption and proposes a theoretical framework in order to identify the critical factors to capture a complete picture of IOS adoption. With our proposed critical factors, we are able to investigate their relative contributions to IOS adoption decisions. We obtain findings that suggested that there are five groups of factors that significantly affect the adoption and use decision of IOS in the Supply Chain Management (SCM) context: 1) interorganizational context, 2) organizational context, 3) technological context, 4) perceived costs, and 5) perceived benefits.

Operative Public Space for Rural Brazil Strategies for Starting Up Economic, Social and Environmental Development of Rural Communities in Emerging Countries: The Goiabeira Case

This article stands in the context of rural communities in Brazil, where, like many others emerging countries, the overwhelming increasing markets and the overcrowded cities are leaving behind informal settlements based on obsolete agricultural economies and techniques. The pilot project for the community of Goiabeira reflects the attempt to imagine a development model that privileges the actual improvement of living conditions, the education and training, the social inclusion and participation of the dwellers of rural communities. Through the inclusion of operative public space, the aim is for them to become self-sustaining, encouraging the use of local resources for appropriate architectural, ecological and energy technologies and devices, that are efficient, affordable and foster community participation, in the respect of the surrounding environment.

Extending the Conceptual Neighborhood Graph of the Relations for the Semantic Adaptation of Multimedia Documents

The recent developments in computing and communication technology permit to users to access multimedia documents with variety of devices (PCs, PDAs, mobile phones...) having heterogeneous capabilities. This diversification of supports has trained the need to adapt multimedia documents according to their execution contexts. A semantic framework for multimedia document adaptation based on the conceptual neighborhood graphs was proposed. In this framework, adapting consists on finding another specification that satisfies the target constraints and which is as close as possible from the initial document. In this paper, we propose a new way of building the conceptual neighborhood graphs to best preserve the proximity between the adapted and the original documents and to deal with more elaborated relations models by integrating the relations relaxation graphs that permit to handle the delays and the distances defined within the relations.

A Methodology to Analyze Technology Convergence: Patent-Citation Based Technology Input-Output Analysis

This research proposes a methodology for patent-citation-based technology input-output analysis by applying the patent information to input-output analysis developed for the dependencies among different industries. For this analysis, a technology relationship matrix and its components, as well as input and technology inducement coefficients, are constructed using patent information. Then, a technology inducement coefficient is calculated by normalizing the degree of citation from certain IPCs to the different IPCs (International patent classification) or to the same IPCs. Finally, we construct a Dependency Structure Matrix (DSM) based on the technology inducement coefficient to suggest a useful application for this methodology.

Investigating the Impact of Wind Speed on Active and Reactive Power Penetration to the Distribution Network

Wind power is among the most actively developing distributed generation (DG) technology. Majority of the wind power based DG technologies employ wind turbine induction generators (WTIG) instead of synchronous generators, for the technical advantages like: reduced size, increased robustness, lower cost, and increased electromechanical damping. However, dynamic changes of wind speed make the amount of active/reactive power injected/drawn to a WTIG embedded distribution network highly variable. This paper analyzes the effect of wind speed changes on the active and reactive power penetration to the wind energy embedded distribution network. Four types of wind speed changes namely; constant, linear change, gust change and random change of wind speed are considered in the analysis. The study is carried out by three-phase, non-linear, dynamic simulation of distribution system component models. Results obtained from the investigation are presented and discussed.

The Usage of Social Networks in Educational Context

Possible advantages of technology in educational context required the defining boundaries of formal and informal learning. Increasing opportunity to ubiquitous learning by technological support has revealed a question of how to discover the potential of individuals in the spontaneous environments such as social networks. This seems to be related with the question of what purposes in social networks have been being used? Social networks provide various advantages in educational context as collaboration, knowledge sharing, common interests, active participation and reflective thinking. As a consequence of these, the purpose of this study is composed of proposing a new model that could determine factors which effect adoption of social network applications for usage in educational context. While developing a model proposal, the existing adoption and diffusion models have been reviewed and they are thought to be suitable on handling an original perspective instead of using completely other diffusion or acceptance models because of different natures of education from other organizations. In the proposed model; social factors, perceived ease of use, perceived usefulness and innovativeness are determined four direct constructs that effect adoption process. Facilitating conditions, image, subjective norms and community identity are incorporated to model as antecedents of these direct four constructs.