Kerma Profile Measurements in CT Chest Scans– a Comparison of Methodologies

The Brazilian legislation has only established diagnostic reference levels (DRLs) in terms of Multiple Scan Average Dose (MSAD) as a quality control parameter for computed tomography (CT) scanners. Compliance with DRLs can be verified by measuring the Computed Tomography Kerma Index (Ca,100) with a pencil ionization chamber or by obtaining the kerma distribution in CT scans with radiochromic films or rod shape lithium fluoride termoluminescent dosimeters (TLD-100). TL dosimeters were used to record kerma profiles and to determine MSAD values of a Bright Speed model GE CT scanner. Measurements were done with radiochromic films and TL dosimeters distributed in cylinders positioned in the center and in four peripheral bores of a standard polymethylmethacrylate (PMMA) body CT dosimetry phantom. Irradiations were done using a protocol for adult chest. The maximum values were found at the midpoint of the longitudinal axis. The MSAD values obtained with three dosimetric techniques were compared.

Peer-to-Peer Epidemic Algorithms for Reliable Multicasting in Ad Hoc Networks

Characteristics of ad hoc networks and even their existence depend on the nodes forming them. Thus, services and applications designed for ad hoc networks should adapt to this dynamic and distributed environment. In particular, multicast algorithms having reliability and scalability requirements should abstain from centralized approaches. We aspire to define a reliable and scalable multicast protocol for ad hoc networks. Our target is to utilize epidemic techniques for this purpose. In this paper, we present a brief survey of epidemic algorithms for reliable multicasting in ad hoc networks, and describe formulations and analytical results for simple epidemics. Then, P2P anti-entropy algorithm for content distribution and our prototype simulation model are described together with our initial results demonstrating the behavior of the algorithm.

The Influence of Socio-Economic Backgrounds towards Satisfaction with Student Housing Facilities

Studies on residential satisfaction have been actively discussed under family house setting. However, limited studies have been conducted on student residential satisfaction. This study is an attempt to fill the research gap. It focuses on the influence of socioeconomic on students- satisfaction with the universities- student housing facilities. The students who stayed at the on-campus student housing were the respondents. This study employed two-stage cluster sampling method in classifying the respondents. Self-administered questionnaires were distributed face-to-face to the students. In general, it is confirmed that students- socio-economic backgrounds have influence on the students- satisfaction with their housing facilities. The main influential factors were the students- economic status, sense of sharing, and ethnicity of their roommates. Furthermore, this study could also provide a useful feedback for the universities in order to improve their student housing facilities.

Replicating Data Objects in Large-scale Distributed Computing Systems using Extended Vickrey Auction

This paper proposes a novel game theoretical technique to address the problem of data object replication in largescale distributed computing systems. The proposed technique draws inspiration from computational economic theory and employs the extended Vickrey auction. Specifically, players in a non-cooperative environment compete for server-side scarce memory space to replicate data objects so as to minimize the total network object transfer cost, while maintaining object concurrency. Optimization of such a cost in turn leads to load balancing, fault-tolerance and reduced user access time. The method is experimentally evaluated against four well-known techniques from the literature: branch and bound, greedy, bin-packing and genetic algorithms. The experimental results reveal that the proposed approach outperforms the four techniques in both the execution time and solution quality.

Svision: Visual Identification of Scanning and Denial of Service Attacks

We propose a novel graphical technique (SVision) for intrusion detection, which pictures the network as a community of hosts independently roaming in a 3D space defined by the set of services that they use. The aim of SVision is to graphically cluster the hosts into normal and abnormal ones, highlighting only the ones that are considered as a threat to the network. Our experimental results using DARPA 1999 and 2000 intrusion detection and evaluation datasets show the proposed technique as a good candidate for the detection of various threats of the network such as vertical and horizontal scanning, Denial of Service (DoS), and Distributed DoS (DDoS) attacks.

Distribution Feeder Reconfiguration Considering Distributed Generators

Recently, distributed generation technologies have received much attention for the potential energy savings and reliability assurances that might be achieved as a result of their widespread adoption. Fueling the attention have been the possibilities of international agreements to reduce greenhouse gas emissions, electricity sector restructuring, high power reliability requirements for certain activities, and concern about easing transmission and distribution capacity bottlenecks and congestion. So it is necessary that impact of these kinds of generators on distribution feeder reconfiguration would be investigated. This paper presents an approach for distribution reconfiguration considering Distributed Generators (DGs). The objective function is summation of electrical power losses A Tabu search optimization is used to solve the optimal operation problem. The approach is tested on a real distribution feeder.

Voltage-Controllable Liquid Crystals Lens

This study investigates a voltage-controllable liquid crystals lens with a Fresnel zone electrode. When applying a proper voltage on the liquid crystal cell, a Fresnel-zone-distributed electric field is induced to direct liquid crystals aligned in a concentric structure. Owing to the concentrically aligned liquid crystals, a Fresnel lens is formed. We probe the Fresnel liquid crystal lens using a polarized incident beam with a wavelength of 632.8 nm, finding that the diffraction efficiency depends on the applying voltage. A remarkable diffraction efficiency of ~39.5 % is measured at the voltage of 0.9V. Additionally, a dual focus lens is fabricated by attaching a plane-convex lens to the Fresnel liquid crystals cell. The Fresnel LC lens and the dual focus lens may be applied for DVD/CD pick-up head, confocal microscopy system, or electrically-controlling optical systems.

Distributed Data-Mining by Probability-Based Patterns

In this paper a new method is suggested for distributed data-mining by the probability patterns. These patterns use decision trees and decision graphs. The patterns are cared to be valid, novel, useful, and understandable. Considering a set of functions, the system reaches to a good pattern or better objectives. By using the suggested method we will be able to extract the useful information from massive and multi-relational data bases.

Free Vibration Analysis of Smart FGM Plates

Analytical investigation of the free vibration behavior of circular functionally graded (FG) plates integrated with two uniformly distributed actuator layers made of piezoelectric (PZT4) material on the top and bottom surfaces of the circular FG plate based on the classical plate theory (CPT) is presented in this paper. The material properties of the functionally graded substrate plate are assumed to be graded in the thickness direction according to the power-law distribution in terms of the volume fractions of the constituents and the distribution of electric potential field along the thickness direction of piezoelectric layers is simulated by a quadratic function. The differential equations of motion are solved analytically for clamped edge boundary condition of the plate. The detailed mathematical derivations are presented and Numerical investigations are performed for FG plates with two surface-bonded piezoelectric layers. Emphasis is placed on investigating the effect of varying the gradient index of FG plate on the free vibration characteristics of the structure. The results are verified by those obtained from threedimensional finite element analyses.

Power System Load Shedding: Key Issues and New Perspectives

Optimal load shedding (LS) design as an emergency plan is one of the main control challenges posed by emerging new uncertainties and numerous distributed generators including renewable energy sources in a modern power system. This paper presents an overview of the key issues and new challenges on optimal LS synthesis concerning the integration of wind turbine units into the power systems. Following a brief survey on the existing LS methods, the impact of power fluctuation produced by wind powers on system frequency and voltage performance is presented. The most LS schemas proposed so far used voltage or frequency parameter via under-frequency or under-voltage LS schemes. Here, the necessity of considering both voltage and frequency indices to achieve a more effective and comprehensive LS strategy is emphasized. Then it is clarified that this problem will be more dominated in the presence of wind turbines.

Energy-Efficient Electrical Power Distribution with Multi-Agent Control at Parallel DC/DC Converters

Consumer electronics are pervasive. It is impossible to imagine a household or office without DVD players, digital cameras, printers, mobile phones, shavers, electrical toothbrushes, etc. All these devices operate at different voltage levels ranging from 1.8 to 20 VDC, in the absence of universal standards. The voltages available are however usually 120/230 VAC at 50/60 Hz. This situation makes an individual electrical energy conversion system necessary for each device. Such converters usually involve several conversion stages and often operate with excessive losses and poor reliability. The aim of the project presented in this paper is to design and implement a multi-channel DC/DC converter system, customizing the output voltage and current ratings according to the requirements of the load. Distributed, multi-agent techniques will be applied for the control of the DC/DC converters.

Computer Aided X-Ray Diffraction Intensity Analysis for Spinels: Hands-On Computing Experience

The mineral having chemical compositional formula MgAl2O4 is called “spinel". The ferrites crystallize in spinel structure are known as spinel-ferrites or ferro-spinels. The spinel structure has a fcc cage of oxygen ions and the metallic cations are distributed among tetrahedral (A) and octahedral (B) interstitial voids (sites). The X-ray diffraction (XRD) intensity of each Bragg plane is sensitive to the distribution of cations in the interstitial voids of the spinel lattice. This leads to the method of determination of distribution of cations in the spinel oxides through XRD intensity analysis. The computer program for XRD intensity analysis has been developed in C language and also tested for the real experimental situation by synthesizing the spinel ferrite materials Mg0.6Zn0.4AlxFe2- xO4 and characterized them by X-ray diffractometry. The compositions of Mg0.6Zn0.4AlxFe2-xO4(x = 0.0 to 0.6) ferrites have been prepared by ceramic method and powder X-ray diffraction patterns were recorded. Thus, the authenticity of the program is checked by comparing the theoretically calculated data using computer simulation with the experimental ones. Further, the deduced cation distributions were used to fit the magnetization data using Localized canting of spins approach to explain the “recovery" of collinear spin structure due to Al3+ - substitution in Mg-Zn ferrites which is the case if A-site magnetic dilution and non-collinear spin structure. Since the distribution of cations in the spinel ferrites plays a very important role with regard to their electrical and magnetic properties, it is essential to determine the cation distribution in spinel lattice.

Measurement of Lead Pollution in the Air of Babylon Governorate/Iraq during Year 2010

This research aims to study the lead pollution in the air of Babylon governorate that resulted generally from vehicles exhausts in addition to industrial and human activities.Vehicles number in Babylon governorate increased significantly after year 2003 that resulted with increase in lead emissions into the air.Measurement of lead emissions was done in seven stations distributed randomly in Babylon governorate. These stations where located in Industrial (Al-Sena'ay) Quarter, 60 street (near to Babylon sewer directorate), 40 Street (near to the first intersection), Al-Hashmia city, Al-Mahaweel city, , Al- Musayab city in addition to another station in Sayd Idris village belong to Abugharaq district (Agricultural station for comparison). The measured concentrations in these stations were compared with the standard limits of Environmental Protection Agency EPA (2 μg /m3). The results of this study showed that the average of lead concentrations ,in Babylon governorate during year 2010, was (3.13 μg/m3) which was greater than standard limits (2 μg/m3). The maximum concentration of lead was (6.41 μg / m3) recorded in the Industrial (Al-Sena'ay) Quarter during April month, while the minimum concentrations was (0.36 μg / m3) recorded in the agricultural station (Abugharaq) during December month.

A Practical Approach for Testing the Process Quality

Process capability index Cpk is the most widely used index in making managerial decisions since it provides bounds on the process yield for normally distributed processes. However, existent methods for assessing process performance which constructed by statistical inference may unfortunately lead to fine results, because uncertainties exist in most real-world applications. Thus, this study adopts fuzzy inference to deal with testing of Cpk . A brief score is obtained for assessing a supplier’s process instead of a severe evaluation.

C@sa: Intelligent Home Control and Simulation

In this paper, we present C@sa, a multiagent system aiming at modeling, controlling and simulating the behavior of an intelligent house. The developed system aims at providing to architects, designers and psychologists a simulation and control tool for understanding which is the impact of embedded and pervasive technology on people daily life. In this vision, the house is seen as an environment made up of independent and distributed devices, controlled by agents, interacting to support user's goals and tasks.

Prioritizing Influential Factors on the Promotion of Virtual Training System

In today's world where everything is rapidly changing and information technology is high in development, many features of culture, society, politic and economy has changed. The advent of information technology and electronic data transmission lead to easy communication and fields like e-learning and e-commerce, are accessible for everyone easily. One of these technologies is virtual training. The "quality" of such kind of education systems is critical. 131 questionnaires were prepared and distributed among university student in Toba University. So the research has followed factors that affect the quality of learning from the perspective of staff, students, professors and this type of university. It is concluded that the important factors in virtual training are the quality of professors, the quality of staff, and the quality of the university. These mentioned factors were the most prior factors in this education system and necessary for improving virtual training.

Time Comparative Simulator for Distributed Process Scheduling Algorithms

In any distributed systems, process scheduling plays a vital role in determining the efficiency of the system. Process scheduling algorithms are used to ensure that the components of the system would be able to maximize its utilization and able to complete all the processes assigned in a specified period of time. This paper focuses on the development of comparative simulator for distributed process scheduling algorithms. The objectives of the works that have been carried out include the development of the comparative simulator, as well as to implement a comparative study between three distributed process scheduling algorithms; senderinitiated, receiver-initiated and hybrid sender-receiver-initiated algorithms. The comparative study was done based on the Average Waiting Time (AWT) and Average Turnaround Time (ATT) of the processes involved. The simulation results show that the performance of the algorithms depends on the number of nodes in the system.

Does Practice Reflect Theory? An Exploratory Study of a Successful Knowledge Management System

To investigate the correspondence of theory and practice, a successfully implemented Knowledge Management System (KMS) is explored through the lens of Alavi and Leidner-s proposed KMS framework for the analysis of an information system in knowledge management (Framework-AISKM). The applied KMS system was designed to manage curricular knowledge in a distributed university environment. The motivation for the KMS is discussed along with the types of knowledge necessary in an academic setting. Elements of the KMS involved in all phases of capturing and disseminating knowledge are described. As the KMS matures the resulting data stores form the precursor to and the potential for knowledge mining. The findings from this exploratory study indicate substantial correspondence between the successful KMS and the theory-based framework providing provisional confirmation for the framework while suggesting factors that contributed to the system-s success. Avenues for future work are described.

Performance Assessment of Computational Gridon Weather Indices from HOAPS Data

Long term rainfall analysis and prediction is a challenging task especially in the modern world where the impact of global warming is creating complications in environmental issues. These factors which are data intensive require high performance computational modeling for accurate prediction. This research paper describes a prototype which is designed and developed on grid environment using a number of coupled software infrastructural building blocks. This grid enabled system provides the demanding computational power, efficiency, resources, user-friendly interface, secured job submission and high throughput. The results obtained using sequential execution and grid enabled execution shows that computational performance has enhanced among 36% to 75%, for decade of climate parameters. Large variation in performance can be attributed to varying degree of computational resources available for job execution. Grid Computing enables the dynamic runtime selection, sharing and aggregation of distributed and autonomous resources which plays an important role not only in business, but also in scientific implications and social surroundings. This research paper attempts to explore the grid enabled computing capabilities on weather indices from HOAPS data for climate impact modeling and change detection.

Distributed Detection and Optimal Traffic-blocking of Network Worms

Despite the recent surge of research in control of worm propagation, currently, there is no effective defense system against such cyber attacks. We first design a distributed detection architecture called Detection via Distributed Blackholes (DDBH). Our novel detection mechanism could be implemented via virtual honeypots or honeynets. Simulation results show that a worm can be detected with virtual honeypots on only 3% of the nodes. Moreover, the worm is detected when less than 1.5% of the nodes are infected. We then develop two control strategies: (1) optimal dynamic trafficblocking, for which we determine the condition that guarantees minimum number of removed nodes when the worm is contained and (2) predictive dynamic traffic-blocking–a realistic deployment of the optimal strategy on scale-free graphs. The predictive dynamic traffic-blocking, coupled with the DDBH, ensures that more than 40% of the network is unaffected by the propagation at the time when the worm is contained.