QoS Management in the Future Internet

The talks about technological convergence had been around for almost twenty years. Today Internet made it possible. And this is not only technical evolution. The way it changed our lives reflected in variety of applications, services and technologies used in day-to-day life. Such benefits imposed even more requirements on heterogeneous and unreliable IP networks. Current paper outlines QoS management system developed in the NetQoS [1] project. It describes an overall architecture of management system for heterogeneous networks and proposes automated multi-layer QoS management. Paper focuses on the structure of the most crucial modules of the system that enable autonomous and multi-layer provisioning and dynamic adaptation.

Investigation of the GFR2400 Reactivity Control System

The presented paper is related to the design methods and neutronic characterization of the reactivity control system in the large power unit of Generation IV Gas cooled Fast Reactor – GFR2400. The reactor core is based on carbide pin fuel type with the application of refractory metallic liners used to enhance the fission product retention of the SiCcladding. The heterogeneous design optimization of control rod is presented and the results of rods worth and their interferences in a core are evaluated. In addition, the idea of reflector removal as an additive reactivity management option is investigated and briefly described.

A CT-based Monte Carlo Dose Calculations for Proton Therapy Using a New Interface Program

The purpose of this study is to introduce a new interface program to calculate a dose distribution with Monte Carlo method in complex heterogeneous systems such as organs or tissues in proton therapy. This interface program was developed under MATLAB software and includes a friendly graphical user interface with several tools such as image properties adjustment or results display. Quadtree decomposition technique was used as an image segmentation algorithm to create optimum geometries from Computed Tomography (CT) images for dose calculations of proton beam. The result of the mentioned technique is a number of nonoverlapped squares with different sizes in every image. By this way the resolution of image segmentation is high enough in and near heterogeneous areas to preserve the precision of dose calculations and is low enough in homogeneous areas to reduce the number of cells directly. Furthermore a cell reduction algorithm can be used to combine neighboring cells with the same material. The validation of this method has been done in two ways; first, in comparison with experimental data obtained with 80 MeV proton beam in Cyclotron and Radioisotope Center (CYRIC) in Tohoku University and second, in comparison with data based on polybinary tissue calibration method, performed in CYRIC. These results are presented in this paper. This program can read the output file of Monte Carlo code while region of interest is selected manually, and give a plot of dose distribution of proton beam superimposed onto the CT images.

Analysis of a Secondary Autothermal Reformer Using a Thermodynamic POX Model

Partial oxidation (POX) of light hydrocarbons (e.g. methane) is occurred in the first part of the autothermal reformer (ATR). The results of the detailed modeling of the reformer based on the thermodynamic model of the POX and 1D heterogeneous catalytic model for the fixed bed section are considered here. According to the results, the overall performance of the ATR can be improved by changing the important feed parameters.

Heterogeneous Attribute Reduction in Noisy System based on a Generalized Neighborhood Rough Sets Model

Neighborhood Rough Sets (NRS) has been proven to be an efficient tool for heterogeneous attribute reduction. However, most of researches are focused on dealing with complete and noiseless data. Factually, most of the information systems are noisy, namely, filled with incomplete data and inconsistent data. In this paper, we introduce a generalized neighborhood rough sets model, called VPTNRS, to deal with the problem of heterogeneous attribute reduction in noisy system. We generalize classical NRS model with tolerance neighborhood relation and the probabilistic theory. Furthermore, we use the neighborhood dependency to evaluate the significance of a subset of heterogeneous attributes and construct a forward greedy algorithm for attribute reduction based on it. Experimental results show that the model is efficient to deal with noisy data.

Dempster-Shafer Information Filtering in Multi-Modality Wireless Sensor Networks

A framework to estimate the state of dynamically varying environment where data are generated from heterogeneous sources possessing partial knowledge about the environment is presented. This is entirely derived within Dempster-Shafer and Evidence Filtering frameworks. The belief about the current state is expressed as belief and plausibility functions. An addition to Single Input Single Output Evidence Filter, Multiple Input Single Output Evidence Filtering approach is introduced. Variety of applications such as situational estimation of an emergency environment can be developed within the framework successfully. Fire propagation scenario is used to justify the proposed framework, simulation results are presented.

Likelihood Estimation for Stochastic Epidemics with Heterogeneous Mixing Populations

We consider a heterogeneously mixing SIR stochastic epidemic process in populations described by a general graph. Likelihood theory is developed to facilitate statistic inference for the parameters of the model under complete observation. We show that these estimators are asymptotically Gaussian unbiased estimates by using a martingale central limit theorem.

Performance Assessment of Computational Gridon Weather Indices from HOAPS Data

Long term rainfall analysis and prediction is a challenging task especially in the modern world where the impact of global warming is creating complications in environmental issues. These factors which are data intensive require high performance computational modeling for accurate prediction. This research paper describes a prototype which is designed and developed on grid environment using a number of coupled software infrastructural building blocks. This grid enabled system provides the demanding computational power, efficiency, resources, user-friendly interface, secured job submission and high throughput. The results obtained using sequential execution and grid enabled execution shows that computational performance has enhanced among 36% to 75%, for decade of climate parameters. Large variation in performance can be attributed to varying degree of computational resources available for job execution. Grid Computing enables the dynamic runtime selection, sharing and aggregation of distributed and autonomous resources which plays an important role not only in business, but also in scientific implications and social surroundings. This research paper attempts to explore the grid enabled computing capabilities on weather indices from HOAPS data for climate impact modeling and change detection.

Rigorous Modeling of Fixed-Bed Reactors Containing Finite Hollow Cylindrical Catalyst with Michaelis-Menten Type of Kinetics

A large number of chemical, bio-chemical and pollution-control processes use heterogeneous fixed-bed reactors. The use of finite hollow cylindrical catalyst pellets can enhance conversion levels in such reactors. The absence of the pellet core can significantly lower the diffusional resistance associated with the solid phase. This leads to a better utilization of the catalytic material, which is reflected in the higher values for the effectiveness factor, leading ultimately to an enhanced conversion level in the reactor. It is however important to develop a rigorous heterogeneous model for the reactor incorporating the two-dimensional feature of the solid phase owing to the presence of the finite hollow cylindrical catalyst pellet. Presently, heterogeneous models reported in the literature invariably employ one-dimension solid phase models meant for spherical catalyst pellets. The objective of the paper is to present a rigorous model of the fixed-bed reactors containing finite hollow cylindrical catalyst pellets. The reaction kinetics considered here is the widely used Michaelis–Menten kinetics for the liquid-phase bio-chemical reactions. The reaction parameters used here are for the enzymatic degradation of urea. Results indicate that increasing the height to diameter ratio helps to improve the conversion level. On the other hand, decreasing the thickness is apparently not as effective. This could however be explained in terms of the higher void fraction of the bed that causes a smaller amount of the solid phase to be packed in the fixed-bed bio-chemical reactor.

Performance Evaluation of TCP Vegas versus Different TCP Variants in Homogeneous and Heterogeneous Wired Networks

A study on the performance of TCP Vegas versus different TCP variants in homogeneous and heterogeneous wired networks are performed via simulation experiment using network simulator (ns-2). This performance evaluation prepared a comparison medium for the performance evaluation of enhanced-TCP Vegas in wired network and for wireless network. In homogeneous network, the performance of TCP Tahoe, TCP Reno, TCP NewReno, TCP Vegas and TCP SACK are analyzed. In heterogeneous network, the performances of TCP Vegas against TCP variants are analyzed. TCP Vegas outperforms other TCP variants in homogeneous wired network. However, TCP Vegas achieves unfair throughput in heterogeneous wired network.

Conversion in Chemical Reactors using Hollow Cylindrical Catalyst Pellet

Heterogeneous catalysis is vital for a number of chemical, refinery and pollution control processes. The use of catalyst pellets of hollow cylindrical shape provide several distinct advantages over other common shapes, and can therefore help to enhance conversion levels in reactors. A better utilization of the catalytic material is probably most notable of these features due to the absence of the pellet core, which helps to significantly lower the effect of the internal transport resistance. This is reflected in the enhancement of the effectiveness factor. For the case of the first order irreversible kinetics, a substantial increase in the effectiveness factor can be obtained by varying shape parameters. Important shape parameters of a finite hollow cylinder are the ratio of the inside to the outside radii (κ) and the height to the diameter ratio (γ). A high value of κ the generally helps to enhance the effectiveness factor. On the other hand, lower values of the effectiveness factors are obtained when the dimension of the height and the diameter are comparable. Thus, the departure of parameter γ from the unity favors higher effectiveness factor. Since a higher effectiveness factor is a measure of a greater utilization of the catalytic material, higher conversion levels can be achieved using the hollow cylindrical pellets possessing optimized shape parameters.

Surface and Guided Waves in Composites with Nematic Coatings

The theoretical prediction of the acoustical polarization effects in the heterogeneous composites, made of thick elastic solids with thin nematic films, is presented. The numericalanalytical solution to the problem of the different wave propagation exhibits some new physical effects in the low frequency domain: the appearance of the critical frequency and the existence of the narrow transition zone where the wave rapidly changes its speed. The associated wave attenuation is highly perturbed in this zone. We also show the possible appearance of the critical frequencies where the attenuation changes the sign. The numerical results of parametrical analysis are presented and discussed.

Study of the Vertical Handoff in Heterogeneous Networks and Implement Based On Opnet

In this document we studied more in detail the Performances of the vertical handover in the networks WLAN, WiMAX, UMTS before studying of it the Procedure of Handoff Vertical, the whole buckled by simulations putting forward the performances of the handover in the heterogeneous networks. The goal of Vertical Handover is to carry out several accesses in real-time in the heterogeneous networks. This makes it possible a user to use several networks (such as WLAN UMTS andWiMAX) in parallel, and the system to commutate automatically at another basic station, without disconnecting itself, as if there were no cut and with little loss of data as possible.

Numerical Investigation into Mixing Performance of Electrokinetically-Driven Power-Law Fluids in Microchannel with Patterned Trapezoid Blocks

The study investigates the mixing performance of electrokinetically-driven power-law fluids in a microchannel containing patterned trapezoid blocks. The effects of the geometry parameters of the patterned trapezoid blocks and the flow behavior index in the power-law model on the mixing efficiency within the microchannel are explored. The results show that the mixing efficiency can be improved by increasing the width of the blocks and extending the length of upper surface of the blocks. In addition, the results show that the mixing efficiency increases with an increasing flow behavior index. Furthermore, it is shown that a heterogeneous patterning of the zeta potential on the upper surfaces of the trapezoid blocks prompts the formation of local flow recirculations, and therefore improves the mixing efficiency. Consequently, it is shown that the mixing performance improves with an increasing magnitude of the heterogeneous surface zeta potential.

High Performance in Parallel Data Integration: An Empirical Evaluation of the Ratio Between Processing Time and Number of Physical Nodes

Many studies have shown that parallelization decreases efficiency [1], [2]. There are many reasons for these decrements. This paper investigates those which appear in the context of parallel data integration. Integration processes generally cannot be allocated to packages of identical size (i. e. tasks of identical complexity). The reason for this is unknown heterogeneous input data which result in variable task lengths. Process delay is defined by the slowest processing node. It leads to a detrimental effect on the total processing time. With a real world example, this study will show that while process delay does initially increase with the introduction of more nodes it ultimately decreases again after a certain point. The example will make use of the cloud computing platform Hadoop and be run inside Amazon-s EC2 compute cloud. A stochastic model will be set up which can explain this effect.

Biological Data Integration using SOA

Nowadays scientific data is inevitably digital and stored in a wide variety of formats in heterogeneous systems. Scientists need to access an integrated view of remote or local heterogeneous data sources with advanced data accessing, analyzing, and visualization tools. This research suggests the use of Service Oriented Architecture (SOA) to integrate biological data from different data sources. This work shows SOA will solve the problems that facing integration process and if the biologist scientists can access the biological data in easier way. There are several methods to implement SOA but web service is the most popular method. The Microsoft .Net Framework used to implement proposed architecture.

Biodegradable Surfactants for Advanced Drug Delivery Strategies

Oxidative stress makes up common incidents in eukaryotic metabolism. The presence of diverse components disturbing the equilibrium during oxygen metabolism increases oxidative damage unspecifically in living cells. Body´s own ubiquinone (Q10) seems to be a promising drug in defending the heightened appearance of reactive oxygen species (ROS). Though, its lipophilic properties require a new strategy in drug formulation to overcome their low bioavailability. Consequently, the manufacture of heterogeneous nanodispersions is in focus for medical applications. The composition of conventional nanodispersions is made up of a drug-consisting core and a surfactive agent, also named as surfactant. Long-termed encapsulation of the surfactive components into tissues might be the consequence of the use during medical therapeutics. The potential of provoking side-effects is given by their nonbiodegradable properties. Further improvements during fabrication process use the incorporation of biodegradable components such as modified γ-polyglutamic acid which decreases the potential of prospective side-effects.

A Temporal Synchronization Model for Heterogeneous Data in Distributed Systems

Multimedia distributed systems deal with heterogeneous data, such as texts, images, graphics, video and audio. The specification of temporal relations among different data types and distributed sources is an open research area. This paper proposes a fully distributed synchronization model to be used in multimedia systems. One original aspect of the model is that it avoids the use of a common reference (e.g. wall clock and shared memory). To achieve this, all possible multimedia temporal relations are specified according to their causal dependencies.

Automatic Real-Patient Medical Data De-Identification for Research Purposes

Our Medicine-oriented research is based on a medical data set of real patients. It is a security problem to share patient private data with peoples other than clinician or hospital staff. We have to remove person identification information from medical data. The medical data without private data are available after a de-identification process for any research purposes. In this paper, we introduce an universal automatic rule-based de-identification application to do all this stuff on an heterogeneous medical data. A patient private identification is replaced by an unique identification number, even in burnedin annotation in pixel data. The identical identification is used for all patient medical data, so it keeps relationships in a data. Hospital can take an advantage of a research feedback based on results.

A Web Text Mining Flexible Architecture

Text Mining is an important step of Knowledge Discovery process. It is used to extract hidden information from notstructured o semi-structured data. This aspect is fundamental because much of the Web information is semi-structured due to the nested structure of HTML code, much of the Web information is linked, much of the Web information is redundant. Web Text Mining helps whole knowledge mining process to mining, extraction and integration of useful data, information and knowledge from Web page contents. In this paper, we present a Web Text Mining process able to discover knowledge in a distributed and heterogeneous multiorganization environment. The Web Text Mining process is based on flexible architecture and is implemented by four steps able to examine web content and to extract useful hidden information through mining techniques. Our Web Text Mining prototype starts from the recovery of Web job offers in which, through a Text Mining process, useful information for fast classification of the same are drawn out, these information are, essentially, job offer place and skills.