Congestion Control for Internet Media Traffic

In this paper we investigated a number of the Internet congestion control algorithms that has been developed in the last few years. It was obviously found that many of these algorithms were designed to deal with the Internet traffic merely as a train of consequent packets. Other few algorithms were specifically tailored to handle the Internet congestion caused by running media traffic that represents audiovisual content. This later set of algorithms is considered to be aware of the nature of this media content. In this context we briefly explained a number of congestion control algorithms and hence categorized them into the two following categories: i) Media congestion control algorithms. ii) Common congestion control algorithms. We hereby recommend the usage of the media congestion control algorithms for the reason of being media content-aware rather than the other common type of algorithms that blindly manipulates such traffic. We showed that the spread of such media content-aware algorithms over Internet will lead to better congestion control status in the coming years. This is due to the observed emergence of the era of digital convergence where the media traffic type will form the majority of the Internet traffic.

CNet Module Design of IMCS

IMCS is Integrated Monitoring and Control System for thermal power plant. This system consists of mainly two parts; controllers and OIS (Operator Interface System). These two parts are connected by Ethernet-based communication. The controller side of communication is managed by CNet module and OIS side is managed by data server of OIS. CNet module sends the data of controller to data server and receives commend data from data server. To minimizes or balance the load of data server, this module buffers data created by controller at every cycle and send buffered data to data server on request of data server. For multiple data server, this module manages the connection line with each data server and response for each request from multiple data server. CNet module is included in each controller of redundant system. When controller fail-over happens on redundant system, this module can provide data of controller to data sever without loss. This paper presents three main features – separation of get task, usage of ring buffer and monitoring communication status –of CNet module to carry out these functions.

Density, Strength, Thermal Conductivity and Leachate Characteristics of Light-Weight Fired Clay Bricks Incorporating Cigarette Butts

Several trillion cigarettes produced worldwide annually lead to many thousands of kilograms of toxic waste. Cigarette butts (CBs) accumulate in the environment due to the poor biodegradability of the cellulose acetate filters. This paper presents some of the results from a continuing study on recycling CBs into fired clay bricks. Physico-mechanical properties of fired clay bricks manufactured with different percentages of CBs are reported and discussed. The results show that the density of fired bricks was reduced by up to 30 %, depending on the percentage of CBs incorporated into the raw materials. Similarly, the compressive strength of bricks tested decreased according to the percentage of CBs included in the mix. The thermal conductivity performance of bricks was improved by 51 and 58 % for 5 and 10 % CBs content respectively. Leaching tests were carried out to investigate the levels of possible leachates of heavy metals from the manufactured clay-CB bricks. The results revealed trace amounts of heavy metals.

Analytical Model Based Evaluation of Human Machine Interfaces Using Cognitive Modeling

Cognitive models allow predicting some aspects of utility and usability of human machine interfaces (HMI), and simulating the interaction with these interfaces. The action of predicting is based on a task analysis, which investigates what a user is required to do in terms of actions and cognitive processes to achieve a task. Task analysis facilitates the understanding of the system-s functionalities. Cognitive models are part of the analytical approaches, that do not associate the users during the development process of the interface. This article presents a study about the evaluation of a human machine interaction with a contextual assistant-s interface using ACTR and GOMS cognitive models. The present work shows how these techniques may be applied in the evaluation of HMI, design and research by emphasizing firstly the task analysis and secondly the time execution of the task. In order to validate and support our results, an experimental study of user performance is conducted at the DOMUS laboratory, during the interaction with the contextual assistant-s interface. The results of our models show that the GOMS and ACT-R models give good and excellent predictions respectively of users performance at the task level, as well as the object level. Therefore, the simulated results are very close to the results obtained in the experimental study.

A Forward Automatic Censored Cell-Averaging Detector for Multiple Target Situations in Log-Normal Clutter

A challenging problem in radar signal processing is to achieve reliable target detection in the presence of interferences. In this paper, we propose a novel algorithm for automatic censoring of radar interfering targets in log-normal clutter. The proposed algorithm, termed the forward automatic censored cell averaging detector (F-ACCAD), consists of two steps: removing the corrupted reference cells (censoring) and the actual detection. Both steps are performed dynamically by using a suitable set of ranked cells to estimate the unknown background level and set the adaptive thresholds accordingly. The F-ACCAD algorithm does not require any prior information about the clutter parameters nor does it require the number of interfering targets. The effectiveness of the F-ACCAD algorithm is assessed by computing, using Monte Carlo simulations, the probability of censoring and the probability of detection in different background environments.

The Heat and Mass Transfer Phenomena in Vacuum Membrane Distillation for Desalination

Vacuum membrane distillation (VMD) process can be used for water purification or the desalination of salt water. The process simply consists of a flat sheet hydrophobic micro porous PTFE membrane and diaphragm vacuum pump without a condenser for the water recovery or trap. The feed was used aqueous NaCl solution. The VMD experiments were performed to evaluate the heat and mass transfer coefficient of the boundary layer in a membrane module. The only operating parameters are feed inlet temperature, and feed flow rate were investigated. The permeate flux was strongly affected by the feed inlet temperature, feed flow rate, and boundary layer heat transfer coefficient. Since lowering the temperature polarization coefficient is essential enhance the process performance considerable and maximizing the heat transfer coefficient for maximizes the mass flux of distillate water. In this paper, the results of VMD experiments are used to measure the boundary layer heat transfer coefficient, and the experimental results are used to reevaluate the empirical constants in the Dittus- Boelter equation.

Life Cycle Assessment of Seawater Desalinization in Western Australia

Perth will run out of available sustainable natural water resources by 2015 if nothing is done to slow usage rates, according to a Western Australian study [1]. Alternative water technology options need to be considered for the long-term guaranteed supply of water for agricultural, commercial, domestic and industrial purposes. Seawater is an alternative source of water for human consumption, because seawater can be desalinated and supplied in large quantities to a very high quality. While seawater desalination is a promising option, the technology requires a large amount of energy which is typically generated from fossil fuels. The combustion of fossil fuels emits greenhouse gases (GHG) and, is implicated in climate change. In addition to environmental emissions from electricity generation for desalination, greenhouse gases are emitted in the production of chemicals and membranes for water treatment. Since Australia is a signatory to the Kyoto Protocol, it is important to quantify greenhouse gas emissions from desalinated water production. A life cycle assessment (LCA) has been carried out to determine the greenhouse gas emissions from the production of 1 gigalitre (GL) of water from the new plant. In this LCA analysis, a new desalination plant that will be installed in Bunbury, Western Australia, and known as Southern Seawater Desalinization Plant (SSDP), was taken as a case study. The system boundary of the LCA mainly consists of three stages: seawater extraction, treatment and delivery. The analysis found that the equivalent of 3,890 tonnes of CO2 could be emitted from the production of 1 GL of desalinated water. This LCA analysis has also identified that the reverse osmosis process would cause the most significant greenhouse emissions as a result of the electricity used if this is generated from fossil fuels

Natural Discovery: Electricity Potential from Vermicompost (Waste to Energy)

Wastages such as grated coconut meat, spent tea and used sugarcane had contributed negative impacts to the environment. Vermicomposting method is fully utilized to manage the wastes towards a more sustainable approach. The worms that are used in the vermicomposting are Eisenia foetida and Eudrillus euginae. This research shows that the vermicompost of wastages has voltage of electrical energy and is able to light up the Light-Emitting Diode (LED) device. Based on the experiment, the use of replicated and double compartments of the component will produce double of voltage. Hence, for conclusion, this harmless and low cost technology of vermicompost can act as a dry cell in order to reduce the usage of hazardous chemicals that can contaminate the environment.

A Complexity Measure for Java Bean based Software Components

The traditional software product and process metrics are neither suitable nor sufficient in measuring the complexity of software components, which ultimately is necessary for quality and productivity improvement within organizations adopting CBSE. Researchers have proposed a wide range of complexity metrics for software systems. However, these metrics are not sufficient for components and component-based system and are restricted to the module-oriented systems and object-oriented systems. In this proposed study it is proposed to find the complexity of the JavaBean Software Components as a reflection of its quality and the component can be adopted accordingly to make it more reusable. The proposed metric involves only the design issues of the component and does not consider the packaging and the deployment complexity. In this way, the software components could be kept in certain limit which in turn help in enhancing the quality and productivity.

Application of Adaptive Neuro-Fuzzy Inference System in the Prediction of Economic Crisis Periods in USA

In this paper discrete choice models, Logit and Probit are examined in order to predict the economic recession or expansion periods in USA. Additionally we propose an adaptive neuro-fuzzy inference system with triangular membership function. We examine the in-sample period 1947-2005 and we test the models in the out-of sample period 2006-2009. The forecasting results indicate that the Adaptive Neuro-fuzzy Inference System (ANFIS) model outperforms significant the Logit and Probit models in the out-of sample period. This indicates that neuro-fuzzy model provides a better and more reliable signal on whether or not a financial crisis will take place.

Empirical Study on the Diffusion of Smartphones and Consumer Behaviour

In this research, the diffusion of innovation regarding smartphone usage is analysed through a consumer behaviour theory. This research aims to determine whether a pattern surrounding the diffusion of innovation exists. As a methodology, an empirical study of the switch from a conventional cell phone to a smartphone was performed. Specifically, a questionnaire survey was completed by general consumers, and the situational and behavioural characteristics of switching from a cell phone to a smartphone were analysed. In conclusion, we found that the speed of the diffusion of innovation, the consumer behaviour characteristics, and the utilities of the product vary according to the stage of the product life cycle.

Measurement of Lead Pollution in the Air of Babylon Governorate/Iraq during Year 2010

This research aims to study the lead pollution in the air of Babylon governorate that resulted generally from vehicles exhausts in addition to industrial and human activities.Vehicles number in Babylon governorate increased significantly after year 2003 that resulted with increase in lead emissions into the air.Measurement of lead emissions was done in seven stations distributed randomly in Babylon governorate. These stations where located in Industrial (Al-Sena'ay) Quarter, 60 street (near to Babylon sewer directorate), 40 Street (near to the first intersection), Al-Hashmia city, Al-Mahaweel city, , Al- Musayab city in addition to another station in Sayd Idris village belong to Abugharaq district (Agricultural station for comparison). The measured concentrations in these stations were compared with the standard limits of Environmental Protection Agency EPA (2 μg /m3). The results of this study showed that the average of lead concentrations ,in Babylon governorate during year 2010, was (3.13 μg/m3) which was greater than standard limits (2 μg/m3). The maximum concentration of lead was (6.41 μg / m3) recorded in the Industrial (Al-Sena'ay) Quarter during April month, while the minimum concentrations was (0.36 μg / m3) recorded in the agricultural station (Abugharaq) during December month.

The Data Mining usage in Production System Management

The paper gives the pilot results of the project that is oriented on the use of data mining techniques and knowledge discoveries from production systems through them. They have been used in the management of these systems. The simulation models of manufacturing systems have been developed to obtain the necessary data about production. The authors have developed the way of storing data obtained from the simulation models in the data warehouse. Data mining model has been created by using specific methods and selected techniques for defined problems of production system management. The new knowledge has been applied to production management system. Gained knowledge has been tested on simulation models of the production system. An important benefit of the project has been proposal of the new methodology. This methodology is focused on data mining from the databases that store operational data about the production process.

Mirror Neuron System Study on Elderly Using Dynamic Causal Modeling fMRI Analysis

Dynamic Causal Modeling (DCM) functional Magnetic Resonance Imaging (fMRI) is a promising technique to study the connectivity among brain regions and effects of stimuli through modeling neuronal interactions from time-series neuroimaging. The aim of this study is to study characteristics of a mirror neuron system (MNS) in elderly group (age: 60-70 years old). Twenty volunteers were MRI scanned with visual stimuli to study a functional brain network. DCM was employed to determine the mechanism of mirror neuron effects. The results revealed major activated areas including precentral gyrus, inferior parietal lobule, inferior occipital gyrus, and supplementary motor area. When visual stimuli were presented, the feed-forward connectivity from visual area to conjunction area was increased and forwarded to motor area. Moreover, the connectivity from the conjunction areas to premotor area was also increased. Such findings can be useful for future diagnostic process for elderly with diseases such as Parkinson-s and Alzheimer-s.

Information Transmission between Large and Small Stocks in the Korean Stock Market

Little attention has been paid to information transmission between the portfolios of large stocks and small stocks in the Korean stock market. This study investigates the return and volatility transmission mechanisms between large and small stocks in the Korea Exchange (KRX). This study also explores whether bad news in the large stock market leads to a volatility of the small stock market that is larger than the good news volatility of the large stock market. By employing the Granger causality test, we found unidirectional return transmissions from the large stocks to medium and small stocks. This evidence indicates that pat information about the large stocks has a better ability to predict the returns of the medium and small stocks in the Korean stock market. Moreover, by using the asymmetric GARCH-BEKK model, we observed the unidirectional relationship of asymmetric volatility transmission from large stocks to the medium and small stocks. This finding suggests that volatility in the medium and small stocks following a negative shock in the large stocks is larger than that following a positive shock in the large stocks.

Improving the Reusability and Interoperability of E-Learning Material

A key requirement for e-learning materials is reusability and interoperability, that is the possibility to use at least part of the contents in different courses, and to deliver them trough different platforms. These features make possible to limit the cost of new packages, but require the development of material according to proper specifications. SCORM (Sharable Content Object Reference Model) is a set of guidelines suitable for this purpose. A specific adaptation project has been started to make possible to reuse existing materials. The paper describes the main characteristics of SCORM specification, and the procedure used to modify the existing material.

Enhancements in Blended e-Learning Management System

A learning management system (commonly abbreviated as LMS) is a software application for the administration, documentation, tracking, and reporting of training programs, classroom and online events, e-learning programs, and training content (Ellis 2009). (Hall 2003) defines an LMS as \"software that automates the administration of training events. All Learning Management Systems manage the log-in of registered users, manage course catalogs, record data from learners, and provide reports to management\". Evidence of the worldwide spread of e-learning in recent years is easy to obtain. In April 2003, no fewer than 66,000 fully online courses and 1,200 complete online programs were listed on the TeleCampus portal from TeleEducation (Paulsen 2003). In the report \" The US market in the Self-paced eLearning Products and Services:2010-2015 Forecast and Analysis\" The number of student taken classes exclusively online will be nearly equal (1% less) to the number taken classes exclusively in physical campuses. Number of student taken online course will increase from 1.37 million in 2010 to 3.86 million in 2015 in USA. In another report by The Sloan Consortium three-quarters of institutions report that the economic downturn has increased demand for online courses and programs.

Solving Inhomogeneous Wave Equation Cauchy Problems using Homotopy Perturbation Method

In this paper, He-s homotopy perturbation method (HPM) is applied to spatial one and three spatial dimensional inhomogeneous wave equation Cauchy problems for obtaining exact solutions. HPM is used for analytic handling of these equations. The results reveal that the HPM is a very effective, convenient and quite accurate to such types of partial differential equations (PDEs).

Bandwidth Allocation for ABR Service in Cellular Networks

Available Bit Rate Service (ABR) is the lower priority service and the better service for the transmission of data. On wireline ATM networks ABR source is always getting the feedback from switches about increase or decrease of bandwidth according to the changing network conditions and minimum bandwidth is guaranteed. In wireless networks guaranteeing the minimum bandwidth is really a challenging task as the source is always in mobile and traveling from one cell to another cell. Re establishment of virtual circuits from start to end every time causes the delay in transmission. In our proposed solution we proposed the mechanism to provide more available bandwidth to the ABR source by re-usage of part of old Virtual Channels and establishing the new ones. We want the ABR source to transmit the data continuously (non-stop) inorderto avoid the delay. In worst case scenario at least minimum bandwidth is to be allocated. In order to keep the data flow continuously, priority is given to the handoff ABR call against new ABR call.

Lessons Learned from Observing User Behavior through Repeated Usability Evaluations

Academic research information service is a must for surveying previous studies in research and development process. OntoFrame is an academic research information service under Semantic Web framework different from simple keyword-based services such as CiteSeer and Google Scholar. The first purpose of this study is for revealing user behavior in their surveys, the objects of using academic research information services, and their needs. The second is for applying lessons learned from the results to OntoFrame.