Abstract: Image compression plays a vital role in today-s
communication. The limitation in allocated bandwidth leads to
slower communication. To exchange the rate of transmission in the
limited bandwidth the Image data must be compressed before
transmission. Basically there are two types of compressions, 1)
LOSSY compression and 2) LOSSLESS compression. Lossy
compression though gives more compression compared to lossless
compression; the accuracy in retrievation is less in case of lossy
compression as compared to lossless compression. JPEG, JPEG2000
image compression system follows huffman coding for image
compression. JPEG 2000 coding system use wavelet transform,
which decompose the image into different levels, where the
coefficient in each sub band are uncorrelated from coefficient of
other sub bands. Embedded Zero tree wavelet (EZW) coding exploits
the multi-resolution properties of the wavelet transform to give a
computationally simple algorithm with better performance compared
to existing wavelet transforms. For further improvement of
compression applications other coding methods were recently been
suggested. An ANN base approach is one such method. Artificial
Neural Network has been applied to many problems in image
processing and has demonstrated their superiority over classical
methods when dealing with noisy or incomplete data for image
compression applications. The performance analysis of different
images is proposed with an analysis of EZW coding system with
Error Backpropagation algorithm. The implementation and analysis
shows approximately 30% more accuracy in retrieved image
compare to the existing EZW coding system.
Abstract: The performance of high-resolution schemes is investigated for unsteady, inviscid and compressible multiphase flows. An Eulerian diffuse interface approach has been chosen for the simulation of multicomponent flow problems. The reduced fiveequation and seven equation models are used with HLL and HLLC approximation. The authors demonstrated the advantages and disadvantages of both seven equations and five equations models studying their performance with HLL and HLLC algorithms on simple test case. The seven equation model is based on two pressure, two velocity concept of Baer–Nunziato [10], while five equation model is based on the mixture velocity and pressure. The numerical evaluations of two variants of Riemann solvers have been conducted for the classical one-dimensional air-water shock tube and compared with analytical solution for error analysis.
Abstract: Lighvan cheese is basically made from sheep milk in
the area of Sahand mountainside which is located in the North West
of Iran. The main objective of this study was to investigate the effect
of enterococci isolated from traditional Lighvan cheese on the quality
of Iranian UF white during ripening. The experimental design was
split plot based on randomized complete blocks, main plots were four
types of starters and subplots were different ripening durations.
Addition of Enterococcus spp. did not significantly (P
Abstract: In high bitrate information hiding techniques, 1 bit is
embedded within each 4 x 4 Discrete Cosine Transform (DCT)
coefficient block by means of vector quantization, then the hidden bit
can be effectively extracted in terminal end. In this paper high bitrate
information hiding algorithms are summarized, and the scheme of
video in video is implemented. Experimental result shows that the host
video which is embedded numerous auxiliary information have little
visually quality decline. Peak Signal to Noise Ratio (PSNR)Y of host
video only degrades 0.22dB in average, while the hidden information
has a high percentage of survives and keeps a high robustness in
H.264/AVC compression, the average Bit Error Rate(BER) of hiding
information is 0.015%.
Abstract: Software development has experienced remarkable progress in the past decade. However, due to the rising complexity and magnitude of the project the development productivity has not been consistently improved. By analyzing the latest ISBSG data repository with 4106 projects, we discovered that software development productivity has actually undergone irregular variations between the years 1995 and 2005. Considering the factors significant to the productivity, we found its variations are primarily caused by the variations of average team size and the unbalanced uses of the less productive language 3GL.
Abstract: Preparation of size controlled nano-particles of silver catalyst on carbon substrate from e-waste has been investigated. Chemical route was developed by extraction of the metals available in nitric acid followed by treatment with hydrofluoric acid. Silver metal particles deposited with an average size 4-10 nm. A stabilizer concentration of 10- 40 g/l was used. The average size of the prepared silver decreased with increase of the anode current density. Size uniformity of the silver nano-particles was improved distinctly at higher current density no more than 20mA... Grain size increased with EK time whereby aggregation of particles was observed after 6 h of reaction.. The chemical method involves adsorption of silver nitrate on the carbon substrate. Adsorbed silver ions were directly reduced to metal particles using hydrazine hydrate. Another alternative method is by treatment with ammonia followed by heating the carbon loaded-silver hydroxide at 980°C. The product was characterized with the help of XRD, XRF, ICP, SEM and TEM techniques.
Abstract: In IETF RFC 2002, Mobile-IP was developed to
enable Laptobs to maintain Internet connectivity while moving
between subnets. However, the packet loss that comes from
switching subnets arises because network connectivity is lost while
the mobile host registers with the foreign agent and this encounters
large end-to-end packet delays. The criterion to initiate a simple and
fast full-duplex connection between the home agent and foreign
agent, to reduce the roaming duration, is a very important issue to be
considered by a work in this paper. State-transition Petri-Nets of the
modeling scenario-based CIA: communication inter-agents procedure
as an extension to the basic Mobile-IP registration process was
designed and manipulated to describe the system in discrete events.
The heuristic of configuration file during practical Setup session for
registration parameters, on Cisco platform Router-1760 using IOS
12.3 (15)T and TFTP server S/W is created. Finally, stand-alone
performance simulations from Simulink Matlab, within each subnet
and also between subnets, are illustrated for reporting better end-toend
packet delays. Results verified the effectiveness of our Mathcad
analytical manipulation and experimental implementation. It showed
lower values of end-to-end packet delay for Mobile-IP using CIA
procedure-based early registration. Furthermore, it reported packets
flow between subnets to improve losses between subnets.
Abstract: For lack of the visualization of the ultrasonic detection
method of partial discharge (PD), the ultrasonic detection technology
combined with the X-ray visual detection method (UXV) is proposed.
The method can conduct qualitative analysis accurately and conduct
reliable positioning diagnosis to the internal insulation defects of
GIS, and while it could make up the blindness of the X-ray visual
detection method and improve the detection rate. In this paper, an
experimental model of GIS is used as the trial platform, a variety of
insulation defects are set inside the GIS cavity. With the proposed
method, the ultrasonic method is used to conduct the preliminary
detection, and then the X-ray visual detection is used to locate and
diagnose precisely. Therefore, the proposed UXV technology is
feasible and practical.
Abstract: The presented work is motivated by a French law
regarding nuclear waste management. A new conceptual Accelerator
Driven System (ADS) designed for the Minor Actinides (MA)
transmutation has been assessed by numerical simulation. The
MUltiple Spallation Target (MUST) ADS combines high thermal power (up to 1.4 GWth) and high specific power. A 30 mA and 1
GeV proton beam is divided into three secondary beams transmitted on three liquid lead-bismuth spallation targets. Neutron and thermalhydraulic
simulations have been performed with the code MURE, based on the Monte-Carlo transport code MCNPX. A methodology has been developed to define characteristic of the MUST ADS concept according to a specific transmutation scenario. The reference
scenario is based on a MA flux (neptunium, americium and curium)
providing from European Fast Reactor (EPR) and a plutonium multireprocessing
strategy is accounted for. The MUST ADS reference
concept is a sodium cooled fast reactor. The MA fuel at equilibrium is mixed with MgO inert matrix to limit the core reactivity and
improve the fuel thermal conductivity. The fuel is irradiated over five
years. Five years of cooling and two years for the fuel fabrication are
taken into account. The MUST ADS reference concept burns about 50% of the initial MA inventory during a complete cycle. In term of
mass, up to 570 kg/year are transmuted in one concept. The methodology to design the MUST ADS and to calculate fuel
composition at equilibrium is precisely described in the paper. A detailed fuel evolution analysis is performed and the reference scenario is compared to a scenario where only americium transmutation is performed.
Abstract: Traditionally, Internet has provided best-effort service to every user regardless of its requirements. However, as Internet becomes universally available, users demand more bandwidth and applications require more and more resources, and interest has developed in having the Internet provide some degree of Quality of Service. Although QoS is an important issue, the question of how it will be brought into the Internet has not been solved yet. Researches, due to the rapid advances in technology are proposing new and more desirable capabilities for the next generation of IP infrastructures. But neither all applications demand the same amount of resources, nor all users are service providers. In this way, this paper is the first of a series of papers that presents an architecture as a first step to the optimization of QoS in the Internet environment as a solution to a SMSE's problem whose objective is to provide public service to internet with certain Quality of Service expectations. The service provides new business opportunities, but also presents new challenges. We have designed and implemented a scalable service framework that supports adaptive bandwidth based on user demands, and the billing based on usage and on QoS. The developed application has been evaluated and the results show that traffic limiting works at optimum and so it does exceeding bandwidth distribution. However, some considerations are done and currently research is under way in two basic areas: (i) development and testing new transfer protocols, and (ii) developing new strategies for traffic improvements based on service differentiation.
Abstract: This paper presents a new problem solving approach
that is able to generate optimal policy solution for finite-state
stochastic sequential decision-making problems with high data
efficiency. The proposed algorithm iteratively builds and improves
an approximate Markov Decision Process (MDP) model along with
cost-to-go value approximates by generating finite length trajectories
through the state-space. The approach creates a synergy between an
approximate evolving model and approximate cost-to-go values to
produce a sequence of improving policies finally converging to the
optimal policy through an intelligent and structured search of the
policy space. The approach modifies the policy update step of the
policy iteration so as to result in a speedy and stable convergence to
the optimal policy. We apply the algorithm to a non-holonomic
mobile robot control problem and compare its performance with
other Reinforcement Learning (RL) approaches, e.g., a) Q-learning,
b) Watkins Q(λ), c) SARSA(λ).
Abstract: The quality of short term load forecasting can improve the efficiency of planning and operation of electric utilities. Artificial Neural Networks (ANNs) are employed for nonlinear short term load forecasting owing to their powerful nonlinear mapping capabilities. At present, there is no systematic methodology for optimal design and training of an artificial neural network. One has often to resort to the trial and error approach. This paper describes the process of developing three layer feed-forward large neural networks for short-term load forecasting and then presents a heuristic search algorithm for performing an important task of this process, i.e. optimal networks structure design. Particle Swarm Optimization (PSO) is used to develop the optimum large neural network structure and connecting weights for one-day ahead electric load forecasting problem. PSO is a novel random optimization method based on swarm intelligence, which has more powerful ability of global optimization. Employing PSO algorithms on the design and training of ANNs allows the ANN architecture and parameters to be easily optimized. The proposed method is applied to STLF of the local utility. Data are clustered due to the differences in their characteristics. Special days are extracted from the normal training sets and handled separately. In this way, a solution is provided for all load types, including working days and weekends and special days. The experimental results show that the proposed method optimized by PSO can quicken the learning speed of the network and improve the forecasting precision compared with the conventional Back Propagation (BP) method. Moreover, it is not only simple to calculate, but also practical and effective. Also, it provides a greater degree of accuracy in many cases and gives lower percent errors all the time for STLF problem compared to BP method. Thus, it can be applied to automatically design an optimal load forecaster based on historical data.
Abstract: In this paper, the periodic surveillance scheme has
been proposed for any convex region using mobile wireless sensor
nodes. A sensor network typically consists of fixed number of
sensor nodes which report the measurements of sensed data such as
temperature, pressure, humidity, etc., of its immediate proximity
(the area within its sensing range). For the purpose of sensing an
area of interest, there are adequate number of fixed sensor
nodes required to cover the entire region of interest. It implies
that the number of fixed sensor nodes required to cover a given
area will depend on the sensing range of the sensor as well as
deployment strategies employed. It is assumed that the sensors to
be mobile within the region of surveillance, can be mounted on
moving bodies like robots or vehicle. Therefore, in our
scheme, the surveillance time period determines the number of
sensor nodes required to be deployed in the region of interest.
The proposed scheme comprises of three algorithms namely:
Hexagonalization, Clustering, and Scheduling, The first algorithm
partitions the coverage area into fixed sized hexagons that
approximate the sensing range (cell) of individual sensor node.
The clustering algorithm groups the cells into clusters, each of
which will be covered by a single sensor node. The later
determines a schedule for each sensor to serve its respective cluster.
Each sensor node traverses all the cells belonging to the cluster
assigned to it by oscillating between the first and the last cell for
the duration of its life time. Simulation results show that our
scheme provides full coverage within a given period of time using
few sensors with minimum movement, less power consumption,
and relatively less infrastructure cost.
Abstract: In this paper, we propose an improvement of pattern
growth-based PrefixSpan algorithm, called I-PrefixSpan. The general idea of I-PrefixSpan is to use sufficient data structure for Seq-Tree
framework and separator database to reduce the execution time and
memory usage. Thus, with I-PrefixSpan there is no in-memory database stored after index set is constructed. The experimental result
shows that using Java 2, this method improves the speed of PrefixSpan up to almost two orders of magnitude as well as the memory usage to more than one order of magnitude.
Abstract: Conventional WBL is effective for meaningful student, because rote student learn by repeating without thinking or trying to understand. It is impossible to have full benefit from conventional WBL. Understanding of rote student-s intention and what influences it becomes important. Poorly designed user interface will discourage rote student-s cultivation and intention to use WBL. Thus, user interface design is an important factor especially when WBL is used as comprehensive replacement of conventional teaching. This research proposes the influencing factors that can enhance student-s intention to use the system. The enhanced TAM is used for evaluating the proposed factors. The research result points out that factors influencing rote student-s intention are Perceived Usefulness of Homepage Content Structure, Perceived User Friendly Interface, Perceived Hedonic Component, and Perceived (homepage) Visual Attractiveness.
Abstract: This paper presents three-phase evolution search methodology to automatically design fuzzy logic controllers (FLCs) that can work in a wide range of operating conditions. These include varying load, parameter variations, and unknown external disturbances. The three-phase scheme consists of an exploration phase, an exploitation phase and a robustness phase. The first two phases search for FLC with high accuracy performances while the last phase aims at obtaining FLC providing the best compromise between the accuracy and robustness performances. Simulations were performed for direct-drive two-axis robot arm. The evolved FLC with the proposed design technique found to provide a very satisfactory performance under the wide range of operation conditions and to overcome problem associated with coupling and nonlinearities characteristics inherent to robot arms.
Abstract: The hybridisation of genetic algorithm with heuristics has been shown to be one of an effective way to improve its performance. In this work, genetic algorithm hybridised with four heuristics including a new heuristic called neighbourhood improvement were investigated through the classical travelling salesman problem. The experimental results showed that the proposed heuristic outperformed other heuristics both in terms of quality of the results obtained and the computational time.
Abstract: This project aims to investigate the potential of
torrefaction to improve the properties of Malaysian palm kernel shell
(PKS) as a solid fuel. A study towards torrefaction of PKS was
performed under various temperature and residence time of 240, 260,
and 280oC and 30, 60, and 90 minutes respectively. The torrefied
PKS was characterized in terms of the mass yield, energy yield,
elemental composition analysis, calorific value analysis, moisture and
volatile matter contents, and ash and fixed carbon contents. The mass
and energy yield changes in the torrefied PKS were observed to
prove that the temperature has more effect compare to residence time
in the torrefaction process. The C content of PKS increases while H
and O contents decrease after torrefaction, which resulted in higher
heating value between 5 to 16%. Meanwhile, torrefaction caused the
ash and fixed carbon content of PKS to increase, and the moisture
and volatile matter to decrease.
Abstract: Due to the rise of aging population, effective utilization
of healthcare resources has become an important issue. With the
advance of ICT technology, the application of tele-healthcare service
has received more attention than ever. The main purpose of this
research is to investigate how to conduct innovative design for
tele-healthcare service based on user-s perspectives. First, the
healthcare service blueprint was used to describe the processes
of tele-healthcare service delivery, and then construct PZB service
quality gap model based on the literature and practitioners-
interviews. Next, TRIZ theory is applied to implement service
innovation. We found the proposed service innovation procedures can
effectively improve the quality of service design.
Abstract: This research explores visitor-s expectations of service
quality in intelligent living space showroom – Living 3.0 in Taiwan.
Based on the five dimensions of PZB service quality, a specialist
questionnaire is utilized to establish a complete service quality
evaluation framework for Living 3.0. In this research, analysis
hierarchy process (AHP) is applied to find the relative weights among
the criteria. Finally, the service quality evaluation framework and
evaluation results can be used as a guide for Living 3.0 proprietors to
review, improve, and enhance service planning and service qualities in
the future.