Abstract: In this study, we propose a network architecture for
providing secure access to information resources of enterprise
network from remote locations in a wireless fashion. Our proposed
architecture offers a very promising solution for organizations which
are in need of a secure, flexible and cost-effective remote access
methodology. Security of the proposed architecture is based on
Virtual Private Network technology and a special role based access
control mechanism with location and time constraints. The flexibility
mainly comes from the use of Internet as the communication medium
and cost-effectiveness is due to the possibility of in-house
implementation of the proposed architecture.
Abstract: With the advent of emerging personal computing paradigms such as ubiquitous and mobile computing, Web contents are becoming accessible from a wide range of mobile devices. Since these devices do not have the same rendering capabilities, Web contents need to be adapted for transparent access from a variety of client agents. Such content adaptation is exploited for either an individual element or a set of consecutive elements in a Web document and results in better rendering and faster delivery to the client device. Nevertheless, Web content adaptation sets new challenges for semantic markup. This paper presents an advanced components platform, called SMC, enabling the development of mobility applications and services according to a channel model based on the principles of Services Oriented Architecture (SOA). It then goes on to describe the potential for integration with the Semantic Web through a novel framework of external semantic annotation that prescribes a scheme for representing semantic markup files and a way of associating Web documents with these external annotations. The role of semantic annotation in this framework is to describe the contents of individual documents themselves, assuring the preservation of the semantics during the process of adapting content rendering. Semantic Web content adaptation is a way of adding value to Web contents and facilitates repurposing of Web contents (enhanced browsing, Web Services location and access, etc).
Abstract: In order to make surfing the internet faster, and to save redundant processing load with each request for the same web page, many caching techniques have been developed to reduce latency of retrieving data on World Wide Web. In this paper we will give a quick overview of existing web caching techniques used for dynamic web pages then we will introduce a design and implementation model that take advantage of “URL Rewriting" feature in some popular web servers, e.g. Apache, to provide an effective approach of caching dynamic web pages.
Abstract: Ramadan requires individuals to abstain from food and fluid intake between sunrise and sunset; physiological considerations predict that poorer mood, physical performance and mental performance will result. In addition, any difficulties will be worsened because preparations for fasting and recovery from it often mean that nocturnal sleep is decreased in length, and this independently affects mood and performance.
A difficulty of interpretation in many studies is that the observed changes could be due to fasting but also to the decreased length of sleep and altered food and fluid intakes before and after the daytime fasting. These factors were separated in this study, which took place over three separate days and compared the effects of different durations of fasting (4, 8 or 16h) upon a wide variety of measures (including subjective and objective assessments of performance, body composition, dehydration and responses to a short bout of exercise) - but with an unchanged amount of nocturnal sleep, controlled supper the previous evening, controlled intakes at breakfast and daytime naps not being allowed. Many of the negative effects of fasting observed in previous studies were present in this experiment also. These findings indicate that fasting was responsible for many of the changes previously observed, though some effect of sleep loss, particularly if occurring on successive days (as would occur in Ramadan) cannot be excluded.
Abstract: Documents clustering become an essential technology
with the popularity of the Internet. That also means that fast and
high-quality document clustering technique play core topics. Text
clustering or shortly clustering is about discovering semantically
related groups in an unstructured collection of documents. Clustering
has been very popular for a long time because it provides unique
ways of digesting and generalizing large amounts of information.
One of the issues of clustering is to extract proper feature (concept)
of a problem domain. The existing clustering technology mainly
focuses on term weight calculation. To achieve more accurate
document clustering, more informative features including concept
weight are important. Feature Selection is important for clustering
process because some of the irrelevant or redundant feature may
misguide the clustering results. To counteract this issue, the proposed
system presents the concept weight for text clustering system
developed based on a k-means algorithm in accordance with the
principles of ontology so that the important of words of a cluster can
be identified by the weight values. To a certain extent, it has resolved
the semantic problem in specific areas.
Abstract: The size, complexity and number of databases used
for protein information have caused bioinformatics to lag behind in
adapting to the need to handle this distributed information.
Integrating all the information from different databases into one
database is a challenging problem. Our main research is to develop a
tool which can be used to access and manipulate protein information
from difference databases. In our approach, we have integrated
difference databases such as Swiss-prot, PDB, Interpro, and EMBL
and transformed these databases in flat file format into relational
form using XML and Bioperl. As a result, we showed this tool can
search different sizes of protein information stored in relational
database and the result can be retrieved faster compared to flat file
database. A web based user interface is provided to allow user to
access or search for protein information in the local database.
Abstract: This paper features the modeling and design of a Fast
Output Sampling (FOS) Feedback control technique for the Active
Vibration Control (AVC) of a smart flexible aluminium cantilever
beam for a Single Input Single Output (SISO) case. Controllers are
designed for the beam by bonding patches of piezoelectric layer as
sensor / actuator to the master structure at different locations along
the length of the beam by retaining the first 2 dominant vibratory
modes. The entire structure is modeled in state space form using the
concept of piezoelectric theory, Euler-Bernoulli beam theory, Finite
Element Method (FEM) and the state space techniques by dividing
the structure into 3, 4, 5 finite elements, thus giving rise to three
types of systems, viz., system 1 (beam divided into 3 finite
elements), system 2 (4 finite elements), system 3 (5 finite elements).
The effect of placing the sensor / actuator at various locations along
the length of the beam for all the 3 types of systems considered is
observed and the conclusions are drawn for the best performance and
for the smallest magnitude of the control input required to control the
vibrations of the beam. Simulations are performed in MATLAB. The
open loop responses, closed loop responses and the tip displacements
with and without the controller are obtained and the performance of
the proposed smart system is evaluated for vibration control.
Abstract: The spiral angle of the elementary cellulose fibril in
the wood cell wall, often called microfibril angle, (MFA). Microfibril
angle in hardwood is one of the key determinants of solid timber
performance due to its strong influence on the stiffness, strength,
shrinkage, swelling, thermal-dynamics mechanical properties and
dimensional stability of wood. Variation of MFA (degree) in the S2
layer of the cell walls among Acacia mangium trees was determined
using small-angle X-ray scattering (SAXS). The length and
orientation of the microfibrils of the cell walls in the irradiated
volume of the thin samples are measured using SAXS and optical
microscope for 3D surface measurement. The undetermined
parameters in the analysis are the MFA, (M) and the standard
deviation (σФ) of the intensity distribution arising from the wandering
of the fibril orientation about the mean value. Nine separate pairs of
values are determined for nine different values of the angle of the
incidence of the X-ray beam relative to the normal to the radial
direction in the sample. The results show good agreement. The
curve distribution of scattered intensity for the real cell wall structure
is compared with that calculated with that assembly of rectangular
cells with the same ratio of transverse to radial cell wall length. It is
demonstrated that for β = 45°, the peaks in the curve intensity
distribution for the real and the rectangular cells coincide. If this
peak position is Ф45, then the MFA can be determined from the
relation M = tan-1 (tan Ф45 / cos 45°), which is precise for rectangular
cells. It was found that 92.93% of the variation of MFA can be
attributed to the distance from pith to bark. Here we shall present our
results of the MFA in the cell wall with respect to its shape, structure
and the distance from pith to park as an important fast check and yet
accurate towards the quality of wood, its uses and application.
Abstract: The topic of surface flattening plays a vital role in the field of computer aided design and manufacture. Surface flattening enables the production of 2D patterns and it can be used in design and manufacturing for developing a 3D surface to a 2D platform, especially in fashion design. This study describes surface flattening based on minimum energy methods according to the property of different fabrics. Firstly, through the geometric feature of a 3D surface, the less transformed area can be flattened on a 2D platform by geodesic. Then, strain energy that has accumulated in mesh can be stably released by an approximate implicit method and revised error function. In some cases, cutting mesh to further release the energy is a common way to fix the situation and enhance the accuracy of the surface flattening, and this makes the obtained 2D pattern naturally generate significant cracks. When this methodology is applied to a 3D mannequin constructed with feature lines, it enhances the level of computer-aided fashion design. Besides, when different fabrics are applied to fashion design, it is necessary to revise the shape of a 2D pattern according to the properties of the fabric. With this model, the outline of 2D patterns can be revised by distributing the strain energy with different results according to different fabric properties. Finally, this research uses some common design cases to illustrate and verify the feasibility of this methodology.
Abstract: The present models and simulation algorithms of intracellular stochastic kinetics are usually based on the premise that diffusion is so fast that the concentrations of all the involved species are homogeneous in space. However, recents experimental measurements of intracellular diffusion constants indicate that the assumption of a homogeneous well-stirred cytosol is not necessarily valid even for small prokaryotic cells. In this work a mathematical treatment of diffusion that can be incorporated in a stochastic algorithm simulating the dynamics of a reaction-diffusion system is presented. The movement of a molecule A from a region i to a region j of the space is represented as a first order reaction Ai k- ! Aj , where the rate constant k depends on the diffusion coefficient. The diffusion coefficients are modeled as function of the local concentration of the solutes, their intrinsic viscosities, their frictional coefficients and the temperature of the system. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the intrinsic reaction kinetics and diffusion dynamics. To demonstrate the method the simulation results of the reaction-diffusion system of chaperoneassisted protein folding in cytoplasm are shown.
Abstract: This paper proposes a fast code acquisition scheme for
optical code division multiple access (O-CDMA) systems. Unlike the
conventional scheme, the proposed scheme employs multiple thresholds
providing a shorter mean acquisition time (MAT) performance.
The simulation results show that the MAT of the proposed scheme
is shorter than that of the conventional scheme.
Abstract: The conjugate gradient optimization algorithm
usually used for nonlinear least squares is presented and is
combined with the modified back propagation algorithm yielding
a new fast training multilayer perceptron (MLP) algorithm
(CGFR/AG). The approaches presented in the paper consist of
three steps: (1) Modification on standard back propagation
algorithm by introducing gain variation term of the activation
function, (2) Calculating the gradient descent on error with
respect to the weights and gains values and (3) the determination
of the new search direction by exploiting the information
calculated by gradient descent in step (2) as well as the previous
search direction. The proposed method improved the training
efficiency of back propagation algorithm by adaptively modifying
the initial search direction. Performance of the proposed method
is demonstrated by comparing to the conjugate gradient algorithm
from neural network toolbox for the chosen benchmark. The
results show that the number of iterations required by the
proposed method to converge is less than 20% of what is required
by the standard conjugate gradient and neural network toolbox
algorithm.
Abstract: This paper presents a novel two-phase hybrid optimization algorithm with hybrid genetic operators to solve the optimal control problem of a single stage hybrid manufacturing system. The proposed hybrid real coded genetic algorithm (HRCGA) is developed in such a way that a simple real coded GA acts as a base level search, which makes a quick decision to direct the search towards the optimal region, and a local search method is next employed to do fine tuning. The hybrid genetic operators involved in the proposed algorithm improve both the quality of the solution and convergence speed. The phase–1 uses conventional real coded genetic algorithm (RCGA), while optimisation by direct search and systematic reduction of the size of search region is employed in the phase – 2. A typical numerical example of an optimal control problem with the number of jobs varying from 10 to 50 is included to illustrate the efficacy of the proposed algorithm. Several statistical analyses are done to compare the validity of the proposed algorithm with the conventional RCGA and PSO techniques. Hypothesis t – test and analysis of variance (ANOVA) test are also carried out to validate the effectiveness of the proposed algorithm. The results clearly demonstrate that the proposed algorithm not only improves the quality but also is more efficient in converging to the optimal value faster. They can outperform the conventional real coded GA (RCGA) and the efficient particle swarm optimisation (PSO) algorithm in quality of the optimal solution and also in terms of convergence to the actual optimum value.
Abstract: In this study, a new and fast algorithm for Ascending
Aorta (AscA) and Descending Aorta (DesA) segmentation is
presented using Computed Tomography Angiography images. This
process is quite important especially at the detection of aortic
plaques, aneurysms, calcification or stenosis. The applied method has
been carried out at four steps. At first step, lung segmentation is
achieved. At the second one, Mediastinum Region (MR) is detected
to use in the segmentation. At the third one, images have been
applied optimal threshold and components which are outside of the
MR were removed. Lastly, identifying and segmentation of AscA and
DesA have been carried out. The performance of the applied method
is found quite well for radiologists and it gives enough results to the
surgeries medically.
Abstract: Face Recognition has always been a fascinating research area. It has drawn the attention of many researchers because of its various potential applications such as security systems, entertainment, criminal identification etc. Many supervised and unsupervised learning techniques have been reported so far. Principal Component Analysis (PCA), Self Organizing Maps (SOM) and Independent Component Analysis (ICA) are the three techniques among many others as proposed by different researchers for Face Recognition, known as the unsupervised techniques. This paper proposes integration of the two techniques, SOM and PCA, for dimensionality reduction and feature selection. Simulation results show that, though, the individual techniques SOM and PCA itself give excellent performance but the combination of these two can also be utilized for face recognition. Experimental results also indicate that for the given face database and the classifier used, SOM performs better as compared to other unsupervised learning techniques. A comparison of two proposed methodologies of SOM, Local and Global processing, shows the superiority of the later but at the cost of more computational time.
Abstract: This paper presents a fast and efficient on-line technique for estimating impedance of unbalanced loads in power systems. The proposed technique is an application of a discrete timedynamic filter based on stochastic estimation theory which is suitable for estimating parameters in noisy environment. The algorithm uses sets of digital samples of the distorted voltage and current waveforms of the non-linear load to estimate the harmonic contents of these two signal. The non-linear load impedance is then calculated from these contents. The method is tested using practical data. Results are reported and compared with those obtained using the conventional least error squares technique. In addition to the very accurate results obtained, the method can detect and reject bad measurements. This can be considered as a very important advantage over the conventional static estimation methods such as the least error square method.
Abstract: In this paper, we employ the approach of linear
programming to propose a new interactive broadcast method. In our
method, a film S is divided into n equal parts and broadcast via k
channels. The user simultaneously downloads these segments from k
channels into the user-s set-top-box (STB) and plays them in order.
Our method assumes that the initial p segments will not have
fast-forwarding capabilities. Every time the user wants to initiate d
times fast-forwarding, according to our broadcasting strategy, the
necessary segments already saved in the user-s STB or are just
download on time for playing. The proposed broadcasting strategy not
only allows the user to pause and rewind, but also to fast-forward.
Abstract: Motion estimation is the most computationally
intensive part in video processing. Many fast motion estimation
algorithms have been proposed to decrease the computational
complexity by reducing the number of candidate motion vectors.
However, these studies are for fast search algorithms themselves while
almost image and video compressions are operated with software
based. Therefore, the timing constraints for running these motion
estimation algorithms not only challenge for the video codec but also
overwhelm for some of processors. In this paper, the performance of
motion estimation is enhanced by using Intel's Streaming SIMD
Extension 2 (SSE2) technology with Intel Pentium 4 processor.
Abstract: This paper presents a novel control strategy of a threephase
four-wire Unified Power Quality (UPQC) for an improvement
in power quality. The UPQC is realized by integration of series and
shunt active power filters (APFs) sharing a common dc bus capacitor.
The shunt APF is realized using a thee-phase, four leg voltage source
inverter (VSI) and the series APF is realized using a three-phase,
three leg VSI. A control technique based on unit vector template
technique (UTT) is used to get the reference signals for series APF,
while instantaneous sequence component theory (ISCT) is used for
the control of Shunt APF. The performance of the implemented
control algorithm is evaluated in terms of power-factor correction,
load balancing, neutral source current mitigation and mitigation of
voltage and current harmonics, voltage sag and swell in a three-phase
four-wire distribution system for different combination of linear and
non-linear loads. In this proposed control scheme of UPQC, the
current/voltage control is applied over the fundamental supply
currents/voltages instead of fast changing APFs currents/voltages,
there by reducing the computational delay and the required sensors.
MATLAB/Simulink based simulations are obtained, which support
the functionality of the UPQC. MATLAB/Simulink based
simulations are obtained, which support the functionality of the
UPQC.
Abstract: The increasing interest on processing data created by
sensor networks has evolved into approaches to implement sensor
networks as databases. The aggregation operator, which calculates a
value from a large group of data such as computing averages or sums,
etc. is an essential function that needs to be provided when
implementing such sensor network databases. This work proposes to
add the DURING clause into TinySQL to calculate values during a
specific long period and suggests a way to implement the aggregation
service in sensor networks by applying materialized view and
incremental view maintenance techniques that is used in data
warehouses. In sensor networks, data values are passed from child
nodes to parent nodes and an aggregation value is computed at the root
node. As such root nodes need to be memory efficient and low
powered, it becomes a problem to recompute aggregate values from all
past and current data. Therefore, applying incremental view
maintenance techniques can reduce the memory consumption and
support fast computation of aggregate values.