Abstract: This study determines the effect of naked and heparinbased
super-paramagnetic iron oxide nanoparticles on the human
cancer cell lines of A2780. Doxorubicin was used as the anticancer
drug, entrapped in the SPIO-NPs. This study aimed to decorate
nanoparticles with heparin, a molecular ligand for 'active' targeting
of cancerous cells and the application of modified-nanoparticles in
cancer treatment. The nanoparticles containing the anticancer drug
DOX were prepared by a solvent evaporation and emulsification
cross-linking method. The physicochemical properties of the
nanoparticles were characterized by various techniques, and uniform
nanoparticles with an average particle size of 110±15 nm with high
encapsulation efficiencies (EE) were obtained. Additionally, a
sustained release of DOX from the SPIO-NPs was successful.
Cytotoxicity tests showed that the SPIO-DOX-HP had higher cell
toxicity than the individual HP and confocal microscopy analysis
confirmed excellent cellular uptake efficiency. These results indicate
that HP based SPIO-NPs have potential uses as anticancer drug
carriers and also have an enhanced anticancer effect.
Abstract: This paper features the modeling and design of a Fast
Output Sampling (FOS) Feedback control technique for the Active
Vibration Control (AVC) of a smart flexible aluminium cantilever
beam for a Single Input Single Output (SISO) case. Controllers are
designed for the beam by bonding patches of piezoelectric layer as
sensor / actuator to the master structure at different locations along
the length of the beam by retaining the first 2 dominant vibratory
modes. The entire structure is modeled in state space form using the
concept of piezoelectric theory, Euler-Bernoulli beam theory, Finite
Element Method (FEM) and the state space techniques by dividing
the structure into 3, 4, 5 finite elements, thus giving rise to three
types of systems, viz., system 1 (beam divided into 3 finite
elements), system 2 (4 finite elements), system 3 (5 finite elements).
The effect of placing the sensor / actuator at various locations along
the length of the beam for all the 3 types of systems considered is
observed and the conclusions are drawn for the best performance and
for the smallest magnitude of the control input required to control the
vibrations of the beam. Simulations are performed in MATLAB. The
open loop responses, closed loop responses and the tip displacements
with and without the controller are obtained and the performance of
the proposed smart system is evaluated for vibration control.
Abstract: A general stochastic spatial MIMO channel model is
proposed for evaluating various MIMO techniques in this paper. It can
generate MIMO channels complying with various MIMO
configurations such as smart antenna, spatial diversity and spatial
multiplexing. The modeling method produces the stochastic fading
involving delay spread, Doppler spread, DOA (direction of arrival),
AS (angle spread), PAS (power azimuth Spectrum) of the scatterers,
antenna spacing and the wavelength. It can be applied in various
MIMO technique researches flexibly with low computing complexity.
Abstract: This paper applies Bayesian Networks to support
information extraction from unstructured, ungrammatical, and
incoherent data sources for semantic annotation. A tool has been
developed that combines ontologies, machine learning, and
information extraction and probabilistic reasoning techniques to
support the extraction process. Data acquisition is performed with the
aid of knowledge specified in the form of ontology. Due to the
variable size of information available on different data sources, it is
often the case that the extracted data contains missing values for
certain variables of interest. It is desirable in such situations to
predict the missing values. The methodology, presented in this paper,
first learns a Bayesian network from the training data and then uses it
to predict missing data and to resolve conflicts. Experiments have
been conducted to analyze the performance of the presented
methodology. The results look promising as the methodology
achieves high degree of precision and recall for information
extraction and reasonably good accuracy for predicting missing
values.
Abstract: The ever increasing use of World Wide Web in the
existing network, results in poor performance. Several techniques
have been developed for reducing web traffic by compressing the size
of the file, saving the web pages at the client side, changing the burst
nature of traffic into constant rate etc. No single method was
adequate enough to access the document instantly through the
Internet. In this paper, adaptive hybrid algorithms are developed for
reducing web traffic. Intelligent agents are used for monitoring the
web traffic. Depending upon the bandwidth usage, user-s preferences,
server and browser capabilities, intelligent agents use the best
techniques to achieve maximum traffic reduction. Web caching,
compression, filtering, optimization of HTML tags, and traffic
dispersion are incorporated into this adaptive selection. Using this
new hybrid technique, latency is reduced to 20 – 60 % and cache hit
ratio is increased 40 – 82 %.
Abstract: In this research paper we have presented control
architecture for robotic arm movement and trajectory planning using
Fuzzy Logic (FL) and Genetic Algorithms (GAs). This architecture is
used to compensate the uncertainties like; movement, friction and
settling time in robotic arm movement. The genetic algorithms and
fuzzy logic is used to meet the objective of optimal control
movement of robotic arm. This proposed technique represents a
general model for redundant structures and may extend to other
structures. Results show optimal angular movement of joints as result
of evolutionary process. This technique has edge over the other
techniques as minimum mathematics complexity used.
Abstract: Load forecasting has always been the essential part of
an efficient power system operation and planning. A novel approach
based on support vector machines is proposed in this paper for annual
power load forecasting. Different kernel functions are selected to
construct a combinatorial algorithm. The performance of the new
model is evaluated with a real-world dataset, and compared with two
neural networks and some traditional forecasting techniques. The
results show that the proposed method exhibits superior performance.
Abstract: e-Government structures permits the government to operate in a more transparent and accountable manner of which it increases the power of the individual in relation to that of the government. This paper identifies the factors that determine customer-s attitude towards e-Government services using a theoretical model based on the Technology Acceptance Model. Data relating to the constructs were collected from 200 respondents. The research model was tested using Structural Equation Modeling (SEM) techniques via the Analysis of Moment Structure (AMOS 16) computer software. SEM is a comprehensive approach to testing hypotheses about relations among observed and latent variables. The proposed model fits the data well. The results demonstrated that e- Government services acceptance can be explained in terms of compatibility and attitude towards e-Government services. The setup of the e-Government services will be compatible with the way users work and are more likely to adopt e-Government services owing to their familiarity with the Internet for various official, personal, and recreational uses. In addition, managerial implications for government policy makers, government agencies, and system developers are also discussed.
Abstract: This paper demonstrates the application of craziness based particle swarm optimization (CRPSO) technique for designing the 8th order low pass Infinite Impulse Response (IIR) filter. CRPSO, the much improved version of PSO, is a population based global heuristic search algorithm which finds near optimal solution in terms of a set of filter coefficients. Effectiveness of this algorithm is justified with a comparative study of some well established algorithms, namely, real coded genetic algorithm (RGA) and particle swarm optimization (PSO). Simulation results affirm that the proposed algorithm CRPSO, outperforms over its counterparts not only in terms of quality output i.e. sharpness at cut-off, pass band ripple, stop band ripple, and stop band attenuation but also in convergence speed with assured stability.
Abstract: This study demonstrates the use of Class F fly ash in
combination with lime or lime kiln dust in the full depth reclamation
(FDR) of asphalt pavements. FDR, in the context of this paper, is a
process of pulverizing a predetermined amount of flexible pavement
that is structurally deficient, blending it with chemical additives and
water, and compacting it in place to construct a new stabilized base
course. Test sections of two structurally deficient asphalt pavements
were reclaimed using Class F fly ash in combination with lime and
lime kiln dust. In addition, control sections were constructed using
cement, cement and emulsion, lime kiln dust and emulsion, and mill
and fill. The service performance and structural behavior of the FDR
pavement test sections were monitored to determine how the fly ash
sections compared to other more traditional pavement rehabilitation
techniques. Service performance and structural behavior were
determined with the use of sensors embedded in the road and Falling
Weight Deflectometer (FWD) tests. Monitoring results of the FWD
tests conducted up to 2 years after reclamation show that the cement,
fly ash+LKD, and fly ash+lime sections exhibited two year resilient
modulus values comparable to open graded cement stabilized
aggregates (more than 750 ksi). The cement treatment resulted in a
significant increase in resilient modulus within 3 weeks of
construction and beyond this curing time, the stiffness increase was
slow. On the other hand, the fly ash+LKD and fly ash+lime test
sections indicated slower shorter-term increase in stiffness. The fly
ash+LKD and fly ash+lime section average resilient modulus values
at two years after construction were in excess of 800 ksi. Additional
longer-term testing data will be available from ongoing pavement
performance and environmental condition data collection at the two
pavement sites.
Abstract: Freeways are originally designed to provide high
mobility to road users. However, the increase in population and
vehicle numbers has led to increasing congestions around the world.
Daily recurrent congestion substantially reduces the freeway capacity
when it is most needed. Building new highways and expanding the
existing ones is an expensive solution and impractical in many
situations. Intelligent and vision-based techniques can, however, be
efficient tools in monitoring highways and increasing the capacity of
the existing infrastructures. The crucial step for highway monitoring
is vehicle detection. In this paper, we propose one of such
techniques. The approach is based on artificial neural networks
(ANN) for vehicles detection and counting. The detection process
uses the freeway video images and starts by automatically extracting
the image background from the successive video frames. Once the
background is identified, subsequent frames are used to detect
moving objects through image subtraction. The result is segmented
using Sobel operator for edge detection. The ANN is, then, used in
the detection and counting phase. Applying this technique to the
busiest freeway in Riyadh (King Fahd Road) achieved higher than
98% detection accuracy despite the light intensity changes, the
occlusion situations, and shadows.
Abstract: Characteristics and sonocatalytic activity of zeolite
Y catalysts loaded with TiO2 using impregnation and ion exchange
methods for the degradation of amaranth dye were investigated.
The Ion-exchange method was used to encapsulate the TiO2 into
the internal pores of the zeolite while the incorporation of TiO2
mostly on the external surface of zeolite was carried out using the
impregnation method. Different characterization techniques were
used to elucidate the physicochemical properties of the produced
catalysts. The framework of zeolite Y remained virtually
unchanged after the encapsulation of TiO2 while the crystallinity of
zeolite decreased significantly after the incorporation of 15 wt% of
TiO2. The sonocatalytic activity was enhanced by TiO2
incorporation with maximum degradation efficiencies of 50% and
68% for the encapsulated titanium and titanium loaded onto the
zeolite, respectively after 120min of reaction. Catalysts
characteristics and sonocatalytic behaviors were significantly
affected by the preparation method and the location of TiO2
introduced with zeolite structure. Behaviors in the sonocatalytic
process were successfully correlated with the characteristics of the
catalysts used.
Abstract: Many computational techniques were applied to
solution of heat conduction problem. Those techniques were the
finite difference (FD), finite element (FE) and recently meshless
methods. FE is commonly used in solution of equation of heat
conduction problem based on the summation of stiffness matrix of
elements and the solution of the final system of equations. Because
of summation process of finite element, convergence rate was
decreased. Hence in the present paper Cellular Automata (CA)
approach is presented for the solution of heat conduction problem.
Each cell considered as a fixed point in a regular grid lead to the
solution of a system of equations is substituted by discrete systems of
equations with small dimensions. Results show that CA can be used
for solution of heat conduction problem.
Abstract: Text-based game is supposed to be a low resource
consumption application that delivers good performances when
compared to graphical-intensive type of games. But, nowadays, some
of the online text-based games are not offering performances that are
acceptable to the users. Therefore, an online text-based game called
Star_Quest has been developed in order to analyze its behavior under
different performance measurements. Performance metrics such as
throughput, scalability, response time and page loading time are
captured to yield the performance of the game. The techniques in
performing the load testing are also disclosed to exhibit the viability
of our work. The comparative assessment between the results
obtained and the accepted level of performances are conducted as to
determine the performance level of the game. The study reveals that
the developed game managed to meet all the performance objectives
set forth.
Abstract: In this paper, we generalize several techniques in
developing Fault Tolerant Software. We introduce property
“Correctness" in evaluating N-version Systems and compare it to
some commonly used properties such as reliability or availability.
We also find out the relation between this property and the number of
versions of system. Our experiments to verify the correctness and the
applicability of the relation are also presented.
Abstract: Evolutionary Algorithms are population-based,
stochastic search techniques, widely used as efficient global
optimizers. However, many real life optimization problems often
require finding optimal solution to complex high dimensional,
multimodal problems involving computationally very expensive
fitness function evaluations. Use of evolutionary algorithms in such
problem domains is thus practically prohibitive. An attractive
alternative is to build meta models or use an approximation of the
actual fitness functions to be evaluated. These meta models are order
of magnitude cheaper to evaluate compared to the actual function
evaluation. Many regression and interpolation tools are available to
build such meta models. This paper briefly discusses the
architectures and use of such meta-modeling tools in an evolutionary
optimization context. We further present two evolutionary algorithm
frameworks which involve use of meta models for fitness function
evaluation. The first framework, namely the Dynamic Approximate
Fitness based Hybrid EA (DAFHEA) model [14] reduces
computation time by controlled use of meta-models (in this case
approximate model generated by Support Vector Machine
regression) to partially replace the actual function evaluation by
approximate function evaluation. However, the underlying
assumption in DAFHEA is that the training samples for the metamodel
are generated from a single uniform model. This does not take
into account uncertain scenarios involving noisy fitness functions.
The second model, DAFHEA-II, an enhanced version of the original
DAFHEA framework, incorporates a multiple-model based learning
approach for the support vector machine approximator to handle
noisy functions [15]. Empirical results obtained by evaluating the
frameworks using several benchmark functions demonstrate their
efficiency
Abstract: This paper presents a novel two-phase hybrid optimization algorithm with hybrid genetic operators to solve the optimal control problem of a single stage hybrid manufacturing system. The proposed hybrid real coded genetic algorithm (HRCGA) is developed in such a way that a simple real coded GA acts as a base level search, which makes a quick decision to direct the search towards the optimal region, and a local search method is next employed to do fine tuning. The hybrid genetic operators involved in the proposed algorithm improve both the quality of the solution and convergence speed. The phase–1 uses conventional real coded genetic algorithm (RCGA), while optimisation by direct search and systematic reduction of the size of search region is employed in the phase – 2. A typical numerical example of an optimal control problem with the number of jobs varying from 10 to 50 is included to illustrate the efficacy of the proposed algorithm. Several statistical analyses are done to compare the validity of the proposed algorithm with the conventional RCGA and PSO techniques. Hypothesis t – test and analysis of variance (ANOVA) test are also carried out to validate the effectiveness of the proposed algorithm. The results clearly demonstrate that the proposed algorithm not only improves the quality but also is more efficient in converging to the optimal value faster. They can outperform the conventional real coded GA (RCGA) and the efficient particle swarm optimisation (PSO) algorithm in quality of the optimal solution and also in terms of convergence to the actual optimum value.
Abstract: Biometric techniques are gaining importance for
personal authentication and identification as compared to the
traditional authentication methods. Biometric templates are
vulnerable to variety of attacks due to their inherent nature. When a
person-s biometric is compromised his identity is lost. In contrast to
password, biometric is not revocable. Therefore, providing security
to the stored biometric template is very crucial. Crypto biometric
systems are authentication systems, which blends the idea of
cryptography and biometrics. Fuzzy vault is a proven crypto
biometric construct which is used to secure the biometric templates.
However fuzzy vault suffer from certain limitations like nonrevocability,
cross matching. Security of the fuzzy vault is affected
by the non-uniform nature of the biometric data. Fuzzy vault when
hardened with password overcomes these limitations. Password
provides an additional layer of security and enhances user privacy.
Retina has certain advantages over other biometric traits. Retinal
scans are used in high-end security applications like access control to
areas or rooms in military installations, power plants, and other high
risk security areas. This work applies the idea of fuzzy vault for
retinal biometric template. Multimodal biometric system
performance is well compared to single modal biometric systems.
The proposed multi modal biometric fuzzy vault includes combined
feature points from retina and fingerprint. The combined vault is
hardened with user password for achieving high level of security.
The security of the combined vault is measured using min-entropy.
The proposed password hardened multi biometric fuzzy vault is
robust towards stored biometric template attacks.
Abstract: Concerning the inpatient care the present situation is
characterized by intense charges of medical technology into the
clinical daily routine and an ever stronger integration of special
techniques into the clinical workflow. Medical technology is by now
an integral part of health care according to consisting general
accepted standards. Purchase and operation thereby represent an
important economic position and both are subject of everyday
optimisation attempts. For this purpose by now exists a huge number
of tools which conduce more likely to a complexness of the problem
by a comprehensive implementation. In this paper the advantages of
an integrative information-workflow on the life-cycle-management in
the region of medical technology are shown.
Abstract: Parsing is important in Linguistics and Natural
Language Processing to understand the syntax and semantics of a
natural language grammar. Parsing natural language text is
challenging because of the problems like ambiguity and inefficiency.
Also the interpretation of natural language text depends on context
based techniques. A probabilistic component is essential to resolve
ambiguity in both syntax and semantics thereby increasing accuracy
and efficiency of the parser. Tamil language has some inherent
features which are more challenging. In order to obtain the solutions,
lexicalized and statistical approach is to be applied in the parsing
with the aid of a language model. Statistical models mainly focus on
semantics of the language which are suitable for large vocabulary
tasks where as structural methods focus on syntax which models
small vocabulary tasks. A statistical language model based on Trigram
for Tamil language with medium vocabulary of 5000 words has
been built. Though statistical parsing gives better performance
through tri-gram probabilities and large vocabulary size, it has some
disadvantages like focus on semantics rather than syntax, lack of
support in free ordering of words and long term relationship. To
overcome the disadvantages a structural component is to be
incorporated in statistical language models which leads to the
implementation of hybrid language models. This paper has attempted
to build phrase structured hybrid language model which resolves
above mentioned disadvantages. In the development of hybrid
language model, new part of speech tag set for Tamil language has
been developed with more than 500 tags which have the wider
coverage. A phrase structured Treebank has been developed with 326
Tamil sentences which covers more than 5000 words. A hybrid
language model has been trained with the phrase structured Treebank
using immediate head parsing technique. Lexicalized and statistical
parser which employs this hybrid language model and immediate
head parsing technique gives better results than pure grammar and
trigram based model.