Abstract: Sedimentation process resulting from soil erosion in
the water basin especially in arid and semi-arid where poor
vegetation cover in the slope of the mountains upstream could
contribute to sediment formation. The consequence of sedimentation
not only makes considerable change in the morphology of the river
and the hydraulic characteristics but would also have a major
challenge for the operation and maintenance of the canal network
which depend on water flow to meet the stakeholder-s requirements.
For this reason mathematical modeling can be used to simulate the
effective factors on scouring, sediment transport and their settling
along the waterways. This is particularly important behind the
reservoirs which enable the operators to estimate the useful life of
these hydraulic structures. The aim of this paper is to simulate the
sedimentation and erosion in the eastern and western water intake
structures of the Dez Diversion weir using GSTARS-3 software. This
is done to estimate the sedimentation and investigate the ways in
which to optimize the process and minimize the operational
problems. Results indicated that the at the furthest point upstream of
the diversion weir, the coarser sediment grains tended to settle. The
reason for this is the construction of the phantom bridge and the
outstanding rocks just upstream of the structure. The construction of
these along the river course has reduced the momentum energy
require to push the sediment loads and make it possible for them to
settle wherever the river regime allows it. Results further indicated a
trend for the sediment size in such a way that as the focus of study
shifts downstream the size of grains get smaller and vice versa. It
was also found that the finding of the GSTARS-3 had a close
proximity with the sets of the observed data. This suggests that the
software is a powerful analytical tool which can be applied in the
river engineering project with a minimum of costs and relatively
accurate results.
Abstract: Oxidative stress and overwhelming free radicals
associated with diabetes mellitus are likely to be linked with
development of certain complication such as retinopathy,
nephropathy and neuropathy. Treatment of diabetic subjects with
antioxidant may be of advantage in attenuating these complications.
Olive leaf (Oleaeuropaea), has been endowed with many beneficial
and health promoting properties mostly linked to its antioxidant
activity. This study aimed to evaluate the significance of
supplementation of Olive leaves extract (OLE) in reducing oxidative
stress, hyperglycemia and hyperlipidemia in Sterptozotocin (STZ)-
induced diabetic rats. After induction of diabetes, a significant rise in
plasma glucose, lipid profiles except High density lipoproteincholestrol
(HDLc), malondialdehyde (MDA) and significant decrease
of plasma insulin, HDLc and Plasma reduced glutathione GSH as
well as alteration in enzymatic antioxidants was observed in all
diabetic animals. During treatment of diabetic rats with 0.5g/kg body
weight of Olive leaves extract (OLE) the levels of plasma (MDA)
,(GSH), insulin, lipid profiles along with blood glucose and
erythrocyte enzymatic antioxidant enzymes were significantly
restored to establish values that were not different from normal
control rats. Untreated diabetic rats on the other hand demonstrated
persistent alterations in the oxidative stress marker (MDA), blood
glucose, insulin, lipid profiles and the antioxidant parameters. These
results demonstrate that OLE may be of advantage in inhibiting
hyperglycemia, hyperlipidemia and oxidative stress induced by
diabetes and suggest that administration of OLE may be helpful in
the prevention or at least reduced of diabetic complications
associated with oxidative stress.
Abstract: A continuum model is presented to study vdW
interaction on buckling analysis of multi-walled walled carbon
nanotube. In previous studies, only the vdW interaction between
adjacent two layers was considered and the vdW interaction between
the other two layers was neglected. The results show that the vdW
interaction cofficients are dependent on the change of interlayer
spacing and the radii of tubes. With increase of radii the vdW
coefficients approach a constant value. The numerical results show
that the effect of vdW interaction on the critical strain for a doublewalled
CNT is negligible when the radius is large enough for the
both the cases of before and after buckling.
Abstract: The quality of short term load forecasting can improve the efficiency of planning and operation of electric utilities. Artificial Neural Networks (ANNs) are employed for nonlinear short term load forecasting owing to their powerful nonlinear mapping capabilities. At present, there is no systematic methodology for optimal design and training of an artificial neural network. One has often to resort to the trial and error approach. This paper describes the process of developing three layer feed-forward large neural networks for short-term load forecasting and then presents a heuristic search algorithm for performing an important task of this process, i.e. optimal networks structure design. Particle Swarm Optimization (PSO) is used to develop the optimum large neural network structure and connecting weights for one-day ahead electric load forecasting problem. PSO is a novel random optimization method based on swarm intelligence, which has more powerful ability of global optimization. Employing PSO algorithms on the design and training of ANNs allows the ANN architecture and parameters to be easily optimized. The proposed method is applied to STLF of the local utility. Data are clustered due to the differences in their characteristics. Special days are extracted from the normal training sets and handled separately. In this way, a solution is provided for all load types, including working days and weekends and special days. The experimental results show that the proposed method optimized by PSO can quicken the learning speed of the network and improve the forecasting precision compared with the conventional Back Propagation (BP) method. Moreover, it is not only simple to calculate, but also practical and effective. Also, it provides a greater degree of accuracy in many cases and gives lower percent errors all the time for STLF problem compared to BP method. Thus, it can be applied to automatically design an optimal load forecaster based on historical data.
Abstract: Gastric ulceration is a discontinuity in gastric mucosa, usually occurs due to imbalance between the gastric mucosal protective factors, that is called gastric mucosal barrier, and the aggressive factors, to which the mucosa is exposed. This study was carried out on sixty male Sprague-Dowely rats (12- 16 weeks old) allocated into two groups. The first control group and the second Gastric lesion group which induced by oral administration of a single daily dose of aspirin at a dose of 300 mg/kg body weight for 7 consecutive-days (6% aspirin solution will be prepared and each rat will be given 5 ml of that solution/kg body weight). Blood is collected 1, 2 and 3 weeks after induction of gastric ulceration. Significant increase in serum copper, nitric oxide, and prostaglandin E2 all over the period of experiment. Significant decrease in erythrocyte superoxide dismutase (t-SOD) activities, serum (calcium, phosphorus, glucose and insulin) levels. Non-significant changes in serum sodium and potassium levels are obtained.
Abstract: Since the advent of the information era, the Internet has
brought various positive effects in everyday life. Nevertheless,
recently, problems and side-effects have been noted. Internet
witch-trials and spread of pornography are only a few of these
problems.In this study, problems and causes of malicious replies on
internet boards were analyzed, using the key ideas of game theory. The
study provides a mathematical model for the internet reply game to
devise three possible plans that could efficiently counteract malicious
replies. Furthermore, seven specific measures that comply with one of
the three plans were proposed and evaluated according to the
importance and utility of each measure using the orthogonal array
survey and SPSS conjoint analysis.The conclusion was that the most
effective measure would be forbidding unsigned user access to
malicious replies. Also notable was that some analytically proposed
measures, when implemented, could backfire and encourage malicious
replies.
Abstract: With a surge of stream processing applications novel
techniques are required for generation and analysis of association
rules in streams. The traditional rule mining solutions cannot handle
streams because they generally require multiple passes over the data
and do not guarantee the results in a predictable, small time. Though
researchers have been proposing algorithms for generation of rules
from streams, there has not been much focus on their analysis.
We propose Association rule profiling, a user centric process for
analyzing association rules and attaching suitable profiles to them
depending on their changing frequency behavior over a previous
snapshot of time in a data stream.
Association rule profiles provide insights into the changing nature
of associations and can be used to characterize the associations. We
discuss importance of characteristics such as predictability of
linkages present in the data and propose metric to quantify it. We
also show how association rule profiles can aid in generation of user
specific, more understandable and actionable rules.
The framework is implemented as SUPAR: System for Usercentric
Profiling of Association Rules in streaming data. The
proposed system offers following capabilities:
i) Continuous monitoring of frequency of streaming item-sets
and detection of significant changes therein for association rule
profiling.
ii) Computation of metrics for quantifying predictability of
associations present in the data.
iii) User-centric control of the characterization process: user
can control the framework through a) constraint specification and b)
non-interesting rule elimination.
Abstract: IEEE 802.16 is a new wireless technology standard, it
has some advantages, including wider coverage, higher bandwidth,
and QoS support. As the new wireless technology for last mile
solution, there are designed two models in IEEE 802.16 standard. One
is PMP (point to multipoint) and the other is Mesh. In this paper we
only focus on IEEE 802.16 Mesh model. According to the IEEE
802.16 standard description, Mesh model has two scheduling modes,
centralized and distributed. Considering the pros and cons of the two
scheduling, we present the combined scheduling QoS framework that
the BS (Base Station) controls time frame scheduling and selects the
shortest path from source to destination directly. On the other hand, we
propose the Expedited Queue mechanism to cut down the transmission
time. The EQ mechanism can reduce a lot of end-to-end delay in our
QoS framework. Simulation study has shown that the average delay is
smaller than contrasts. Furthermore, our proposed scheme can also
achieve higher performance.
Abstract: Discretization of spatial derivatives is an important
issue in meshfree methods especially when the derivative terms
contain non-linear coefficients. In this paper, various methods used
for discretization of second-order spatial derivatives are investigated
in the context of Smoothed Particle Hydrodynamics. Three popular
forms (i.e. "double summation", "second-order kernel derivation",
and "difference scheme") are studied using one-dimensional unsteady
heat conduction equation. To assess these schemes, transient response
to a step function initial condition is considered. Due to parabolic
nature of the heat equation, one can expect smooth and monotone
solutions. It is shown, however in this paper, that regardless of
the type of kernel function used and the size of smoothing radius,
the double summation discretization form leads to non-physical
oscillations which persist in the solution. Also, results show that when
a second-order kernel derivative is used, a high-order kernel function
shall be employed in such a way that the distance of inflection
point from origin in the kernel function be less than the nearest
particle distance. Otherwise, solutions may exhibit oscillations near
discontinuities unlike the "difference scheme" which unconditionally
produces monotone results.
Abstract: In this paper, the periodic surveillance scheme has
been proposed for any convex region using mobile wireless sensor
nodes. A sensor network typically consists of fixed number of
sensor nodes which report the measurements of sensed data such as
temperature, pressure, humidity, etc., of its immediate proximity
(the area within its sensing range). For the purpose of sensing an
area of interest, there are adequate number of fixed sensor
nodes required to cover the entire region of interest. It implies
that the number of fixed sensor nodes required to cover a given
area will depend on the sensing range of the sensor as well as
deployment strategies employed. It is assumed that the sensors to
be mobile within the region of surveillance, can be mounted on
moving bodies like robots or vehicle. Therefore, in our
scheme, the surveillance time period determines the number of
sensor nodes required to be deployed in the region of interest.
The proposed scheme comprises of three algorithms namely:
Hexagonalization, Clustering, and Scheduling, The first algorithm
partitions the coverage area into fixed sized hexagons that
approximate the sensing range (cell) of individual sensor node.
The clustering algorithm groups the cells into clusters, each of
which will be covered by a single sensor node. The later
determines a schedule for each sensor to serve its respective cluster.
Each sensor node traverses all the cells belonging to the cluster
assigned to it by oscillating between the first and the last cell for
the duration of its life time. Simulation results show that our
scheme provides full coverage within a given period of time using
few sensors with minimum movement, less power consumption,
and relatively less infrastructure cost.
Abstract: In this paper, we propose an improvement of pattern
growth-based PrefixSpan algorithm, called I-PrefixSpan. The general idea of I-PrefixSpan is to use sufficient data structure for Seq-Tree
framework and separator database to reduce the execution time and
memory usage. Thus, with I-PrefixSpan there is no in-memory database stored after index set is constructed. The experimental result
shows that using Java 2, this method improves the speed of PrefixSpan up to almost two orders of magnitude as well as the memory usage to more than one order of magnitude.
Abstract: This paper presents a solution for the behavioural
animation of autonomous virtual agent navigation in virtual environments.
We focus on using Dempster-Shafer-s Theory of Evidence
in developing visual sensor for virtual agent. The role of the visual
sensor is to capture the information about the virtual environment
or identifie which part of an obstacle can be seen from the position
of the virtual agent. This information is require for vitual agent to
coordinate navigation in virtual environment. The virual agent uses
fuzzy controller as a navigation system and Fuzzy α - level for
the action selection method. The result clearly demonstrates the path
produced is reasonably smooth even though there is some sharp turn
and also still not diverted too far from the potential shortest path.
This had indicated the benefit of our method, where more reliable
and accurate paths produced during navigation task.
Abstract: the aim of that work is to study the proton transfer
phenomenon which takes place in the elastic scattering of 12C on 11B
at energies near the coulomb barrier. This reaction was studied at four
different energies 16, 18, 22, 24 MeV. The experimental data of the
angular distribution at these energies were compared to the
calculation prediction using the optical potential codes such as
ECIS88 and SPIVAL. For the raising in the cross section at backward
angles due to the transfer process we could use Distorted Wave Born
Approximation (DWUCK5). Our analysis showed that SPIVAL code
with l-dependent imaginary potential could be used effectively.
Abstract: It is important to remove manganese from water
because of its effects on human and the environment. Human
activities are one of the biggest contributors for excessive manganese
concentration in the environment. The proposed method to remove
manganese in aqueous solution by using adsorption as in carbon
nanotubes (CNT) at different parameters: The parameters are CNT
dosage, pH, agitation speed and contact time. Different pHs are pH
6.0, pH 6.5, pH 7.0, pH 7.5 and pH 8.0, CNT dosages are 5mg,
6.25mg, 7.5mg, 8.75mg or 10mg, contact time are 10 min, 32.5 min,
55 min, 87.5 min and 120 min while the agitation speeds are 100rpm,
150rpm, 200rpm, 250rpm and 300rpm. The parameters chosen for
experiments are based on experimental design done by using Central
Composite Design, Design Expert 6.0 with 4 parameters, 5 levels and
2 replications. Based on the results, condition set at pH 7.0, agitation
speed of 300 rpm, 7.5mg and contact time 55 minutes gives the
highest removal with 75.5%. From ANOVA analysis in Design
Expert 6.0, the residual concentration will be very much affected by
pH and CNT dosage. Initial manganese concentration is 1.2mg/L
while the lowest residual concentration achieved is 0.294mg/L,
which almost satisfy DOE Malaysia Standard B requirement.
Therefore, further experiments must be done to remove manganese
from model water to the required standard (0.2 mg/L) with the initial
concentration set to 0.294 mg/L.
Abstract: The Application of e-health solutions has brought superb advancements in the health care industry. E-health solutions have already been embraced in the industrialized countries. In an effort to catch up with the growth, the developing countries have strived to revolutionize the healthcare industry by use of Information technology in different ways. Based on a technology assessment carried out in Kenya – one of the developing countries – and using multiple case studies in Nyanza Province, this work focuses on an investigation on how five rural hospitals are adapting to the technology shift. The issues examined include the ICT infrastructure and e-health technologies in place, the knowledge of participants in terms of benefits gained through the use of ICT and the challenges posing barriers to the use of ICT technologies in these hospitals. The results reveal that the ICT infrastructure in place is inadequate for e-health implementations as a result to various challenges that exist. Consequently, suggestions on how to tackle the various challenges have been addressed in this paper.
Abstract: In non destructive testing by radiography, a perfect
knowledge of the weld defect shape is an essential step to
appreciate the quality of the weld and make decision on its
acceptability or rejection. Because of the complex nature of the
considered images, and in order that the detected defect region
represents the most accurately possible the real defect, the choice
of thresholding methods must be done judiciously. In this paper,
performance criteria are used to conduct a comparative study of
four non parametric histogram thresholding methods for automatic
extraction of weld defect in radiographic images.
Abstract: Recently in the field of bridges that are newly built or
repaired, fast construction is required more than ever. For these
reasons, precast prefabricated bridge that enables rapid construction is
actively discussed and studied today. In South Korea, it is called
modular bridge. Cross beam is an integral component of modular
bridge. It functions for load distribution, reduction of bending
moment, resistance of horizontal strength on lateral upper structure. In
this study, the structural characteristics of domestic and foreign cross
beam types were compared. Based on this, alternative cross beam
connection types suitable for modular bridge were selected. And
bulb-T girder specimens were fabricated with each type of connection.
The behavior of each specimen was analyzed under static loading, and
cross beam connection type which is expected to be best suited to
modular bridge proposed.
Abstract: In this paper, SFQ (Start Time Fair Queuing)
algorithm is analyzed when this is applied in computer networks to
know what kind of behavior the traffic in the net has when different
data sources are managed by the scheduler. Using the NS2 software
the computer networks were simulated to be able to get the graphs
showing the performance of the scheduler. Different traffic sources
were introduced in the scripts, trying to establish the real scenario.
Finally the results were that depending on the data source, the traffic
can be affected in different levels, when Constant Bite Rate is
applied, the scheduler ensures a constant level of data sent and
received, but the truth is that in the real life it is impossible to ensure
a level that resists the changes in work load.
Abstract: Conventional WBL is effective for meaningful student, because rote student learn by repeating without thinking or trying to understand. It is impossible to have full benefit from conventional WBL. Understanding of rote student-s intention and what influences it becomes important. Poorly designed user interface will discourage rote student-s cultivation and intention to use WBL. Thus, user interface design is an important factor especially when WBL is used as comprehensive replacement of conventional teaching. This research proposes the influencing factors that can enhance student-s intention to use the system. The enhanced TAM is used for evaluating the proposed factors. The research result points out that factors influencing rote student-s intention are Perceived Usefulness of Homepage Content Structure, Perceived User Friendly Interface, Perceived Hedonic Component, and Perceived (homepage) Visual Attractiveness.
Abstract: In this paper, an automatic detecting algorithm for
QRS complex detecting was applied for analyzing ECG recordings
and five criteria for dangerous arrhythmia diagnosing are applied for a
protocol type of automatic arrhythmia diagnosing system. The
automatic detecting algorithm applied in this paper detected the
distribution of QRS complexes in ECG recordings and related
information, such as heart rate and RR interval. In this investigation,
twenty sampled ECG recordings of patients with different pathologic
conditions were collected for off-line analysis. A combinative
application of four digital filters for bettering ECG signals and
promoting detecting rate for QRS complex was proposed as
pre-processing. Both of hardware filters and digital filters were
applied to eliminate different types of noises mixed with ECG
recordings. Then, an automatic detecting algorithm of QRS complex
was applied for verifying the distribution of QRS complex. Finally,
the quantitative clinic criteria for diagnosing arrhythmia were
programmed in a practical application for automatic arrhythmia
diagnosing as a post-processor. The results of diagnoses by automatic
dangerous arrhythmia diagnosing were compared with the results of
off-line diagnoses by experienced clinic physicians. The results of
comparison showed the application of automatic dangerous
arrhythmia diagnosis performed a matching rate of 95% compared
with an experienced physician-s diagnoses.