Abstract: A handful of propagation textbooks that discuss radio frequency (RF) propagation models merely list out the models and perhaps discuss them rather briefly; this may well be frustrating for the potential first time modeller who's got no idea on how these models could have been derived. This paper fundamentally provides an overture in modelling the radio channel. Explicitly, for the modelling practice discussed here, signal strength field measurements had to be conducted beforehand (this was done at 469 MHz); to be precise, this paper primarily concerns empirically/statistically modelling the radio channel, and thus provides results obtained from empirically modelling the environments in question. This paper, on the whole, proposes three propagation models, corresponding to three experimented environments. Perceptibly, the models have been derived by way of making the most use of statistical measures. Generally speaking, the first two models were derived via simple linear regression analysis, whereas the third have been originated using multiple regression analysis (with five various predictors). Additionally, as implied by the title of this paper, both indoor and outdoor environments have been experimented; however, (somewhat) two of the environments are neither entirely indoor nor entirely outdoor. The other environment, however, is completely indoor.
Abstract: Thailand is one of the world-s leaders of rice
producers and exporters. Farmers have to increase the rice cultivation
frequency for serving the national increasing of export-s demand. It
leads to an elimination of rice residues by open burning which is the
quickest and costless management method. The open burning of rice
residue is one of the major causes of air pollutants and greenhouse
gas (GHG) emission. Under ASEAN agreement on trans-boundary
haze, Thailand set the master plan to mitigate air pollutant emission
from open burning of agricultural residues. In this master plan,
residues incorporation is promoted as alternative management
method to open burning. However, the assessment of both options in
term of GHG emission in order to investigate their contribution to
long-term global warming is still scarce or inexistent. In this study, a
method on rice residues assessment was first developed in order to
estimate and compare GHG emissions from rice cultivation under
rice residues open burning and the case with incorporation of the
same amount of rice residues, using 2006 IPCC guidelines for
emission estimation and Life Cycle Analysis technique. The
emission from rice cultivation in different preparing area practice
was also discussed.
Abstract: This paper presents the first reflexions about Margaret Mascarenhas-s novel, “Skin", based on post-colonial critic perception of History and its agents. By doing so, this study will put light on a literary corpus of Indian Literatures: the Goan Literature whose cultural basis creates an unique historiographic metafiction conducted by different characters that one by one plays the narrator role.
Abstract: Pharmacology curriculum plays an integral role in
medical education. Learning pharmacology to choose and prescribe
drugs is a major challenge encountered by students. We developed
pharmacology applied learning activities for first year medical
students that included realistic clinical situations with escalating
complications which required the students to analyze the situation
and think critically to choose a safe drug. Tutor feedback was
provided at the end of session. Evaluation was done to assess the
students- level of interest and usefulness of the sessions in rational
selection of drugs. Majority (98 %) of the students agreed that the
session was an extremely useful learning exercise and agreed that
similar sessions would help in rational selection of drugs. Applied
learning sessions in the early years of medical program may promote
deep learning and bridge the gap between pharmacology theory and
clinical practice. Besides, it may also enhance safe prescribing skills.
Abstract: The rheological properties of light crude oil and its mixture with water were investigated experimentally. These rheological properties include steady flow behavior, yield stress, transient flow behavior, and viscoelastic behavior. A RheoStress RS600 rheometer was employed in all of the rheological examination tests. The light crude oil exhibits a Newtonian and for emulsion exhibits a non-Newtonian shear thinning behavior over the examined shear rate range of 0.1–120 s-1. In first time, a series of samples of crude oil from the Algerian Sahara has been tested and the results expressed in terms of τ=f(γ) have demonstrated their Newtonian character for the temperature included in [20°C, 70°C]. In second time and at T=20°C, the oil-water emulsions (30%, 50% and 70%) by volume of water), thermodynamically stable, have demonstrated a non-Newtonian rheological behavior that is to say, Herschel-Bulkley and Bingham types. For each type of crude oil-water emulsion, the rheological parameters are calculated by numerical treatment of results.
Abstract: This article presents a detailed analysis and comparative
performance evaluation of model reference adaptive control systems.
In contrast to classical control theory, adaptive control methods allow
to deal with time-variant processes. Inspired by the works [1] and
[2], two methods based on the MIT rule and Lyapunov rule are
applied to a linear first order system. The system is simulated and
it is investigated how changes to the adaptation gain affect the
system performance. Furthermore, variations in the reference model
parameters, that is changing the desired closed-loop behaviour are
examinded.
Abstract: The choice of finite element to use in order to predict
nonlinear static or dynamic response of complex structures becomes
an important factor. Then, the main goal of this research work is to
focus a study on the effect of the in-plane rotational degrees of
freedom in linear and geometrically non linear static and dynamic
analysis of thin shell structures by flat shell finite elements. In this
purpose: First, simple triangular and quadrilateral flat shell finite
elements are implemented in an incremental formulation based on the
updated lagrangian corotational description for geometrically
nonlinear analysis. The triangular element is a combination of DKT
and CST elements, while the quadrilateral is a combination of DKQ
and the bilinear quadrilateral membrane element. In both elements,
the sixth degree of freedom is handled via introducing fictitious
stiffness. Secondly, in the same code, the sixth degrees of freedom in
these elements is handled differently where the in-plane rotational
d.o.f is considered as an effective d.o.f in the in-plane filed
interpolation. Our goal is to compare resulting shell elements. Third,
the analysis is enlarged to dynamic linear analysis by direct
integration using Newmark-s implicit method. Finally, the linear
dynamic analysis is extended to geometrically nonlinear dynamic
analysis where Newmark-s method is used to integrate equations of
motion and the Newton-Raphson method is employed for iterating
within each time step increment until equilibrium is achieved. The
obtained results demonstrate the effectiveness and robustness of the
interpolation of the in-plane rotational d.o.f. and present deficiencies
of using fictitious stiffness in dynamic linear and nonlinear analysis.
Abstract: The modified Claus process is the major technology
for the recovery of elemental sulfur from hydrogen sulfide. The
chemical reactions that can occur in the reaction furnace are
numerous and many byproducts such as carbon disulfide and carbon
carbonyl sulfide are produced. These compounds can often contribute
from 20 to 50% of the pollutants and therefore, should be hydrolyzed
in the catalytic converter. The inlet temperature of the first catalytic
reactor should be maintained over than 250 °C, to hydrolyze COS
and CS2. In this paper, the various configurations for the first
converter reheating of sulfur recovery unit are investigated. As a
result, the performance of each method is presented for a typical
clause unit. The results show that the hot gas method seems to be
better than the other methods.
Abstract: In this paper we used data mining techniques to
identify outlier patients who are using large amount of drugs over a
long period of time. Any healthcare or health insurance system
should deal with the quantities of drugs utilized by chronic diseases
patients. In Kingdom of Bahrain, about 20% of health budget is spent
on medications. For the managers of healthcare systems, there is no
enough information about the ways of drug utilization by chronic
diseases patients, is there any misuse or is there outliers patients. In
this work, which has been done in cooperation with information
department in the Bahrain Defence Force hospital; we select the data
for Cardiac patients in the period starting from 1/1/2008 to
December 31/12/2008 to be the data for the model in this paper. We
used three techniques for finding the drug utilization for cardiac
patients. First we applied a clustering technique, followed by
measuring of clustering validity, and finally we applied a decision
tree as classification algorithm. The clustering results is divided into
three clusters according to the drug utilization, for 1603 patients, who
received 15,806 prescriptions during this period can be partitioned
into three groups, where 23 patients (2.59%) who received 1316
prescriptions (8.32%) are classified to be outliers. The classification
algorithm shows that the use of average drug utilization and the age,
and the gender of the patient can be considered to be the main
predictive factors in the induced model.
Abstract: The incidence of oral cancer in Taiwan increased year
by year. It replaced the nasopharyngeal as the top incurrence among
head and neck cancers since 1994. Early examination and earlier
identification for earlier treatment is the most effective medical
treatment for these cancers. Although the government fully subsidized
the expenses with tremendous promotion program for oral cancer
screening, the citizen-s participation remained low. Purpose of this
study is to understand the factors affecting the citizens- behavior
intensions of taking an oral cancer screening. Based on the Theory of
Planned Behavior, this study adopted four distinctive variables in
explaining the captioned behavior intentions.700 questionnaires were
dispatched with 500 valid responses or 71.4% returned by the citizens
with an age 30 or above from the eastern counties of Taiwan. Test
results has shown that attitude toward, subjective norms of, and
perceived behavioral control over the oral cancer screening varied
from some demographic factors to another. The study proofed that
attitude toward, subjective norms of, and perceived behavioral control
over the oral cancer screening had positive impacts on the
corresponding behavior intention. The test concluded that the theory
of planned behavior was appropriate as a theoretical framework in
explaining the influencing factors of intentions of taking oral cancer
screening. This study suggested the healthcare professional should
provide high accessibility of screening services other than just
delivering knowledge on oral cancer to promote the citizens-
intentions of taking the captioned screening. This research also
provided a practical implication to the healthcare professionals when
formulating and implementing promotion instruments for lifting the
screening rate of oral cancer.
Abstract: The field of biomedical materials plays an imperative
requisite and a critical role in manufacturing a variety of biological
artificial replacements in a modern world. Recently, titanium (Ti)
materials are being used as biomaterials because of their superior
corrosion resistance and tremendous specific strength, free- allergic
problems and the greatest biocompatibility compared to other
competing biomaterials such as stainless steel, Co-Cr alloys,
ceramics, polymers, and composite materials. However, regardless of
these excellent performance properties, Implantable Ti materials have
poor shear strength and wear resistance which limited their
applications as biomaterials. Even though the wear properties of Ti
alloys has revealed some improvements, the crucial effectiveness of
biomedical Ti alloys as wear components requires a comprehensive
deep understanding of the wear reasons, mechanisms, and techniques
that can be used to improve wear behavior. This review examines
current information on the effect of thermal and thermomechanical
processing of implantable Ti materials on the long-term prosthetic
requirement which related with wear behavior. This paper focuses
mainly on the evolution, evaluation and development of effective
microstructural features that can improve wear properties of bio
grade Ti materials using thermal and thermomechanical treatments.
Abstract: Wood as a natural renewable material is vulnerable to
degradation by microorganisms and susceptible to change in
dimension by water. In order to effectively improve the durability of
wood, an active reagent, maleic anhydride (Man) was selected for
wood modification. Man was first dissolved into a solvent, and then
penetrated into wood porous structure under a vacuum/pressure
condition. After a final catalyst-thermal treatment, wood modification
was finished. The test results indicate that acetone is a good solvent for
transporting Man into wood matrix. SEM observation proved that
wood samples treated by Man kept a good cellular structure, indicating
a well penetration of Man into wood cell walls. FTIR analysis
suggested that Man reacted with hydroxyl groups on wood cell walls
by its ring-ether group, resulting in reduction of amount of hydroxyl
groups and resultant good dimensional stability as well as fine decay
resistance. Consequently, Man modifying wood to improve its
durability is an effective method.
Abstract: Accurately predicting non-peak traffic is crucial to
daily traffic for all forecasting models. In the paper, least squares
support vector machines (LS-SVMs) are investigated to solve such a
practical problem. It is the first time to apply the approach and analyze
the forecast performance in the domain. For comparison purpose, two
parametric and two non-parametric techniques are selected because of
their effectiveness proved in past research. Having good
generalization ability and guaranteeing global minima, LS-SVMs
perform better than the others. Providing sufficient improvement in
stability and robustness reveals that the approach is practically
promising.
Abstract: Cognitive Science appeared about 40 years ago,
subsequent to the challenge of the Artificial Intelligence, as common
territory for several scientific disciplines such as: IT, mathematics,
psychology, neurology, philosophy, sociology, and linguistics. The
new born science was justified by the complexity of the problems
related to the human knowledge on one hand, and on the other by the
fact that none of the above mentioned sciences could explain alone
the mental phenomena. Based on the data supplied by the
experimental sciences such as psychology or neurology, models of
the human mind operation are built in the cognition science. These
models are implemented in computer programs and/or electronic
circuits (specific to the artificial intelligence) – cognitive systems –
whose competences and performances are compared to the human
ones, leading to the psychology and neurology data reinterpretation,
respectively to the construction of new models. During these
processes if psychology provides the experimental basis, philosophy
and mathematics provides the abstraction level utterly necessary for
the intermission of the mentioned sciences.
The ongoing general problematic of the cognitive approach
provides two important types of approach: the computational one,
starting from the idea that the mental phenomenon can be reduced to
1 and 0 type calculus operations, and the connection one that
considers the thinking products as being a result of the interaction
between all the composing (included) systems. In the field of
psychology measurements in the computational register use classical
inquiries and psychometrical tests, generally based on calculus
methods. Deeming things from both sides that are representing the
cognitive science, we can notice a gap in psychological product
measurement possibilities, regarded from the connectionist
perspective, that requires the unitary understanding of the quality –
quantity whole. In such approach measurement by calculus proves to
be inefficient. Our researches, deployed for longer than 20 years,
lead to the conclusion that measuring by forms properly fits to the
connectionism laws and principles.
Abstract: This work has been carried out in order to provide an understanding of the physical behaviors of the flow variation of pressure and temperature in a vortex tube. A computational fluid dynamics model is used to predict the flow fields and the associated temperature separation within a Ranque–Hilsch vortex tube. The CFD model is a steady axisymmetric model (with swirl) that utilizes the standard k-ε turbulence model. The second–order numerical schemes, was used to carry out all the computations. Vortex tube with a circumferential inlet stream and an axial (cold) outlet stream and a circumferential (hot) outlet stream was considered. Performance curves (temperature separation versus cold outlet mass fraction) were obtained for a specific vortex tube with a given inlet mass flow rate. Simulations have been carried out for varying amounts of cold outlet mass flow rates. The model results have a good agreement with experimental data.
Abstract: Designing, implementing, and debugging concurrency
control algorithms in a real system is a complex, tedious, and errorprone
process. Further, understanding concurrency control
algorithms and distributed computations is itself a difficult task.
Visualization can help with both of these problems. Thus, we have
developed an exploratory environment in which people can prototype
and test various versions of concurrency control algorithms, study
and debug distributed computations, and view performance statistics
of distributed systems. In this paper, we describe the exploratory
environment and show how it can be used to explore concurrency
control algorithms for the interactive steering of distributed
computations.
Abstract: Majority of pepper farmers in Malaysia are using the
open-sun method for drying the pepper berries. This method is time
consuming and exposed the berries to rain and contamination. A
maintenance-friendly and properly enclosed dryer is therefore
desired. A dryer design with a solar collector and a chimney was
studied and adapted to suit the needs of small-scale pepper farmers in
Malaysia. The dryer will provide an environment with an optimum
operating temperature meant for drying pepper berries. The dryer
model was evaluated by using commercially available computational
fluid dynamic (CFD) software in order to understand the heat and
mass transfer inside the dryer. Natural convection was the only mode
of heat transportation considered in this study as in accordance to the
idea of having a simple and maintenance-friendly design. To
accommodate the effect of low buoyancy found in natural convection
driers, a biomass burner was integrated into the solar dryer design.
Abstract: If organizations like Mellat Bank want to identify its
customer market completely to reach its specified goals, it can
segment the market to offer the product package to the right segment.
Our objective is to offer a segmentation model for Iran banking
market in Mellat bank view. The methodology of this project is
combined by “segmentation on the basis of four part-quality
variables" and “segmentation on the basis of different in means".
Required data are gathered from E-Systems and researcher personal
observation. Finally, the research offers the organization that at first
step form a four dimensional matrix with 756 segments using four
variables named value-based, behavioral, activity style, and activity
level, and at the second step calculate the means of profit for every
cell of matrix in two distinguished work level (levels α1:normal
condition and α2: high pressure condition) and compare the segments
by checking two conditions that are 1- homogeneity every segment
with its sub segment and 2- heterogeneity with other segments, and
so it can do the necessary segmentation process. After all, the last
offer (more explained by an operational example and feedback
algorithm) is to test and update the model because of dynamic
environment, technology, and banking system.
Abstract: Data mining incorporates a group of statistical
methods used to analyze a set of information, or a data set. It operates
with models and algorithms, which are powerful tools with the great
potential. They can help people to understand the patterns in certain
chunk of information so it is obvious that the data mining tools have
a wide area of applications. For example in the theoretical chemistry
data mining tools can be used to predict moleculeproperties or
improve computer-assisted drug design. Classification analysis is one
of the major data mining methodologies. The aim of thecontribution
is to create a classification model, which would be able to deal with a
huge data set with high accuracy. For this purpose logistic regression,
Bayesian logistic regression and random forest models were built
using R software. TheBayesian logistic regression in Latent GOLD
software was created as well. These classification methods belong to
supervised learning methods.
It was necessary to reduce data matrix dimension before construct
models and thus the factor analysis (FA) was used. Those models
were applied to predict the biological activity of molecules, potential
new drug candidates.
Abstract: A new genetic algorithm, termed the 'optimum individual monogenetic genetic algorithm' (OIMGA), is presented whose properties have been deliberately designed to be well suited to hardware implementation. Specific design criteria were to ensure fast access to the individuals in the population, to keep the required silicon area for hardware implementation to a minimum and to incorporate flexibility in the structure for the targeting of a range of applications. The first two criteria are met by retaining only the current optimum individual, thereby guaranteeing a small memory requirement that can easily be stored in fast on-chip memory. Also, OIMGA can be easily reconfigured to allow the investigation of problems that normally warrant either large GA populations or individuals many genes in length. Local convergence is achieved in OIMGA by retaining elite individuals, while population diversity is ensured by continually searching for the best individuals in fresh regions of the search space. The results given in this paper demonstrate that both the performance of OIMGA and its convergence time are superior to those of a range of existing hardware GA implementations.