Abstract: This paper represents an investigation on how exploiting multiple transmit antennas by OFDM based wireless LAN subscribers can mitigate physical layer error rate. Then by comparing the Wireless LANs that utilize spatial diversity techniques with the conventional ones it will reveal how PHY and TCP throughputs behaviors are ameliorated. In the next step it will assess the same issues based on a cellular context operation which is mainly introduced as an innovated solution that beside a multi cell operation scenario benefits spatio-temporal signaling schemes as well. Presented simulations will shed light on the improved performance of the wide range and high quality wireless LAN services provided by the proposed approach.
Abstract: Building life cycle will never be excused from the existence of defects and deterioration. They are common problems in building, existed in newly build or in aged building. Buildings constructed from wood are indeed affected by its agent and serious defects and damages can reduce values to a building. In repair works, it is important to identify the causes and repair techniques that best suites with the condition. This paper reviews the conservation of traditional timber mosque in Malaysia comprises the concept, principles and approaches of mosque conservation in general. As in conservation practice, wood in historic building can be conserved by using various restoration and conservation techniques which this can be grouped as Fully and Partial Replacement, Mechanical Reinforcement, Consolidation by Impregnation and Reinforcement, Removing Paint and also Preservation of Wood and Control Insect Invasion, as to prolong and extended the function of a timber in a building. It resulted that the common techniques adopted in timber mosque conservation are from the conventional ways and the understanding of the repair technique requires the use of only preserve wood to prevent the future immature defects.
Abstract: Proteins or genes that have similar sequences are likely to perform the same function. One of the most widely used techniques for sequence comparison is sequence alignment. Sequence alignment allows mismatches and insertion/deletion, which represents biological mutations. Sequence alignment is usually performed only on two sequences. Multiple sequence alignment, is a natural extension of two-sequence alignment. In multiple sequence alignment, the emphasis is to find optimal alignment for a group of sequences. Several applicable techniques were observed in this research, from traditional method such as dynamic programming to the extend of widely used stochastic optimization method such as Genetic Algorithms (GAs) and Simulated Annealing. A framework with combination of Genetic Algorithm and Simulated Annealing is presented to solve Multiple Sequence Alignment problem. The Genetic Algorithm phase will try to find new region of solution while Simulated Annealing can be considered as an alignment improver for any near optimal solution produced by GAs.
Abstract: This research contribution propels the idea of collaborating environment for the execution of student satellite projects in the backdrop of project management principles. The recent past has witnessed a technological shift in the aerospace industry from the big satellite projects to the small spacecrafts especially for the earth observation and communication purposes. This vibrant shift has vitalized the academia and industry to share their resources and to create a win-win paradigm of mutual success and technological development along with the human resource development in the field of aerospace. Small student satellites are the latest jargon of academia and more than 100 CUBESAT projects have been executed successfully all over the globe and many new student satellite projects are in the development phase. The small satellite project management requires the application of specific knowledge, skills, tools and techniques to achieve the defined mission requirements. The Authors have presented the detailed outline for the project management of student satellites and presented the role of industry to collaborate with the academia to get the optimized results in academic environment.
Abstract: This paper focuses on operational risk measurement
techniques and on economic capital estimation methods. A data
sample of operational losses provided by an anonymous Central
European bank is analyzed using several approaches. Loss
Distribution Approach and scenario analysis method are considered.
Custom plausible loss events defined in a particular scenario are
merged with the original data sample and their impact on capital
estimates and on the financial institution is evaluated. Two main
questions are assessed – What is the most appropriate statistical
method to measure and model operational loss data distribution? and
What is the impact of hypothetical plausible events on the financial
institution? The g&h distribution was evaluated to be the most
suitable one for operational risk modeling. The method based on the
combination of historical loss events modeling and scenario analysis
provides reasonable capital estimates and allows for the measurement
of the impact of extreme events on banking operations.
Abstract: Speckled images arise when coherent microwave,
optical, and acoustic imaging techniques are used to image an object, surface or scene. Examples of coherent imaging systems include synthetic aperture radar, laser imaging systems, imaging sonar
systems, and medical ultrasound systems. Speckle noise is a form of object or target induced noise that results when the surface of the object is Rayleigh rough compared to the wavelength of the illuminating radiation. Detection and estimation in images corrupted
by speckle noise is complicated by the nature of the noise and is not
as straightforward as detection and estimation in additive noise. In
this work, we derive stochastic models for speckle noise, with an emphasis on speckle as it arises in medical ultrasound images. The
motivation for this work is the problem of segmentation and tissue classification using ultrasound imaging. Modeling of speckle in this
context involves partially developed speckle model where an underlying Poisson point process modulates a Gram-Charlier series
of Laguerre weighted exponential functions, resulting in a doubly
stochastic filtered Poisson point process. The statistical distribution of partially developed speckle is derived in a closed canonical form.
It is observed that as the mean number of scatterers in a resolution cell is increased, the probability density function approaches an
exponential distribution. This is consistent with fully developed speckle noise as demonstrated by the Central Limit theorem.
Abstract: With the hardware technology advancing, the cost of
storing is decreasing. Thus there is an urgent need for new techniques
and tools that can intelligently and automatically assist us in
transferring this data into useful knowledge. Different techniques of
data mining are developed which are helpful for handling these large
size databases [7]. Data mining is also finding its role in the field of
biotechnology. Pedigree means the associated ancestry of a crop
variety. Genetic diversity is the variation in the genetic composition
of individuals within or among species. Genetic diversity depends
upon the pedigree information of the varieties. Parents at lower
hierarchic levels have more weightage for predicting genetic
diversity as compared to the upper hierarchic levels. The weightage
decreases as the level increases. For crossbreeding, the two varieties
should be more and more genetically diverse so as to incorporate the
useful characters of the two varieties in the newly developed variety.
This paper discusses the searching and analyzing of different possible
pairs of varieties selected on the basis of morphological characters,
Climatic conditions and Nutrients so as to obtain the most optimal
pair that can produce the required crossbreed variety. An algorithm
was developed to determine the genetic diversity between the
selected wheat varieties. Cluster analysis technique is used for
retrieving the results.
Abstract: Environmental decision making, particularly about
hazardous waste management, is inherently exposed to a high
potential conflict, principally because of the trade-off between sociopolitical,
environmental, health and economic factors. The need to
plan complex contexts has led to an increasing request for decision
analytic techniques as support for the decision process. In this work,
alternative systems of asbestos-containing waste management
(ACW) in Puglia (Southern Italy) were explored by a multi-criteria
decision analysis. In particular, through Analytic Hierarchy Process
five alternatives management have been compared and ranked
according to their performance and efficiency, taking into account
environmental, health and socio-economic aspects. A separated
valuation has been performed for different temporal scale. For short
period results showed a narrow deviation between the disposal
alternatives “mono-material landfill in public quarry" and “dedicate
cells in existing landfill", with the best performance of the first one.
While for long period “treatment plant to eliminate hazard from
asbestos-containing waste" was prevalent, although high energy
demand required to achieve the change of crystalline structure. A
comparison with results from a participative approach in valuation
process might be considered as future development of method
application to ACW management.
Abstract: This paper presents the automated methods employed
for extracting craniofacial landmarks in white light images as part of
a registration framework designed to support three neurosurgical
procedures. The intraoperative space is characterised by white light
stereo imaging while the preoperative plan is performed on CT scans.
The registration aims at aligning these two modalities to provide a
calibrated environment to enable image-guided solutions. The
neurosurgical procedures can then be carried out by mapping the
entry and target points from CT space onto the patient-s space. The
registration basis adopted consists of natural landmarks (eye corner
and ear tragus). A 5mm accuracy is deemed sufficient for these three
procedures and the validity of the selected registration basis in
achieving this accuracy has been assessed by simulation studies. The
registration protocol is briefly described, followed by a presentation
of the automated techniques developed for the extraction of the
craniofacial features and results obtained from tests on the AR and
FERET databases. Since the three targeted neurosurgical procedures
are routinely used for head injury management, the effect of
bruised/swollen faces on the automated algorithms is assessed. A
user-interactive method is proposed to deal with such unpredictable
circumstances.
Abstract: This article combines two techniques: data
envelopment analysis (DEA) and Factor analysis (FA) to data
reduction in decision making units (DMU). Data envelopment
analysis (DEA), a popular linear programming technique is useful to
rate comparatively operational efficiency of decision making units
(DMU) based on their deterministic (not necessarily stochastic)
input–output data and factor analysis techniques, have been proposed
as data reduction and classification technique, which can be applied
in data envelopment analysis (DEA) technique for reduction input –
output data. Numerical results reveal that the new approach shows a
good consistency in ranking with DEA.
Abstract: Design of a constant chord propeller is presented in
this paper in order to reduce propeller-s design procedure-s costs. The
design process was based on Lock and Goldstein-s techniques of
propeller design and analysis. In order to calculate optimum chord of
propeller, chord of a referential element is generalized as whole
blades chord. The design outcome which named CS-X-1 is modeled
& analyzed by CFD methods using K-ε: R.N.G turbulence model.
Convergence of results of two codes proved that outcome results of
design process are reliable. Design result is a two-blade propeller
with a total diameter of 1.1 meter, radial velocity of 3000 R.P.M,
efficiency above .75 and power coefficient near 1.05.
Abstract: Recently, genetic algorithms (GA) and particle swarm optimization (PSO) technique have attracted considerable attention among various modern heuristic optimization techniques. Since the two approaches are supposed to find a solution to a given objective function but employ different strategies and computational effort, it is appropriate to compare their performance. This paper presents the application and performance comparison of PSO and GA optimization techniques, for Thyristor Controlled Series Compensator (TCSC)-based controller design. The design objective is to enhance the power system stability. The design problem of the FACTS-based controller is formulated as an optimization problem and both the PSO and GA optimization techniques are employed to search for optimal controller parameters. The performance of both optimization techniques in terms of computational time and convergence rate is compared. Further, the optimized controllers are tested on a weakly connected power system subjected to different disturbances, and their performance is compared with the conventional power system stabilizer (CPSS). The eigenvalue analysis and non-linear simulation results are presented and compared to show the effectiveness of both the techniques in designing a TCSC-based controller, to enhance power system stability.
Abstract: Academics and researchers are interested in the effects of social media on college students, with a specific focus on the most popular social media website; Facebook. Previous studied have found contradictory result on the relationship between Facebook usage and the student engagement with positive, detrimental and no significant relationships. However, these studies were limited to western higher education system. This paper fills a gap in the literature by using a sample (300) of Sri Lankan management undergraduates to examine the relationship between Facebook usage and student engagement. Student engagement was measured 35 item scale based on the National Survey of Student Engagement and Facebook usage by Facebook intensity scale. Descriptive statistics, path analysis and structural equation modeling were applied as statistical tools and techniques. Results indicate that student engagement scale was significantly negatively related with the Facebook usage with the influence from student engagement on Facebook usage.
Abstract: The passive electrical properties of a tissue depends
on the intrinsic constituents and its structure, therefore by measuring
the complex electrical impedance of the tissue it might be possible to
obtain indicators of the tissue state or physiological activity [1].
Complete bio-impedance information relative to physiology and
pathology of a human body and functional states of the body tissue or
organs can be extracted by using a technique containing a fourelectrode
measurement setup. This work presents the estimation
measurement setup based on the four-electrode technique. First, the
complex impedance is estimated by three different estimation
techniques: Fourier, Sine Correlation and Digital De-convolution and
then estimation errors for the magnitude, phase, reactance and
resistance are calculated and analyzed for different levels of
disturbances in the observations. The absolute values of relative
errors are plotted and the graphical performance of each technique is
compared.
Abstract: Predicting protein-protein interactions represent a key step in understanding proteins functions. This is due to the fact that proteins usually work in context of other proteins and rarely function alone. Machine learning techniques have been applied to predict protein-protein interactions. However, most of these techniques address this problem as a binary classification problem. Although it is easy to get a dataset of interacting proteins as positive examples, there are no experimentally confirmed non-interacting proteins to be considered as negative examples. Therefore, in this paper we solve this problem as a one-class classification problem using one-class support vector machines (SVM). Using only positive examples (interacting protein pairs) in training phase, the one-class SVM achieves accuracy of about 80%. These results imply that protein-protein interaction can be predicted using one-class classifier with comparable accuracy to the binary classifiers that use artificially constructed negative examples.
Abstract: Question answering (QA) aims at retrieving precise information from a large collection of documents. Most of the Question Answering systems composed of three main modules: question processing, document processing and answer processing. Question processing module plays an important role in QA systems to reformulate questions. Moreover answer processing module is an emerging topic in QA systems, where these systems are often required to rank and validate candidate answers. These techniques aiming at finding short and precise answers are often based on the semantic relations and co-occurrence keywords. This paper discussed about a new model for question answering which improved two main modules, question processing and answer processing which both affect on the evaluation of the system operations. There are two important components which are the bases of the question processing. First component is question classification that specifies types of question and answer. Second one is reformulation which converts the user's question into an understandable question by QA system in a specific domain. The objective of an Answer Validation task is thus to judge the correctness of an answer returned by a QA system, according to the text snippet given to support it. For validating answers we apply candidate answer filtering, candidate answer ranking and also it has a final validation section by user voting. Also this paper described new architecture of question and answer processing modules with modeling, implementing and evaluating the system. The system differs from most question answering systems in its answer validation model. This module makes it more suitable to find exact answer. Results show that, from total 50 asked questions, evaluation of the model, show 92% improving the decision of the system.
Abstract: Entrepreneurship has become an important and
extensively researched concept in business studies. Research on
foreign direct investment (FDI) has become widespread due to the
growth of FDI and its importance in globalization. Most
entrepreneurship studies examined the importance and influence of
entrepreneurial orientation in a micro-level context. On the other
hand, studies and research concerning FDI used statistical techniques
to analyze the effect, determinants, and motives of FDI on a
macroeconomic level, ignoring empirical studies on other noneconomic
determinants. In order to bridge the gap between the theory
and empirical evidence on FDI and the theory and research on
entrepreneurship, this study examines the impact of entrepreneurship
on inward foreign direct investment. The relationship between
entrepreneurship and foreign direct investment is investigated
through regression analysis of pooled time-series and cross-sectional
data. The results suggest that entrepreneurship has a significant effect
on FDI.
Abstract: In this work, we present a comparison between
different techniques of image compression. First, the image is
divided in blocks which are organized according to a certain scan.
Later, several compression techniques are applied, combined or
alone. Such techniques are: wavelets (Haar's basis), Karhunen-Loève
Transform, etc. Simulations show that the combined versions are the
best, with minor Mean Squared Error (MSE), and higher Peak Signal
to Noise Ratio (PSNR) and better image quality, even in the presence
of noise.
Abstract: Appeared toward 1986, the object-oriented databases
management systems had not known successes knew five years after
their birth. One of the major difficulties is the query optimization.
We propose in this paper a new approach that permits to enrich
techniques of query optimization existing in the object-oriented
databases. Seen success that knew the query optimization in the
relational model, our approach inspires itself of these optimization
techniques and enriched it so that they can support the new concepts
introduced by the object databases.
Abstract: The Kansei engineering is a technology which
converts human feelings into quantitative terms and helps designers
develop new products that meet customers- expectation. Standard
Kansei engineering procedure involves finding relationships between
human feelings and design elements of which many researchers have
found forward and backward relationship through various soft
computing techniques. In this paper, we proposed the framework of
Kansei engineering linking relationship not only between human
feelings and design elements, but also the whole part of product, by
constructing association rules. In this experiment, we obtain input
from emotion score that subjects rate when they see the whole part of
the product by applying semantic differentials. Then, association
rules are constructed to discover the combination of design element
which affects the human feeling. The results of our experiment
suggest the pattern of relationship of design elements according to
human feelings which can be derived from the whole part of product.