Abstract: Little research has examined working memory
capacity (WMC) in signed language interpreters and deaf signers.
This paper presents the findings of a study that investigated WMC in
professional Australian Sign Language (Auslan)/English interpreters
and deaf signers. Thirty-one professional Auslan/English interpreters
(14 hearing native signers and 17 hearing non-native signers)
completed an English listening span task and then an Auslan working
memory span task, which tested their English WMC and their Auslan
WMC, respectively. Moreover, 26 deaf signers (6 deaf native signers
and 20 deaf non-native signers) completed the Auslan working
memory span task. The results revealed a non-significant difference
between the hearing native signers and the hearing non-native signers
in their English WMC, and a non-significant difference between the
hearing native signers and the hearing non-native signers in their
Auslan WMC. Moreover, the results yielded a non-significant
difference between the hearing native signers- English WMC and
their Auslan WMC, and a non-significant difference between the
hearing non-native signers- English WMC and their Auslan WMC.
Furthermore, a non-significant difference was found between the deaf
native signers and the deaf non-native signers in their Auslan WMC.
Abstract: A large amount of valuable information is available in
plain text clinical reports. New techniques and technologies are
applied to extract information from these reports. In this study, we
developed a domain based software system to transform 600
Otorhinolaryngology discharge notes to a structured form for
extracting clinical data from the discharge notes. In order to decrease
the system process time discharge notes were transformed into a data
table after preprocessing. Several word lists were constituted to
identify common section in the discharge notes, including patient
history, age, problems, and diagnosis etc. N-gram method was used
for discovering terms co-Occurrences within each section. Using this
method a dataset of concept candidates has been generated for the
validation step, and then Predictive Apriori algorithm for Association
Rule Mining (ARM) was applied to validate candidate concepts.
Abstract: Sweet cherries (Prunus avium L.) contain various
phenolic compounds which contribute to total antioxidant activity.
Total polyphenols, tannins, flavonoids and anthocyanins, and
antioxidant capacity in a fruits of a number of selected sweet cherry
genotypes were investigated. Total polyphenols content ranged from
4.12 to 8.34 mg gallic acid equivantents/g dry fruit weight and total
tannins content ranged from 0.19 to 1.95 mg gallic acid equivalent/g
dry fruit weight. Total flavonoids were within the range 0.42-1.56 mg
of rutin equivalents/g dry fruit weight and total anthocyanins content
were between 0.35 and 0.69 mg cyanidin 3-glucoside equivalent/ g
dry fruit weight. Although sweet cherry fruits are a significant source
of different phenolic compounds, antioxidant activity of sweet
cherries is not related only with the total polyphenolics, flavonoids or
anthocyanins.
Abstract: Microarray data profiles gene expression on a whole
genome scale, therefore, it provides a good way to study associations
between gene expression and occurrence or progression of cancer.
More and more researchers realized that microarray data is helpful
to predict cancer sample. However, the high dimension of gene
expressions is much larger than the sample size, which makes this
task very difficult. Therefore, how to identify the significant genes
causing cancer becomes emergency and also a hot and hard research
topic. Many feature selection algorithms have been proposed in
the past focusing on improving cancer predictive accuracy at the
expense of ignoring the correlations between the features. In this
work, a novel framework (named by SGS) is presented for stable gene
selection and efficient cancer prediction . The proposed framework
first performs clustering algorithm to find the gene groups where
genes in each group have higher correlation coefficient, and then
selects the significant genes in each group with Bayesian Lasso and
important gene groups with group Lasso, and finally builds prediction
model based on the shrinkage gene space with efficient classification
algorithm (such as, SVM, 1NN, Regression and etc.). Experiment
results on real world data show that the proposed framework often
outperforms the existing feature selection and prediction methods,
say SAM, IG and Lasso-type prediction model.
Abstract: Traditionally, VLSI implementations of spiking
neural nets have featured large neuron counts for fixed computations
or small exploratory, configurable nets. This paper presents the
system architecture of a large configurable neural net system
employing a dedicated mapping algorithm for projecting the targeted
biology-analog nets and dynamics onto the hardware with its
attendant constraints.
Abstract: The aim of this paper is to know the sociodemographic
and operational-financial determinants of the services
quality perceived by users of the national health services. Through
the use of an inquiry conducted by the Ministry of Health,
comprehending 16.936 interviews in 2006, we intend to find out if
there is any characteristic that determines the 2006 inquiry results.
With the revision of the literature we also want to know if the
operational-financial results have implications in hospitals users-
perception on the quality of the received services. In order to achieve
our main goals we will make use of the regression analysis to find out
the possible dimensions that determine those results.
Abstract: In this paper, we address the problem of reducing the
switching activity (SA) in on-chip buses through the use of a bus
binding technique in high-level synthesis. While many binding
techniques to reduce the SA exist, we present yet another technique for
further reducing the switching activity. Our proposed method
combines bus binding and data sequence reordering to explore a wider
solution space. The problem is formulated as a multiple traveling
salesman problem and solved using simulated annealing technique.
The experimental results revealed that a binding solution obtained
with the proposed method reduces 5.6-27.2% (18.0% on average) and
2.6-12.7% (6.8% on average) of the switching activity when compared
with conventional binding-only and hybrid binding-encoding
methods, respectively.
Abstract: The reliable results of an insulated oval duct
considering heat radiation are obtained basing on accurate oval
perimeter obtained by integral method as well as one-dimensional
Plane Wedge Thermal Resistance (PWTR) model. This is an extension
study of former paper of insulated oval duct neglecting heat radiation.
It is found that in the practical situations with long-short-axes ratio a/b
4.5% while t/R2
Abstract: Recently, grid computing has been widely focused on
the science, industry, and business fields, which are required a vast
amount of computing. Grid computing is to provide the environment
that many nodes (i.e., many computers) are connected with each
other through a local/global network and it is available for many
users. In the environment, to achieve data processing among nodes
for any applications, each node executes mutual authentication by
using certificates which published from the Certificate Authority
(for short, CA). However, if a failure or fault has occurred in the
CA, any new certificates cannot be published from the CA. As
a result, a new node cannot participate in the gird environment.
In this paper, an off-the-shelf scheme for dependable grid systems
using virtualization techniques is proposed and its implementation is
verified. The proposed approach using the virtualization techniques
is to restart an application, e.g., the CA, if it has failed. The system
can tolerate a failure or fault if it has occurred in the CA. Since
the proposed scheme is implemented at the application level easily,
the cost of its implementation by the system builder hardly takes
compared it with other methods. Simulation results show that the
CA in the system can recover from its failure or fault.
Abstract: The curriculum of the primary school science course was redesigned on the basis of constructivism in 2005-2006 academic years, in Turkey. In this context, the name of this course has been changed as “Science and Technology"; and both content and course books, students workbooks for this course have been redesigned in light of constructivism. The aim of this study is to determine whether the Science and Technology course books and student work books for primary school 5th grade are appropriate for the constructivism by evaluating them in terms of the fundamental principles of constructivism. In this study, out of qualitative research methods, documentation technique (i.e. document analysis) is applied; while selecting samples, criterion-sampling is used out of purposeful sampling techniques. When the Science and Technology course book and workbook for the 5th grade in primary education are examined, it is seen that both books complete each other in certain areas. Consequently, it can be claimed that in spite of some inadequate and missing points in the course book and workbook of the primary school Science and Technology course for the 5th grade students, these books are attempted to be designed in terms of the principles of constructivism. To overcome the inadequacies in the books, it can be suggested to redesign them. In addition to them, not to ignore the technology dimension of the course, the activities that encourage the students to prepare projects using technology cycle should be included.
Abstract: An attempt in this paper proposes a re-modification to
the minimum moment approach of resource leveling which is a modified minimum moment approach to the traditional method by
Harris. The method is based on critical path method. The new approach suggests the difference between the methods in the
selection criteria of activity which needs to be shifted for leveling resource histogram. In traditional method, the improvement factor
found first to select the activity for each possible day of shifting. In
modified method maximum value of the product of Resources Rate
and Free Float was found first and improvement factor is then
calculated for that activity which needs to be shifted. In the proposed
method the activity to be selected first for shifting is based on the largest value of resource rate. The process is repeated for all the
remaining activities for possible shifting to get updated histogram.
The proposed method significantly reduces the number of iterations
and is easier for manual computations.
Abstract: Nuclear matrix protein 22 (NMP22) is a FDA approved
biomarker for bladder cancer. The objective of this study is to develop
a simple NMP22 immumosensor (NMP22-IMS) for accurate
measurement of NMP22. The NMP22-IMS was constructed with
NMP22 antibody immobilized on screen-printed carbon electrodes.
The construction procedures and antibody immobilization are simple.
Results showed that the NMP22-IMS has an excellent (r2³0.95)
response range (20 – 100 ng/mL). In conclusion, a simple and reliable
NMP22-IMS was developed, capable of precisely determining urine
NMP22 level.
Abstract: Ireland developed a National Strategy 2030 that
argued for the creation of a new form of higher education institution,
a Technological University. The research reported here reviews the
first stage of this partnership development. The study found that
national policy can create system capacity and change, but that
individual partners may have more to gain or lose in collaborating.
When presented as a zero-sum activity, fear among partners is high.
The level of knowledge and networking within the higher education
system possessed by each partner contributed to decisions to
participate or not in a joint proposal for collaboration. Greater
success resulted when there were gains for all partners. This research
concludes that policy mandates can provide motivation to
collaborate, but that the partnership needs to be built more on shared
values versus coercion by mandates.
Abstract: Wireless channels are characterized by more serious
bursty and location-dependent errors. Many packet scheduling
algorithms have been proposed for wireless networks to guarantee
fairness and delay bounds. However, most existing schemes do not
consider the difference of traffic natures among packet flows. This
will cause the delay-weight coupling problem. In particular, serious
queuing delays may be incurred for real-time flows. In this paper, it
is proposed a scheduling algorithm that takes traffic types of flows
into consideration when scheduling packets and also it is provided
scheduling flexibility by trading off video quality to meet the
playback deadline.
Abstract: The protection of the contents of digital products is
referred to as content authentication. In some applications, to be able
to authenticate a digital product could be extremely essential. For
example, if a digital product is used as a piece of evidence in the
court, its integrity could mean life or death of the accused. Generally,
the problem of content authentication can be solved using semifragile
digital watermarking techniques. Recently many authors have
proposed Computer Generated Hologram Watermarking (CGHWatermarking)
techniques. Starting from these studies, in this paper
a semi-fragile Computer Generated Hologram coding technique is
proposed, which is able to detect malicious tampering while
tolerating some incidental distortions. The proposed technique uses
as watermark an encrypted image, and it is well suitable for digital
image authentication.
Abstract: The feature of HIV genome is in a wide range because
of it is highly heterogeneous. Hence, the infection ability of the virus changes related with different chemokine receptors. From this point,
R5 and X4 HIV viruses use CCR5 and CXCR5 coreceptors respectively while R5X4 viruses can utilize both coreceptors. Recently, in Bioinformatics, R5X4 viruses have been studied to
classify by using the coreceptors of HIV genome.
The aim of this study is to develop the optimal Multilayer
Perceptron (MLP) for high classification accuracy of HIV sub-type viruses. To accomplish this purpose, the unit number in hidden layer
was incremented one by one, from one to a particular number. The statistical data of R5X4, R5 and X4 viruses was preprocessed by the
signal processing methods. Accessible residues of these virus sequences were extracted and modeled by Auto-Regressive Model
(AR) due to the dimension of residues is large and different from each other. Finally the pre-processed dataset was used to evolve MLP with various number of hidden units to determine R5X4
viruses. Furthermore, ROC analysis was used to figure out the optimal MLP structure.
Abstract: The objectives were to identify cyanide-degrading
bacteria and study cyanide removal efficiency. Agrobacterium
tumefaciens SUTS 1 was isolated. This is a new strain of
microorganisms for cyanide degradation. The maximum growth rate
of SUTS 1 obtained 4.7 × 108 CFU/ml within 4 days. The cyanide
removal efficiency was studied at 25, 50, and 150 mg/L cyanide. The
residual cyanide, ammonia, nitrate, nitrite, pH, and cell counts were
analyzed. At 25 and 50 mg/L cyanide, SUTS 1 obtained similar
removal efficiency approximately 87.50%. At 150 mg/L cyanide,
SUTS 1 enhanced the cyanide removal efficiency up to 97.90%. Cell
counts of SUTS 1 increased when the cyanide concentration was set
at lower. The ammonia increased when the removal efficiency
increased. The nitrate increased when the ammonia decreased but the
nitrite did not detect in all experiments. pH values also increased
when the cyanide concentrations were set at higher.
Abstract: Laser beam forming is a novel technique developed for the joining of metallic components. In this study, an overview of the laser beam forming process, areas of application, the basic mechanisms of the laser beam forming process, some recent research
studies and the need to focus more research effort on improving the
laser-material interaction of laser beam forming of titanium and its
alloys are presented.
Abstract: Systems Analysis and Design is a key subject in
Information Technology courses, but students do not find it easy to
cope with, since it is not “precise" like programming and not exact
like Mathematics. It is a subject working with many concepts,
modeling ideas into visual representations and then translating the
pictures into a real life system. To complicate matters users who are
not necessarily familiar with computers need to give their inputs to
ensure that they get the system the need. Systems Analysis and
Design also covers two fields, namely Analysis, focusing on the
analysis of the existing system and Design, focusing on the design of
the new system. To be able to test the analysis and design of a
system, it is necessary to develop a system or at least a prototype of
the system to test the validity of the analysis and design. The skills
necessary in each aspect differs vastly. Project Management Skills,
Database Knowledge and Object Oriented Principles are all
necessary. In the context of a developing country where students
enter tertiary education underprepared and the digital divide is alive
and well, students need to be motivated to learn the necessary skills,
get an opportunity to test it in a “live" but protected environment –
within the framework of a university. The purpose of this article is to
improve the learning experience in Systems Analysis and Design
through reviewing the underlying teaching principles used, the
teaching tools implemented, the observations made and the
reflections that will influence future developments in Systems
Analysis and Design. Action research principles allows the focus to
be on a few problematic aspects during a particular semester.
Abstract: In this paper we address the issue of classifying the fluorescent intensity of a sample in Indirect Immuno-Fluorescence (IIF). Since IIF is a subjective, semi-quantitative test in its very nature, we discuss a strategy to reliably label the image data set by using the diagnoses performed by different physicians. Then, we discuss image pre-processing, feature extraction and selection. Finally, we propose two ANN-based classifiers that can separate intrinsically dubious samples and whose error tolerance can be flexibly set. Measured performance shows error rates less than 1%, which candidates the method to be used in daily medical practice either to perform pre-selection of cases to be examined, or to act as a second reader.