Abstract: the elastic scattering of protons, deuterons and 3He on 6Li at different incident energies have been analyzed in the framework of the optical model using ECIS88 as well as SPI GENOA codes. The potential parameters were extracted in the phenomenological treatment of measured by us angular distributions and literature data. A good agreement between theoretical and experimental differential cross sections was obtained in whole angular range. Parameters for real part of potential have been also calculated microscopically with singleand double-folding model for the p and d, 3He scattering, respectively, using DFPOT code. For best agreement with experiment the normalization factor N for the potential depth is obtained in the range of 0.7-0.9.
Abstract: Re-entrant scheduling is an important search problem
with many constraints in the flow shop. In the literature, a number of
approaches have been investigated from exact methods to
meta-heuristics. This paper presents a genetic algorithm that encodes
the problem as multi-level chromosomes to reflect the dependent
relationship of the re-entrant possibility and resource consumption.
The novel encoding way conserves the intact information of the data
and fastens the convergence to the near optimal solutions. To test the
effectiveness of the method, it has been applied to the
resource-constrained re-entrant flow shop scheduling problem.
Computational results show that the proposed GA performs better than
the simulated annealing algorithm in the measure of the makespan
Abstract: Automatic reusability appraisal could be helpful in
evaluating the quality of developed or developing reusable software
components and in identification of reusable components from
existing legacy systems; that can save cost of developing the software
from scratch. But the issue of how to identify reusable components
from existing systems has remained relatively unexplored. In this
paper, we have mentioned two-tier approach by studying the
structural attributes as well as usability or relevancy of the
component to a particular domain. Latent semantic analysis is used
for the feature vector representation of various software domains. It
exploits the fact that FeatureVector codes can be seen as documents
containing terms -the idenifiers present in the components- and so
text modeling methods that capture co-occurrence information in
low-dimensional spaces can be used. Further, we devised Neuro-
Fuzzy hybrid Inference System, which takes structural metric values
as input and calculates the reusability of the software component.
Decision tree algorithm is used to decide initial set of fuzzy rules for
the Neuro-fuzzy system. The results obtained are convincing enough
to propose the system for economical identification and retrieval of
reusable software components.
Abstract: The corrugated steel cladding used to cover most of
steel buildings is considered as non-structural element. This research
will reflect the effect of cladding as a shear diaphragm in increasing
the normal elastic capacity of columns. This study is important
because of the lack of information of the behavior of cladding and
secondary members in various codes. Mathematical models for six
different cases are carried by software. The results extracted from the
program have been plotted showing the effects of different variables
on the ultimate load of column. The variables considered in our
research are the spacing between columns and the thickness of the
corrugated sheet representing the sheet stiffness.
Abstract: Conventional approaches in the implementation of logic programming applications on embedded systems are solely of software nature. As a consequence, a compiler is needed that transforms the initial declarative logic program to its equivalent procedural one, to be programmed to the microprocessor. This approach increases the complexity of the final implementation and reduces the overall system's performance. On the contrary, presenting hardware implementations which are only capable of supporting logic programs prevents their use in applications where logic programs need to be intertwined with traditional procedural ones, for a specific application. We exploit HW/SW codesign methods to present a microprocessor, capable of supporting hybrid applications using both programming approaches. We take advantage of the close relationship between attribute grammar (AG) evaluation and knowledge engineering methods to present a programmable hardware parser that performs logic derivations and combine it with an extension of a conventional RISC microprocessor that performs the unification process to report the success or failure of those derivations. The extended RISC microprocessor is still capable of executing conventional procedural programs, thus hybrid applications can be implemented. The presented implementation is programmable, supports the execution of hybrid applications, increases the performance of logic derivations (experimental analysis yields an approximate 1000% increase in performance) and reduces the complexity of the final implemented code. The proposed hardware design is supported by a proposed extended C-language called C-AG.
Abstract: Adolescents in Northern Uganda are at risk of teenage
pregnancies, unsafe abortions and sexually transmitted infections
(STIs). There is silence on sex both at home and school. This cross
sectional descriptive analytical study interviews a random sample of
827 students and 13 teachers on knowledge, perception and
acceptability to a comprehensive adolescent sexual and reproductive
health education in “O” and “A” level secondary schools in Gulu
District. Quantitative data was analyzed using SPSS 16.0. Directed
content analysis of themes of transcribed qualitative data was
conducted manually for common codes, sub-categories and
categories. Of the 827 students; 54.3% (449) reported being in a
sexual relationship especially those aged 15-17 years. Majority
96.1% (807) supported the teaching of a comprehensive ASRHE,
citing no negative impact 71.5% (601). Majority 81.6% (686) agreed
that such education could help prevention of STIs, abortions and
teenage pregnancies, and that it should be taught by health workers
69.0% (580). Majority 76.6% (203) reported that ASRHE was not
currently being taught in their schools. Students had low knowledge
levels and misconceptions about ASRHE. ASRHE was highly
acceptable though not being emphasized; its success in school
settings requires multidisciplinary culturally sensitive approaches
amongst which health workers should be frontiers.
Abstract: A method is presented for obtaining the error probability for block codes. The method is based on the eigenvalueeigenvector properties of the code correlation matrix. It is found that under a unary transformation and for an additive white Gaussian noise environment, the performance evaluation of a block code becomes a one-dimensional problem in which only one eigenvalue and its corresponding eigenvector are needed in the computation. The obtained error rate results show remarkable agreement between simulations and analysis.
Abstract: This paper gives an overview of the mapping
mechanism of SEAM-a methodology for the automatic generation of
knowledge models and its mapping onto Java codes. It discusses the
rules that will be used to map the different components in the
knowledge model automatically onto Java classes, properties and
methods. The aim of developing this mechanism is to help in the
creation of a prototype which will be used to validate the knowledge
model which has been generated automatically. It will also help to
link the modeling phase with the implementation phase as existing
knowledge engineering methodologies do not provide for proper
guidelines for the transition from the knowledge modeling phase to
development phase. This will decrease the development overheads
associated to the development of Knowledge Based Systems.
Abstract: We investigate efficient spreading codes for transmitter based techniques of code division multiple access (CDMA) systems. The channel is considered to be known at the transmitter which is usual in a time division duplex (TDD) system where the channel is assumed to be the same on uplink and downlink. For such a TDD/CDMA system, both bitwise and blockwise multiuser transmission schemes are taken up where complexity is transferred to the transmitter side so that the receiver has minimum complexity. Different spreading codes are considered at the transmitter to spread the signal efficiently over the entire spectrum. The bit error rate (BER) curves portray the efficiency of the codes in presence of multiple access interference (MAI) as well as inter symbol interference (ISI).
Abstract: This paper focuses on the Mega-Sub Controlled
Structure Systems (MSCSS) performances and characteristics
regarding the new control principle contained in MSCSS subjected to
strong earthquake excitations. The adopted control scheme consists of
modulated sub-structures where the control action is achieved by
viscous dampers and sub-structure own configuration. The
elastic-plastic time history analysis under severe earthquake excitation
is analyzed base on the Finite Element Analysis Method (FEAM), and
some comparison results are also given in this paper. The result shows
that the MSCSS systems can remarkably reduce vibrations effects
more than the mega-sub structure (MSS). The study illustrates that the
improved MSCSS presents good seismic resistance ability even at 1.2g
and can absorb seismic energy in the structure, thus imply that
structural members cross section can be reduce and achieve to good
economic characteristics. Furthermore, the elasto-plastic analysis
demonstrates that the MSCSS is accurate enough regarding
international building evaluation and design codes. This paper also
shows that the elasto-plastic dynamic analysis method is a reasonable
and reliable analysis method for structures subjected to strong
earthquake excitations and that the computed results are more precise.
Abstract: Scale Time Offset Robust Modulation (STORM) [1]–
[3] is a high bandwidth waveform design that adds time-scale
to embedded reference modulations using only time-delay [4]. In
an environment where each user has a specific delay and scale,
identification of the user with the highest signal power and that
user-s phase is facilitated by the STORM processor. Both of these
parameters are required in an efficient multiuser detection algorithm.
In this paper, the STORM modulation approach is evaluated with
a direct sequence spread quadrature phase shift keying (DS-QPSK)
system. A misconception of the STORM time scale modulation is that
a fine temporal resolution is required at the receiver. STORM will
be applied to a QPSK code division multiaccess (CDMA) system
by modifying the spreading codes. Specifically, the in-phase code
will use a typical spreading code, and the quadrature code will
use a time-delayed and time-scaled version of the in-phase code.
Subsequently, the same temporal resolution in the receiver is required
before and after the application of STORM. In this paper, the bit error
performance of STORM in a synchronous CDMA system is evaluated
and compared to theory, and the bit error performance of STORM
incorporated in a single user WCDMA downlink is presented to
demonstrate the applicability of STORM in a modern communication
system.
Abstract: In the context of channel coding, the Generalized Belief Propagation (GBP) is an iterative algorithm used to recover the transmission bits sent through a noisy channel. To ensure a reliable transmission, we apply a map on the bits, that is called a code. This code induces artificial correlations between the bits to send, and it can be modeled by a graph whose nodes are the bits and the edges are the correlations. This graph, called Tanner graph, is used for most of the decoding algorithms like Belief Propagation or Gallager-B. The GBP is based on a non unic transformation of the Tanner graph into a so called region-graph. A clear advantage of the GBP over the other algorithms is the freedom in the construction of this graph. In this article, we explain a particular construction for specific graph topologies that involves relevant performance of the GBP. Moreover, we investigate the behavior of the GBP considered as a dynamic system in order to understand the way it evolves in terms of the time and in terms of the noise power of the channel. To this end we make use of classical measures and we introduce a new measure called the hyperspheres method that enables to know the size of the attractors.
Abstract: Reverse Engineering is a very important process in
Software Engineering. It can be performed backwards from system
development life cycle (SDLC) in order to get back the source data
or representations of a system through analysis of its structure,
function and operation. We use reverse engineering to introduce an
automatic tool to generate system requirements from its program
source codes. The tool is able to accept the Cµ programming source
codes, scan the source codes line by line and parse the codes to
parser. Then, the engine of the tool will be able to generate system
requirements for that specific program to facilitate reuse and
enhancement of the program. The purpose of producing the tool is to
help recovering the system requirements of any system when the
system requirements document (SRD) does not exist due to
undocumented support of the system.
Abstract: Two constructions of unit-memory binary convolutional
codes from linear block codes over the finite semi-local ring F2r +vF2r , where v2 = v, are presented. In both cases, if the linear block code is systematic, then the resulting convolutional encoder
is systematic, minimal, basic and non-catastrophic. The Hamming
free distance of the convolutional code is bounded below by the
minimum Hamming distance of the block code. New examples of
binary convolutional codes that meet the Heller upper bound for
systematic codes are given.
Abstract: Technology of thin film deposition is of interest in
many engineering fields, from electronic manufacturing to corrosion
protective coating. A typical deposition process, like that developed
at the University of Eindhoven, considers the deposition of a thin,
amorphous film of C:H or of Si:H on the substrate, using the
Expanding Thermal arc Plasma technique. In this paper a computing
procedure is proposed to simulate the flow field in a deposition
chamber similar to that at the University of Eindhoven and a
sensitivity analysis is carried out in terms of: precursor mass flow
rate, electrical power, supplied to the torch and fluid-dynamic
characteristics of the plasma jet, using different nozzles. To this
purpose a deposition chamber similar in shape, dimensions and
operating parameters to the above mentioned chamber is considered.
Furthermore, a method is proposed for a very preliminary evaluation
of the film thickness distribution on the substrate. The computing
procedure relies on two codes working in tandem; the output from
the first code is the input to the second one. The first code simulates
the flow field in the torch, where Argon is ionized according to the
Saha-s equation, and in the nozzle. The second code simulates the
flow field in the chamber. Due to high rarefaction level, this is a
(commercial) Direct Simulation Monte Carlo code. Gas is a mixture
of 21 chemical species and 24 chemical reactions from Argon plasma
and Acetylene are implemented in both codes. The effects of the
above mentioned operating parameters are evaluated and discussed
by 2-D maps and profiles of some important thermo-fluid-dynamic
parameters, as per Mach number, velocity and temperature. Intensity,
position and extension of the shock wave are evaluated and the
influence of the above mentioned test conditions on the film
thickness and uniformity of distribution are also evaluated.
Abstract: In automatic manufacturing and assembling of mechanical, electrical and electronic parts one needs to reliably identify the position of components and to extract the information of these components. Data Matrix Codes (DMC) are established by these days in many areas of industrial manufacturing thanks to their concentration of information on small spaces. In today’s usually order-related industry, where increased tracing requirements prevail, they offer further advantages over other identification systems. This underlines in an impressive way the necessity of a robust code reading system for detecting DMC on the components in factories. This paper compares two methods for estimating the angle of orientation of Data Matrix Codes: one method based on the Hough Transform and the other based on the Mean Shift Algorithm. We concentrate on Data Matrix Codes in industrial environment, punched, milled, lasered or etched on different materials in arbitrary orientation.
Abstract: This paper maps the structure of the social network of
the 2011 class ofsixty graduate students of the Masters of Science
(Knowledge Management) programme at the Nanyang Technological
University, based on their friending relationships on Facebook. To
ensure anonymity, actual names were not used. Instead, they were
replaced with codes constructed from their gender, nationality, mode
of study, year of enrollment and a unique number. The relationships
between friends within the class, and among the seniors and alumni
of the programme wereplotted. UCINet and Pajek were used to plot
the sociogram, to compute the density, inclusivity, and degree,
global, betweenness, and Bonacich centralities, to partition the
students into two groups, namely, active and peripheral, and to
identify the cut-points. Homophily was investigated, and it was
observed for nationality and study mode. The groups students formed
on Facebook were also studied, and of fifteen groups, eight were
classified as dead, which we defined as those that have been inactive
for over two months.
Abstract: The purpose of this research is to develop and apply the
RSCMAC to enhance the dynamic accuracy of Global Positioning
System (GPS). GPS devices provide services of accurate positioning,
speed detection and highly precise time standard for over 98% area on
the earth. The overall operation of Global Positioning System includes
24 GPS satellites in space; signal transmission that includes 2
frequency carrier waves (Link 1 and Link 2) and 2 sets random
telegraphic codes (C/A code and P code), on-earth monitoring stations
or client GPS receivers. Only 4 satellites utilization, the client position
and its elevation can be detected rapidly. The more receivable
satellites, the more accurate position can be decoded. Currently, the
standard positioning accuracy of the simplified GPS receiver is greatly
increased, but due to affected by the error of satellite clock, the
troposphere delay and the ionosphere delay, current measurement
accuracy is in the level of 5~15m. In increasing the dynamic GPS
positioning accuracy, most researchers mainly use inertial navigation
system (INS) and installation of other sensors or maps for the
assistance. This research utilizes the RSCMAC advantages of fast
learning, learning convergence assurance, solving capability of
time-related dynamic system problems with the static positioning
calibration structure to improve and increase the GPS dynamic
accuracy. The increasing of GPS dynamic positioning accuracy can be
achieved by using RSCMAC system with GPS receivers collecting
dynamic error data for the error prediction and follows by using the
predicted error to correct the GPS dynamic positioning data. The
ultimate purpose of this research is to improve the dynamic positioning
error of cheap GPS receivers and the economic benefits will be
enhanced while the accuracy is increased.
Abstract: The advances in multimedia and networking technologies
have created opportunities for Internet pirates, who can easily
copy multimedia contents and illegally distribute them on the Internet,
thus violating the legal rights of content owners. This paper describes
how a simple and well-known watermarking procedure based on a
spread spectrum method and a watermark recovery by correlation can
be improved to effectively and adaptively protect MPEG-2 videos
distributed on the Internet. In fact, the procedure, in its simplest
form, is vulnerable to a variety of attacks. However, its security
and robustness have been increased, and its behavior has been
made adaptive with respect to the video terminals used to open
the videos and the network transactions carried out to deliver them
to buyers. In fact, such an adaptive behavior enables the proposed
procedure to efficiently embed watermarks, and this characteristic
makes the procedure well suited to be exploited in web contexts,
where watermarks usually generated from fingerprinting codes have
to be inserted into the distributed videos “on the fly", i.e. during the
purchase web transactions.
Abstract: OpenMP is an API for parallel programming model of shared memory multiprocessors. Novice OpenMP programmers often produce the code that compiler cannot find human errors. It was investigated how compiler coped with the common mistakes that can occur in OpenMP code. The latest version(4.4.3) of GCC is used for this research. It was found that GCC compiled the codes without any errors or warnings. In this paper the programming aid tool is presented for OpenMP programs. It can check 12 common mistakes that novice programmer can commit during the programming of OpenMP. It was demonstrated that the programming aid tool can detect the various common mistakes that GCC failed to detect.