Abstract: Procurement is an important component in the field of
operating resource management and e-procurement is the golden key
to optimizing the supply chains system. Global firms are optimistic
on the level of savings that can be achieved through full
implementation of e-procurement strategies. E-procurement is an
Internet-based business process for obtaining materials and services
and managing their inflow into the organization. In this paper, the
subjects of supply chains and e-procurement and its benefits to
organizations have been studied. Also, e-procurement in construction
and its drivers and barriers have been discussed and a framework of
supplier selection in an e-procurement environment has been
demonstrated. This paper also has addressed critical success factors
in adopting e-procurement in supply chains.
Abstract: Conventional approaches in the implementation of logic programming applications on embedded systems are solely of software nature. As a consequence, a compiler is needed that transforms the initial declarative logic program to its equivalent procedural one, to be programmed to the microprocessor. This approach increases the complexity of the final implementation and reduces the overall system's performance. On the contrary, presenting hardware implementations which are only capable of supporting logic programs prevents their use in applications where logic programs need to be intertwined with traditional procedural ones, for a specific application. We exploit HW/SW codesign methods to present a microprocessor, capable of supporting hybrid applications using both programming approaches. We take advantage of the close relationship between attribute grammar (AG) evaluation and knowledge engineering methods to present a programmable hardware parser that performs logic derivations and combine it with an extension of a conventional RISC microprocessor that performs the unification process to report the success or failure of those derivations. The extended RISC microprocessor is still capable of executing conventional procedural programs, thus hybrid applications can be implemented. The presented implementation is programmable, supports the execution of hybrid applications, increases the performance of logic derivations (experimental analysis yields an approximate 1000% increase in performance) and reduces the complexity of the final implemented code. The proposed hardware design is supported by a proposed extended C-language called C-AG.
Abstract: Since 1992, year where Hugo de Garis has published
the first paper on Evolvable Hardware (EHW), a period of intense
creativity has followed. It has been actively researched, developed
and applied to various problems. Different approaches have been
proposed that created three main classifications: extrinsic, mixtrinsic
and intrinsic EHW. Each of these solutions has a real interest.
Nevertheless, although the extrinsic evolution generates some
excellent results, the intrinsic systems are not so advanced. This
paper suggests 3 possible solutions to implement the run-time
configuration intrinsic EHW system: FPGA-based Run-Time
Configuration system, JBits-based Run-Time Configuration system
and Multi-board functional-level Run-Time Configuration system.
The main characteristic of the proposed architectures is that they are
implemented on Field Programmable Gate Array. A comparison of
proposed solutions demonstrates that multi-board functional-level
run-time configuration is superior in terms of scalability, flexibility
and the implementation easiness.
Abstract: Wavelet transforms are multiresolution
decompositions that can be used to analyze signals and images.
Image compression is one of major applications of wavelet
transforms in image processing. It is considered as one of the most
powerful methods that provides a high compression ratio. However,
its implementation is very time-consuming. At the other hand,
parallel computing technologies are an efficient method for image
compression using wavelets. In this paper, we propose a parallel
wavelet compression algorithm based on quadtrees. We implement
the algorithm using MatlabMPI (a parallel, message passing version
of Matlab), and compute its isoefficiency function, and show that it is
scalable. Our experimental results confirm the efficiency of the
algorithm also.
Abstract: Service identification is one of the main activities in
the modeling of a service-oriented solution, and therefore errors
made during identification can flow down through detailed design
and implementation activities that may necessitate multiple
iterations, especially in building composite applications. Different
strategies exist for how to identify candidate services that each of
them has its own benefits and trade offs. The approach presented in
this paper proposes a selective identification of services approach,
based on in depth business process analysis coupled with use cases
and existing assets analysis and goal service modeling. This article
clearly emphasizes the key activities need for the analysis and
service identification to build a optimized service oriented
architecture. In contrast to other approaches this article mentions
some best practices and steps, wherever appropriate, to point out the
vagueness involved in service identification.
Abstract: This paper presents the theoretical background and
the real implementation of an automated computer system to
introduce machine vision in flower, fruit and vegetable processing
for recollection, cutting, packaging, classification, or fumigation
tasks. The considerations and implementation issues presented in this
work can be applied to a wide range of varieties of flowers, fruits and
vegetables, although some of them are especially relevant due to the
great amount of units that are manipulated and processed each year
over the world. The computer vision algorithms developed in this
work are shown in detail, and can be easily extended to other
applications. A special attention is given to the electromagnetic
compatibility in order to avoid noisy images. Furthermore, real
experimentation has been carried out in order to validate the
developed application. In particular, the tests show that the method
has good robustness and high success percentage in the object
characterization.
Abstract: In this paper, we evaluate the choice of suitable
quantization characteristics for both the decoder messages and the
received samples in Low Density Parity Check (LDPC) coded
systems using M-QAM (Quadrature Amplitude Modulation)
schemes. The analysis involves the demapper block that provides
initial likelihood values for the decoder, by relating its quantization
strategy of the decoder. A mapping strategy refers to the grouping of
bits within a codeword, where each m-bit group is used to select a
2m-ary signal in accordance with the signal labels. Further we
evaluate the system with mapping strategies like Consecutive-Bit
(CB) and Bit-Reliability (BR). A new demapper version, based on
approximate expressions, is also presented to yield a low complexity
hardware implementation.
Abstract: Various intelligences and inspirations have been
adopted into the iterative searching process called as meta-heuristics.
They intelligently perform the exploration and exploitation in the
solution domain space aiming to efficiently seek near optimal
solutions. In this work, the bee algorithm, inspired by the natural
foraging behaviour of honey bees, was adapted to find the near
optimal solutions of the transportation management system, dynamic
multi-zone dispatching. This problem prepares for an uncertainty and
changing customers- demand. In striving to remain competitive,
transportation system should therefore be flexible in order to cope
with the changes of customers- demand in terms of in-bound and outbound
goods and technological innovations. To remain higher service
level but lower cost management via the minimal imbalance scenario,
the rearrangement penalty of the area, in each zone, including time
periods are also included. However, the performance of the algorithm
depends on the appropriate parameters- setting and need to be
determined and analysed before its implementation. BEE parameters
are determined through the linear constrained response surface
optimisation or LCRSOM and weighted centroid modified simplex
methods or WCMSM. Experimental results were analysed in terms
of best solutions found so far, mean and standard deviation on the
imbalance values including the convergence of the solutions
obtained. It was found that the results obtained from the LCRSOM
were better than those using the WCMSM. However, the average
execution time of experimental run using the LCRSOM was longer
than those using the WCMSM. Finally a recommendation of proper
level settings of BEE parameters for some selected problem sizes is
given as a guideline for future applications.
Abstract: A novel three-phase active power filter (APF) circuit with photovoltaic (PV) system to improve the quality of service and to reduce the capacity of energy storage capacitor is presented. The energy balance concept and sampling technique were used to simplify the calculation algorithm for the required utility source current and to control the voltage of the energy storage capacitor. The feasibility was verified by using the Pspice simulations and experiments. When the APF mode was used during non-operational period, not only the utilization rate, power factor and power quality could be improved, but also the capacity of energy storage capacitor could sparing. As the results, the advantages of the APF circuit are simplicity of control circuits, low cost, and good transient response.
Abstract: The paper describes design of an ontology in the
financial domain for mutual funds. The design of this ontology
consists of four steps, namely, specification, knowledge acquisition,
implementation and semantic query. Specification includes a
description of the taxonomy and different types mutual funds and
their scope. Knowledge acquisition involves the information
extraction from heterogeneous resources. Implementation describes
the conceptualization and encoding of this data. Finally, semantic
query permits complex queries to integrated data, mapping of these
database entities to ontological concepts.
Abstract: With the turn of this century, many researchers
started showing interest in Embedded Firewall (EF) implementations.
These are not the usual firewalls that are used as checkpoints at network gateways. They are, rather, applied near those hosts that need protection. Hence by using them, individual or grouped network
components can be protected from the inside as well as from external attacks.
This paper presents a study of EF-s, looking at their architecture and problems. A comparative study assesses how practical each kind is. It particularly focuses on the architecture, weak points, and
portability of each kind. A look at their use by different categories of users is also presented.
Abstract: Truncated multiplier is a good candidate for digital
signal processing (DSP) applications including finite impulse
response (FIR) and discrete cosine transform (DCT). Through
truncated multiplier a significant reduction in Field Programmable
Gate Array (FPGA) resources can be achieved. This paper presents
for the first time a comparison of resource utilization of Spartan-3AN
and Virtex-5 implementation of standard and truncated multipliers
using Very High Speed Integrated Circuit Hardware Description
Language (VHDL). The Virtex-5 FPGA shows significant
improvement as compared to Spartan-3AN FPGA device. The
Virtex-5 FPGA device shows better performance with a percentage
ratio of number of occupied slices for standard to truncated
multipliers is increased from 40% to 73.86% as compared to Spartan-
3AN is decreased from 68.75% to 58.78%. Results show that the
anomaly in Spartan-3AN FPGA device average connection and
maximum pin delay have been efficiently reduced in Virtex-5 FPGA
device.
Abstract: The implementation of single-electron tunneling
(SET) simulators based on the master-equation (ME) formalism
requires the efficient and accurate identification of an exhaustive list
of active states and related tunnel events. Dynamic simulations also
require the control of the emerging states and guarantee the safe
elimination of decaying states. This paper describes algorithms for
use in the stationary and dynamic control of the lists of active states
and events. The paper presents results obtained using these
algorithms with different SET structures.
Abstract: This paper presents and evaluates a new classification
method that aims to improve classifiers performances and speed up
their training process. The proposed approach, called labeled
classification, seeks to improve convergence of the BP (Back
propagation) algorithm through the addition of an extra feature
(labels) to all training examples. To classify every new example, tests
will be carried out each label. The simplicity of implementation is the
main advantage of this approach because no modifications are
required in the training algorithms. Therefore, it can be used with
others techniques of acceleration and stabilization. In this work, two
models of the labeled classification are proposed: the LMLP
(Labeled Multi Layered Perceptron) and the LNFC (Labeled Neuro
Fuzzy Classifier). These models are tested using Iris, wine, texture
and human thigh databases to evaluate their performances.
Abstract: This paper presents a procedure of forming the
mathematical model of radial electric power systems for simulation
of both transient and steady-state conditions. The research idea has
been based on nodal voltages technique and on differentiation of
Kirchhoff's current law (KCL) applied to each non-reference node of
the radial system, the result of which the nodal voltages has been
calculated by solving a system of algebraic equations. Currents of the
electric power system components have been determined by solving
their respective differential equations. Transforming the three-phase
coordinate system into Cartesian coordinate system in the model
decreased the overall number of equations by one third. The use of
Cartesian coordinate system does not ignore the DC component
during transient conditions, but restricts the model's implementation
for symmetrical modes of operation only. An example of the input
data for a four-bus radial electric power system has been calculated.
Abstract: The study of proteomics reached unexpected levels of
interest, as a direct consequence of its discovered influence over some
complex biological phenomena, such as problematic diseases like
cancer. This paper presents the latest authors- achievements regarding
the analysis of the networks of proteins (interactome networks), by
computing more efficiently the betweenness centrality measure. The
paper introduces the concept of betweenness centrality, and then
describes how betweenness computation can help the interactome net-
work analysis. Current sequential implementations for the between-
ness computation do not perform satisfactory in terms of execution
times. The paper-s main contribution is centered towards introducing
a speedup technique for the betweenness computation, based on
modified shortest path algorithms for sparse graphs. Three optimized
generic algorithms for betweenness computation are described and
implemented, and their performance tested against real biological
data, which is part of the IntAct dataset.
Abstract: Pressures for urban redevelopment are intensifying in
all large cities. A new logic for urban development is required –
green urbanism – that provides a spatial framework for directing
population and investment inwards to brownfields and greyfields
precincts, rather than outwards to the greenfields. This represents
both a major opportunity and a major challenge for city planners in
pluralist liberal democracies. However, plans for more compact
forms of urban redevelopment are stalling in the face of community
resistance. A new paradigm and spatial planning platform is required
that will support timely multi-level and multi-actor stakeholder
engagement, resulting in the emergence of consensus plans for
precinct-level urban regeneration capable of more rapid
implementation. Using Melbourne, Australia as a case study, this
paper addresses two of the urban intervention challenges – where and
how – via the application of a 21st century planning tool ENVISION
created for this purpose.
Abstract: The risk sphere in business is fast changing and expanding. Almost anything has become a risk factor that will have potent, direct, and far reaching impacts on business. This paper examines the intensity of enterprise risk management (ERM) practices among the Malaysian public listed companies. The paper espouses a ERM framework comprising fourteen important implementation elements and processes. Results of the analysis indicate that the intensity of ERM implementation among the respondents is in the ‘good’ category of the semantic scale, which is deemed encouraging vis-à-vis the country’s regulatory regime.
Abstract: Biometric measures of one kind or another have been
used to identify people since ancient times, with handwritten
signatures, facial features, and fingerprints being the traditional
methods. Of late, Systems have been built that automate the task of
recognition, using these methods and newer ones, such as hand
geometry, voiceprints and iris patterns. These systems have different
strengths and weaknesses. This work is a two-section composition. In
the starting section, we present an analytical and comparative study
of common biometric techniques. The performance of each of them
has been viewed and then tabularized as a result. The latter section
involves the actual implementation of the techniques under
consideration that has been done using a state of the art tool called,
MATLAB. This tool aids to effectively portray the corresponding
results and effects.
Abstract: Starting from a biologically inspired framework, Gabor filters were built up from retinal filters via LMSE algorithms. Asubset of retinal filter kernels was chosen to form a particular Gabor filter by using a weighted sum. One-dimensional optimization approaches were shown to be inappropriate for the problem. All model parameters were fixed with biological or image processing constraints. Detailed analysis of the optimization procedure led to the introduction of a minimization constraint. Finally, quantization of weighting factors was investigated. This resulted in an optimized cascaded structure of a Gabor filter bank implementation with lower computational cost.