Abstract: In a previous work, we presented the numerical
solution of the two dimensional second order telegraph partial
differential equation discretized by the centred and rotated five-point
finite difference discretizations, namely the explicit group (EG) and
explicit decoupled group (EDG) iterative methods, respectively. In
this paper, we utilize a domain decomposition algorithm on these
group schemes to divide the tasks involved in solving the same
equation. The objective of this study is to describe the development
of the parallel group iterative schemes under OpenMP programming
environment as a way to reduce the computational costs of the
solution processes using multicore technologies. A detailed
performance analysis of the parallel implementations of points and
group iterative schemes will be reported and discussed.
Abstract: This paper examines the influence of communication
form on employee uncertainty during mergers and acquisitions
(M&As). Specifically, the author uses narrative theory to analyze
how narrative organizational communication affects the three
components of uncertainty – decreased predictive, explanatory, and
descriptive ability. It is hypothesized that employees whose
organizations use narrative M&A communication will have greater
predictive, explanatory, and descriptive abilities than employees of
organizations using non-narrative M&A communication. This paper
contributes to the stream of research examining uncertainty during
mergers and acquisitions and argues that narratives are an effective
means of managing uncertainty in the mergers and acquisitions
context.
Abstract: This document describes an electronic speed control designed to drive two DC motors from a 6 V battery pack to be controlled by a commercial universal infrared remote control hand set. Conceived for a tank-like vehicle, one motor drives the left side wheels or tracks and the other motor drives the right side. As it is shown here, there is a left-right steering input and a forward– backward throttles input, like would be used on a model car. It is designed using a microcontroller PIC16F873A.
Abstract: The promises of component-based technology can only be fully realized when the system contains in its design a necessary level of separation of concerns. The authors propose to focus on the concerns that emerge throughout the life cycle of the system and use them as an architectural foundation for the design of a component-based framework. The proposed model comprises a set of superimposed views of the system describing its functional and non-functional concerns. This approach is illustrated by the design of a specific framework for data analysis and data acquisition and supplemented with experiences from using the systems developed with this framework at the Fermi National Accelerator Laboratory.
Abstract: This paper provides an introduction into the evolution
of information and communication technology and illustrates its
usage in the work domain. The paper is sub-divided into two parts.
The first part gives an overview over the different phases of
information processing in the work domain. It starts by charting the
past and present usage of computers in work environments and shows
current technological trends, which are likely to influence future
business applications. The second part starts by briefly describing,
how the usage of computers changed business processes in the past,
and presents first Ambient Intelligence applications based on
identification and localization information, which are already used in
the production and retail sector. Based on current systems and
prototype applications, the paper gives an outlook of how Ambient
Intelligence technologies could change business processes in the
future.
Abstract: Using spatial models as a shared common basis of
information about the environment for different kinds of contextaware
systems has been a heavily researched topic in the last years.
Thereby the research focused on how to create, to update, and to
merge spatial models so as to enable highly dynamic, consistent and
coherent spatial models at large scale. In this paper however, we
want to concentrate on how context-aware applications could use this
information so as to adapt their behavior according to the situation
they are in. The main idea is to provide the spatial model
infrastructure with a situation recognition component based on
generic situation templates. A situation template is – as part of a
much larger situation template library – an abstract, machinereadable
description of a certain basic situation type, which could be
used by different applications to evaluate their situation. In this
paper, different theoretical and practical issues – technical, ethical
and philosophical ones – are discussed important for understanding
and developing situation dependent systems based on situation
templates. A basic system design is presented which allows for the
reasoning with uncertain data using an improved version of a
learning algorithm for the automatic adaption of situation templates.
Finally, for supporting the development of adaptive applications, we
present a new situation-aware adaptation concept based on
workflows.
Abstract: This paper presents a hybrid approach for solving nqueen problem by combination of PSO and SA. PSO is a population based heuristic method that sometimes traps in local maximum. To solve this problem we can use SA. Although SA suffer from many iterations and long time convergence for solving some problems, By good adjusting initial parameters such as temperature and the length of temperature stages SA guarantees convergence. In this article we use discrete PSO (due to nature of n-queen problem) to achieve a good local maximum. Then we use SA to escape from local maximum. The experimental results show that our hybrid method in comparison of SA method converges to result faster, especially for high dimensions n-queen problems.
Abstract: Incompressible Navier-Stokes equations are reviewed
in this work. Three-dimensional Navier-Stokes equations are solved
analytically. The Mathematical derivation shows that the solutions
for the zero and constant pressure gradients are similar. Descriptions
of the proposed formulation and validation against two laminar
experiments and three different turbulent flow cases are reported in
this paper. Even though, the analytical solution is derived for nonreacting
flows, it could reproduce trends for cases including
combustion.
Abstract: We consider linear regression models where both input data (the values of independent variables) and output data (the observations of the dependent variable) are interval-censored. We introduce a possibilistic generalization of the least squares estimator, so called OLS-set for the interval model. This set captures the impact of the loss of information on the OLS estimator caused by interval censoring and provides a tool for quantification of this effect. We study complexity-theoretic properties of the OLS-set. We also deal with restricted versions of the general interval linear regression model, in particular the crisp input – interval output model. We give an argument that natural descriptions of the OLS-set in the crisp input – interval output cannot be computed in polynomial time. Then we derive easily computable approximations for the OLS-set which can be used instead of the exact description. We illustrate the approach by an example.
Abstract: The mechanical behavior of porous media is governed by the interaction between its solid skeleton and the fluid existing inside its pores. The interaction occurs through the interface of gains and fluid. The traditional analysis methods of porous media, based on the effective stress and Darcy's law, are unable to account for these interactions. For an accurate analysis, the porous media is represented in a fluid-filled porous solid on the basis of the Biot theory of wave propagation in poroelastic media. In Biot formulation, the equations of motion of the soil mixture are coupled with the global mass balance equations to describe the realistic behavior of porous media. Because of irregular geometry, the domain is generally treated as an assemblage of fmite elements. In this investigation, the numerical formulation for the field equations governing the dynamic response of fluid-saturated porous media is analyzed and employed for the study of transient wave motion. A finite element model is developed and implemented into a computer code called DYNAPM for dynamic analysis of porous media. The weighted residual method with 8-node elements is used for developing of a finite element model and the analysis is carried out in the time domain considering the dynamic excitation and gravity loading. Newmark time integration scheme is developed to solve the time-discretized equations which are an unconditionally stable implicit method Finally, some numerical examples are presented to show the accuracy and capability of developed model for a wide variety of behaviors of porous media.
Abstract: Most of the real queuing systems include special properties and constraints, which can not be analyzed directly by using the results of solved classical queuing models. Lack of Markov chains features, unexponential patterns and service constraints, are the mentioned conditions. This paper represents an applied general algorithm for analysis and optimizing the queuing systems. The algorithm stages are described through a real case study. It is consisted of an almost completed non-Markov system with limited number of customers and capacities as well as lots of common exception of real queuing networks. Simulation is used for optimizing this system. So introduced stages over the following article include primary modeling, determining queuing system kinds, index defining, statistical analysis and goodness of fit test, validation of model and optimizing methods of system with simulation.
Abstract: Innovations in technology have created new ethical
challenges. Essential use of electronic communication in the
workplace has escalated at an astronomical rate over the past decade.
As such, legal and ethical dilemmas confronted by both the employer
and the employee concerning managerial control and ownership of einformation
have increased dramatically in the USA. From the
employer-s perspective, ownership and control of all information
created for the workplace is an undeniable source of economic
advantage and must be monitored zealously. From the perspective of
the employee, individual rights, such as privacy, freedom of speech,
and freedom from unreasonable search and seizure, continue to be
stalwart legal guarantees that employers are not legally or ethically
entitled to abridge in the workplace. These issues have been the
source of great debate and the catalyst for legal reform. The fine line
between ethical and legal has been complicated by emerging
technologies. This manuscript will identify and discuss a number of
specific legal and ethical issues raised by the dynamic electronic
workplace and conclude with suggestions that employers should
follow to respect the delicate balance between employees- legal
rights to privacy and the employer's right to protect its knowledge
systems and infrastructure.
Abstract: The purpose of this study was to present a reliable mean for human-computer interfacing based on finger gestures made in two dimensions, which could be interpreted and adequately used in controlling a remote robot's movement. The gestures were captured and interpreted using an algorithm based on trigonometric functions, in calculating the angular displacement from one point of touch to another as the user-s finger moved within a time interval; thereby allowing for pattern spotting of the captured gesture. In this paper the design and implementation of such a gesture based user interface was presented, utilizing the aforementioned algorithm. These techniques were then used to control a remote mobile robot's movement. A resistive touch screen was selected as the gesture sensor, then utilizing a programmed microcontroller to interpret them respectively.
Abstract: Cosmic showers, during the transit through space, produce
sub - products as a result of interactions with the intergalactic
or interstellar medium which after entering earth generate secondary
particles called Extensive Air Shower (EAS). Detection and analysis
of High Energy Particle Showers involve a plethora of theoretical and
experimental works with a host of constraints resulting in inaccuracies
in measurements. Therefore, there exist a necessity to develop a
readily available system based on soft-computational approaches
which can be used for EAS analysis. This is due to the fact that soft
computational tools such as Artificial Neural Network (ANN)s can be
trained as classifiers to adapt and learn the surrounding variations. But
single classifiers fail to reach optimality of decision making in many
situations for which Multiple Classifier System (MCS) are preferred
to enhance the ability of the system to make decisions adjusting
to finer variations. This work describes the formation of an MCS
using Multi Layer Perceptron (MLP), Recurrent Neural Network
(RNN) and Probabilistic Neural Network (PNN) with data inputs
from correlation mapping Self Organizing Map (SOM) blocks and
the output optimized by another SOM. The results show that the setup
can be adopted for real time practical applications for prediction
of primary energy and location of EAS from density values captured
using detectors in a circular grid.
Abstract: In Image processing the Image compression can improve
the performance of the digital systems by reducing the cost and
time in image storage and transmission without significant reduction
of the Image quality. This paper describes hardware architecture of
low complexity Discrete Cosine Transform (DCT) architecture for
image compression[6]. In this DCT architecture, common computations
are identified and shared to remove redundant computations
in DCT matrix operation. Vector processing is a method used for
implementation of DCT. This reduction in computational complexity
of 2D DCT reduces power consumption. The 2D DCT is performed
on 8x8 matrix using two 1-Dimensional Discrete cosine transform
blocks and a transposition memory [7]. Inverse discrete cosine
transform (IDCT) is performed to obtain the image matrix and
reconstruct the original image. The proposed image compression
algorithm is comprehended using MATLAB code. The VLSI design
of the architecture is implemented Using Verilog HDL. The proposed
hardware architecture for image compression employing DCT was
synthesized using RTL complier and it was mapped using 180nm
standard cells. . The Simulation is done using Modelsim. The
simulation results from MATLAB and Verilog HDL are compared.
Detailed analysis for power and area was done using RTL compiler
from CADENCE. Power consumption of DCT core is reduced to
1.027mW with minimum area[1].
Abstract: In this paper, an efficient local appearance feature
extraction method based the multi-resolution Curvelet transform is
proposed in order to further enhance the performance of the well
known Linear Discriminant Analysis(LDA) method when applied
to face recognition. Each face is described by a subset of band
filtered images containing block-based Curvelet coefficients. These
coefficients characterize the face texture and a set of simple statistical
measures allows us to form compact and meaningful feature vectors.
The proposed method is compared with some related feature extraction
methods such as Principal component analysis (PCA), as well
as Linear Discriminant Analysis LDA, and independent component
Analysis (ICA). Two different muti-resolution transforms, Wavelet
(DWT) and Contourlet, were also compared against the Block Based
Curvelet-LDA algorithm. Experimental results on ORL, YALE and
FERET face databases convince us that the proposed method provides
a better representation of the class information and obtains much
higher recognition accuracies.
Abstract: The design of a pattern classifier includes an attempt
to select, among a set of possible features, a minimum subset of
weakly correlated features that better discriminate the pattern classes.
This is usually a difficult task in practice, normally requiring the
application of heuristic knowledge about the specific problem
domain. The selection and quality of the features representing each
pattern have a considerable bearing on the success of subsequent
pattern classification. Feature extraction is the process of deriving
new features from the original features in order to reduce the cost of
feature measurement, increase classifier efficiency, and allow higher
classification accuracy. Many current feature extraction techniques
involve linear transformations of the original pattern vectors to new
vectors of lower dimensionality. While this is useful for data
visualization and increasing classification efficiency, it does not
necessarily reduce the number of features that must be measured
since each new feature may be a linear combination of all of the
features in the original pattern vector. In this paper a new approach is
presented to feature extraction in which feature selection, feature
extraction, and classifier training are performed simultaneously using
a genetic algorithm. In this approach each feature value is first
normalized by a linear equation, then scaled by the associated weight
prior to training, testing, and classification. A knn classifier is used to
evaluate each set of feature weights. The genetic algorithm optimizes
a vector of feature weights, which are used to scale the individual
features in the original pattern vectors in either a linear or a nonlinear
fashion. By this approach, the number of features used in classifying
can be finely reduced.
Abstract: This paper describes a three-dimensional thermal
model of the current path included in the low voltage power circuit
breakers. The model can be used to analyse the thermal behaviour of
the current path during both steady-state and transient conditions.
The current path lengthwise temperature distribution and timecurrent
characteristic of the terminal connections of the power circuit
breaker have been obtained. The influence of the electric current and
voltage drop on main electric contact of the circuit breaker has been
investigated. To validate the three-dimensional thermal model, some
experimental tests have been done. There is a good correlation
between experimental and simulation results.
Abstract: This paper describes the experimental efficiency of a
compact organic Rankine cycle (ORC) system with a compact
rotary-vane-type expander. The compact ORC system can be used for
power generation from low-temperature heat sources such as waste
heat from various small-scale heat engines, fuel cells, electric devices,
and solar thermal energy. The purpose of this study is to develop an
ORC system with a low power output of less than 1 kW with a hot
temperature source ranging from 60°C to 100°C and a cold
temperature source ranging from 10°C to 30°C. The power output of
the system is rather less due to limited heat efficiency. Therefore, the
system should have an economically optimal efficiency. In order to
realize such a system, an efficient and low-cost expander is
indispensable. An experimental ORC system was developed using the
rotary-vane-type expander which is one of possible candidates of the
expander. The experimental results revealed the expander
performance for various rotation speeds, expander efficiencies, and
thermal efficiencies. Approximately 30 W of expander power output
with 48% expander efficiency and 4% thermal efficiency with a
temperature difference between the hot and cold sources of 80°C was
achieved.
Abstract: We consider the problem of bandwidth allocation in a
substrate network as an optimization problem for the aggregate utility
of multiple applications with diverse requirements and describe a
simulation scheme for dynamically adaptive bandwidth allocation
protocols. The proposed simulation model based on Coloured Petri
Nets (CPN) is realized using CPN Tools.