Abstract: Some meta-schedulers query the information system of individual supercomputers in order to submit jobs to the least busy supercomputer on a computational Grid. However, this information can become outdated by the time a job starts due to changes in scheduling priorities. The MSR scheme is based on Multiple Simultaneous Requests and can take advantage of opportunities resulting from these priorities changes. This paper presents the SWARM meta-scheduler, which can speed up the execution of large sets of tasks by minimizing the job queuing time through the submission of multiple requests. Performance tests have shown that this new meta-scheduler is faster than an implementation of the MSR scheme and the gLite meta-scheduler. SWARM has been used through the GridQTL project beta-testing portal during the past year. Statistics are provided for this usage and demonstrate its capacity to achieve reliably a substantial reduction of the execution time in production conditions.
Abstract: This study aims to propose three evaluation methods to
evaluate the Tokyo Cap and Trade Program when emissions trading is
performed virtually among enterprises, focusing on carbon dioxide
(CO2), which is the only emitted greenhouse gas that tends to increase.
The first method clarifies the optimum reduction rate for the highest
cost benefit, the second discusses emissions trading among enterprises
through market trading, and the third verifies long-term emissions
trading during the term of the plan (2010-2019), checking the validity
of emissions trading partly using Geographic Information Systems
(GIS). The findings of this study can be summarized in the following
three points.
1. Since the total cost benefit is the greatest at a 44% reduction rate, it
is possible to set it more highly than that of the Tokyo Cap and
Trade Program to get more total cost benefit.
2. At a 44% reduction rate, among 320 enterprises, 8 purchasing
enterprises and 245 sales enterprises gain profits from emissions
trading, and 67 enterprises perform voluntary reduction without
conducting emissions trading. Therefore, to further promote
emissions trading, it is necessary to increase the sales volumes of
emissions trading in addition to sales enterprises by increasing the
number of purchasing enterprises.
3. Compared to short-term emissions trading, there are few enterprises
which benefit in each year through the long-term emissions trading
of the Tokyo Cap and Trade Program. Only 81 enterprises at the
most can gain profits from emissions trading in FY 2019. Therefore,
by setting the reduction rate more highly, it is necessary to increase
the number of enterprises that participate in emissions trading and
benefit from the restraint of CO2 emissions.
Abstract: Information on weed distribution within the field is necessary to implement spatially variable herbicide application. Since hand labor is costly, an automated weed control system could be feasible. This paper deals with the development of an algorithm for real time specific weed recognition system based on Histogram Maxima with threshold of an image that is used for the weed classification. This algorithm is specifically developed to classify images into broad and narrow class for real-time selective herbicide application. The developed system has been tested on weeds in the lab, which have shown that the system to be very effectiveness in weed identification. Further the results show a very reliable performance on images of weeds taken under varying field conditions. The analysis of the results shows over 95 percent classification accuracy over 140 sample images (broad and narrow) with 70 samples from each category of weeds.
Abstract: In recent years, everything is trending toward digitalization
and with the rapid development of the Internet technologies,
digital media needs to be transmitted conveniently over the network.
Attacks, misuse or unauthorized access of information is of great
concern today which makes the protection of documents through
digital media a priority problem. This urges us to devise new data
hiding techniques to protect and secure the data of vital significance.
In this respect, steganography often comes to the fore as a tool for
hiding information. Steganography is a process that involves hiding
a message in an appropriate carrier like image or audio. It is of
Greek origin and means "covered or hidden writing". The goal of
steganography is covert communication. Here the carrier can be sent
to a receiver without any one except the authenticated receiver only
knows existence of the information. Considerable amount of work
has been carried out by different researchers on steganography. In this
work the authors propose a novel Steganographic method for hiding
information within the spatial domain of the gray scale image. The
proposed approach works by selecting the embedding pixels using
some mathematical function and then finds the 8 neighborhood of
the each selected pixel and map each bit of the secret message in
each of the neighbor pixel coordinate position in a specified manner.
Before embedding a checking has been done to find out whether the
selected pixel or its neighbor lies at the boundary of the image or not.
This solution is independent of the nature of the data to be hidden
and produces a stego image with minimum degradation.
Abstract: The least mean square (LMS) algorithmis one of the
most well-known algorithms for mobile communication systems
due to its implementation simplicity. However, the main limitation
is its relatively slow convergence rate. In this paper, a booster
using the concept of Markov chains is proposed to speed up the
convergence rate of LMS algorithms. The nature of Markov
chains makes it possible to exploit the past information in the
updating process. Moreover, since the transition matrix has a
smaller variance than that of the weight itself by the central limit
theorem, the weight transition matrix converges faster than the
weight itself. Accordingly, the proposed Markov-chain based
booster thus has the ability to track variations in signal
characteristics, and meanwhile, it can accelerate the rate of
convergence for LMS algorithms. Simulation results show that the
LMS algorithm can effectively increase the convergence rate and
meantime further approach the Wiener solution, if the
Markov-chain based booster is applied. The mean square error is
also remarkably reduced, while the convergence rate is improved.
Abstract: Extensive use of the Internet coupled with the
marvelous growth in e-commerce and m-commerce has created a
huge demand for information security. The Secure Socket Layer
(SSL) protocol is the most widely used security protocol in the
Internet which meets this demand. It provides protection against
eaves droppings, tampering and forgery. The cryptographic
algorithms RC4 and HMAC have been in use for achieving security
services like confidentiality and authentication in the SSL. But recent
attacks against RC4 and HMAC have raised questions in the
confidence on these algorithms. Hence two novel cryptographic
algorithms MAJE4 and MACJER-320 have been proposed as
substitutes for them. The focus of this work is to demonstrate the
performance of these new algorithms and suggest them as dependable
alternatives to satisfy the need of security services in SSL. The
performance evaluation has been done by using practical
implementation method.
Abstract: Deformable active contours are widely used in
computer vision and image processing applications for image
segmentation, especially in biomedical image analysis. The active
contour or “snake" deforms towards a target object by controlling the
internal, image and constraint forces. However, if the contour
initialized with a lesser number of control points, there is a high
probability of surpassing the sharp corners of the object during
deformation of the contour. In this paper, a new technique is
proposed to construct the initial contour by incorporating prior
knowledge of significant corners of the object detected using the
Harris operator. This new reconstructed contour begins to deform, by
attracting the snake towards the targeted object, without missing the
corners. Experimental results with several synthetic images show the
ability of the new technique to deal with sharp corners with a high
accuracy than traditional methods.
Abstract: In this work a surgical simulator is produced which
enables a training otologist to conduct a virtual, real-time prosthetic
insertion. The simulator provides the Ear, Nose and Throat surgeon
with real-time visual and haptic responses during virtual cochlear
implantation into a 3D model of the human Scala Tympani (ST). The
parametric model is derived from measured data as published in the
literature and accounts for human morphological variance, such as
differences in cochlear shape, enabling patient-specific pre- operative
assessment. Haptic modeling techniques use real physical data and
insertion force measurements, to develop a force model which
mimics the physical behavior of an implant as it collides with the ST
walls during an insertion. Output force profiles are acquired from the
insertion studies conducted in the work, to validate the haptic model.
The simulator provides the user with real-time, quantitative insertion
force information and associated electrode position as user inserts the
virtual implant into the ST model. The information provided by this
study may also be of use to implant manufacturers for design
enhancements as well as for training specialists in optimal force
administration, using the simulator. The paper reports on the methods
for anatomical modeling and haptic algorithm development, with
focus on simulator design, development, optimization and validation.
The techniques may be transferrable to other medical applications
that involve prosthetic device insertions where user vision is
obstructed.
Abstract: The fundamental defect inherent to the thermoforming
technology is wall-thickness variation of the products due to
inadequate thermal processing during production of polymer. A
nonlinear viscoelastic rheological model is implemented for
developing the process model. This model describes deformation
process of a sheet in thermoforming process. Because of relaxation
pause after plug-assist stage and also implementation of two stage
thermoforming process have minor wall-thickness variation and
consequently better mechanical properties of polymeric articles. For
model validation, a comparative analysis of the theoretical and
experimental data is presented.
Abstract: Color constancy algorithms are generally based on the
simplified assumption about the spectral distribution or the reflection
attributes of the scene surface. However, in reality, these assumptions
are too restrictive. The methodology is proposed to extend existing
algorithm to applying color constancy locally to image patches rather
than globally to the entire images.
In this paper, a method based on low-level image features using
superpixels is proposed. Superpixel segmentation partition an image
into regions that are approximately uniform in size and shape. Instead
of using entire pixel set for estimating the illuminant, only superpixels
with the most valuable information are used. Based on large scale
experiments on real-world scenes, it can be derived that the estimation
is more accurate using superpixels than when using the entire image.
Abstract: In projects like waterpower, transportation and
mining, etc., proving up the rock-mass structure and hidden tectonic
to estimate the geological body-s activity is very important.
Integrating the seismic results, drilling and trenching data,
CSAMT method was carried out at a planning dame site in southwest
China to evaluate the stability of a deformation. 2D and imitated 3D
inversion resistivity results of CSAMT method were analyzed. The
results indicated that CSAMT was an effective method for defining
an outline of deformation body to several hundred meters deep; the
Lung Pan Deformation was stable in natural conditions; but uncertain
after the future reservoir was impounded.
This research presents a good case study of the fine surveying and
research on complex geological structure and hidden tectonic in
engineering project.
Abstract: A new digital transceiver circuit for asynchronous frame detection is proposed where both the transmitter and receiver contain all digital components, thereby avoiding possible use of conventional devices like monostable multivibrators with unstable external components such as resistances and capacitances. The proposed receiver circuit, in particular, uses a combinational logic block yielding an output which changes its state as soon as the start bit of a new frame is detected. This, in turn, helps in generating an efficient receiver sampling clock. A data latching circuit is also used in the receiver to latch the recovered data bits in any new frame. The proposed receiver structure is also extended from 4- bit information to any general n data bits within a frame with a common expression for the output of the combinational logic block. Performance of the proposed hardware design is evaluated in terms of time delay, reliability and robustness in comparison with the standard schemes using monostable multivibrators. It is observed from hardware implementation that the proposed circuit achieves almost 33 percent speed up over any conventional circuit.
Abstract: The rheological properties, structure and potential synergistic interactions of whey proteins (1-6%) and inulin (20%) in mixed gels in the presence of CaCl2 was the aim of this study. Whey proteins have a strong influence on inulin gel formation. At low concentrations (2%) whey proteins did not impair in inulin gel formation. At higher concentration (4%) whey proteins impaired inulin gelation and inulin impaired the formation of a Ca2+-induced whey protein network. The presence of whey proteins at a level allowing for protein gel network formation (6%) significantly increased the rheological parameters values of the gels. SEM micrographs showed that whey protein structure was coated by inulin moieties which could make the mixed gels firmer. The protein surface hydrophobicity measurements did not exclude synergistic interactions between inulin and whey proteins, however. The use of an electrophoretic technique did not show any stable inulin-whey protein complexes.
Abstract: The resident registration number was adopted for the purposes of enhanced services for resident convenience and effective performance of governmental administrative affairs. However, it has been used for identification purposes customarily and irrationally in line with the development and spread of the Internet. In response to the growing concern about the leakage of collected RRNs and possible abuses of stolen RRNs, e.g. identity theft, for crimes, the Korean Communications Commission began to take legal/regulatory actions in 2011 to minimize the online collection and use of resident registration numbers. As the use of the RRN was limited after the revision of the Act on Promotion of Information and Communications Network Utilization and Information Protection, etc., online business providers were required to have alternatives to the RRN for the purpose of identifying the user's identity and age, in compliance with the law, and settling disputes with customers. This paper presents means of verifying the personal identity by taking advantage of the commonly used infrastructure and simply replacing personal information entered and stored, without requiring users to enter their RRNs.
Abstract: Modern organizations operate under the pressure of
dynamic and often unpredictable changes, both in external and
internal environment. Market success, in this context, requires a
particular competence in the form of flexibility, interpreted here both
on the level of individuals and on the level of organization. This
paper addresses the changes taking place in the sphere of
employment, as observed in economic entities operating on Polish
market. Based on own empirical studies, the authors focus on the
progressing trend of ‘flexibilization’ of employment, particularly in
the context of transformations in organizational structure, designed to
facilitate the transition into management by projects and
differentiation of labor forms.
Abstract: Artificial Neural Network (ANN) has been
extensively used for classification of heart sounds for its
discriminative training ability and easy implementation. However, it
suffers from overparameterization if the number of nodes is not
chosen properly. In such cases, when the dataset has redundancy
within it, ANN is trained along with this redundant information that
results in poor validation. Also a larger network means more
computational expense resulting more hardware and time related
cost. Therefore, an optimum design of neural network is needed
towards real-time detection of pathological patterns, if any from heart
sound signal. The aims of this work are to (i) select a set of input
features that are effective for identification of heart sound signals and
(ii) make certain optimum selection of nodes in the hidden layer for a
more effective ANN structure. Here, we present an optimization
technique that involves Singular Value Decomposition (SVD) and
QR factorization with column pivoting (QRcp) methodology to
optimize empirically chosen over-parameterized ANN structure.
Input nodes present in ANN structure is optimized by SVD followed
by QRcp while only SVD is required to prune undesirable hidden
nodes. The result is presented for classifying 12 common
pathological cases and normal heart sound.
Abstract: Integration of system process information obtained
through an image processing system with an evolving knowledge
database to improve the accuracy and predictability of wear particle
analysis is the main focus of the paper. The objective is to automate
intelligently the analysis process of wear particle using classification
via self organizing maps. This is achieved using relationship
measurements among corresponding attributes of various
measurements for wear particle. Finally, visualization technique is
proposed that helps the viewer in understanding and utilizing these
relationships that enable accurate diagnostics.
Abstract: Integration of system process information obtained
through an image processing system with an evolving knowledge
database to improve the accuracy and predictability of wear debris
analysis is the main focus of the paper. The objective is to automate
intelligently the analysis process of wear particle using classification
via self-organizing maps. This is achieved using relationship
measurements among corresponding attributes of various
measurements for wear debris. Finally, visualization technique is
proposed that helps the viewer in understanding and utilizing these
relationships that enable accurate diagnostics.
Abstract: In this paper we present a generic approach for the problem of the blind estimation of the parameters of linear and convolutional error correcting codes. In a non-cooperative context, an adversary has only access to the noised transmission he has intercepted. The intercepter has no knowledge about the parameters used by the legal users. So, before having acess to the information he has first to blindly estimate the parameters of the error correcting code of the communication. The presented approach has the main advantage that the problem of reconstruction of such codes can be expressed in a very simple way. This allows us to evaluate theorical bounds on the complexity of the reconstruction process but also bounds on the estimation rate. We show that some classical reconstruction techniques are optimal and also explain why some of them have theorical complexities greater than these experimentally observed.
Abstract: Knowing the geometrical object pose of products in manufacturing line before robot manipulation is required and less time consuming for overall shape measurement. In order to perform it, the information of shape representation and matching of objects is become required. Objects are compared with its descriptor that conceptually subtracted from each other to form scalar metric. When the metric value is smaller, the object is considered closed to each other. Rotating the object from static pose in some direction introduce the change of value in scalar metric value of boundary information after feature extraction of related object. In this paper, a proposal method for indexing technique for retrieval of 3D geometrical models based on similarity between boundaries shapes in order to measure 3D CAD object pose using object shape feature matching for Computer Aided Testing (CAT) system in production line is proposed. In experimental results shows the effectiveness of proposed method.