Abstract: The technique of k-anonymization has been proposed to obfuscate private data through associating it with at least k identities. This paper investigates the basic tabular structures that
underline the notion of k-anonymization using cell suppression.
These structures are studied under idealized conditions to identify the
essential features of the k-anonymization notion. We optimize data kanonymization
through requiring a minimum number of anonymized
values that are balanced over all columns and rows. We study the
relationship between the sizes of the anonymized tables, the value k, and the number of attributes. This study has a theoretical value through contributing to develop a mathematical foundation of the kanonymization
concept. Its practical significance is still to be
investigated.
Abstract: Success is a European project that will implement several clean transport offers in three European cities and evaluate the environmental impacts. The goal of these measures is to improve urban mobility or the displacement of residents inside cities. For e.g. park and ride, electric vehicles, hybrid bus and bike sharing etc. A list of 28 criteria and 60 measures has been established for evaluation of these transport projects. The evaluation criteria can be grouped into: Transport, environment, social, economic and fuel consumption. This article proposes a decision support system based that encapsulates a hybrid approach based on fuzzy logic, multicriteria analysis and belief theory for the evaluation of impacts of urban mobility solutions. A web-based tool called DeSSIA (Decision Support System for Impacts Assessment) has been developed that treats complex data. The tool has several functionalities starting from data integration (import of data), evaluation of projects and finishes by graphical display of results. The tool development is based on the concept of MVC (Model, View, and Controller). The MVC is a conception model adapted to the creation of software's which impose separation between data, their treatment and presentation. Effort is laid on the ergonomic aspects of the application. It has codes compatible with the latest norms (XHTML, CSS) and has been validated by W3C (World Wide Web Consortium). The main ergonomic aspect focuses on the usability of the application, ease of learning and adoption. By the usage of technologies such as AJAX (XML and Java Script asynchrones), the application is more rapid and convivial. The positive points of our approach are that it treats heterogeneous data (qualitative, quantitative) from various information sources (human experts, survey, sensors, model etc.).
Abstract: This paper regards the phenomena of intensive suburbanization and urbanization in Olomouc city and in Olomouc region in general for the period of 1986–2009. A Remote Sensing approach that involves tracking of changes in Land Cover units is proposed to quantify the urbanization state and trends in temporal and spatial aspects. It actually consisted of two approaches, Experiment 1 and Experiment 2 which implied two different image classification solutions in order to provide Land Cover maps for each 1986–2009 time split available in the Landsat image set. Experiment 1 dealt with the unsupervised classification, while Experiment 2 involved semi- supervised classification, using a combination of object-based and pixel-based classifiers. The resulting Land Cover maps were subsequently quantified for the proportion of urban area unit and its trend through time, and also for the urban area unit stability, yielding the relation of spatial and temporal development of the urban area unit. Some outcomes seem promising but there is indisputably room for improvements of source data and also processing and filtering.
Abstract: Inferring the network structure from time series data
is a hard problem, especially if the time series is short and noisy.
DNA microarray is a technology allowing to monitor the mRNA
concentration of thousands of genes simultaneously that produces
data of these characteristics. In this study we try to investigate the
influence of the experimental design on the quality of the result.
More precisely, we investigate the influence of two different types of
random single gene perturbations on the inference of genetic networks
from time series data. To obtain an objective quality measure for
this influence we simulate gene expression values with a biologically
plausible model of a known network structure. Within this framework
we study the influence of single gene knock-outs in opposite to
linearly controlled expression for single genes on the quality of the
infered network structure.
Abstract: The medical data statistical analysis often requires the
using of some special techniques, because of the particularities of
these data. The principal components analysis and the data clustering
are two statistical methods for data mining very useful in the medical
field, the first one as a method to decrease the number of studied
parameters, and the second one as a method to analyze the
connections between diagnosis and the data about the patient-s
condition. In this paper we investigate the implications obtained from
a specific data analysis technique: the data clustering preceded by a
selection of the most relevant parameters, made using the principal
components analysis. Our assumption was that, using the principal
components analysis before data clustering - in order to select and to
classify only the most relevant parameters – the accuracy of
clustering is improved, but the practical results showed the opposite
fact: the clustering accuracy decreases, with a percentage
approximately equal with the percentage of information loss reported
by the principal components analysis.
Abstract: Researches on the general rules of temperature field
changing and their effects on the bridge in construction are necessary.
This paper investigated the rules of temperature field changing and its
effects on bridge using onsite measurement and computational
analysis. Guanyinsha Bridge was used as a case study in this research.
The temperature field was simulated in analyses. The effects of certain
boundary conditions such as sun radiance, wind speed, and model
parameters such as heat factor and specific heat on temperature field
are investigated. Recommended values for these parameters are
proposed. The simulated temperature field matches the measured
observations with high accuracy. At the same time, the stresses and
deflections of the bridge computed with the simulated temperature
field matches measured values too. As a conclusion, the temperature
effect analysis of reinforced concrete box girder can be conducted
directly based on the reliable weather data of the concerned area.
Abstract: This paper proposed classification models that would
be used as a proxy for hard disk drive (HDD) functional test equitant
which required approximately more than two weeks to perform the
HDD status classification in either “Pass" or “Fail". These models
were constructed by using committee network which consisted of a
number of single neural networks. This paper also included the
method to solve the problem of sparseness data in failed part, which
was called “enforce learning method". Our results reveal that the
constructed classification models with the proposed method could
perform well in the sparse data conditions and thus the models,
which used a few seconds for HDD classification, could be used to
substitute the HDD functional tests.
Abstract: Motion capture devices have been utilized in
producing several contents, such as movies and video games. However,
since motion capture devices are expensive and inconvenient to use,
motions segmented from captured data was recycled and synthesized
to utilize it in another contents, but the motions were generally
segmented by contents producers in manual. Therefore, automatic
motion segmentation is recently getting a lot of attentions. Previous
approaches are divided into on-line and off-line, where on-line
approaches segment motions based on similarities between
neighboring frames and off-line approaches segment motions by
capturing the global characteristics in feature space. In this paper, we
propose a graph-based high-level motion segmentation method. Since
high-level motions consist of several repeated frames within temporal
distances, we consider all similarities among all frames within the
temporal distance. This is achieved by constructing a graph, where
each vertex represents a frame and the edges between the frames are
weighted by their similarity. Then, normalized cuts algorithm is used
to partition the constructed graph into several sub-graphs by globally
finding minimum cuts. In the experiments, the results using the
proposed method showed better performance than PCA-based method
in on-line and GMM-based method in off-line, as the proposed method
globally segment motions from the graph constructed based
similarities between neighboring frames as well as similarities among
all frames within temporal distances.
Abstract: Evapotranspiration (ET) is a major component of the hydrologic cycle and its accurate estimation is essential for hydrological studies. In past, various estimation methods have been developed for different climatological data, and the accuracy of these methods varies with climatic conditions. Reference crop evapotranspiration (ET0) is a key variable in procedures established for estimating evapotranspiration rates of agricultural crops. Values of ET0 are used with crop coefficients for many aspects of irrigation and water resources planning and management. Numerous methods are used for estimating ET0. As per internationally accepted procedures outlined in the United Nations Food and Agriculture Organization-s Irrigation and Drainage Paper No. 56(FAO-56), use of Penman-Monteith equation is recommended for computing ET0 from ground based climatological observations. In the present study, seven methods have been selected for performance evaluation. User friendly software has been developed using programming language visual basic. The visual basic has ability to create graphical environment using less coding. For given data availability the developed software estimates reference evapotranspiration for any given area and period for which data is available. The accuracy of the software has been checked by the examples given in FAO-56.The developed software is a user friendly tool for estimating ET0 under different data availability and climatic conditions.
Abstract: Wireless Sensor Networks consist of inexpensive, low power sensor nodes deployed to monitor the environment and collect
data. Gathering information in an energy efficient manner is a critical aspect to prolong the network lifetime. Clustering algorithms have an advantage of enhancing the network lifetime. Current clustering algorithms usually focus on global re-clustering and local re-clustering separately. This paper, proposed a combination of those two reclustering methods to reduce the energy consumption of the network. Furthermore, the proposed algorithm can apply to homogeneous as well as heterogeneous wireless sensor networks. In addition, the cluster head rotation happens, only when its energy drops below a dynamic threshold value computed by the algorithm. The simulation result shows that the proposed algorithm prolong the network lifetime compared to existing algorithms.
Abstract: In this paper, we study statistical multiplexing of VBR
video in ATM networks. ATM promises to provide high speed realtime
multi-point to central video transmission for telemedicine
applications in rural hospitals and in emergency medical services.
Video coders are known to produce variable bit rate (VBR) signals
and the effects of aggregating these VBR signals need to be
determined in order to design a telemedicine network infrastructure
capable of carrying these signals. We first model the VBR video
signal and simulate it using a generic continuous-data autoregressive
(AR) scheme. We carry out the queueing analysis by the Fluid
Approximation Model (FAM) and the Markov Modulated Poisson
Process (MMPP). The study has shown a trade off: multiplexing
VBR signals reduces burstiness and improves resource utilization,
however, the buffer size needs to be increased with an associated
economic cost. We also show that the MMPP model and the Fluid
Approximation model fit best, respectively, the cell region and the
burst region. Therefore, a hybrid MMPP and FAM completely
characterizes the overall performance of the ATM statistical
multiplexer. The ramifications of this technology are clear: speed,
reliability (lower loss rate and jitter), and increased capacity in video
transmission for telemedicine. With migration to full IP-based
networks still a long way to achieving both high speed and high
quality of service, the proposed ATM architecture will remain of
significant use for telemedicine.
Abstract: Power line channel is proposed as an alternative for broadband data transmission especially in developing countries like Tanzania [1]. However the channel is affected by stochastic attenuation and deep notches which can lead to the limitation of channel capacity and achievable data rate. Various studies have characterized the channel without giving exactly the maximum performance and limitation in data transfer rate may be this is due to complexity of channel modeling being used. In this paper the channel performance of medium voltage, low voltage and indoor power line channel is presented. In the investigations orthogonal frequency division multiplexing (OFDM) with phase shift keying (PSK) as carrier modulation schemes is considered, for indoor, medium and low voltage channels with typical ten branches and also Golay coding is applied for medium voltage channel. From channels, frequency response deep notches are observed in various frequencies which can lead to reduce the achievable data rate. However, is observed that data rate up to 240Mbps is realized for a signal to noise ratio of about 50dB for indoor and low voltage channels, however for medium voltage a typical link with ten branches is affected by strong multipath and coding is required for feasible broadband data transfer.
Abstract: We present a hardware oriented method for real-time
measurements of object-s position in video. The targeted application
area is light spots used as references for robotic navigation. Different
algorithms for dynamic thresholding are explored in combination
with component labeling and Center Of Gravity (COG) for highest
possible precision versus Signal-to-Noise Ratio (SNR). This method
was developed with a low hardware cost in focus having only one
convolution operation required for preprocessing of data.
Abstract: Accurate software cost estimates are critical to both
developers and customers. They can be used for generating request
for proposals, contract negotiations, scheduling, monitoring and
control. The exact relationship between the attributes of the effort
estimation is difficult to establish. A neural network is good at
discovering relationships and pattern in the data. So, in this paper a
comparative analysis among existing Halstead Model, Walston-Felix
Model, Bailey-Basili Model, Doty Model and Neural Network
Based Model is performed. Neural Network has outperformed the
other considered models. Hence, we proposed Neural Network
system as a soft computing approach to model the effort estimation
of the software systems.
Abstract: OLAP uses multidimensional structures, to provide
access to data for analysis. Traditionally, OLAP operations are more
focused on retrieving data from a single data mart. An exception is
the drill across operator. This, however, is restricted to retrieving
facts on common dimensions of the multiple data marts. Our concern
is to define further operations while retrieving data from multiple
data marts. Towards this, we have defined six operations which
coalesce data marts. While doing so we consider the common as well
as the non-common dimensions of the data marts.
Abstract: Many advanced Routing protocols for wireless sensor networks have been implemented for the effective routing of data. Energy awareness is an essential design issue and almost all of these routing protocols are considered as energy efficient and its ultimate objective is to maximize the whole network lifetime. However, the introductions of video and imaging sensors have posed additional challenges. Transmission of video and imaging data requires both energy and QoS aware routing in order to ensure efficient usage of the sensors and effective access to the gathered measurements. In this paper, the performance of the energy-aware QoS routing Protocol are analyzed in different performance metrics like average lifetime of a node, average delay per packet and network throughput. The parameters considered in this study are end-to-end delay, real time data generation/capture rates, packet drop probability and buffer size. The network throughput for realtime and non-realtime data was also has been analyzed. The simulation has been done in NS2 simulation environment and the simulation results were analyzed with respect to different metrics.
Abstract: This paper considers the benefits gained by using an
efficient quality of service management such as DiffServ technique to
improve the performance of military communications. Low delay and
no blockage must be achieved especially for real time tactical data.
All traffic flows generated by different applications do not need same
bandwidth, same latency, same error ratio and this scalable technique
of packet management based on priority levels is analysed. End to
end architectures supporting various traffic flows and including lowbandwidth
and high-delay HF or SHF military links as well as
unprotected Internet sub domains are studied. A tuning of Diffserv
parameters is proposed in accordance with different loads of various
traffic and different operational situations.
Abstract: The African Great Lakes Region refers to the zone
around lakes Victoria, Tanganyika, Albert, Edward, Kivu, and
Malawi. The main source of electricity in this region is hydropower
whose systems are generally characterized by relatively weak,
isolated power schemes, poor maintenance and technical deficiencies
with limited electricity infrastructures. Most of the hydro sources are
rain fed, and as such there is normally a deficiency of water during
the dry seasons and extended droughts. In such calamities fossil fuels
sources, in particular petroleum products and natural gas, are
normally used to rescue the situation but apart from them being nonrenewable,
they also release huge amount of green house gases to our
environment which in turn accelerates the global warming that has at
present reached an amazing stage. Wind power is ample, renewable,
widely distributed, clean, and free energy source that does not
consume or pollute water. Wind generated electricity is one of the
most practical and commercially viable option for grid quality and
utility scale electricity production. However, the main shortcoming
associated with electric wind power generation is fluctuation in its
output both in space and time. Before making a decision to establish
a wind park at a site, the wind speed features there should therefore
be known thoroughly as well as local demand or transmission
capacity. The main objective of this paper is to utilise monthly
average wind speed data collected from one prospective site within
the African Great Lakes Region to demonstrate that the available
wind power there is high enough to generate electricity. The mean
monthly values were calculated from records gathered on hourly
basis for a period of 5 years (2001 to 2005) from a site in Tanzania.
The documentations that were collected at a height of 2 m were
projected to a height of 50 m which is the standard hub height of
wind turbines. The overall monthly average wind speed was found to
be 12.11 m/s whereas June to November was established to be the
windy season as the wind speed during the session is above the
overall monthly wind speed. The available wind power density
corresponding to the overall mean monthly wind speed was evaluated
to be 1072 W/m2, a potential that is worthwhile harvesting for the
purpose of electric generation.
Abstract: Advances in processors architecture, such as multicore,
increase the size of complexity of parallel computer systems.
With multi-core architecture there are different parallel languages
that can be used to run parallel programs. One of these languages is
OpenMP which embedded in C/Cµ or FORTRAN. Because of this
new architecture and the complexity, it is very important to evaluate
the performance of OpenMP constructs, kernels, and application
program on multi-core systems. Performance is the activity of
collecting the information about the execution characteristics of a
program. Performance tools consists of at least three interfacing
software layers, including instrumentation, measurement, and
analysis. The instrumentation layer defines the measured
performance events. The measurement layer determines what
performance event is actually captured and how it is measured by the
tool. The analysis layer processes the performance data and
summarizes it into a form that can be displayed in performance tools.
In this paper, a number of OpenMP performance tools are surveyed,
explaining how each is used to collect, analyse, and display data
collection.
Abstract: Multicarrier transmission system such as Orthogonal
Frequency Division Multiplexing (OFDM) is a promising technique
for high bit rate transmission in wireless communication systems.
OFDM is a spectrally efficient modulation technique that can achieve
high speed data transmission over multipath fading channels without
the need for powerful equalization techniques. A major drawback
of OFDM is the high Peak-to-Average Power Ratio (PAPR) of the
transmit signal which can significantly impact the performance of the
power amplifier. In this paper we have compared the PAPR reduction
performance of Golay and Reed-Muller coded OFDM signal. From
our simulation it has been found that the PAPR reduction performance
of Golay coded OFDM is better than the Reed-Muller coded OFDM
signal. Moreover, for the optimum PAPR reduction performance, code
configuration for Golay and Reed-Muller codes has been identified.