Abstract: Data mining is the procedure of determining interesting patterns from the huge amount of data. With the intention of accessing the data faster the most supporting processes needed is clustering. Clustering is the process of identifying similarity between data according to the individuality present in the data and grouping associated data objects into clusters. Cluster ensemble is the technique to combine various runs of different clustering algorithms to obtain a general partition of the original dataset, aiming for consolidation of outcomes from a collection of individual clustering outcomes. The performances of clustering ensembles are mainly affecting by two principal factors such as diversity and quality. This paper presents the overview about the different cluster ensemble algorithm along with their methods used in cluster ensemble to improve the diversity and quality in the several cluster ensemble related papers and shows the comparative analysis of different cluster ensemble also summarize various cluster ensemble methods. Henceforth this clear analysis will be very useful for the world of clustering experts and also helps in deciding the most appropriate one to determine the problem in hand.
Abstract: This study will provide a brief description of the stress in Najdi Arabic dialect as well as Modern Standard Arabic. Beyond the analysis of stress patterns, this paper will also attempt to deal with two important phenomena that affect stress, namely epenthesis/insertion, vowel shortening, and consonant (the glottal stop) deletion.
Abstract: Wireless Body Area Network (WBAN) is a short-range
wireless communication around human body for various applications
such as wearable devices, entertainment, military, and especially
medical devices. WBAN attracts the attention of continuous health
monitoring system including diagnostic procedure, early detection of
abnormal conditions, and prevention of emergency situations.
Compared to cellular network, WBAN system is more difficult to
control inter- and inner-cell interference due to the limited power,
limited calculation capability, mobility of patient, and
non-cooperation among WBANs.
In this paper, we compare the performance of resource allocation
scheme based on several Pseudo Orthogonal Codewords (POCs) to
mitigate inter-WBAN interference. Previously, the POCs are widely
exploited for a protocol sequence and optical orthogonal code. Each
POCs have different properties of auto- and cross-correlation and
spectral efficiency according to its construction of POCs. To identify
different WBANs, several different pseudo orthogonal patterns based
on POCs exploits for resource allocation of WBANs. By simulating
these pseudo orthogonal resource allocations of WBANs on
MATLAB, we obtain the performance of WBANs according to
different POCs and can analyze and evaluate the suitability of POCs
for the resource allocation in the WBANs system.
Abstract: This study attempts to consider the linkage between management and computer sciences in order to develop the software named “IntelSymb” as a demo application to prove data analysis of non-energy* fields’ diversification, which will positively influence on energy dependency mitigation of countries. Afterward, we analyzed 18 years of economic fields of development (5 sectors) of 13 countries by identifying which patterns mostly prevailed and which can be dominant in the near future. To make our analysis solid and plausible, as a future work, we suggest developing a gateway or interface, which will be connected to all available on-line data bases (WB, UN, OECD, U.S. EIA) for countries’ analysis by fields. Sample data consists of energy (TPES and energy import indicators) and non-energy industries’ (Main Science and Technology Indicator, Internet user index, and Sales and Production indicators) statistics from 13 OECD countries over 18 years (1995-2012). Our results show that the diversification of non-energy industries can have a positive effect on energy sector dependency (energy consumption and import dependence on crude oil) deceleration. These results can provide empirical and practical support for energy and non-energy industries diversification’ policies, such as the promoting of Information and Communication Technologies (ICTs), services and innovative technologies efficiency and management, in other OECD and non-OECD member states with similar energy utilization patterns and policies. Industries, including the ICT sector, generate around 4 percent of total GHG, but this is much higher — around 14 percent — if indirect energy use is included. The ICT sector itself (excluding the broadcasting sector) contributes approximately 2 percent of global GHG emissions, at just under 1 gigatonne of carbon dioxide equivalent (GtCO2eq). Ergo, this can be a good example and lesson for countries which are dependent and independent on energy, and mainly emerging oil-based economies, as well as to motivate non-energy industries diversification in order to be ready to energy crisis and to be able to face any economic crisis as well.
Abstract: The very well-known stacked sets of numbers referred
to as Pascal’s triangle present the coefficients of the binomial
expansion of the form (x+y)n. This paper presents an approach (the
Staircase Horizontal Vertical, SHV-method) to the generalization of
planar Pascal’s triangle for polynomial expansion of the form
(x+y+z+w+r+⋯)n. The presented generalization of Pascal’s triangle
is different from other generalizations of Pascal’s triangles given in
the literature. The coefficients of the generalized Pascal’s triangles,
presented in this work, are generated by inspection, using embedded
Pascal’s triangles. The coefficients of I-variables expansion are
generated by horizontally laying out the Pascal’s elements of (I-1)
variables expansion, in a staircase manner, and multiplying them with
the relevant columns of vertically laid out classical Pascal’s elements,
hence avoiding factorial calculations for generating the coefficients
of the polynomial expansion. Furthermore, the classical Pascal’s
triangle has some pattern built into it regarding its odd and even
numbers. Such pattern is known as the Sierpinski’s triangle. In this
study, a presentation of Sierpinski-like patterns of the generalized
Pascal’s triangles is given. Applications related to those coefficients
of the binomial expansion (Pascal’s triangle), or polynomial
expansion (generalized Pascal’s triangles) can be in areas of
combinatorics, and probabilities.
Abstract: Energy consumption data, in particular those involving
public buildings, are impacted by many factors: the building structure,
climate/environmental parameters, construction, system operating
condition, and user behavior patterns. Traditional methods for data
analysis are insufficient. This paper delves into the data mining
technology to determine its application in the analysis of building
energy consumption data including energy consumption prediction,
fault diagnosis, and optimal operation. Recent literature are reviewed
and summarized, the problems faced by data mining technology in the
area of energy consumption data analysis are enumerated, and research
points for future studies are given.
Abstract: Landfill waste is a common problem as it has an
economic and environmental impact even if it is closed. Landfill
waste contains a high density of various persistent compounds such
as heavy metals, organic and inorganic materials. As persistent
compounds are slowly-degradable or even non-degradable in the
environment, they often produce sublethal or even lethal effects on
aquatic organisms. The aims of the present study were to estimate
sublethal effects of the Kairiai landfill (WGS: 55°55‘46.74“,
23°23‘28.4“) leachate on the locomotor activity of rainbow trout
Oncorhynchus mykiss juveniles using the original system package
developed in our laboratory for automated monitoring, recording and
analysis of aquatic organisms’ activity, and to determine patterns of
fish behavioral response to sublethal effects of leachate. Four
different concentrations of leachate were chosen: 0.125; 0.25; 0.5 and
1.0 mL/L (0.0025; 0.005; 0.01 and 0.002 as part of 96-hour LC50,
respectively). Locomotor activity was measured after 5, 10 and 30
minutes of exposure during 1-minute test-periods of each fish (7 fish
per treatment). The threshold-effect-concentration amounted to 0.18
mL/L (0.0036 parts of 96-hour LC50). This concentration was found
to be even 2.8-fold lower than the concentration generally assumed to
be “safe” for fish. At higher concentrations, the landfill leachate
solution elicited behavioral response of test fish to sublethal levels of
pollutants. The ability of the rainbow trout to detect and avoid
contaminants occurred after 5 minutes of exposure. The intensity of
locomotor activity reached a peak within 10 minutes, evidently
decreasing after 30 minutes. This could be explained by the
physiological and biochemical adaptation of fish to altered
environmental conditions. It has been established that the locomotor
activity of juvenile trout depends on leachate concentration and
exposure duration. Modeling of these parameters showed that the
activity of juveniles increased at higher leachate concentrations, but
slightly decreased with the increasing exposure duration. Experiment
results confirm that the behavior of rainbow trout juveniles is a
sensitive and rapid biomarker that can be used in combination with
the system for fish behavior monitoring, registration and analysis to
determine sublethal concentrations of pollutants in ambient water.
Further research should be focused on software improvement aimed
to include more parameters of aquatic organisms’ behavior and to
investigate the most rapid and appropriate behavioral responses in
different species. In practice, this study could be the basis for the
development and creation of biological early-warning systems
(BEWS).
Abstract: In language learning, second language learners as well
as Native speakers commit errors in their attempt to achieve
competence in the target language. The realm of collocation has to do
with meaning relation between lexical items. In all human language,
there is a kind of ‘natural order’ in which words are arranged or relate
to one another in sentences so much so that when a word occurs in a
given context, the related or naturally co-occurring word will
automatically come to the mind. It becomes an error, therefore, if
students inappropriately pair or arrange such ‘naturally’ co–occurring
lexical items in a text. It has been observed that most of the second
language learners in this research group commit collocation errors. A
study of this kind is very significant as it gives insight into the kinds
of errors committed by learners. This will help the language teacher
to be able to identify the sources and causes of such errors as well as
correct them thereby guiding, helping and leading the learners
towards achieving some level of competence in the language. The
aim of the study is to understand the nature of these errors as
stumbling blocks to effective essay writing. The objective of the
study is to identify the errors, analyze their structural compositions so
as to determine whether there are similarities between students in this
regard and to find out whether there are patterns to these kinds of
errors which will enable the researcher to understand their sources
and causes. As a descriptive research, the researcher samples some
nine hundred essays collected from three hundred undergraduate
learners of English as a second language in the Federal College of
Education, Kano, North- West Nigeria, i.e. three essays per each
student. The essays which were given on three different lecture times
were of similar thematic preoccupations (i.e. same topics) and length
(i.e. same number of words). The essays were written during the
lecture hour at three different lecture occasions. The errors were
identified in a systematic manner whereby errors so identified were
recorded only once even if they occur severally in students’ essays.
The data was collated using percentages in which the identified
numbers of occurrences were converted accordingly in percentages.
The findings from the study indicate that there are similarities as well
as regular and repeated errors which provided a pattern. Based on the
pattern identified, the conclusion is that students’ collocation errors
are attributable to poor teaching and learning which resulted in wrong
generalization of rules.
Abstract: In this research, we propose to conduct diagnostic and
predictive analysis about the key factors and consequences of urban
population relocation. To achieve this goal, urban simulation models
extract the urban development trends as land use change patterns from
a variety of data sources. The results are treated as part of urban big
data with other information such as population change and economic
conditions. Multiple data mining methods are deployed on this data to
analyze nonlinear relationships between parameters. The result
determines the driving force of population relocation with respect to
urban sprawl and urban sustainability and their related parameters.
This work sets the stage for developing a comprehensive urban
simulation model for catering to specific questions by targeted users. It
contributes towards achieving sustainability as a whole.
Abstract: Web Usage Mining is the application of data mining
techniques to find usage patterns from web log data, so as to grasp
required patterns and serve the requirements of Web-based
applications. User’s expertise on the internet may be improved by
minimizing user’s web access latency. This may be done by
predicting the future search page earlier and the same may be prefetched
and cached. Therefore, to enhance the standard of web
services, it is needed topic to research the user web navigation
behavior. Analysis of user’s web navigation behavior is achieved
through modeling web navigation history. We propose this technique
which cluster’s the user sessions, based on the K-medoids technique.
Abstract: The purpose of this study is to analyze the temporal
and spatial variability of thermal conditions in the Republic of
Armenia. The paper describes annual fluctuations in air temperature.
Research has been focused on case study region of Armenia and
surrounding areas, where long–term measurements and observations
of weather conditions have been performed within the National
Meteorological Service of Armenia and its surrounding areas. The
study contains yearly air temperature data recorded between 1961-
2012. Mann-Kendal test and the autocorrelation function were
applied to detect the change trend of annual mean temperature, as
well as other parametric and non-parametric tests searching to find
the presence of some breaks in the long term evolution of
temperature. The analysis of all records reveals a tendency mostly
towards warmer years, with increased temperatures especially in
valleys and inner basins. The maximum temperature increase is up to
1,5°C. Negative results have not been observed in Armenia. The
patterns of temperature change have been observed since the 1990’s
over much of the Armenian territory. The climate in Armenia was
influenced by global change in the last 2 decades, as results from the
methods employed within the study.
Abstract: A sign pattern is a matrix whose entries belong to the set
{+,−, 0}. An n-by-n sign pattern A is said to allow an eventually
positive matrix if there exist some real matrices A with the same
sign pattern as A and a positive integer k0 such that Ak > 0 for all
k ≥ k0. It is well known that identifying and classifying the n-by-n
sign patterns that allow an eventually positive matrix are posed as two
open problems. In this article, the tree sign patterns of small order
that allow an eventually positive matrix are classified completely.
Abstract: The paper focuses on the benefits of business process
modeling. Although this discipline is developing for many years,
there is still necessity of creating new opportunities to meet the ever
increasing users’ needs. Because one of these needs is related to the
conversion of business process models from one standard to another,
the authors have developed a converter between BPMN and EPC
standards using workflow patterns as intermediate tool. Nowadays
there are too many systems for business process modeling. The
variety of output formats is almost the same as the systems
themselves. This diversity additionally hampers the conversion of the
models. The presented study is aimed at discussing problems due to
differences in the output formats of various modeling environments.
Abstract: In the Solid-State-Drive (SSD) performance, whether
the data has been well parallelized is an important factor. SSD
parallelization is affected by allocation scheme and it is directly
connected to SSD performance. There are dynamic allocation and
static allocation in representative allocation schemes. Dynamic
allocation is more adaptive in exploiting write operation parallelism,
while static allocation is better in read operation parallelism.
Therefore, it is hard to select the appropriate allocation scheme when
the workload is mixed read and write operations. We simulated
conditions on a few mixed data patterns and analyzed the results to
help the right choice for better performance. As the results, if data
arrival interval is long enough prior operations to be finished and
continuous read intensive data environment static allocation is more
suitable. Dynamic allocation performs the best on write performance
and random data patterns.
Abstract: This paper is concerned with the ways in which
Assisted Reproductive Technologies (ARTs) affect women’s lives
and perceptions regarding their infertility, contraception and
reproductive health. The paper is based on a qualitative feminist
survey study to explore and analyze issues arising from the use of
ARTs by women in New Delhi, the capital of India. A rapid growth in
the number of fertility clinics has been noticed recently. A critical
analysis of interviews revealed that these technologies are used and
developed for making profits at the cost of women’s lives. In this
way, these technologies are influencing and changing the traditional
patterns of motherhood requiring a rethinking about new ways of
reproduction introduced through the use of ARTs.
Abstract: A Multi-dimensional computational fluid dynamics
(CFD) two-phase model was developed with the aim to simulate
the in-core coolant circuit of a pressurized heavy water reactor
(PHWR) of a commercial nuclear power plant (NPP). Due to the
fact that this PHWR is a Reactor Pressure Vessel type (RPV),
three-dimensional (3D) detailed modelling of the large reservoirs of
the RPV (the upper and lower plenums and the downcomer) were
coupled with an in-house finite volume one-dimensional (1D) code
in order to model the 451 coolant channels housing the nuclear fuel.
Regarding the 1D code, suitable empirical correlations for taking into
account the in-channel distributed (friction losses) and concentrated
(spacer grids, inlet and outlet throttles) pressure losses were used.
A local power distribution at each one of the coolant channels
was also taken into account. The heat transfer between the coolant
and the surrounding moderator was accurately calculated using a
two-dimensional theoretical model. The implementation of subcooled
boiling and condensation models in the 1D code along with the use
of functions for representing the thermal and dynamic properties of
the coolant and moderator (heavy water) allow to have estimations
of the in-core steam generation under nominal flow conditions for a
generic fission power distribution. The in-core mass flow distribution
results for steady state nominal conditions are in agreement with the
expected from design, thus getting a first assessment of the coupled
1/3D model. Results for nominal condition were compared with
those obtained with a previous 1/3D single-phase model getting more
realistic temperature patterns, also allowing visualize low values of
void fraction inside the upper plenum. It must be mentioned that the
current results were obtained by imposing prescribed fission power
functions from literature. Therefore, results are showed with the aim
of point out the potentiality of the developed model.
Abstract: This research tested the performance of alternative
warehouse designs concerning the picking process. The chosen
performance measures were Travel Distance and Total Fulfilment
Time. An explanatory case study was built up around a model
implemented with SIMUL8. Hypotheses were set by selecting
outcomes from the literature survey matching popular empirical
findings. 17.4% reductions were found for Total Fulfilment Time and
Resource Utilisation. The latter was then used as a proxy for
operational efficiency. Literal replication of theoretical data-patterns
was considered as an internal validity sign. Assessing the estimated
changes benefits ahead of implementation was found to be a
contribution to practice.
Abstract: Background: To compare the thinning patterns of the
ganglion cell-inner plexiform layer (GCIPL) and peripapillary retinal
nerve fiber layer (pRNFL) as measured using Cirrus high-definition
optical coherence tomography (HD-OCT) in patients with visual field
(VF) defects that respect the vertical meridian. Methods: Twenty eyes of eleven patients with VF defects that
respect the vertical meridian were enrolled retrospectively. The
thicknesses of the macular GCIPL and pRNFL were measured using
Cirrus HD-OCT. The 5% and 1% thinning area index (TAI) was
calculated as the proportion of abnormally thin sectors at the 5% and
1% probability level within the area corresponding to the affected VF.
The 5% and 1% TAI were compared between the GCIPL and pRNFL
measurements. Results: The color-coded GCIPL deviation map showed a
characteristic vertical thinning pattern of the GCIPL, which is also
seen in the VF of patients with brain lesions. The 5% and 1% TAI
were significantly higher in the GCIPL measurements than in the
pRNFL measurements (all P < 0.01). Conclusions: Macular GCIPL analysis clearly visualized a
characteristic topographic pattern of retinal ganglion cell (RGC) loss
in patients with VF defects that respect the vertical meridian, unlike
pRNFL measurements. Macular GCIPL measurements provide more
valuable information than pRNFL measurements for detecting the
loss of RGCs in patients with retrograde degeneration of the optic
nerve fibers.
Abstract: The traditional rhythms of the West African country
of Guinea have played a centuries-long role in defining the different
people groups that make up the country. Throughout their history,
before and since colonization by the French, the different ethnicities
have used their traditional music as a distinct part of their historical
identities. That is starting to change. Guinea is an impoverished
nation created in the early twentieth-century with little regard for the
history and cultures of the people who were included. The traditional
rhythms of the different people groups and their heritages have
remained. Fifteen individual traditional Guinean rhythms were
chosen to represent popular rhythms from the four geographical
regions of Guinea. Each rhythm was traced back to its native village
and video recorded on-site by as many different local performing
groups as could be located. The cyclical patterns rhythms were
transcribed via a circular, spatial design and then copied into a box
notation system where sounds happening at the same time could be
studied. These rhythms were analyzed for their consistency-overperformance
in a Fundamental Rhythm Pattern analysis so rhythms
could be compared for how they are changing through different
performances. The analysis showed that the traditional rhythm
performances of the Middle and Forest Guinea regions were the most
cohesive and showed the least evidence of change between
performances. The role of music in each of these regions is both
limited and focused. The Coastal and High Guinea regions have
much in common historically through their ethnic history and
modern-day trade connections, but the rhythm performances seem to
be less consistent and demonstrate more changes in how they are
performed today. In each of these regions the role and usage of music
is much freer and wide-spread. In spite of advances being made as a
country, different ethnic groups still frequently only respond and
participate (dance and sing) to the music of their native ethnicity.
There is some evidence that this self-imposed musical barrier is
beginning to change and evolve, partially through the development of
better roads, more access to electricity and technology, the nationwide
Ebola health crisis, and a growing self-identification as a
unified nation.
Abstract: This paper integrates Octagon and Square Search
pattern (OCTSS) motion estimation algorithm into H.264/AVC
(Advanced Video Coding) video codec in Adaptive Group of Pictures
(AGOP) mode. AGOP structure is computed based on scene change
in the video sequence. Octagon and square search pattern block-based
motion estimation method is implemented in inter-prediction process
of H.264/AVC. Both these methods reduce bit rate and computational
complexity while maintaining the quality of the video sequence
respectively. Experiments are conducted for different types of video
sequence. The results substantially proved that the bit rate,
computation time and PSNR gain achieved by the proposed method
is better than the existing H.264/AVC with fixed GOP and AGOP.
With a marginal gain in quality of 0.28dB and average gain in bitrate
of 132.87kbps, the proposed method reduces the average computation
time by 27.31 minutes when compared to the existing state-of-art
H.264/AVC video codec.