Abstract: Feature selection and attribute reduction are crucial
problems, and widely used techniques in the field of machine
learning, data mining and pattern recognition to overcome the
well-known phenomenon of the Curse of Dimensionality. This paper
presents a feature selection method that efficiently carries out attribute
reduction, thereby selecting the most informative features of a dataset.
It consists of two components: 1) a measure for feature subset
evaluation, and 2) a search strategy. For the evaluation measure,
we have employed the fuzzy-rough dependency degree (FRFDD)
of the lower approximation-based fuzzy-rough feature selection
(L-FRFS) due to its effectiveness in feature selection. As for the
search strategy, a modified version of a binary shuffled frog leaping
algorithm is proposed (B-SFLA). The proposed feature selection
method is obtained by hybridizing the B-SFLA with the FRDD. Nine
classifiers have been employed to compare the proposed approach
with several existing methods over twenty two datasets, including
nine high dimensional and large ones, from the UCI repository.
The experimental results demonstrate that the B-SFLA approach
significantly outperforms other metaheuristic methods in terms of the
number of selected features and the classification accuracy.
Abstract: Focus on reducing energy consumption in existing
buildings at large scale, e.g. in cities or countries, has been
increasing in recent years. In order to reduce energy consumption
in existing buildings, political incentive schemes are put in place and
large scale investments are made by utility companies. Prioritising
these investments requires a comprehensive overview of the energy
consumption in the existing building stock, as well as potential
energy-savings. However, a building stock comprises thousands
of buildings with different characteristics making it difficult to
model energy consumption accurately. Moreover, the complexity of
the building stock makes it difficult to convey model results to
policymakers and other stakeholders. In order to manage the complexity of the building stock, building
archetypes are often employed in building stock energy models
(BSEMs). Building archetypes are formed by segmenting the building
stock according to specific characteristics. Segmenting the building
stock according to building type and building age is common, among
other things because this information is often easily available. This
segmentation makes it easy to convey results to non-experts. However, using a single archetypical building to represent all
buildings in a segment of the building stock is associated with
loss of detail. Thermal characteristics are aggregated while other
characteristics, which could affect the energy efficiency of a building,
are disregarded. Thus, using a simplified representation of the
building stock could come at the expense of the accuracy of the
model. The present study evaluates the accuracy of a conventional
archetype-based BSEM that segments the building stock according
to building type- and age. The accuracy is evaluated in terms of the
archetypes’ ability to accurately emulate the average energy demands
of the corresponding buildings they were meant to represent. This is
done for the buildings’ energy demands as a whole as well as for
relevant sub-demands. Both are evaluated in relation to the type- and
the age of the building. This should provide researchers, who use
archetypes in BSEMs, with an indication of the expected accuracy
of the conventional archetype model, as well as the accuracy lost in
specific parts of the calculation, due to use of the archetype method.
Abstract: Knee collateral ligaments play a significant role in restraining excessive frontal motion (varus/valgus rotations). In this investigation, a multiscale frame was developed based on structural hierarchies of the collateral ligaments starting from the bottom (tropocollagen molecule) to up where the fibred reinforced structure established. Experimental data of failure tensile test were considered as the principal driver of the developed model. This model was calibrated statistically using Bayesian calibration due to the high number of unknown parameters. Then the model is scaled up to fit the real structure of the collateral ligaments and simulated under realistic boundary conditions. Predications have been successful in describing the observed transient response of the collateral ligaments during tensile test under pre- and post-damage loading conditions. Collateral ligaments maximum stresses and strengths were observed near to the femoral insertions, a results that is in good agreement with experimental investigations. Also for the first time, damage initiation and propagation were documented with this model as a function of the cross-link density between tropocollagen molecules.
Abstract: In the present study, the reactions involving 4-bromo-2,6-bis-hydroxymethyl-phenol (BBHMP) and nicotinamide (NA) in the presence of Pd(II) ions were investigated. Optimum conditions for the reactions were established as pH = 7 and λ = 450 nm. According to absorbance measurements, the molar ratio of BBHMP: NA: Pd2+ was found to be 1: 2: 2. As a result of physicochemical, spectrophotometric and thermal analyses, the reactions of BBHMP and NA with Pd(II) are complexation reactions and one molecule of BBHMP and two molecules of NA react with two molecules of the Pd(II) ion.
Abstract: One of the major shortcomings of widely used
scientometric indicators is that different disciplines cannot be
compared with each other. The issue of cross-disciplinary
normalization has been long discussed, but even the classification
of publications into scientific domains poses problems. Structural
properties of citation networks offer new possibilities, however, the
large size and constant growth of these networks asks for precaution.
Here we present a new tool that in order to perform cross-field
normalization of scientometric indicators of individual publications
relays on the structural properties of citation networks. Due to the
large size of the networks, a systematic procedure for identifying
scientific domains based on a local community detection algorithm
is proposed. The algorithm is tested with different benchmark
and real-world networks. Then, by the use of this algorithm, the
mechanism of the scientometric indicator normalization process is
shown for a few indicators like the citation number, P-index and
a local version of the PageRank indicator. The fat-tail trend of the
article indicator distribution enables us to successfully perform the
indicator normalization process.
Abstract: This paper deals with measuring and modelling of the quality of the machined surface of the metal machining process. The average surface roughness (Ra) which represents the quality of the machined part was measured during the dry turning of the AISI 4140 steel. A large number of factors with the unknown relations among them influences this parameter, and that is why mathematical modelling is extremely complicated. Different values of cutting speed, feed rate, depth of cut (cutting regime) and workpiece hardness causes different surface roughness values. Modelling with soft computing techniques may be very useful in such cases. This paper presents the usage of the fuzzy logic-based system for determining metal machining process parameter in order to find the proper values of cutting regimes.
Abstract: Many research projects require accurate delineation of the different land cover type of the agricultural area. Especially it is critically important for the definition of specific plants like cannabis. However, the complexity of vegetation stands structure, abundant vegetation species, and the smooth transition between different seconder section stages make vegetation classification difficult when using traditional approaches such as the maximum likelihood classifier. Most of the time, classification distinguishes only between trees/annual or grain. It has been difficult to accurately determine the cannabis mixed with other plants. In this paper, a mixed distribution models approach is applied to classify pure and mix cannabis parcels using Worldview-2 imagery in the Lakes region of Turkey. Five different land use types (i.e. sunflower, maize, bare soil, and cannabis) were identified in the image. A constrained Gaussian mixture discriminant analysis (GMDA) was used to unmix the image. In the study, 255 reflectance ratios derived from spectral signatures of seven bands (Blue-Green-Yellow-Red-Rededge-NIR1-NIR2) were randomly arranged as 80% for training and 20% for test data. Gaussian mixed distribution model approach is proved to be an effective and convenient way to combine very high spatial resolution imagery for distinguishing cannabis vegetation. Based on the overall accuracies of the classification, the Gaussian mixed distribution model was found to be very successful to achieve image classification tasks. This approach is sensitive to capture the illegal cannabis planting areas in the large plain. This approach can also be used for monitoring and determination with spectral reflections in illegal cannabis planting areas.
Abstract: The study of static dielectric properties in a binary mixture of 1,2 dichloroethane (DE) and n,n dimethylformamide (DMF) polar liquids has been carried out in the frequency range of 10 MHz to 30 GHz for 11 different concentration using time domain reflectometry technique at 10ºC temperature. The dielectric relaxation study of solute-solvent mixture at microwave frequencies gives information regarding the creation of monomers and multimers as well as interaction between the molecules of the binary mixture. The least squares fit method is used to determine the values of dielectric parameters such as static dielectric constant (ε0), dielectric constant at high frequency (ε∞) and relaxation time (τ).
Abstract: Human motion recognition has been extensively increased in recent years due to its importance in a wide range of applications, such as human-computer interaction, intelligent surveillance, augmented reality, content-based video compression and retrieval, etc. However, it is still regarded as a challenging task especially in realistic scenarios. It can be seen as a general machine learning problem which requires an effective human motion representation and an efficient learning method. In this work, we introduce a descriptor based on Laban Movement Analysis technique, a formal and universal language for human movement, to capture both quantitative and qualitative aspects of movement. We use Discrete Hidden Markov Model (DHMM) for training and classification motions. We improve the classification algorithm by proposing two DHMMs for each motion class to process the motion sequence in two different directions, forward and backward. Such modification allows avoiding the misclassification that can happen when recognizing similar motions. Two experiments are conducted. In the first one, we evaluate our method on a public dataset, the Microsoft Research Cambridge-12 Kinect gesture data set (MSRC-12) which is a widely used dataset for evaluating action/gesture recognition methods. In the second experiment, we build a dataset composed of 10 gestures(Introduce yourself, waving, Dance, move, turn left, turn right, stop, sit down, increase velocity, decrease velocity) performed by 20 persons. The evaluation of the system includes testing the efficiency of our descriptor vector based on LMA with basic DHMM method and comparing the recognition results of the modified DHMM with the original one. Experiment results demonstrate that our method outperforms most of existing methods that used the MSRC-12 dataset, and a near perfect classification rate in our dataset.
Abstract: What people say on social media has turned into a
rich source of information to understand social behavior. Specifically,
the growing use of Twitter social media for political communication
has arisen high opportunities to know the opinion of large numbers
of politically active individuals in real time and predict the global
political tendencies of a specific country. It has led to an increasing
body of research on this topic. The majority of these studies have
been focused on polarized political contexts characterized by only
two alternatives. Unlike them, this paper tackles the challenge
of forecasting Spanish political trends, characterized by multiple
political parties, by means of analyzing the Twitters Users political
tendency. According to this, a new strategy, named Tweets Analysis
Strategy (TAS), is proposed. This is based on analyzing the users
tweets by means of discovering its sentiment (positive, negative or
neutral) and classifying them according to the political party they
support. From this individual political tendency, the global political
prediction for each political party is calculated. In order to do this,
two different strategies for analyzing the sentiment analysis are
proposed: one is based on Positive and Negative words Matching
(PNM) and the second one is based on a Neural Networks Strategy
(NNS). The complete TAS strategy has been performed in a Big-Data
environment. The experimental results presented in this paper reveal
that NNS strategy performs much better than PNM strategy to analyze
the tweet sentiment. In addition, this research analyzes the viability
of the TAS strategy to obtain the global trend in a political context
make up by multiple parties with an error lower than 23%.
Abstract: Various processes are modelled using a discrete phase,
where particles are seeded from a source. Such particles can represent
liquid water droplets, which are affecting the continuous phase by
exchanging thermal energy, momentum, species etc. Discrete phases
are typically modelled using parcel, which represents a collection of
particles, which share properties such as temperature, velocity etc.
When coupling the phases, the exchange rates are integrated over
the cell, in which the parcel is located. This can cause spikes and
fluctuating exchange rates. This paper presents an alternative method of coupling a discrete
and a continuous plug flow phase. This is done using triangular
parcels, which span between nodes following the dynamics of single
droplets. Thus, the triangular parcels are propagated using the corner
nodes. At each time step, the exchange rates are spatially integrated
over the surface of the triangular parcels, which yields a smooth
continuous exchange rate to the continuous phase. The results shows that the method is more stable, converges
slightly faster and yields smooth exchange rates compared with
the steam tube approach. However, the computational requirements
are about five times greater, so the applicability of the alternative
method should be limited to processes, where the exchange rates are
important. The overall balances of the exchanged properties did not
change significantly using the new approach.
Abstract: Classification of high resolution polarimetric Synthetic Aperture Radar (PolSAR) images plays an important role in land cover and land use management. Recently, classification algorithms based on Bag of Visual Words (BOVW) model have attracted significant interest among scholars and researchers in and out of the field of remote sensing. In this paper, BOVW model with pixel based low-level features has been implemented to classify a subset of San Francisco bay PolSAR image, acquired by RADARSAR 2 in C-band. We have used segment-based decision-making strategy and compared the result with the result of traditional Support Vector Machine (SVM) classifier. 90.95% overall accuracy of the classification with the proposed algorithm has shown that the proposed algorithm is comparable with the state-of-the-art methods. In addition to increase in the classification accuracy, the proposed method has decreased undesirable speckle effect of SAR images.
Abstract: Over the past decade, there has been a steep rise in
the data-driven analysis in major areas of medicine, such as clinical
decision support system, survival analysis, patient similarity analysis,
image analytics etc. Most of the data in the field are well-structured
and available in numerical or categorical formats which can be used
for experiments directly. But on the opposite end of the spectrum,
there exists a wide expanse of data that is intractable for direct
analysis owing to its unstructured nature which can be found in the
form of discharge summaries, clinical notes, procedural notes which
are in human written narrative format and neither have any relational
model nor any standard grammatical structure. An important step
in the utilization of these texts for such studies is to transform
and process the data to retrieve structured information from the
haystack of irrelevant data using information retrieval and data mining
techniques. To address this problem, the authors present Q-Map in
this paper, which is a simple yet robust system that can sift through
massive datasets with unregulated formats to retrieve structured
information aggressively and efficiently. It is backed by an effective
mining technique which is based on a string matching algorithm
that is indexed on curated knowledge sources, that is both fast
and configurable. The authors also briefly examine its comparative
performance with MetaMap, one of the most reputed tools for medical
concepts retrieval and present the advantages the former displays over
the latter.
Abstract: The difficulties in riding small wheel bicycles and their lesser stability have been perceived for a long time. Although small wheel bicycles are designed using the similar approach and guidelines that have worked well for big wheel bicycles, the performance of the big wheelers and the smaller wheelers are markedly different. Since both the big wheelers and small wheelers have same fundamental geometry, most blame the small wheel for this discrepancy in the performance. This paper reviews existing guidelines for bicycle design, especially the front steering geometry for the bicycle, and provides a systematic and quantitative analysis of different wheel sized bicycles. A validated mathematical model has been used as a tool to assess the dynamic performance of the bicycles in term of their self-stability. The results obtained were found to corroborate the subjective perception of cyclists for small wheel bicycles. The current approach for small wheel bicycle design requires higher speed to be self-stable. However, it was found that increasing the headtube angle and selecting a proper trail could improve the dynamic performance of small wheel bicycles. A range of parameters for front steering geometry has been identified for small wheel bicycles that have comparable stability as big wheel bicycles. Interestingly, most of the identified geometries are found to be beyond the ISO recommended range and seem to counter the current approach of small wheel bicycle design. Therefore, it was successfully shown that the guidelines for big wheelers do not translate directly to small wheelers, but careful selection of the front geometry could make small wheel bicycles as stable as big wheel bicycles.
Abstract: This study has the objective of assessing the effect of kaizen (5S, Muda elimination and Quality Control Circle (QCC) on employees’ affective attitude (job satisfaction, commitment and job stress) in Kombolcha Textile Share Company. A conceptual model was developed to describe the relationship between Kaizen and Employees’ Affective Attitude (EAA) factors. The three factors of Employee Affective Attitude were measured using questionnaire derived from other validated questionnaire. In the data collection to conduct this study; questionnaire, unstructured interview, written documents and direct observations are used. To analyze the data, SPSS and Microsoft Excel were used. In addition, the internal consistency of similar items in the questionnaire instrument was measured for their equivalence by using the cronbach’s alpha test. In this study, the effect of 5S, Muda elimination and QCC on job satisfaction, commitment and job stress in Kombolcha Textile Share Company is assessed and factors that reduce employees’ job satisfaction with respect to kaizen implementation are identified. The total averages of means from the questionnaire are 3.1 for job satisfaction, 4.31 for job commitment and 4.2 for job stress. And results from interview and secondary data show that kaizen implementation have effect on EAA. In general, based on the thesis results it was concluded that kaizen (5S, muda elimination and QCC) have positive effect for improving EAA factors at KTSC. Finally, recommendations for improvement are given based on the results.
Abstract: During the last 10 years, in spite of the economic crisis, the number of tourists arriving in Greece has increased, particularly during the tourist season from April to October. In this paper, the number of annual tourist arrivals is studied to explore their preferences with regard to the month of travel, the selected destinations, as well the amount of money spent. The collected data are processed with statistical methods, yielding numerical and graphical results. From the computation of statistical parameters and the forecasting with exponential smoothing, useful conclusions are arrived at that can be used by the Greek tourism authorities, as well as by tourist organizations, for planning purposes for the coming years. The results of this paper and the computed forecast can also be used for decision making by private tourist enterprises that are investing in Greece. With regard to the statistical methods, the method of Simple Exponential Smoothing of time series of data is employed. The search for a best forecast for 2017 and 2018 provides the value of the smoothing coefficient. For all statistical computations and graphics Microsoft Excel is used.
Abstract: The water-based bioconvection of a nanofluid
containing motile gyrotactic micro-organisms over nonlinear
inclined stretching sheet has been investigated. The governing
nonlinear boundary layer equations of the model are reduced to a
system of ordinary differential equations via Oberbeck-Boussinesq
approximation and similarity transformations. Further, the modified
set of equations with associated boundary conditions are solved using
Finite Element Method. The impact of various pertinent parameters
on the velocity, temperature, nanoparticles concentration, density of
motile micro-organisms profiles are obtained and analyzed in details.
The results show that with the increase in angle of inclination δ,
velocity decreases while temperature, nanoparticles concentration,
a density of motile micro-organisms increases. Additionally, the
skin friction coefficient, Nusselt number, Sherwood number, density
number are computed for various thermophysical parameters. It
is noticed that increasing Brownian motion and thermophoresis
parameter leads to an increase in temperature of fluid which results
in a reduction in Nusselt number. On the contrary, Sherwood number
rises with an increase in Brownian motion and thermophoresis
parameter. The findings have been validated by comparing the
results of special cases with existing studies.
Abstract: In Chile, since 2002, there are extracurricular enrichment programs complementary to regular education for students with high intellectual capacity. This paper describes a model for the educational inclusion of students, with special educational needs associated with high intellectual capacity, developed at the University of Concepción and its effects on its students, academics and undergraduate students that collaborate with the program. The Talentos UdeC Program was created in 2003 and is intended for 240 children and youth from 11 to 18 years old, from 15 communes of the Biobio region. The case Talentos UdeC is analyzed from a mixed qualitative study in which those participating in the educational model are considered. The sample was composed of 30 students, 30 academics, and 30 undergraduate students. In the case of students, pre and post program measurements were made to analyze their socio-emotional adaptation, academic motivation and socially responsible behavior. The mentioned variables are measured through questionnaires designed and validated by the University of Concepcion that included: The Socially Responsible Behavior Questionnaire (CCSR); the Academic Motivation Questionnaire (CMA) and the Socio-Emotional Adaptation Questionnaire (CASE). The information obtained by these questionnaires was analyzed through a quantitative analysis. Academics and undergraduate students were interviewed to learn their perception of the effects of the program on themselves, on students and on society. The information obtained is analyzed using qualitative analysis based on the identification of common themes and descriptors for the construction of conceptual categories of answers. Quantitative results show differences in the first three variables analyzed in the students, after their participation for two years in Talentos UdeC. Qualitative results demonstrate perception of effects in the vision of world, project of life and in other areas of the students’ development; perception of effects in a personal, professional and organizational plane by academics and a perception of effects in their personal-social development and training in generic competencies by undergraduates students.
Abstract: Ambient air pollution with fine particulate matter (PM10) is a systematic permanent problem in many countries around the world. The accumulation of a large number of measurements of both the PM10 concentrations and the accompanying atmospheric factors allow for their statistical modeling to detect dependencies and forecast future pollution. This study applies the classification and regression trees (CART) method for building and analyzing PM10 models. In the empirical study, average daily air data for the city of Pleven, Bulgaria for a period of 5 years are used. Predictors in the models are seven meteorological variables, time variables, as well as lagged PM10 variables and some lagged meteorological variables, delayed by 1 or 2 days with respect to the initial time series, respectively. The degree of influence of the predictors in the models is determined. The selected best CART models are used to forecast future PM10 concentrations for two days ahead after the last date in the modeling procedure and show very accurate results.
Abstract: Poor air quality is one of the main environmental causes of premature deaths worldwide, and mainly in cities, where the majority of the population lives. It is a consequence of successive land cover (LC) and use changes, as a result of the intensification of human activities. Knowing these landscape modifications in a comprehensive spatiotemporal dimension is, therefore, essential for understanding variations in air pollutant concentrations. In this sense, the use of air quality models is very useful to simulate the physical and chemical processes that affect the dispersion and reaction of chemical species into the atmosphere. However, the modelling performance should always be evaluated since the resolution of the input datasets largely dictates the reliability of the air quality outcomes. Among these data, the updated LC is an important parameter to be considered in atmospheric models, since it takes into account the Earth’s surface changes due to natural and anthropic actions, and regulates the exchanges of fluxes (emissions, heat, moisture, etc.) between the soil and the air. This work aims to evaluate the performance of the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem), when different LC classifications are used as an input. The influence of two LC classifications was tested: i) the 24-classes USGS (United States Geological Survey) LC database included by default in the model, and the ii) CLC (Corine Land Cover) and specific high-resolution LC data for Portugal, reclassified according to the new USGS nomenclature (33-classes). Two distinct WRF-Chem simulations were carried out to assess the influence of the LC on air quality over Europe and Portugal, as a case study, for the year 2015, using the nesting technique over three simulation domains (25 km2, 5 km2 and 1 km2 horizontal resolution). Based on the 33-classes LC approach, particular emphasis was attributed to Portugal, given the detail and higher LC spatial resolution (100 m x 100 m) than the CLC data (5000 m x 5000 m). As regards to the air quality, only the LC impacts on tropospheric ozone concentrations were evaluated, because ozone pollution episodes typically occur in Portugal, in particular during the spring/summer, and there are few research works relating to this pollutant with LC changes. The WRF-Chem results were validated by season and station typology using background measurements from the Portuguese air quality monitoring network. As expected, a better model performance was achieved in rural stations: moderate correlation (0.4 – 0.7), BIAS (10 – 21µg.m-3) and RMSE (20 – 30 µg.m-3), and where higher average ozone concentrations were estimated. Comparing both simulations, small differences grounded on the Leaf Area Index and air temperature values were found, although the high-resolution LC approach shows a slight enhancement in the model evaluation. This highlights the role of the LC on the exchange of atmospheric fluxes, and stresses the need to consider a high-resolution LC characterization combined with other detailed model inputs, such as the emission inventory, to improve air quality assessment.