Investigating the Pedestrian Willingness to Pay to Choose Appropriate Policies for Improving the Safety of Pedestrian Facilities

Road traffic accidents lead to a higher rate of death and injury, especially in vulnerable road users such as pedestrians. Improving the safety of facilities for pedestrians is a major concern for policymakers because of the high number of pedestrian fatalities and direct and indirect costs which are imposed to the society. This study focuses on the idea of determining the willingness to pay of pedestrians for increasing their safety while crossing the street. In this study, three different scenarios including crossing the street with zebra crossing facilities, crossing the street with zebra crossing facilities and installing a pedestrian traffic light and constructing a pedestrian bridge with escalator are presented. The research was conducted based on stated preferences method. The required data were collected from a questionnaire that consisted of three parts: pedestrian’s demographic characteristics, travel characteristics and scenarios. Four different payment amounts are presented for each scenario and a logit model has been built for each proposed payment. The results show that sex, age, education, average household income and individual salary have significant effect on choosing a scenario. Among the policies that have been mentioned through the questionnaire scenarios, the scenario of crossing the street with zebra crossing facilities and installing a traffic lights is the most frequent, with willingness to pay 10,000 Rials and the scenario of crossing the street with a zebra crossing with a willingness to pay 100,000 Rials having the least frequency. For all scenarios, as the payment is increasing, the willingness to pay decreases.

The Importance of Conserving Pre-Historical, Historical and Cultural Heritage and Its Tourist Exploitation

Tourism in the present is the largest industry in the world, being an important global activity that has grown a lot in recent times. In this context, the activity of cultural tourism is growing, being seen as an important source of knowledge and information enjoyed by visitors. This article aims to discuss the cultural tourism, archaeological records and indigenous communities and the importance of preserving these invaluable sources of information, focusing on the records of the first peoples inhabiting the South American and North American lands. The study was based on discussions, theoretical studies, bibliographical research. Archaeological records are an important source of knowledge and information. Indigenous ethnic tourism represents a rescue of the authenticity of indigenous traditional cultures and their relation to the natural habitat. Cultural and indigenous tourism activity requires long-term planning to make it a sustainable activity.

Rain Cell Ratio Technique in Path Attenuation for Terrestrial Radio Links

A rain cell ratio model is proposed that computes attenuation of the smallest rain cell which represents the maximum rain rate value i.e. the cell size when rainfall rate is exceeded 0.01% of the time, R0.01 and predicts attenuation for other cells as the ratio with this maximum. This model incorporates the dependence of the path factor r on the ellipsoidal path variation of the Fresnel zone at different frequencies. In addition, the inhomogeneity of rainfall is modeled by a rain drop packing density factor. In order to derive the model, two empirical methods that can be used to find rain cell size distribution Dc are presented. Subsequently, attenuation measurements from different climatic zones for terrestrial radio links with frequencies F in the range 7-38 GHz are used to test the proposed model. Prediction results show that the path factor computed from the rain cell ratio technique has improved reliability when compared with other path factor and effective rain rate models, including the current ITU-R 530-15 model of 2013.

The U.S. Missile Defense Shield and Global Security Destabilization: An Inconclusive Link

Missile proliferation and global stability are intrinsically linked. Missile threats continually appear at the forefront of global security issues. North Korea’s recently demonstrated nuclear and intercontinental ballistic missile (ICBM) capabilities, for the first time since the Cold War, renewed public interest in strategic missile defense capabilities. To protect from limited ICBM attacks from so-called rogue actors, the United States developed the Ground-based Midcourse Defense (GMD) system. This study examines if the GMD missile defense shield has contributed to a safer world or triggered a new arms race. Based upon increased missile-related developments and the lack of adherence to international missile treaties, it is generally perceived that the GMD system is a destabilizing factor for global security. By examining the current state of arms control treaties as well as existing missile arsenals and ongoing efforts in technologies to overcome U.S. missile defenses, this study seeks to analyze the contribution of GMD to global stability. A thorough investigation cannot ignore that, through the establishment of this limited capability, the U.S. violated longstanding, successful weapons treaties and caused concern among states that possess ICBMs. GMD capability contributes to the perception that ICBM arsenals could become ineffective, creating an imbalance in favor of the United States, leading to increased global instability and tension. While blame for the deterioration of global stability and non-adherence to arms control treaties is often placed on U.S. missile defense, the facts do not necessarily support this view. The notion of a renewed arms race due to GMD is supported neither by current missile arsenals nor by the inevitable development of new and enhanced missile technology, to include multiple independently targeted reentry vehicles (MIRVs), maneuverable reentry vehicles (MaRVs), and hypersonic glide vehicles (HGVs). The methodology in this study encapsulates a period of time, pre- and post-GMD introduction, while analyzing international treaty adherence, missile counts and types, and research in new missile technologies. The decline in international treaty adherence, coupled with a measurable increase in the number and types of missiles or research in new missile technologies during the period after the introduction of GMD, could be perceived as a clear indicator of GMD contributing to global instability. However, research into improved technology (MIRV, MaRV and HGV) prior to GMD, as well as a decline of various global missile inventories and testing of systems during this same period, would seem to invalidate this theory. U.S. adversaries have exploited the perception of the U.S. missile defense shield as a destabilizing factor as a pretext to strengthen and modernize their militaries and justify their policies. As a result, it can be concluded that global stability has not significantly decreased due to GMD; but rather, the natural progression of technological and missile development would inherently include innovative and dynamic approaches to target engagement, deterrence, and national defense.

A 1H NMR-Linked PCR Modelling Strategy for Tracking the Fatty Acid Sources of Aldehydic Lipid Oxidation Products in Culinary Oils Exposed to Simulated Shallow-Frying Episodes

Objectives/Hypotheses: The adverse health effect potential of dietary lipid oxidation products (LOPs) has evoked much clinical interest. Therefore, we employed a 1H NMR-linked Principal Component Regression (PCR) chemometrics modelling strategy to explore relationships between data matrices comprising (1) aldehydic LOP concentrations generated in culinary oils/fats when exposed to laboratory-simulated shallow frying practices, and (2) the prior saturated (SFA), monounsaturated (MUFA) and polyunsaturated fatty acid (PUFA) contents of such frying media (FM), together with their heating time-points at a standard frying temperature (180 oC). Methods: Corn, sunflower, extra virgin olive, rapeseed, linseed, canola, coconut and MUFA-rich algae frying oils, together with butter and lard, were heated according to laboratory-simulated shallow-frying episodes at 180 oC, and FM samples were collected at time-points of 0, 5, 10, 20, 30, 60, and 90 min. (n = 6 replicates per sample). Aldehydes were determined by 1H NMR analysis (Bruker AV 400 MHz spectrometer). The first (dependent output variable) PCR data matrix comprised aldehyde concentration scores vectors (PC1* and PC2*), whilst the second (predictor) one incorporated those from the fatty acid content/heating time variables (PC1-PC4) and their first-order interactions. Results: Structurally complex trans,trans- and cis,trans-alka-2,4-dienals, 4,5-epxy-trans-2-alkenals and 4-hydroxy-/4-hydroperoxy-trans-2-alkenals (group I aldehydes predominantly arising from PUFA peroxidation) strongly and positively loaded on PC1*, whereas n-alkanals and trans-2-alkenals (group II aldehydes derived from both MUFA and PUFA hydroperoxides) strongly and positively loaded on PC2*. PCR analysis of these scores vectors (SVs) demonstrated that PCs 1 (positively-loaded linoleoylglycerols and [linoleoylglycerol]:[SFA] content ratio), 2 (positively-loaded oleoylglycerols and negatively-loaded SFAs), 3 (positively-loaded linolenoylglycerols and [PUFA]:[SFA] content ratios), and 4 (exclusively orthogonal sampling time-points) all powerfully contributed to aldehydic PC1* SVs (p 10-3 to < 10-9), as did all PC1-3 x PC4 interaction ones (p 10-5 to < 10-9). PC2* was also markedly dependent on all the above PC SVs (PC2 > PC1 and PC3), and the interactions of PC1 and PC2 with PC4 (p < 10-9 in each case), but not the PC3 x PC4 contribution. Conclusions: NMR-linked PCR analysis is a valuable strategy for (1) modelling the generation of aldehydic LOPs in heated cooking oils and other FM, and (2) tracking their unsaturated fatty acid (UFA) triacylglycerol sources therein.

Effect of Fatty Acids in Feed on Levels of Antibody Titers and CD4 and CD8 T-Lymphocyte against Newcastle Disease Virus of Vaccinated Broiler Chicken

400 one-day-old male broiler chicks (Ross-308) randomly divided to 2 main groups, 1st main group (GA) was feeding basal diet with medium chain fatty acid (MCFA) at rate of 0.15% and divided to four subgroups, 3 subgroups vaccinated with different routes with Newcastle Disease Virus (NDV) and non-vaccinated group. The 2nd main group (GB) feeding basal diet without MCFA and divided the same as 1st main group. The parameters used in this study included: ND antibody titers at 1, 10, 21, 28, 35 and 42 days of age and values of CD4 and CD8 at 1, 20, 30 and 42 days of age. This experiment detected increase in ND antibodies titers in (G1, G2, G3) groups were fed on basal diet MCFA comparing to groups were fed without adding MCFA (G5, G6, G7) and control groups (G4, G8). The results of cellular immune response (CD4 and CD8) T-cells in broiler chicks indicated that there was obviously significant relationship between dietary Fatty Acid (FA) versus the diet without FA on the level of CD4 parameter, for the entire experimental period. The effect of different ages was statistically significant in creating different values of CD4 level, whereas the CD4 level decreases markedly with age. However, analyzing the data of different vaccination methods, oculonasal method of vaccination led to the highest value of CD4 compared with the oral, S/C and control groups. There were statistical differences in CD8 values due to supplementation of FA versus the basal diet and due to the effect of different age periods. As for the age effect, the CD8 value at 20 days of age was significantly higher than at 42 and 30 days.

Feasibility Study of Friction Stir Welding Application for Kevlar Material

Friction stir welding (FSW) is a joining process in the solid state, which eliminates problems associated with the material melting and solidification, such as cracks, residual stresses and distortions generated during conventional welding. Among the most important advantages of FSW are; easy automation, less distortion, lower residual stress and good mechanical properties in the joining region. FSW is a recent approach to metal joining and although originally intended for aluminum alloys, it is investigated in a variety of metallic materials. The basic concept of FSW is a rotating tool, made of non-consumable material, specially designed with a geometry consisting of a pin and a recess (shoulder). This tool is inserted as spinning on its axis at the adjoining edges of two sheets or plates to be joined and then it travels along the joining path line. The tool rotation axis defines an angle of inclination with which the components to be welded. This angle is used for receiving the material to be processed at the tool base and to promote the gradual forge effect imposed by the shoulder during the passage of the tool. This prevents the material plastic flow at the tool lateral, ensuring weld closure on the back of the pin. In this study, two 4 mm Kevlar® plates which were produced with the Kevlar® fabrics, are analyzed with COMSOL Multiphysics in order to investigate the weldability via FSW. Thereafter, some experimental investigation is done with an appropriate workbench in order to compare them with the analysis results.

Maximum Wind Power Extraction Strategy and Decoupled Control of DFIG Operating in Variable Speed Wind Generation Systems

This paper appraises the performances of two control scenarios, for doubly fed induction generator (DFIG) operating in wind generation system (WGS), which are the direct decoupled control (DDC) and indirect decoupled control (IDC). Both control scenarios studied combines vector control and Maximum Power Point Tracking (MPPT) control theory so as to maximize the captured power through wind turbine. Modeling of DFIG based WGS and details of both control scenarios have been presented, a proportional integral controller is employed in the active and reactive power control loops for both control methods. The performance of the both control scenarios in terms of power reference tracking and robustness against machine parameters inconstancy has been shown, analyzed and compared, which can afford a reference to the operators and engineers of a wind farm. All simulations have been implemented via MATLAB/Simulink.

Use of Hierarchical Temporal Memory Algorithm in Heart Attack Detection

In order to reduce the number of deaths due to heart problems, we propose the use of Hierarchical Temporal Memory Algorithm (HTM) which is a real time anomaly detection algorithm. HTM is a cortical learning algorithm based on neocortex used for anomaly detection. In other words, it is based on a conceptual theory of how the human brain can work. It is powerful in predicting unusual patterns, anomaly detection and classification. In this paper, HTM have been implemented and tested on ECG datasets in order to detect cardiac anomalies. Experiments showed good performance in terms of specificity, sensitivity and execution time.

Predictive Semi-Empirical NOx Model for Diesel Engine

Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model.  Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.

dynr.mi: An R Program for Multiple Imputation in Dynamic Modeling

Assessing several individuals intensively over time yields intensive longitudinal data (ILD). Even though ILD provide rich information, they also bring other data analytic challenges. One of these is the increased occurrence of missingness with increased study length, possibly under non-ignorable missingness scenarios. Multiple imputation (MI) handles missing data by creating several imputed data sets, and pooling the estimation results across imputed data sets to yield final estimates for inferential purposes. In this article, we introduce dynr.mi(), a function in the R package, Dynamic Modeling in R (dynr). The package dynr provides a suite of fast and accessible functions for estimating and visualizing the results from fitting linear and nonlinear dynamic systems models in discrete as well as continuous time. By integrating the estimation functions in dynr and the MI procedures available from the R package, Multivariate Imputation by Chained Equations (MICE), the dynr.mi() routine is designed to handle possibly non-ignorable missingness in the dependent variables and/or covariates in a user-specified dynamic systems model via MI, with convergence diagnostic check. We utilized dynr.mi() to examine, in the context of a vector autoregressive model, the relationships among individuals’ ambulatory physiological measures, and self-report affect valence and arousal. The results from MI were compared to those from listwise deletion of entries with missingness in the covariates. When we determined the number of iterations based on the convergence diagnostics available from dynr.mi(), differences in the statistical significance of the covariate parameters were observed between the listwise deletion and MI approaches. These results underscore the importance of considering diagnostic information in the implementation of MI procedures.

A Mathematical Model Approach Regarding the Children’s Height Development with Fractional Calculus

The study aims to use a mathematical approach with the fractional calculus which is developed to have the ability to continuously analyze the factors related to the children’s height development. Until now, tracking the development of the child is getting more important and meaningful. Knowing and determining the factors related to the physical development of the child any desired time would provide better, reliable and accurate results for childcare. In this frame, 7 groups for height percentile curve (3th, 10th, 25th, 50th, 75th, 90th, and 97th) of Turkey are used. By using discrete height data of 0-18 years old children and the least squares method, a continuous curve is developed valid for any time interval. By doing so, in any desired instant, it is possible to find the percentage and location of the child in Percentage Chart. Here, with the help of the fractional calculus theory, a mathematical model is developed. The outcomes of the proposed approach are quite promising compared to the linear and the polynomial method. The approach also yields to predict the expected values of children in the sense of height.

Settlement Analysis of Axially Loaded Bored Piles: A Case History

Pile load tests should be applied to check the bearing capacity calculations and to determine the settlement of the pile corresponding to test load. Strain gauges can be installed into pile in order to determine the shaft resistance of the piles for every soil layer respectively. Detailed results can be obtained by means of strain gauges placed at certain levels into test piles. In the scope of this study, pile load test data obtained from two different projects are examined.  Instrumented static pile load tests were applied on totally 7 test bored piles of different diameters (80 cm, 150 cm, and 200 cm) and different lengths (between 30-76 m) in two different project site. Settlement analysis of test piles is done by using some of load transfer methods and finite element method. Plaxis 3D which is a three-dimensional finite element program is also used for settlement analysis of the test piles. In this study, firstly bearing capacity of test piles are determined and compared with strain gauge data which is required for settlement analysis. Then, settlement values of the test piles are estimated by using load transfer methods developed in recent years and finite element method. The aim of this study is to show similarities and differences between the results obtained from settlement analysis methods and instrumented pile load tests.

Effects of Ophiocordyceps dipterigena BCC 2073 β-Glucan as a Prebiotic on the in vitro Growth of Probiotic and Pathogenic Bacteria

The  β-glucan produced by Ophiocordyceps dipterigena BCC 2073 is a (1, 3)-β-D-glucan with highly branching O-6-linkedside chains that is resistant to acid hydrolysis (by hydrochloric acid and porcine pancreatic alpha-amylase). This β-glucan can be utilized as a prebiotic due to its advantageous structural and biological properties. The effects of using this β-glucan as the sole carbon source for the in vitro growth of two probiotic bacteria (L. acidophilus BCC 13938 and B. animalis ATCC 25527) were investigated. Compared with the effect of using 1% glucose or fructo-oligosaccharide (FOS) as the sole carbon source, using 1% β-glucan for this purpose showed that this prebiotic supported and stimulated the growth of both types of probiotic bacteria and induced them to produce the highest levels of metabolites during their growth. The highest levels of lactic and acetic acid, 10.04 g·L-1 and 2.82 g·L-1, respectively, were observed at 2 h of cultivation using glucose as the sole carbon source. Furthermore, the fermentation broth obtained using 1% β-glucan as the sole carbon source had greater antibacterial activity against selected pathogenic bacteria (B. subtilis TISTR 008, E. coli TISTR 780, and S. typhimurium TISTR 292) than did the broths prepared using glucose or FOS as the sole carbon source. The fermentation broth obtained by growing L. acidophilus BCC 13938 in the presence of β-glucan inhibited the growth of B. subtilis TISTR 008 by more than 70% and inhibited the growth of both S. typhimurium TISTR 292 and E. coli TISTR 780 by more than 90%. In conclusion, O. dipterigena BCC 2073 is a potential source of a β-glucan prebiotic that could be used for commercial production in the near future.

Experimental Evaluation of 10 Ecotypes of Toxic and Non-Toxic Jatropha curcas as Raw Material to Produce Biodiesel in Morelos State, Mexico

Jatropha curcas is a perennial oleaginous plant that is currently considered an energy crop with high potential as an environmentally sustainable biofuel. During the last decades, research in biofuels has grown in tropical and subtropical regions in Latin America. However, as far we know, there are no reports on the growth and yield patterns of Jatropha curcas under the specific agro climatic scenarios of the State of Morelos, Mexico. This study presents the results of 52 months monitoring of 10 toxic and non-toxic ecotypes of Jatropha curcas (E1M, E2M, E3M, E4M, E5M, E6O, E7O, E8O, E9C, E10C) in an experimental plantation with minimum watering and fertilization resources. The main objective is to identify the ecotypes with the highest potential as biodiesel raw material in the select region, by developing experimental information. Specifically, we monitored biophysical and growth parameters, including plant survival and seed production (at the end of month 52), to study the performance of each ecotype and to establish differences among the variables of morphological growth, net seed oil content, and toxicity. To analyze the morphological growth, a statistical approach to the biophysical parameters was used; the net seed oil content -80 to 192 kg/ha- was estimated with the first harvest; and the toxicity was evaluated by examining the phorbol ester concentration (µg/L) in the oil extracted from the seeds. The comparison and selection of ecotypes was performed through a methodology developed based on the normalization of results. We identified four outstanding ecotypes (E1M, E2M, E3M, and E4M) that can be used to establish Jatropha curcas as energy crops in the state of Morelos for feasible agro-industrial production of biodiesel and other products related to the use of biomass.

Comparison of Irradiance Decomposition and Energy Production Methods in a Solar Photovoltaic System

Installations of solar photovoltaic systems have increased considerably in the last decade. Therefore, it has been noticed that monitoring of meteorological data (solar irradiance, air temperature, wind velocity, etc.) is important to predict the potential of a given geographical area in solar energy production. In this sense, the present work compares two computational tools that are capable of estimating the energy generation of a photovoltaic system through correlation analyzes of solar radiation data: PVsyst software and an algorithm based on the PVlib package implemented in MATLAB. In order to achieve the objective, it was necessary to obtain solar radiation data (measured and from a solarimetric database), analyze the decomposition of global solar irradiance in direct normal and horizontal diffuse components, as well as analyze the modeling of the devices of a photovoltaic system (solar modules and inverters) for energy production calculations. Simulated results were compared with experimental data in order to evaluate the performance of the studied methods. Errors in estimation of energy production were less than 30% for the MATLAB algorithm and less than 20% for the PVsyst software.

Elastic and Plastic Collision Comparison Using Finite Element Method

The prevision of post-impact conditions and the behavior of the bodies during the impact have been object of several collision models. The formulation from Hertz’s theory is generally used dated from the 19th century. These models consider the repulsive force as proportional to the deformation of the bodies under contact and may consider it proportional to the rate of deformation. The objective of the present work is to analyze the behavior of the bodies during impact using the Finite Element Method (FEM) with elastic and plastic material models. The main parameters to evaluate are, the contact force, the time of contact and the deformation of the bodies. An advantage of using the FEM approach is the possibility to apply a plastic deformation to the model according to the material definition: there will be used Johnson–Cook plasticity model whose parameters are obtained through empirical tests of real materials. This model allows analyzing the permanent deformation caused by impact, phenomenon observed in real world depending on the forces applied to the body. These results are compared between them and with the model-based Hertz theory.

Diagnosis on Environmental Impacts of Tourism at Caju Beach in Palmas, Tocantins, Brazil

Environmental impacts are the changes in the physical, chemical or biological properties of natural areas that are most often caused by human actions on the environment and which have consequences for human health, society and the elements of nature. The identification of the environmental impacts is important so that they are mitigated, and above all that the mitigating measures are applied in the area. This work aims to identify the environmental impacts generated in the Praia do Caju area in the city of Palmas/Brazil and show that the lack of structure on the beach intensifies the environmental impacts. The present work was carried out having as parameter, the typologies of exploratory and descriptive and quantitative research through a matrix of environmental impacts through direct observation and registration. The study took place during the holidays from August to December 2016 and photographic record of impacts. From the collected data it was possible to verify that Caju beach suffers constant degradation due to irregular deposition.

Characteristics of the Long-Term Regional Tourism Development in Georgia

Tourism industry development is one of the key priorities in Georgia, as it has positive influence on economic activities. Its contribution is very important for the different regions, as well as for the national economy. Benefits of the tourism industry include new jobs, service development, and increasing tax revenues, etc. The main aim of this research is to review and analyze the potential of the Georgian tourism industry with its long-term strategy and current challenges. To plan activities in a long-term development, it is required to evaluate several factors on the regional and on the national level. Factors include activities, transportation, services, lodging facilities, infrastructure and institutions. The major research contributions are practical estimates about regional tourism development which plays an important role in the integration process with global markets.

Social Semantic Web-Based Analytics Approach to Support Lifelong Learning

The purpose of this paper is to describe how learning analytics approaches based on social semantic web techniques can be applied to enhance the lifelong learning experiences in a connectivist perspective. For this reason, a prototype of a system called SoLearn (Social Learning Environment) that supports this approach. We observed and studied literature related to lifelong learning systems, social semantic web and ontologies, connectivism theory, learning analytics approaches and reviewed implemented systems based on these fields to extract and draw conclusions about necessary features for enhancing the lifelong learning process. The semantic analytics of learning can be used for viewing, studying and analysing the massive data generated by learners, which helps them to understand through recommendations, charts and figures their learning and behaviour, and to detect where they have weaknesses or limitations. This paper emphasises that implementing a learning analytics approach based on social semantic web representations can enhance the learning process. From one hand, the analysis process leverages the meaning expressed by semantics presented in the ontology (relationships between concepts). From the other hand, the analysis process exploits the discovery of new knowledge by means of inferring mechanism of the semantic web.