Production of Bioethanol through Hydrolysis of Agro-Industrial Banana Crop Residues

Nowadays, the main biofuels source production as bioethanol is food crops. This means a high competition between foods and energy production. For this reason, it is necessary to take into account the use of new raw materials friendly to the environment. The main objective of this paper is to evaluate the potential of the agro-industrial banana crop residues in the production of bioethanol. A factorial design of 24 was used, the design has variables such as pH, time and concentration of hydrolysis, another variable is the time of fermentation that is of 7 or 15 days. In the hydrolysis phase, the pH is acidic (H2SO4) or basic (NaOH), the time is 30 or 15 minutes and the concentration is 0.1 or 0.5 M. It was observed that basic media, low concentrations, fermentation, and higher pretreatment times produced better performance in terms of biofuel obtained.

Sparse Unmixing of Hyperspectral Data by Exploiting Joint-Sparsity and Rank-Deficiency

In this work, we exploit two assumed properties of the abundances of the observed signatures (endmembers) in order to reconstruct the abundances from hyperspectral data. Joint-sparsity is the first property of the abundances, which assumes the adjacent pixels can be expressed as different linear combinations of same materials. The second property is rank-deficiency where the number of endmembers participating in hyperspectral data is very small compared with the dimensionality of spectral library, which means that the abundances matrix of the endmembers is a low-rank matrix. These assumptions lead to an optimization problem for the sparse unmixing model that requires minimizing a combined l2,p-norm and nuclear norm. We propose a variable splitting and augmented Lagrangian algorithm to solve the optimization problem. Experimental evaluation carried out on synthetic and real hyperspectral data shows that the proposed method outperforms the state-of-the-art algorithms with a better spectral unmixing accuracy.

Latent Factors of Severity in Truck-Involved and Non-Truck-Involved Crashes on Freeways

Truck-involved crashes have higher crash severity than non-truck-involved crashes. There have been many studies about the frequency of crashes and the development of severity models, but those studies only analyzed the relationship between observed variables. To identify why more people are injured or killed when trucks are involved in the crash, we must examine to quantify the complex causal relationship between severity of the crash and risk factors by adopting the latent factors of crashes. The aim of this study was to develop a structural equation or model based on truck-involved and non-truck-involved crashes, including five latent variables, i.e. a crash factor, environmental factor, road factor, driver’s factor, and severity factor. To clarify the unique characteristics of truck-involved crashes compared to non-truck-involved crashes, a confirmatory analysis method was used. To develop the model, we extracted crash data from 10,083 crashes on Korean freeways from 2008 through 2014. The results showed that the most significant variable affecting the severity of a crash is the crash factor, which can be expressed by the location, cause, and type of the crash. For non-truck-involved crashes, the crash and environment factors increase severity of the crash; conversely, the road and driver factors tend to reduce severity of the crash. For truck-involved crashes, the driver factor has a significant effect on severity of the crash although its effect is slightly less than the crash factor. The multiple group analysis employed to analyze the differences between the heterogeneous groups of drivers.

Performance Analysis of Deterministic Stable Election Protocol Using Fuzzy Logic in Wireless Sensor Network

In Wireless Sensor Network (WSN), the sensor containing motes (nodes) incorporate batteries that can lament at some extent. To upgrade the energy utilization, clustering is one of the prototypical approaches for split sensor motes into a number of clusters where one mote (also called as node) proceeds as a Cluster Head (CH). CH selection is one of the optimization techniques for enlarging stability and network lifespan. Deterministic Stable Election Protocol (DSEP) is an effectual clustering protocol that makes use of three kinds of nodes with dissimilar residual energy for CH election. Fuzzy Logic technology is used to expand energy level of DSEP protocol by using fuzzy inference system. This paper presents protocol DSEP using Fuzzy Logic (DSEP-FL) CH by taking into account four linguistic variables such as energy, concentration, centrality and distance to base station. Simulation results show that our proposed method gives more effective results in term of a lifespan of network and stability as compared to the performance of other clustering protocols.

A Memetic Algorithm for an Energy-Costs-Aware Flexible Job-Shop Scheduling Problem

In this article, the flexible job-shop scheduling problem is extended by consideration of energy costs which arise owing to the power peak, and further decision variables such as work in process and throughput time are incorporated into the objective function. This enables a production plan to be simultaneously optimized in respect of the real arising energy and logistics costs. The energy-costs-aware flexible job-shop scheduling problem (EFJSP) which arises is described mathematically, and a memetic algorithm (MA) is presented as a solution. In the MA, the evolutionary process is supplemented with a local search. Furthermore, repair procedures are used in order to rectify any infeasible solutions that have arisen in the evolutionary process. The potential for lowering the real arising costs of a production plan through consideration of energy consumption levels is highlighted.

Quantifying Freeway Capacity Reductions by Rainfall Intensities Based on Stochastic Nature of Flow Breakdown

This study quantifies a decrement in freeway capacity during rainfall. Traffic and rainfall data were gathered from Highway Agencies and Wunderground weather service. Three inter-urban freeway sections and its nearest weather stations were selected as experimental sites. Capacity analysis found reductions of maximum and mean pre-breakdown flow rates due to rainfall. The Kruskal-Wallis test also provided some evidence to suggest that the variance in the pre-breakdown flow rate is statistically insignificant. Potential application of this study lies in the operation of real time traffic management schemes such as Variable Speed Limits (VSL), Hard Shoulder Running (HSR), and Ramp Metering System (RMS), where speed or flow limits could be set based on a number of factors, including rainfall events and their intensities.

Investigation of Main Operating Parameters Affecting Gas Turbine Efficiency and Gas Releases

This work presents a study on the influence of the main operating variables on the gas turbine cycle. A numerical simulation of a gas turbine cycle is performed for a real net power of 100 MW. A calculation code is developed using EES software. The operating variables are taken in conformity with the local environmental conditions adopted by the Tunisian Society of Electricity and Gas. Results show that the increase of ambient temperature leads to an increase of Tpz and NOx emissions rate and a decrease of cycle efficiency and UHC emissions. The CO emissions decrease with the raise of residence time, while NOx emissions rate increases and UHC emissions rate decreases. Furthermore, both of cycle efficiency and NOx emissions increase with the increase of the pressure ratio.

Low Energy Technology for Leachate Valorisation

Landfills present long-term threats to soil, air, groundwater and surface water due to the formation of greenhouse gases (methane gas and carbon dioxide) and leachate from decomposing garbage. The composition of leachate differs from site to site and also within the landfill. The leachates alter with time (from weeks to years) since the landfilled waste is biologically highly active and their composition varies. Mainly, the composition of the leachate depends on factors such as characteristics of the waste, the moisture content, climatic conditions, degree of compaction and the age of the landfill. Therefore, the leachate composition cannot be generalized and the traditional treatment models should be adapted in each case. Although leachate composition is highly variable, what different leachates have in common is hazardous constituents and their potential eco-toxicological effects on human health and on terrestrial ecosystems. Since leachate has distinct compositions, each landfill or dumping site would represent a different type of risk on its environment. Nevertheless, leachates consist always of high organic concentration, conductivity, heavy metals and ammonia nitrogen. Leachate could affect the current and future quality of water bodies due to uncontrolled infiltrations. Therefore, control and treatment of leachate is one of the biggest issues in urban solid waste treatment plants and landfills design and management. This work presents a treatment model that will be carried out "in-situ" using a cost-effective novel technology that combines solar evaporation/condensation plus forward osmosis. The plant is powered by renewable energies (solar energy, biomass and residual heat), which will minimize the carbon footprint of the process. The final effluent quality is very high, allowing reuse (preferred) or discharge into watercourses. In the particular case of this work, the final effluents will be reused for cleaning and gardening purposes. A minority semi-solid residual stream is also generated in the process. Due to its special composition (rich in metals and inorganic elements), this stream will be valorized in ceramic industries to improve the final products characteristics.

Resilience Assessment for Power Distribution Systems

Power distribution systems are essential and crucial infrastructures for the development and maintenance of a sustainable society. These systems are extremely vulnerable to various types of natural and man-made disasters. The assessment of resilience focuses on preparedness and mitigation actions under pre-disaster conditions. It also concentrates on response and recovery actions under post-disaster situations. The aim of this study is to present a methodology to assess the resilience of electric power distribution poles against wind-related events. The proposed methodology can improve the accuracy and rapidity of the evaluation of the conditions and the assessment of the resilience of poles. The methodology provides a metric for the evaluation of the resilience of poles under pre-disaster and post-disaster conditions. The metric was developed using mathematical expressions for physical forces that involve various variables, such as physical dimensions of the pole, the inclination of the pole, and wind speed. A three-dimensional imaging technology (photogrammetry) was used to determine the inclination of poles. Based on expert opinion, the proposed metric was used to define zones to visualize resilience. Visual representation of resilience is helpful for decision makers to prioritize their resources before and after experiencing a wind-related disaster. Multiple electric poles in the City of Beaumont, TX were used in a case study to evaluate the proposed methodology.  

Achievements of Healthcare Services Vis-À-Vis the Millennium Development Goals Targets: Evidence from Pakistan

This study investigates the impact of public healthcare facilities and socio-economic circumstances on the status of child health in Pakistan. The complete analysis is carried out in correspondence with fourth and sixth millennium development goals. Further, the health variables chosen are also inherited from targeted indicators of the mentioned goals (MDGs). Trends in the Human Opportunity Index (HOI) for both health inequalities and coverage are analyzed using the Pakistan Social and Living Standards Measurement (PLSM) data set for 2001-02 to 2012-13 at the national and provincial level. To reveal the relative importance of each circumstance in achieving the targeted values for child health, Shorrocks decomposition is applied on HOI. The annual point average growth rate of HOI is used to simulate the time period for the achievement of target set by MDGs and universal access also. The results indicate an improvement in HOI for a reduction in child mortality rates from 52.1% in 2001-02 to 67.3% in 2012-13, which confirms the availability of healthcare opportunities to a larger segment of society. Similarly, immunization against measles and other diseases such as Diphtheria, Polio, Bacillus Calmette-Guerin (BCG), and Hepatitis has also registered an improvement from 51.6% to 69.9% during the period of study at the national level. On a positive note, no gender disparity has been found for child health indicators and that health outcome is mostly affected by the parental and geographical features and availability of health infrastructure. However, the study finds that this achievement has been uneven across provinces. Pakistan is not only lagging behind in achieving its health goals, disappointingly with the current rate of health care provision, but it will take many additional years to achieve its targets.

A Transform Domain Function Controlled VSSLMS Algorithm for Sparse System Identification

The convergence rate of the least-mean-square (LMS) algorithm deteriorates if the input signal to the filter is correlated. In a system identification problem, this convergence rate can be improved if the signal is white and/or if the system is sparse. We recently proposed a sparse transform domain LMS-type algorithm that uses a variable step-size for a sparse system identification. The proposed algorithm provided high performance even if the input signal is highly correlated. In this work, we investigate the performance of the proposed TD-LMS algorithm for a large number of filter tap which is also a critical issue for standard LMS algorithm. Additionally, the optimum value of the most important parameter is calculated for all experiments. Moreover, the convergence analysis of the proposed algorithm is provided. The performance of the proposed algorithm has been compared to different algorithms in a sparse system identification setting of different sparsity levels and different number of filter taps. Simulations have shown that the proposed algorithm has prominent performance compared to the other algorithms.

Load Forecasting in Microgrid Systems with R and Cortana Intelligence Suite

Energy production optimization has been traditionally very important for utilities in order to improve resource consumption. However, load forecasting is a challenging task, as there are a large number of relevant variables that must be considered, and several strategies have been used to deal with this complex problem. This is especially true also in microgrids where many elements have to adjust their performance depending on the future generation and consumption conditions. The goal of this paper is to present a solution for short-term load forecasting in microgrids, based on three machine learning experiments developed in R and web services built and deployed with different components of Cortana Intelligence Suite: Azure Machine Learning, a fully managed cloud service that enables to easily build, deploy, and share predictive analytics solutions; SQL database, a Microsoft database service for app developers; and PowerBI, a suite of business analytics tools to analyze data and share insights. Our results show that Boosted Decision Tree and Fast Forest Quantile regression methods can be very useful to predict hourly short-term consumption in microgrids; moreover, we found that for these types of forecasting models, weather data (temperature, wind, humidity and dew point) can play a crucial role in improving the accuracy of the forecasting solution. Data cleaning and feature engineering methods performed in R and different types of machine learning algorithms (Boosted Decision Tree, Fast Forest Quantile and ARIMA) will be presented, and results and performance metrics discussed.

Coastal Vulnerability Index and Its Projection for Odisha Coast, East Coast of India

Tropical cyclone is one among the worst natural hazards that results in a trail of destruction causing enormous damage to life, property, and coastal infrastructures. In a global perspective, the Indian Ocean is considered as one of the cyclone prone basins in the world. Specifically, the frequency of cyclogenesis in the Bay of Bengal is higher compared to the Arabian Sea. Out of the four maritime states in the East coast of India, Odisha is highly susceptible to tropical cyclone landfall. Historical records clearly decipher the fact that the frequency of cyclones have reduced in this basin. However, in the recent decades, the intensity and size of tropical cyclones have increased. This is a matter of concern as the risk and vulnerability level of Odisha coast exposed to high wind speed and gusts during cyclone landfall have increased. In this context, there is a need to assess and evaluate the severity of coastal risk, area of exposure under risk, and associated vulnerability with a higher dimension in a multi-risk perspective. Changing climate can result in the emergence of a new hazard and vulnerability over a region with differential spatial and socio-economic impact. Hence there is a need to have coastal vulnerability projections in a changing climate scenario. With this motivation, the present study attempts to estimate the destructiveness of tropical cyclones based on Power Dissipation Index (PDI) for those cyclones that made landfall along Odisha coast that exhibits an increasing trend based on historical data. The study also covers the futuristic scenarios of integral coastal vulnerability based on the trends in PDI for the Odisha coast. This study considers 11 essential and important parameters; the cyclone intensity, storm surge, onshore inundation, mean tidal range, continental shelf slope, topo-graphic elevation onshore, rate of shoreline change, maximum wave height, relative sea level rise, rainfall distribution, and coastal geomorphology. The study signifies that over a decadal scale, the coastal vulnerability index (CVI) depends largely on the incremental change in variables such as cyclone intensity, storm surge, and associated inundation. In addition, the study also performs a critical analysis on the modulation of PDI on storm surge and inundation characteristics for the entire coastal belt of Odisha State. Interestingly, the study brings to light that a linear correlation exists between the storm-tide with PDI. The trend analysis of PDI and its projection for coastal Odisha have direct practical applications in effective coastal zone management and vulnerability assessment.

A Study of Mode Choice Model Improvement Considering Age Grouping

The purpose of this study is providing an improved mode choice model considering parameters including age grouping of prime-aged and old age. In this study, 2010 Household Travel Survey data were used and improper samples were removed through the analysis. Chosen alternative, date of birth, mode, origin code, destination code, departure time, and arrival time are considered from Household Travel Survey. By preprocessing data, travel time, travel cost, mode, and ratio of people aged 45 to 55 years, 55 to 65 years and over 65 years were calculated. After the manipulation, the mode choice model was constructed using LIMDEP by maximum likelihood estimation. A significance test was conducted for nine parameters, three age groups for three modes. Then the test was conducted again for the mode choice model with significant parameters, travel cost variable and travel time variable. As a result of the model estimation, as the age increases, the preference for the car decreases and the preference for the bus increases. This study is meaningful in that the individual and households characteristics are applied to the aggregate model.

The Impact of Information and Communication Technology on Bilateral Trade in Goods

This paper investigates the impact of Information and Communication Technology (ICT) on bilateral trade in goods. Empirical analysis is performed on the United States and 34 partnering countries from 2000 to 2013. Our econometric model fits the data well, explaining 52% of the variation in trade flows for goods trade, 53.2% of the variation in trade flows for goods export and 48% of the variation in trade flows for goods import. For every 10% increase in fixed broadband Internet subscribers per 100 people increases, goods trade by 7.9% and for every 5% increase in fixed broadband Internet subscribers per 100 people, goods export increases by 11%. For every 1% increase in fixed telephone line penetration per 100 people, goods trade increases by 26.3%, goods export increases by 24.4% and goods import increases by 24.8%. For every 1% increase in mobile-cellular telephone subscriptions, goods trade decreases by 29.6% and goods export decreases by 27.1%, whilst for every 0.01% increase in mobile-cellular telephone subscriptions, goods import decreases by 34.3%. For every 1% increase in the percentage of population who used the Internet from any location in the last 12 months Internet, goods trade increases by 32.5%, goods export increases by 38.9%, goods import increases by 33%. All our trade determinants as well as our ICT variables have significances on goods exports for the US. We can also draw from our study that the US relies more rather heavily on ICT for its goods export compared to goods import.

Factors Affecting Slot Machine Performance in an Electronic Gaming Machine Facility

A facility exploiting only electronic gambling machines (EGMs) opened in 2007 in Quebec City, Canada under the name of Salons de Jeux du Québec (SdjQ). This facility is one of the first worldwide to rely on that business model. This paper models the performance of such EGMs. The interest from a managerial point of view is to identify the variables that can be controlled or influenced so that a comprehensive model can help improve the overall performance of the business. The EGM individual performance model contains eight different variables under study (Game Title, Progressive jackpot, Bonus Round, Minimum Coin-in, Maximum Coin-in, Denomination, Slant Top and Position). Using data from Quebec City’s SdjQ, a linear regression analysis explains 90.80% of the EGM performance. Moreover, results show a behavior slightly different than that of a casino. The addition of GameTitle as a factor to predict the EGM performance is one of the main contributions of this paper. The choice of the game (GameTitle) is very important. Games having better position do not have significantly better performance than games located elsewhere on the gaming floor. Progressive jackpots have a positive and significant effect on the individual performance of EGMs. The impact of BonusRound on the dependent variable is significant but negative. The effect of Denomination is significant but weakly negative. As expected, the Language of an EGMS does not impact its individual performance. This paper highlights some possible improvements by indicating which features are performing well. Recommendations are given to increase the performance of the EGMs performance.

Study on Hydrophilicity of Anodic Aluminum Oxide Templates with TiO2-NTs

This paper aims to discuss the hydrophilicity about the anodic aluminum oxide (AAO) template with titania nanotubes (NTs). The AAO templates with pore size diameters of 20-250 nm were generated by anodizing 6061 aluminum alloy substrates in acid solution of sulfuric acid (H2SO4), oxalic acid (COOH)2, and phosphoric acid (H3PO4), respectively. TiO2-NTs were grown on AAO templates by the sol-gel deposition process successfully. The water contact angle on AAO/TiO2-NTs surface was lower compared to the water contact angle on AAO surface. So, the characteristic of hydrophilicity was significantly associated with the AAO pore size and what kinds of materials were immersed variables.

An Investigation on Material Removal Rate of EDM Process: A Response Surface Methodology Approach

In the present work response surface methodology (RSM) based central composite design (CCD) is used for analyzing the electrical discharge machining (EDM) process. For experimentation, mild steel is selected as work piece and copper is used as electrode. Three machining parameters namely current (I), spark on time (Ton) and spark off time (Toff) are selected as the input variables. The output or response chosen is material removal rate (MRR) which is to be maximized. To reduce the number of runs face centered central composite design (FCCCD) was used. ANOVA was used to determine the significance of parameter and interactions. The suitability of model is tested using Anderson darling (AD) plot. The results conclude that different parameters considered i.e. current, pulse on and pulse off time; all have dominant effect on the MRR. At last, the optimized parameter setting for maximizing MRR is found through main effect plot analysis.

Asymmetrical Informative Estimation for Macroeconomic Model: Special Case in the Tourism Sector of Thailand

This paper used an asymmetric informative concept to apply in the macroeconomic model estimation of the tourism sector in Thailand. The variables used to statistically analyze are Thailand international and domestic tourism revenues, the expenditures of foreign and domestic tourists, service investments by private sectors, service investments by the government of Thailand, Thailand service imports and exports, and net service income transfers. All of data is a time-series index which was observed between 2002 and 2015. Empirically, the tourism multiplier and accelerator were estimated by two statistical approaches. The first was the result of the Generalized Method of Moments model (GMM) based on the assumption which the tourism market in Thailand had perfect information (Symmetrical data). The second was the result of the Maximum Entropy Bootstrapping approach (MEboot) based on the process that attempted to deal with imperfect information and reduced uncertainty in data observations (Asymmetrical data). In addition, the tourism leakages were investigated by a simple model based on the injections and leakages concept. The empirical findings represented the parameters computed from the MEboot approach which is different from the GMM method. However, both of the MEboot estimation and GMM model suggests that Thailand’s tourism sectors are in a period capable of stimulating the economy.

Effects of an Educative Model in Socially Responsible Behavior and Other Psychological Variables

The eudaimonic perspective in philosophy and psychology suggests that a good life is closely related to developing oneself in order to contribute to the well-being and happiness of other people and of the world as a whole. Educational psychology can help to achieve this through the design and validation of educative models. Since 2004, the University of Concepcion and other Chilean universities apply an educative model to train socially responsible professionals, people that in the exercise of their profession contribute to generate equity for the development and assess the impacts of their decisions, opting for those that serve the common good. The main aim is to identify if a relationship exists between achieved learning, attitudes toward social responsibility, self-attribution of socially responsible behavior, value type, professional behavior observed and, participation in a specific model to train socially responsible (SR) professionals. The Achieved Learning and Attitudes Toward Social Responsibility Questionnaire, interview with employers and Values Questionnaire and Self-attribution of SR Behavior Questionnaire is applied to 394 students and graduates, divided into experimental and control groups (trained and not trained under the educative model), in order to identify the professional behavior of the graduates. The results show that students and graduates perceive cognitive, affective and behavioral learning, with significant differences in attitudes toward social responsibility and self-attribution of SR behavior, between experimental and control. There are also differences in employers' perceptions about the professional practice of those who were trained under the model and those who were not. It is concluded that the educative model has an impact on the learning of social responsibility and educates for a full life. It is also concluded that it is necessary to identify mediating variables of the model effect.