Phenolic-Based Chemical Production from Catalytic Depolymerization of Alkaline Lignin over Fumed Silica Catalyst

Lignin depolymerization into phenolic-based chemicals is an interesting process for utilizing and upgrading a benefit and value of lignin. In this study, the depolymerization reaction was performed to convert alkaline lignin into smaller molecule compounds. Fumed SiO₂ was used as a catalyst to improve catalytic activity in lignin decomposition. The important parameters in depolymerization process (i.e., reaction temperature, reaction time, etc.) were also investigated. In addition, gas chromatography with mass spectrometry (GC-MS), flame-ironized detector (GC-FID), and Fourier transform infrared spectroscopy (FT-IR) were used to analyze and characterize the lignin products. It was found that fumed SiO₂ catalyst led the good catalytic activity in lignin depolymerization. The main products from catalytic depolymerization were guaiacol, syringol, vanillin, and phenols. Additionally, metal supported on fumed SiO₂ such as Cu/SiO₂ and Ni/SiO₂ increased the catalyst activity in terms of phenolic products yield.

Assessment of Menus in a Selected Social Welfare Home with Regard to Nutritional Recommendations

The aim of the study was to assess diets of residents of nursing homes. Provided by social welfare home, 10 day menus were introduced into the computer program Diet 5 and analyzed in respect of protein, fats, carbohydrates, energy, vitamin D and calcium. The resulting mean values of 10-day menus were compared with the existing Nutrition Standards for Polish population. The analysis menus showed that the average amount of energy supplied from food is not sufficient. Carbohydrates in food supply are too high and represent 257% of normal. The average value of fats and proteins supplied with food is adequate 85.2 g/day and 75.2 g/day. The calcium content of the diet is 513.9 mg/day. The amount of vitamin D supplied in the age group 51-65 years is 2.3 µg/day. Dietary errors that have been shown are due to the lack of detailed nutritional guidelines for nursing homes, as well as state-owned care facilities in general.

Current Status and Future Trends of Mechanized Fruit Thinning Devices and Sensor Technology

This paper reviews the different concepts that have been investigated concerning the mechanization of fruit thinning as well as multiple working principles and solutions that have been developed for feature extraction of horticultural products, both in the field and industrial environments. The research should be committed towards selective methods, which inevitably need to incorporate some kinds of sensor technology. Computer vision often comes out as an obvious solution for unstructured detection problems, although leaves despite the chosen point of view frequently occlude fruits. Further research on non-traditional sensors that are capable of object differentiation is needed. Ultrasonic and Near Infrared (NIR) technologies have been investigated for applications related to horticultural produce and show a potential to satisfy this need while simultaneously providing spatial information as time of flight sensors. Light Detection and Ranging (LIDAR) technology also shows a huge potential but it implies much greater costs and the related equipment is usually much larger, making it less suitable for portable devices, which may serve a purpose on smaller unstructured orchards. Portable devices may serve a purpose on these types of orchards. In what concerns sensor methods, on-tree fruit detection, major challenge is to overcome the problem of fruits’ occlusion by leaves and branches. Hence, nontraditional sensors capable of providing some type of differentiation should be investigated.

Standalone Docking Station with Combined Charging Methods for Agricultural Mobile Robots

One of the biggest concerns in the field of agriculture is around the energy efficiency of robots that will perform agriculture’s activity and their charging methods. In this paper, two different charging methods for agricultural standalone docking stations are shown that will take into account various variants as field size and its irregularities, work’s nature to which the robot will perform, deadlines that have to be respected, among others. Its features also are dependent on the orchard, season, battery type and its technical specifications and cost. First charging base method focuses on wireless charging, presenting more benefits for small field. The second charging base method relies on battery replacement being more suitable for large fields, thus avoiding the robot stop for recharge. Existing many methods to charge a battery, the CC CV was considered the most appropriate for either simplicity or effectiveness. The choice of the battery for agricultural purposes is if most importance. While the most common battery used is Li-ion battery, this study also discusses the use of graphene-based new type of batteries with 45% over capacity to the Li-ion one. A Battery Management Systems (BMS) is applied for battery balancing. All these approaches combined showed to be a promising method to improve a lot of technical agricultural work, not just in terms of plantation and harvesting but also about every technique to prevent harmful events like plagues and weeds or even to reduce crop time and cost.

Algorithm for Path Recognition in-between Tree Rows for Agricultural Wheeled-Mobile Robots

Machine vision has been widely used in recent years in agriculture, as a tool to promote the automation of processes and increase the levels of productivity. The aim of this work is the development of a path recognition algorithm based on image processing to guide a terrestrial robot in-between tree rows. The proposed algorithm was developed using the software MATLAB, and it uses several image processing operations, such as threshold detection, morphological erosion, histogram equalization and the Hough transform, to find edge lines along tree rows on an image and to create a path to be followed by a mobile robot. To develop the algorithm, a set of images of different types of orchards was used, which made possible the construction of a method capable of identifying paths between trees of different heights and aspects. The algorithm was evaluated using several images with different characteristics of quality and the results showed that the proposed method can successfully detect a path in different types of environments.

Reaction Rate of Olive Stone during Combustion in a Bubbling Fluidized Bed

Combustion of biomass is a promising alternative to reduce the high pollutant emission levels associated to the combustion of fossil flues due to the net null emission of CO2 attributed to biomass. However, the biomass selected should also have low contents of nitrogen and sulfur to limit the NOx and SOx emissions derived from its combustion. In this sense, olive stone is an excellent fuel to power combustion reactors with reduced levels of pollutant emissions. In this work, the combustion of olive stone particles is analyzed experimentally in a thermogravimetric analyzer (TGA) and in a bubbling fluidized bed reactor (BFB). The bubbling fluidized bed reactor was installed over a scale, conforming a macro-TGA. In both equipment, the evolution of the mass of the samples was registered as the combustion process progressed. The results show a much faster combustion process in the bubbling fluidized bed reactor compared to the thermogravimetric analyzer measurements, due to the higher heat transfer coefficient and the abrasion of the fuel particles by the bed material in the BFB reactor.

The Study of Formal and Semantic Errors of Lexis by Persian EFL Learners

Producing a text in a language which is not one’s mother tongue can be a demanding task for language learners. Examining lexical errors committed by EFL learners is a challenging area of investigation which can shed light on the process of second language acquisition. Despite the considerable number of investigations into grammatical errors, few studies have tackled formal and semantic errors of lexis committed by EFL learners. The current study aimed at examining Persian learners’ formal and semantic errors of lexis in English. To this end, 60 students at three different proficiency levels were asked to write on 10 different topics in 10 separate sessions. Finally, 600 essays written by Persian EFL learners were collected, acting as the corpus of the study. An error taxonomy comprising formal and semantic errors was selected to analyze the corpus. The formal category covered misselection and misformation errors, while the semantic errors were classified into lexical, collocational and lexicogrammatical categories. Each category was further classified into subcategories depending on the identified errors. The results showed that there were 2583 errors in the corpus of 9600 words, among which, 2030 formal errors and 553 semantic errors were identified. The most frequent errors in the corpus included formal error commitment (78.6%), which were more prevalent at the advanced level (42.4%). The semantic errors (21.4%) were more frequent at the low intermediate level (40.5%). Among formal errors of lexis, the highest number of errors was devoted to misformation errors (98%), while misselection errors constituted 2% of the errors. Additionally, no significant differences were observed among the three semantic error subcategories, namely collocational, lexical choice and lexicogrammatical. The results of the study can shed light on the challenges faced by EFL learners in the second language acquisition process.

Using TRACE, PARCS, and SNAP Codes to Analyze the Load Rejection Transient of ABWR

The purpose of the study is to analyze the load rejection transient of ABWR by using TRACE, PARCS, and SNAP codes. This study has some steps. First, using TRACE, PARCS, and SNAP codes establish the model of ABWR. Second, the key parameters are identified to refine the TRACE/PARCS/SNAP model further in the frame of a steady state analysis. Third, the TRACE/PARCS/SNAP model is used to perform the load rejection transient analysis. Finally, the FSAR data are used to compare with the analysis results. The results of TRACE/PARCS are consistent with the FSAR data for the important parameters. It indicates that the TRACE/PARCS/SNAP model of ABWR has a good accuracy in the load rejection transient.

A Comparative Study of Single- and Multi-Walled Carbon Nanotube Incorporation to Indium Tin Oxide Electrodes for Solar Cells

Alternative electrode materials for optoelectronic devices have been widely investigated in recent years. Since indium tin oxide (ITO) is the most preferred transparent conductive electrode, producing ITO films by simple and cost-effective solution-based techniques with enhanced optical and electrical properties has great importance. In this study, single- and multi-walled carbon nanotubes (SWCNT and MWCNT) incorporated into the ITO structure to increase electrical conductivity, mechanical strength, and chemical stability. Carbon nanotubes (CNTs) were firstly functionalized by acid treatment (HNO3:H2SO4), and the thermal resistance of CNTs after functionalization was determined by thermogravimetric analysis (TGA). Thin films were then prepared by spin coating technique and characterized by scanning electron microscopy (SEM), X-ray diffraction (XRD), four-point probe measurement system and UV-Vis spectrophotometer. The effects of process parameters were compared for ITO, MWCNT-ITO, and SWCNT-ITO films. Two factors including CNT concentration and annealing temperature were considered. The UV-Vis measurements demonstrated that the transmittance of ITO films was 83.58% at 550 nm, which was decreased depending on the concentration of CNT dopant. On the other hand, both CNT dopants provided an enhancement in the crystalline structure and electrical conductivity. Due to compatible diameter and better dispersibility of SWCNTs in the ITO solution, the best result in terms of electrical conductivity was obtained by SWCNT-ITO films with the 0.1 g/L SWCNT dopant concentration and heat-treatment at 550 °C for 1 hour.

Impact of Quality Assurance Mechanisms on the Work Efficiency of Staff in the Educational Space of Georgia

At this stage, Georgia is a country which is actively involved in the European integration process, for which the primary priority is effective integration in the European education system. The modern Georgian higher education system is the process of establishing a new sociocultural reality, whose main priorities are determined by the Quality System as a continuous cycle of planning, implementation, checking and acting. Obviously, in this situation, the issue of management of education institutions comes out in the foreground, since the proper planning and implementation of personnel management processes is one of the main determinants of the company's performance. At the same time, one of the most important factors is the psychological comfort of the personnel, ensuring their protection and efficiency of stress management policy. The purpose of this research is to determine how intensely the relationship is between the psychological comfort of the personnel and the efficiency of the quality system in the institution as the quality assurance mechanisms of educational institutions affect the stability of personnel, prevention and management of the stressful situation. The research was carried out within the framework of the Internal Grant Project «The Role of Organizational Culture in the Process of Settlement of Management of Stress and Conflict, Georgian Reality and European Experience » of the Batumi Navigation Teaching University, based on the analysis of the survey results of target groups. The small-scale research conducted by us has revealed that the introduction of quality assurance system and its active implementation increased the quality of management of Georgian educational institutions, increased the level of universal engagement in internal and external processes and as a result, it has improved the quality of education as well as social and psychological comfort indicators of the society.

Statistical Texture Analysis

This paper presents an overview of the methodologies and algorithms for statistical texture analysis of 2D images. Methods for digital-image texture analysis are reviewed based on available literature and research work either carried out or supervised by the authors.

Estimating the Traffic Impacts of Green Light Optimal Speed Advisory Systems Using Microsimulation

Even though signalised intersections are necessary for urban road traffic management, they can act as bottlenecks and disrupt traffic operations. Interrupted traffic flow causes congestion, delays, stop-and-go conditions (i.e. excessive acceleration/deceleration) and longer journey times. Vehicle and infrastructure connectivity offers the potential to provide improved new services with additional functions of assisting drivers. This paper focuses on one of the applications of vehicle-to-infrastructure communication namely Green Light Optimal Speed Advisory (GLOSA). To assess the effectiveness of GLOSA in the urban road network, an integrated microscopic traffic simulation framework is built into VISSIM software. Vehicle movements and vehicle-infrastructure communications are simulated through the interface of External Driver Model. A control algorithm is developed for recommending an optimal speed that is continuously updated in every time step for all vehicles approaching a signal-controlled point. This algorithm allows vehicles to pass a traffic signal without stopping or to minimise stopping times at a red phase. This study is performed with all connected vehicles at 100% penetration rate. Conventional vehicles are also simulated in the same network as a reference. A straight road segment composed of two opposite directions with two traffic lights per lane is studied. The simulation is implemented under 150 vehicles per hour and 200 per hour traffic volume conditions to identify how different traffic densities influence the benefits of GLOSA. The results indicate that traffic flow is improved by the application of GLOSA. According to this study, vehicles passed through the traffic lights more smoothly, and waiting times were reduced by up to 28 seconds. Average delays decreased for the entire network by 86.46% and 83.84% under traffic densities of 150 vehicles per hour per lane and 200 vehicles per hour per lane, respectively.

Effect of Columns Stiffness's and Number of Floors on the Accuracy of the Tributary Area Method

The using of finite element programs in analyzing and designing buildings are becoming very popular, but there are many engineers still using the tributary area method (TAM) in designing the structural members such as columns. This study is an attempt to investigate the accuracy of the TAM results with different load condition (gravity and lateral load), different floors numbers, and different columns stiffness's. To conduct this study, linear elastic analysis in ETABS program is used. The results from finite element method are compared to those obtained from TAM. According to the analysis of the data obtained, it can be seen that there is significance difference between the real load carried by columns and the load which is calculated by using the TAM. Thus, using 3-D models are the best choice to calculate the real load effected on columns and design these columns according to this load.

Application of AIMSUN Microscopic Simulation Model in Evaluating Side Friction Impacts on Traffic Stream Performance

Side friction factors can be defined as all activities taking place at the side of the road and within the traffic stream, which would negatively affect the traffic stream performance. If the effect of these factors is adequately addressed and managed, traffic stream performance and capacity could be improved. The main objective of this paper is to identify and assess the impact of different side friction factors on traffic stream performance of a hypothesized urban arterial road. Hypothetical data were assumed mainly because there is no road operating under ideal conditions, with zero side friction, in the developing countries. This is important for the creation of the base model which is important for comparison purposes. For this purpose, three essential steps were employed. Step one, a hypothetical base model was developed under ideal traffic and geometric conditions. Step two, 18 hypothetical alternative scenarios were developed including side friction factors such as on-road parking, pedestrian movement, and the presence of trucks in the traffic stream. These scenarios were evaluated for one, two, and three lane configurations and under different traffic volumes ranging from low to high. Step three, the impact of side friction, of each scenario, on speed-flow models was evaluated using AIMSUN microscopic traffic simulation software. Generally, it was found that, a noticeable negative shift in the speed flow curves from the base conditions was observed for all scenarios. This indicates negative impact of the side friction factors on free flow speed and traffic stream average speed as well as on capacity.

The Establishment of Probabilistic Risk Assessment Analysis Methodology for Dry Storage Concrete Casks Using SAPHIRE 8

To understand the risk for dry storage concrete casks in the cask loading, transfer, and storage phase, the purpose of this research is to establish the probabilistic risk assessment (PRA) analysis methodology for dry storage concrete casks by using SAPHIRE 8 code. This analysis methodology is used to perform the study of Taiwan nuclear power plants (NPPs) dry storage system. The process of research has three steps. First, the data of the concrete casks and Taiwan NPPs are collected. Second, the PRA analysis methodology is developed by using SAPHIRE 8. Third, the PRA analysis is performed by using this methodology. According to the analysis results, the maximum risk is the multipurpose canister (MPC) drop case.

Data Recording for Remote Monitoring of Autonomous Vehicles

Autonomous vehicles offer the possibility of significant benefits to social welfare. However, fully automated cars might not be going to happen in the near further. To speed the adoption of the self-driving technologies, many governments worldwide are passing laws requiring data recorders for the testing of autonomous vehicles. Currently, the self-driving vehicle, (e.g., shuttle bus) has to be monitored from a remote control center. When an autonomous vehicle encounters an unexpected driving environment, such as road construction or an obstruction, it should request assistance from a remote operator. Nevertheless, large amounts of data, including images, radar and lidar data, etc., have to be transmitted from the vehicle to the remote center. Therefore, this paper proposes a data compression method of in-vehicle networks for remote monitoring of autonomous vehicles. Firstly, the time-series data are rearranged into a multi-dimensional signal space. Upon the arrival, for controller area networks (CAN), the new data are mapped onto a time-data two-dimensional space associated with the specific CAN identity. Secondly, the data are sampled based on differential sampling. Finally, the whole set of data are encoded using existing algorithms such as Huffman, arithmetic and codebook encoding methods. To evaluate system performance, the proposed method was deployed on an in-house built autonomous vehicle. The testing results show that the amount of data can be reduced as much as 1/7 compared to the raw data.

Visualizing Imaging Pathways after Anatomy-Specific Follow-Up Imaging Recommendations

Radiologists routinely make follow-up imaging recommendations, usually based on established clinical practice guidelines, such as the Fleischner Society guidelines for managing lung nodules. In order to ensure optimal care, it is important to make guideline-compliant recommendations, and also for patients to follow-up on these imaging recommendations in a timely manner. However, determining such compliance rates after a specific finding has been observed usually requires many time-consuming manual steps. To address some of these limitations with current approaches, in this paper we discuss a methodology to automatically detect finding-specific follow-up recommendations from radiology reports and create a visualization for relevant subsequent exams showing the modality transitions. Nearly 5% of patients who had a lung related follow-up recommendation continued to have at least eight subsequent outpatient CT exams during a seven year period following the recommendation. Radiologist and section chiefs can use the proposed tool to better understand how a specific patient population is being managed, identify possible deviations from established guideline recommendations and have a patient-specific graphical representation of the imaging pathways for an abstract view of the overall treatment path thus far.

Development of Equivalent Inelastic Springs to Model C-Devices

'C' shape yielding devices (C-devices) are effective tools for introducing supplemental sources of energy dissipation by hysteresis. Studies have shown that C-devices made of mild steel can be successfully applied as integral parts of seismic retrofitting schemes. However, explicit modelling of these devices can become cumbersome, expensive and time consuming. The device under study in this article has been previously used in non-invasive dissipative bracing for seismic retrofitting. The device is cut from a mild steel plate and has an overall shape that resembles that of a rectangular portal frame with circular interior corner transitions to avoid stress concentration and to control the extension of the dissipative region of the device. A number of inelastic finite element (FE) analyses using either inelastic 2D plane stress elements or inelastic fibre frame elements are reported and used to calibrate a 1D equivalent inelastic spring model that effectively reproduces the cyclic response of the device. The more elaborate FE model accounts for the frictional forces developed between the steel plate and the bolts used to connect the C-device to structural members. FE results also allow the visualization of the inelastic regions of the device where energy dissipation is expected to occur. FE analysis results are in a good agreement with experimental observations.

Forecasting Issues in Energy Markets within a Reg-ARIMA Framework

Electricity markets throughout the world have undergone substantial changes. Accurate, reliable, clear and comprehensible modeling and forecasting of different variables (loads and prices in the first instance) have achieved increasing importance. In this paper, we describe the actual state of the art focusing on reg-SARMA methods, which have proven to be flexible enough to accommodate the electricity price/load behavior satisfactory. More specifically, we will discuss: 1) The dichotomy between point and interval forecasts; 2) The difficult choice between stochastic (e.g. climatic variation) and non-deterministic predictors (e.g. calendar variables); 3) The confrontation between modelling a single aggregate time series or creating separated and potentially different models of sub-series. The noteworthy point that we would like to make it emerge is that prices and loads require different approaches that appear irreconcilable even though must be made reconcilable for the interests and activities of energy companies.

Distributed Cost-Based Scheduling in Cloud Computing Environment

Cloud computing can be defined as one of the prominent technologies that lets a user change, configure and access the services online. it can be said that this is a prototype of computing that helps in saving cost and time of a user practically the use of cloud computing can be found in various fields like education, health, banking etc.  Cloud computing is an internet dependent technology thus it is the major responsibility of Cloud Service Providers(CSPs) to care of data stored by user at data centers. Scheduling in cloud computing environment plays a vital role as to achieve maximum utilization and user satisfaction cloud providers need to schedule resources effectively.  Job scheduling for cloud computing is analyzed in the following work. To complete, recreate the task calculation, and conveyed scheduling methods CloudSim3.0.3 is utilized. This research work discusses the job scheduling for circulated processing condition also by exploring on this issue we find it works with minimum time and less cost. In this work two load balancing techniques have been employed: ‘Throttled stack adjustment policy’ and ‘Active VM load balancing policy’ with two brokerage services ‘Advanced Response Time’ and ‘Reconfigure Dynamically’ to evaluate the VM_Cost, DC_Cost, Response Time, and Data Processing Time. The proposed techniques are compared with Round Robin scheduling policy.