Rotor Dynamic Analysis for a Shaft Train by Using Finite Element Method

In the present paper, a large turbo-generator shaft train including a heavy-duty gas turbine engine, a coupling, and a generator is established. The method of analysis is based on finite element simplified model for lateral and torsional vibration calculation. The basic elements of rotor are the shafts and the disks which are represented as circular cross section flexible beams and rigid body elements, respectively. For more accurate results, the gyroscopic effect and bearing dynamics coefficients and function of rotation are taken into account, and for the influence of shear effect, rotor has been modeled in the form of Timoshenko beam. Lateral critical speeds, critical speed map, damped mode shapes, Campbell diagram, zones of instability, amplitudes, phase angles response due to synchronous forces of excitation and amplification factor are calculated. Also, in the present paper, the effect of imbalanced rotor and effects of changing in internal force and temperature are studied.

Multi-Objective Optimization of Run-of-River Small-Hydropower Plants Considering Both Investment Cost and Annual Energy Generation

This paper presents the techno-economic evaluation of run-of-river small-hydropower plants. In this regard, a multi-objective optimization procedure is proposed for the optimal sizing of the hydropower plants, and NSGAII is employed as the optimization algorithm. Annual generated energy and investment cost are considered as the objective functions, and number of generator units (n) and nominal turbine flow rate (QT) constitute the decision variables. Site of Yeripao in Benin is considered as the case study. We have categorized the river of this site using its environmental characteristics: gross head, and first quartile, median, third quartile and mean of flow. Effects of each decision variable on the objective functions are analysed. The results gave Pareto Front which represents the trade-offs between annual energy generation and the investment cost of hydropower plants, as well as the recommended optimal solutions. We noted that with the increase of the annual energy generation, the investment cost rises. Thus, maximizing energy generation is contradictory with minimizing the investment cost. Moreover, we have noted that the solutions of Pareto Front are grouped according to the number of generator units (n). The results also illustrate that the costs per kWh are grouped according to the n and rise with the increase of the nominal turbine flow rate. The lowest investment costs per kWh are obtained for n equal to one and are between 0.065 and 0.180 €/kWh. Following the values of n (equal to 1, 2, 3 or 4), the investment cost and investment cost per kWh increase almost linearly with increasing the nominal turbine flowrate while annual generated. Energy increases logarithmically with increasing of the nominal turbine flowrate. This study made for the Yeripao river can be applied to other rivers with their own characteristics.

The Effect of Foreign Owned Firms and Licensed Manufacturing Agreements on Innovation: Case of Pharmaceutical Firms in Developing Countries

Given the fact that the pharmaceutical industry is a commonly studied sector in the context of innovation, the majority of innovation research is devoted to the developed markets known by high research and development (R&D) assets and intensive innovation. In contrast, in developing countries where R&D assets are very low, there is relatively little research to mention in the area of pharmaceutical sector innovation, characterized mainly by two principal elements which are the presence of foreign-owned firms and licensed manufacturing agreements between local firms and multinationals. With the scarcity of research in this field, this paper attempts to study the effect of these two elements on the firms’ innovation tendencies. Other traditional factors that influence innovation, which are the age and the size of the firm, the R&D activities and the market structure, revealed in the literature review, will be included in the study in order to try to make this work more exhaustive. The study starts by examining innovation tendency in pharmaceutical firms located in developing countries before analyzing the effect of foreign-owned firms and licensed manufacturing agreements between local firms and multinationals on technological, organizational and marketing innovation. Based on the related work and on the theoretical framework developed, there is a probability that foreign-owned firms and licensed manufacturing agreements between local firms and multinationals have a negative influence on technological innovation. The opposite effect is possible in the case of organizational and marketing innovation.

Disaggregation the Daily Rainfall Dataset into Sub-Daily Resolution in the Temperate Oceanic Climate Region

High resolution rain data are very important to fulfill the input of hydrological models. Among models of high-resolution rainfall data generation, the temporal disaggregation was chosen for this study. The paper attempts to generate three different rainfall resolutions (4-hourly, hourly and 10-minutes) from daily for around 20-year record period. The process was done by DiMoN tool which is based on random cascade model and method of fragment. Differences between observed and simulated rain dataset are evaluated with variety of statistical and empirical methods: Kolmogorov-Smirnov test (K-S), usual statistics, and Exceedance probability. The tool worked well at preserving the daily rainfall values in wet days, however, the generated data are cumulated in a shorter time period and made stronger storms. It is demonstrated that the difference between generated and observed cumulative distribution function curve of 4-hourly datasets is passed the K-S test criteria while in hourly and 10-minutes datasets the P-value should be employed to prove that their differences were reasonable. The results are encouraging considering the overestimation of generated high-resolution rainfall data.

Phenolic-Based Chemical Production from Catalytic Depolymerization of Alkaline Lignin over Fumed Silica Catalyst

Lignin depolymerization into phenolic-based chemicals is an interesting process for utilizing and upgrading a benefit and value of lignin. In this study, the depolymerization reaction was performed to convert alkaline lignin into smaller molecule compounds. Fumed SiO₂ was used as a catalyst to improve catalytic activity in lignin decomposition. The important parameters in depolymerization process (i.e., reaction temperature, reaction time, etc.) were also investigated. In addition, gas chromatography with mass spectrometry (GC-MS), flame-ironized detector (GC-FID), and Fourier transform infrared spectroscopy (FT-IR) were used to analyze and characterize the lignin products. It was found that fumed SiO₂ catalyst led the good catalytic activity in lignin depolymerization. The main products from catalytic depolymerization were guaiacol, syringol, vanillin, and phenols. Additionally, metal supported on fumed SiO₂ such as Cu/SiO₂ and Ni/SiO₂ increased the catalyst activity in terms of phenolic products yield.

Assessment of Menus in a Selected Social Welfare Home with Regard to Nutritional Recommendations

The aim of the study was to assess diets of residents of nursing homes. Provided by social welfare home, 10 day menus were introduced into the computer program Diet 5 and analyzed in respect of protein, fats, carbohydrates, energy, vitamin D and calcium. The resulting mean values of 10-day menus were compared with the existing Nutrition Standards for Polish population. The analysis menus showed that the average amount of energy supplied from food is not sufficient. Carbohydrates in food supply are too high and represent 257% of normal. The average value of fats and proteins supplied with food is adequate 85.2 g/day and 75.2 g/day. The calcium content of the diet is 513.9 mg/day. The amount of vitamin D supplied in the age group 51-65 years is 2.3 µg/day. Dietary errors that have been shown are due to the lack of detailed nutritional guidelines for nursing homes, as well as state-owned care facilities in general.

Reaction Rate of Olive Stone during Combustion in a Bubbling Fluidized Bed

Combustion of biomass is a promising alternative to reduce the high pollutant emission levels associated to the combustion of fossil flues due to the net null emission of CO2 attributed to biomass. However, the biomass selected should also have low contents of nitrogen and sulfur to limit the NOx and SOx emissions derived from its combustion. In this sense, olive stone is an excellent fuel to power combustion reactors with reduced levels of pollutant emissions. In this work, the combustion of olive stone particles is analyzed experimentally in a thermogravimetric analyzer (TGA) and in a bubbling fluidized bed reactor (BFB). The bubbling fluidized bed reactor was installed over a scale, conforming a macro-TGA. In both equipment, the evolution of the mass of the samples was registered as the combustion process progressed. The results show a much faster combustion process in the bubbling fluidized bed reactor compared to the thermogravimetric analyzer measurements, due to the higher heat transfer coefficient and the abrasion of the fuel particles by the bed material in the BFB reactor.

Computation of Natural Logarithm Using Abstract Chemical Reaction Networks

Recent researches has focused on nucleic acids as a substrate for designing biomolecular circuits for in situ monitoring and control. A common approach is to express them by a set of idealised abstract chemical reaction networks (ACRNs). Here, we present new results on how abstract chemical reactions, viz., catalysis, annihilation and degradation, can be used to implement circuit that accurately computes logarithm function using the method of Arithmetic-Geometric Mean (AGM), which has not been previously used in conjunction with ACRNs.

The Study of Formal and Semantic Errors of Lexis by Persian EFL Learners

Producing a text in a language which is not one’s mother tongue can be a demanding task for language learners. Examining lexical errors committed by EFL learners is a challenging area of investigation which can shed light on the process of second language acquisition. Despite the considerable number of investigations into grammatical errors, few studies have tackled formal and semantic errors of lexis committed by EFL learners. The current study aimed at examining Persian learners’ formal and semantic errors of lexis in English. To this end, 60 students at three different proficiency levels were asked to write on 10 different topics in 10 separate sessions. Finally, 600 essays written by Persian EFL learners were collected, acting as the corpus of the study. An error taxonomy comprising formal and semantic errors was selected to analyze the corpus. The formal category covered misselection and misformation errors, while the semantic errors were classified into lexical, collocational and lexicogrammatical categories. Each category was further classified into subcategories depending on the identified errors. The results showed that there were 2583 errors in the corpus of 9600 words, among which, 2030 formal errors and 553 semantic errors were identified. The most frequent errors in the corpus included formal error commitment (78.6%), which were more prevalent at the advanced level (42.4%). The semantic errors (21.4%) were more frequent at the low intermediate level (40.5%). Among formal errors of lexis, the highest number of errors was devoted to misformation errors (98%), while misselection errors constituted 2% of the errors. Additionally, no significant differences were observed among the three semantic error subcategories, namely collocational, lexical choice and lexicogrammatical. The results of the study can shed light on the challenges faced by EFL learners in the second language acquisition process.

Using TRACE, PARCS, and SNAP Codes to Analyze the Load Rejection Transient of ABWR

The purpose of the study is to analyze the load rejection transient of ABWR by using TRACE, PARCS, and SNAP codes. This study has some steps. First, using TRACE, PARCS, and SNAP codes establish the model of ABWR. Second, the key parameters are identified to refine the TRACE/PARCS/SNAP model further in the frame of a steady state analysis. Third, the TRACE/PARCS/SNAP model is used to perform the load rejection transient analysis. Finally, the FSAR data are used to compare with the analysis results. The results of TRACE/PARCS are consistent with the FSAR data for the important parameters. It indicates that the TRACE/PARCS/SNAP model of ABWR has a good accuracy in the load rejection transient.

The Establishment of Probabilistic Risk Assessment Analysis Methodology for Dry Storage Concrete Casks Using SAPHIRE 8

To understand the risk for dry storage concrete casks in the cask loading, transfer, and storage phase, the purpose of this research is to establish the probabilistic risk assessment (PRA) analysis methodology for dry storage concrete casks by using SAPHIRE 8 code. This analysis methodology is used to perform the study of Taiwan nuclear power plants (NPPs) dry storage system. The process of research has three steps. First, the data of the concrete casks and Taiwan NPPs are collected. Second, the PRA analysis methodology is developed by using SAPHIRE 8. Third, the PRA analysis is performed by using this methodology. According to the analysis results, the maximum risk is the multipurpose canister (MPC) drop case.

Data Recording for Remote Monitoring of Autonomous Vehicles

Autonomous vehicles offer the possibility of significant benefits to social welfare. However, fully automated cars might not be going to happen in the near further. To speed the adoption of the self-driving technologies, many governments worldwide are passing laws requiring data recorders for the testing of autonomous vehicles. Currently, the self-driving vehicle, (e.g., shuttle bus) has to be monitored from a remote control center. When an autonomous vehicle encounters an unexpected driving environment, such as road construction or an obstruction, it should request assistance from a remote operator. Nevertheless, large amounts of data, including images, radar and lidar data, etc., have to be transmitted from the vehicle to the remote center. Therefore, this paper proposes a data compression method of in-vehicle networks for remote monitoring of autonomous vehicles. Firstly, the time-series data are rearranged into a multi-dimensional signal space. Upon the arrival, for controller area networks (CAN), the new data are mapped onto a time-data two-dimensional space associated with the specific CAN identity. Secondly, the data are sampled based on differential sampling. Finally, the whole set of data are encoded using existing algorithms such as Huffman, arithmetic and codebook encoding methods. To evaluate system performance, the proposed method was deployed on an in-house built autonomous vehicle. The testing results show that the amount of data can be reduced as much as 1/7 compared to the raw data.

Development of Equivalent Inelastic Springs to Model C-Devices

'C' shape yielding devices (C-devices) are effective tools for introducing supplemental sources of energy dissipation by hysteresis. Studies have shown that C-devices made of mild steel can be successfully applied as integral parts of seismic retrofitting schemes. However, explicit modelling of these devices can become cumbersome, expensive and time consuming. The device under study in this article has been previously used in non-invasive dissipative bracing for seismic retrofitting. The device is cut from a mild steel plate and has an overall shape that resembles that of a rectangular portal frame with circular interior corner transitions to avoid stress concentration and to control the extension of the dissipative region of the device. A number of inelastic finite element (FE) analyses using either inelastic 2D plane stress elements or inelastic fibre frame elements are reported and used to calibrate a 1D equivalent inelastic spring model that effectively reproduces the cyclic response of the device. The more elaborate FE model accounts for the frictional forces developed between the steel plate and the bolts used to connect the C-device to structural members. FE results also allow the visualization of the inelastic regions of the device where energy dissipation is expected to occur. FE analysis results are in a good agreement with experimental observations.

Distributed Cost-Based Scheduling in Cloud Computing Environment

Cloud computing can be defined as one of the prominent technologies that lets a user change, configure and access the services online. it can be said that this is a prototype of computing that helps in saving cost and time of a user practically the use of cloud computing can be found in various fields like education, health, banking etc.  Cloud computing is an internet dependent technology thus it is the major responsibility of Cloud Service Providers(CSPs) to care of data stored by user at data centers. Scheduling in cloud computing environment plays a vital role as to achieve maximum utilization and user satisfaction cloud providers need to schedule resources effectively.  Job scheduling for cloud computing is analyzed in the following work. To complete, recreate the task calculation, and conveyed scheduling methods CloudSim3.0.3 is utilized. This research work discusses the job scheduling for circulated processing condition also by exploring on this issue we find it works with minimum time and less cost. In this work two load balancing techniques have been employed: ‘Throttled stack adjustment policy’ and ‘Active VM load balancing policy’ with two brokerage services ‘Advanced Response Time’ and ‘Reconfigure Dynamically’ to evaluate the VM_Cost, DC_Cost, Response Time, and Data Processing Time. The proposed techniques are compared with Round Robin scheduling policy.

A Recommendation to Oncologists for Cancer Treatment by Immunotherapy: Quantitative and Qualitative Analysis

Today, the treatment of cancer, in a relatively short period, with minimum adverse effects is a great concern for oncologists. In this paper, based on a recently used mathematical model for cancer, a guideline has been proposed for the amount and duration of drug doses for cancer treatment by immunotherapy. Dynamically speaking, the mathematical ordinary differential equation (ODE) model of cancer has different equilibrium points; one of them is unstable, which is called the no tumor equilibrium point. In this paper, based on the number of tumor cells an intelligent soft computing controller (a combination of fuzzy logic controller and genetic algorithm), decides regarding the amount and duration of drug doses, to eliminate the tumor cells and stabilize the unstable point in a relatively short time. Two different immunotherapy approaches; active and adoptive, have been studied and presented. It is shown that the rate of decay of tumor cells is faster and the doses of drug are lower in comparison with the result of some other literatures. It is also shown that the period of treatment and the doses of drug in adoptive immunotherapy are significantly less than the active method. A recommendation to oncologists has also been presented.

Project Management at University: Towards an Evaluation Process around Cooperative Learning

The enrollment in current Master's degree programs usually pursues gaining the expertise required in real-life workplaces. The experience we present here concerns the learning process of "Project Management Methodology (PMM)", around a cooperative/collaborative mechanism aimed at affording students measurable learning goals and providing the teacher with the ability of focusing on the weaknesses detected. We have designed a mixed summative/formative evaluation, which assures curriculum engage while enriches the comprehension of PMM key concepts. In this experience we converted the students into active actors in the evaluation process itself and we endowed ourselves as teachers with a flexible process in which along with qualifications (score), other attitudinal feedback arises. Despite the high level of self-affirmation on their discussion within the interactive assessment sessions, they ultimately have exhibited a great ability to review and correct the wrong reasoning when that was the case.

A Risk Assessment for the Small Hive Beetle Based on Meteorological Standard Measurements

The Small Hive Beetle, Aethina tumida (Coleoptera: Nitidulidae) is a parasite for honey bee colonies, Apis mellifera, and was recently introduced to the European continent, accidentally. Based on the literature, a model was developed by using regional meteorological variables (daily values of minimum, maximum and mean air temperature as well as mean soil temperature at 50 mm depth) to calculate the time-point of hive invasion by A. tumida in springtime, the development duration of pupae as well as the number of generations of A. tumida per year. Luxembourg was used as a test region for our model for 2005 to 2013. The model output indicates a successful surviving of the Small Hive Beetle in Luxembourg with two up to three generations per year. Additionally, based on our meteorological data sets a first migration of SHB to apiaries can be expected from mid of March up to April. Our approach can be transferred easily to other countries to estimate the risk potential for a successful introduction and spreading of A. tumida in Western Europe.

Simulation on Fuel Metering Unit Used for TurboShaft Engine Model

Fuel Metering Unit (FMU) in fuel system of an aeroengine sometimes has direct influence on the engine performance, which is neglected for the sake of easy access to mathematical model of the engine in most cases. In order to verify the influence of FMU on an engine model, this paper presents a co-simulation of a stepping motor driven FMU (digital FMU) in a turboshaft aeroengine, using AMESim and MATLAB to obtain the steady and dynamic characteristics of the FMU. For this method, mechanical and hydraulic section of the unit is modeled through AMESim, while the stepping motor is mathematically modeled through MATLAB/Simulink. Combining these two sub-models yields an AMESim/MATLAB co-model of the FMU. A simplified component level model for the turboshaft engine is established and connected with the FMU model. Simulation results on the full model show that the engine model considering FMU characteristics describes the engine more precisely especially in its transition state. An FMU dynamics will cut down the rotation speed of the high pressure shaft and the inlet pressure of the combustor during the step response. The work in this paper reveals the impact of FMU on engine operation characteristics and provides a reference to an engine model for ground tests.

Automatic Thresholding for Data Gap Detection for a Set of Sensors in Instrumented Buildings

Building systems are highly vulnerable to different kinds of faults and failures. In fact, various faults, failures and human behaviors could affect the building performance. This paper tackles the detection of unreliable sensors in buildings. Different literature surveys on diagnosis techniques for sensor grids in buildings have been published but all of them treat only bias and outliers. Occurences of data gaps have also not been given an adequate span of attention in the academia. The proposed methodology comprises the automatic thresholding for data gap detection for a set of heterogeneous sensors in instrumented buildings. Sensor measurements are considered to be regular time series. However, in reality, sensor values are not uniformly sampled. So, the issue to solve is from which delay each sensor become faulty? The use of time series is required for detection of abnormalities on the delays. The efficiency of the method is evaluated on measurements obtained from a real power plant: an office at Grenoble Institute of technology equipped by 30 sensors.

A Review on Image Segmentation Techniques and Performance Measures

Image segmentation is a method to extract regions of interest from an image. It remains a fundamental problem in computer vision. The increasing diversity and the complexity of segmentation algorithms have led us firstly, to make a review and classify segmentation techniques, secondly to identify the most used measures of segmentation performance and thirdly, discuss deeply on segmentation philosophy in order to help the choice of adequate segmentation techniques for some applications. To justify the relevance of our analysis, recent algorithms of segmentation are presented through the proposed classification.