Analysis of Air-Water Two-Phase Flow in a 3x3 Rod Bundle

This study investigated the void fraction characteristics under low superficial gas velocity (Jg) and low superficial fluid velocity (Jf) conditions in a 3x3 rod bundle geometry. Three arrangements of conductivity probes were set to measure the void fraction at various cross-sectional regions, including rod-gap, sub-channel and rod-wall regions. The experimental tests were performed under the flow conditions of Jg = 0-0.236 m/s and Jf = 0-0.142 m/s, and the time-averaged void fractions were recorded at each flow condition. It was observed that while the superficial gas velocity increases, the small bubbles started to cluster together and become big bubbles. As the superficial fluid velocity increases, the local void fractions of the three test regions will get closer and the bubble distribution will be more uniform across the cross section.

Chemical Leaching of Metals from Landfill’s Fine Fraction

Leaching of heavy metals (chromium, zinc, copper) from the fine fraction of the Torma landfill (Estonia) was investigated. The leaching kinetics studies have determined the dependence of some metal’s concentration on the leaching time. Metals were leached with Aqua Regia, distilled water and EDTA (Ethylenediaminetetraacetic acid); process was most intensive 2 hours after the start of the experiment, except for copper with EDTA (0.5 h) and lead with EDTA (4 h). During leaching, steady concentrations of Fe, Mn, Cd and Pb were fully stabilized after 8 h; however concentrations of Cu and Ni were not stabilized after 10 h.

Effect of Halo Protection Device on the Aerodynamic Performance of Formula Racecar

This paper explores the aerodynamics of the formula racecar when a ‘halo’ driver-protection device is added to the chassis. The halo protection device was introduced at the start of the 2018 racing season as a safety measure against foreign object impacts that a driver may encounter when driving an open-wheel racecar. In the one-year since its introduction, the device has received wide acclaim for protecting the driver on two separate occasions. The benefit of such a safety device certainly cannot be disputed. However, by adding the halo device to a car, it changes the airflow around the vehicle, and most notably, to the engine air-intake and the rear wing. These negative effects in the air supply to the engine, and equally to the downforce created by the rear wing are studied in this paper using numerical technique, and the resulting CFD outputs are presented and discussed. Comparing racecar design prior to and after the introduction of the halo device, it is shown that the design of the air intake and the rear wing has not followed suit since the addition of the halo device. The reduction of engine intake mass flow due to the halo device is computed and presented for various speeds the car may be going. Because of the location of the halo device in relation to the air intake, airflow is directed away from the engine, making the engine perform less than optimal. The reduction is quantified in this paper to show the correspondence to reduce the engine output when compared to a similar car without the halo device. This paper shows that through aerodynamic arguments, the engine in a halo car will not receive unobstructed, clean airflow that a non-halo car does. Another negative effect is on the downforce created by the rear wing. Because the amount of downforce created by the rear wing is influenced by every component that comes before it, when a halo device is added upstream to the rear wing, airflow is obstructed, and less is available for making downforce. This reduction in downforce is especially dramatic as the speed is increased. This paper presents a graph of downforce over a range of speeds for a car with and without the halo device. Acknowledging that although driver safety is paramount, the negative effect of this safety device on the performance of the car should still be well understood so that any possible redesign to mitigate these negative effects can be taken into account in next year’s rules regulation.

Sustainable Hydrogel Nanocomposites Based on Grafted Chitosan and Clay for Effective Adsorption of Cationic Dye

Contamination of water, due to the discharge of untreated industrial wastewaters into the ecosystem, has become a serious problem for many countries. In this study, bioadsorbents based on chitosan-g-poly(acrylamide) and montmorillonite (MMt) clay (CTS-g-PAAm/MMt) hydrogel nanocomposites were prepared via free‐radical grafting copolymerization and crosslinking of acrylamide monomer (AAm) onto natural polysaccharide chitosan (CTS) as backbone, in presence of various contents of MMt clay as nanofiller. Then, they were hydrolyzed to obtain highly functionalized pH‐sensitive nanomaterials with uppermost swelling properties. Their structure characterization was conducted by X-Ray Diffraction (XRD) and Scanning Electron Microscopy (SEM) analyses. The adsorption performances of the developed nanohybrids were examined for removal of methylene blue (MB) cationic dye from aqueous solutions. The factors affecting the removal of MB, such as clay content, pH medium, adsorbent dose, initial dye concentration and temperature were explored. The adsorption process was found to be highly pH dependent. From adsorption kinetic results, the prepared adsorbents showed remarkable adsorption capacity and fast adsorption rate, mainly more than 88% of MB removal efficiency was reached after 50 min in 200 mg L-1 of dye solution. In addition, the incorporating of various content of clay has enhanced adsorption capacity of CTS-g-PAAm matrix from 1685 to a highest value of 1749 mg g-1 for the optimized nanocomposite containing 2 wt.% of MMt. The experimental kinetic data were well described by the pseudo-second-order model, while the equilibrium data were represented perfectly by Langmuir isotherm model. The maximum Langmuir equilibrium adsorption capacity (qm) was found to increase from 2173 mg g−1 until 2221 mg g−1 by adding 2 wt.% of clay nanofiller. Thermodynamic parameters revealed the spontaneous and endothermic nature of the process. In addition, the reusability study revealed that these bioadsorbents could be well regenerated with desorption efficiency overhead 87% and without any obvious decrease of removal efficiency as compared to starting ones even after four consecutive adsorption/desorption cycles, which exceeded 64%. These results suggest that the optimized nanocomposites are promising as low cost bioadsorbents.

Understanding the Architecture of Hindu Temples: A Philosophical Interpretation

Vedic philosophy is one of the oldest existing philosophies of the world. Started around 6500 BC, in Western Indian subcontinent, the Indus valley Civilizations developed a theology which, gradually developed into a well-established philosophy of beliefs, popularly known as ‘Hindu religion’. In Vedic theology, the abstract concept of God was formulated mostly by close observation of the dynamicity and the recurrence of natural and universal phenomena. Through the ages, the philosophy of this theology went through various discursions, debates, and questionings and the abstract concept of God was, in time, formalized into more representational forms by the means of various signs and symbols. Often, these symbols were used in more subtle ways in the construction of “sacred” sculptures and structures. Apparently, two different philosophies were developed from the Vedic philosophy and these two philosophies are mostly seen in the northern part and southern part of the Indian subcontinent. This paper tries to summarize the complex philosophical treaties of Hinduism of northern and southern India and seeks to understand the meanings of the various signs and symbolisms that were incorporated in the architecture of Hindu temples, including the names given to various parts of the temples. The Hindu temples are not only places of worship or ‘houses of Gods’ like the Greek and Roman temples but are also structures that symbolize the dynamicity and also spiritual upliftment of human beings.

Developing Improvements to Multi-Hazard Risk Assessments

This paper outlines the approaches taken to assess multi-hazard assessments. There is currently confusion in assessing multi-hazard impacts, and so this study aims to determine which of the available options are the most useful. The paper uses an international literature search, and analysis of current multi-hazard assessments and a case study to illustrate the effectiveness of the chosen method. Findings from this study will help those wanting to assess multi-hazards to undertake a straightforward approach. The paper is significant as it helps to interpret the various approaches and concludes with the preferred method. Many people in the world live in hazardous environments and are susceptible to disasters. Unfortunately, when a disaster strikes it is often compounded by additional cascading hazards, thus people would confront more than one hazard simultaneously. Hazards include natural hazards (earthquakes, floods, etc.) or cascading human-made hazards (for example, Natural Hazard Triggering Technological disasters (Natech) such as fire, explosion, toxic release). Multi-hazards have a more destructive impact on urban areas than one hazard alone. In addition, climate change is creating links between different disasters such as causing landslide dams and debris flows leading to more destructive incidents. Much of the prevailing literature deals with only one hazard at a time. However, recently sophisticated multi-hazard assessments have started to appear. Given that multi-hazards occur, it is essential to take multi-hazard risk assessment under consideration. This paper aims to review the multi-hazard assessment methods through articles published to date and categorize the strengths and disadvantages of using these methods in risk assessment. Napier City is selected as a case study to demonstrate the necessity of using multi-hazard risk assessments. In order to assess multi-hazard risk assessments, first, the current multi-hazard risk assessment methods were described. Next, the drawbacks of these multi-hazard risk assessments were outlined. Finally, the improvements to current multi-hazard risk assessments to date were summarised. Generally, the main problem of multi-hazard risk assessment is to make a valid assumption of risk from the interactions of different hazards. Currently, risk assessment studies have started to assess multi-hazard situations, but drawbacks such as uncertainty and lack of data show the necessity for more precise risk assessment. It should be noted that ignoring or partial considering multi-hazards in risk assessment will lead to an overestimate or overlook in resilient and recovery action managements.

A Design of Anisotropic Wet Etching System to Reduce Hillocks on Etched Surface of Silicon Substrate

This research aims to design and build a wet etching system, which is suitable for anisotropic wet etching, in order to reduce etching time, to reduce hillocks on the etched surface (to reduce roughness), and to create a 45-degree wall angle (micro-mirror). This study would start by designing a wet etching system. There are four main components in this system: an ultrasonic cleaning, a condenser, a motor and a substrate holder. After that, an ultrasonic machine was modified by applying a condenser to maintain the consistency of the solution concentration during the etching process and installing a motor for improving the roughness. This effect on the etch rate and the roughness showed that the etch rate increased and the roughness was reduced.

Exporting Physiochemical Changes during the Fermentation of Aloe Vera

Aloe Vera is a short-stemmed succulent plant which is commonly used in Myanmar traditional medicine. A. vera gel was also used as food addictive. This study aims to improve the Myanmar folk medicine to a functional beverage. In this research, Aloe vera was fermented with Saccharomyces cerevisiae for 6 months. Three different processes were carried out. Process I contains A. vera 10%, sugar 30%, water 50%, and starter culture 10%, process II contains A. vera 10%, sugar 15%, honey 15%, and water 50%, starter culture 10%; process III contains A. vera 10%, honey 30%, water 50%, starter culture 10%. During wine fermentation, the wine parameters such as alcohol content, total soluble solid (ºBrix), pH, color and cell population were analyzed. After 30 days of fermentation, total cell population remained 2.8x106 in P-I, P-II and 3.2x106 in P-III. Total soluble solid content dropped to 15.8 in P-I, P-II and 15.7 in P-III. After 30 days, clear wine was transferred to other vassals for racking. After 6 months of racking, microbial population reached under detectable level and alcohol content was round about 11% but not significantly different among these processes. P-II was found to have the highest color intensity at 450 nm and it got the most taster satisfaction when sensory evaluation was carried out using five hedonic scales after 6 month of racking.

Development of Requirements Analysis Tool for Medical Autonomy in Long-Duration Space Exploration Missions

Improving resources for medical autonomy of astronauts in prolonged space missions, such as a Mars mission, requires not only technology development, but also decision-making support systems. The Advanced Crew Medical System - Medical Condition Requirements study, funded by the Canadian Space Agency, aimed to create knowledge content and a scenario-based query capability to support medical autonomy of astronauts. The key objective of this study was to create a prototype tool for identifying medical infrastructure requirements in terms of medical knowledge, skills and materials. A multicriteria decision-making method was used to prioritize the highest risk medical events anticipated in a long-term space mission. Starting with those medical conditions, event sequence diagrams (ESDs) were created in the form of decision trees where the entry point is the diagnosis and the end points are the predicted outcomes (full recovery, partial recovery, or death/severe incapacitation). The ESD formalism was adapted to characterize and compare possible outcomes of medical conditions as a function of available medical knowledge, skills, and supplies in a given mission scenario. An extensive literature review was performed and summarized in a medical condition database. A PostgreSQL relational database was created to allow query-based evaluation of health outcome metrics with different medical infrastructure scenarios. Critical decision points, skill and medical supply requirements, and probable health outcomes were compared across chosen scenarios. The three medical conditions with the highest risk rank were acute coronary syndrome, sepsis, and stroke. Our efforts demonstrate the utility of this approach and provide insight into the effort required to develop appropriate content for the range of medical conditions that may arise.

Analysis of Network Performance Using Aspect of Quantum Cryptography

Quantum cryptography is described as a point-to-point secure key generation technology that has emerged in recent times in providing absolute security. Researchers have started studying new innovative approaches to exploit the security of Quantum Key Distribution (QKD) for a large-scale communication system. A number of approaches and models for utilization of QKD for secure communication have been developed. The uncertainty principle in quantum mechanics created a new paradigm for QKD. One of the approaches for use of QKD involved network fashioned security. The main goal was point-to-point Quantum network that exploited QKD technology for end-to-end network security via high speed QKD. Other approaches and models equipped with QKD in network fashion are introduced in the literature as. A different approach that this paper deals with is using QKD in existing protocols, which are widely used on the Internet to enhance security with main objective of unconditional security. Our work is towards the analysis of the QKD in Mobile ad-hoc network (MANET).

Review of the Road Crash Data Availability in Iraq

Iraq is a middle income country where the road safety issue is considered one of the leading causes of deaths. To control the road risk issue, the Iraqi Ministry of Planning, General Statistical Organization started to organise a collection system of traffic accidents data with details related to their causes and severity. These data are published as an annual report. In this paper, a review of the available crash data in Iraq will be presented. The available data represent the rate of accidents in aggregated level and classified according to their types, road users’ details, and crash severity, type of vehicles, causes and number of causalities. The review is according to the types of models used in road safety studies and research, and according to the required road safety data in the road constructions tasks. The available data are also compared with the road safety dataset published in the United Kingdom as an example of developed country. It is concluded that the data in Iraq are suitable for descriptive and exploratory models, aggregated level comparison analysis, and evaluation and monitoring the progress of the overall traffic safety performance. However, important traffic safety studies require disaggregated level of data and details related to the factors of the likelihood of traffic crashes. Some studies require spatial geographic details such as the location of the accidents which is essential in ranking the roads according to their level of safety, and name the most dangerous roads in Iraq which requires tactic plan to control this issue. Global Road safety agencies interested in solve this problem in low and middle-income countries have designed road safety assessment methodologies which are basing on the road attributes data only. Therefore, in this research it is recommended to use one of these methodologies.

Religion versus Secularism on Women’s Liberation: The Question of Women Liberation and Modern Education

The nineteenth century was characterized by major educational reforms in the Arab World. One of the unintended outcomes of colonization in Arab countries was the initiation of women liberation as well as the introduction of modern education and its application in sensitizing people on the rights of women and their liberation. The reforms were often attributed to various undercurrents that took place at different levels within the Ottoman Empire, and particularly the arrival and influence of the Christian missionaries were supported by the American and European governments. These trends were also significantly attributed to the increase in the presence of Europeans in the region, as well as the introduction of secular ideas and approaches related to the meaning of modernity. Using literary analysis as a method, this paper examines the role of an important male figure like the political activist and writer Qāsim Amīn and the religious reformer Muḥammad ʻAbduh in starting this discourse and shows their impact on the emancipation of women movement (Taḥrīr), and how later women led the movement with their published work. This paper explores Arab Salons and the initiation of women’s literary circles. Women from wealthy families in Egypt and Syria who had studied in Europe or interacted with European counterparts began these circles. These salons acted as central locations where people could meet and hold discussions on political, social, and literary trends as they happened each day. The paper concludes with a discussion of current debates between the Islamist and the secularist branches of the movement today. While the Islamists believe that adhering to the core of Islam with some of its contested position on women is a modern ideology of liberation that fits the current culture of modern time Egypt; the secularists argue that the influence that Islam has on the women’s liberation movement in Egypt has been a threat to the natural success and progress of the movement, which was initiated in the early nineteenth century independent of the more recent trends towards religiosity in the country.

Spatial Disparity in Education and Medical Facilities: A Case Study of Barddhaman District, West Bengal, India

The economic scenario of any region does not show the real picture for the measurement of overall development. Therefore, economic development must be accompanied by social development to be able to make an assessment to measure the level of development. The spatial variation with respect to social development has been discussed taking into account the quality of functioning of a social system in a specific area. In this paper, an attempt has been made to study the spatial distribution of social infrastructural facilities and analyze the magnitude of regional disparities at inter- block level in Barddhman district. It starts with the detailed account of the selection process of social infrastructure indicators and describes the methodology employed in the empirical analysis. Analyzing the block level data, this paper tries to identify the disparity among the blocks in the levels of social development. The results have been subsequently explained using both statistical analysis and geo spatial technique. The paper reveals that the social development is not going on at the same rate in every part of the district. Health facilities and educational facilities are concentrated at some selected point. So overall development activities come to be concentrated in a few centres and the disparity is seen over the blocks.

Assessment and Uncertainty Analysis of ROSA/LSTF Test on Pressurized Water Reactor 1.9% Vessel Upper Head Small-Break Loss-of-Coolant Accident

An experiment utilizing the ROSA/LSTF (rig of safety assessment/large-scale test facility) simulated a 1.9% vessel upper head small-break loss-of-coolant accident with an accident management (AM) measure under the total failure of high-pressure injection system of emergency core cooling system in a pressurized water reactor. Steam generator (SG) secondary-side depressurization on the AM measure was started by fully opening relief valves in both SGs when the maximum core exit temperature rose to 623 K. A large increase took place in the cladding surface temperature of simulated fuel rods on account of a late and slow response of core exit thermocouples during core boil-off. The author analyzed the LSTF test by reference to the matrix of an integral effect test for the validation of a thermal-hydraulic system code. Problems remained in predicting the primary coolant distribution and the core exit temperature with the RELAP5/MOD3.3 code. The uncertainty analysis results of the RELAP5 code confirmed that the sample size with respect to the order statistics influences the value of peak cladding temperature with a 95% probability at a 95% confidence level, and the Spearman’s rank correlation coefficient.

Simultaneous Optimization of Design and Maintenance through a Hybrid Process Using Genetic Algorithms

In general, issues related to design and maintenance are considered in an independent manner. However, the decisions made in these two sets influence each other. The design for maintenance is considered an opportunity to optimize the life cycle cost of a product, particularly in the nuclear or aeronautical field, where maintenance expenses represent more than 60% of life cycle costs. The design of large-scale systems starts with product architecture, a choice of components in terms of cost, reliability, weight and other attributes, corresponding to the specifications. On the other hand, the design must take into account maintenance by improving, in particular, real-time monitoring of equipment through the integration of new technologies such as connected sensors and intelligent actuators. We noticed that different approaches used in the Design For Maintenance (DFM) methods are limited to the simultaneous characterization of the reliability and maintainability of a multi-component system. This article proposes a method of DFM that assists designers to propose dynamic maintenance for multi-component industrial systems. The term "dynamic" refers to the ability to integrate available monitoring data to adapt the maintenance decision in real time. The goal is to maximize the availability of the system at a given life cycle cost. This paper presents an approach for simultaneous optimization of the design and maintenance of multi-component systems. Here the design is characterized by four decision variables for each component (reliability level, maintainability level, redundancy level, and level of monitoring data). The maintenance is characterized by two decision variables (the dates of the maintenance stops and the maintenance operations to be performed on the system during these stops). The DFM model helps the designers choose technical solutions for the large-scale industrial products. Large-scale refers to the complex multi-component industrial systems and long life-cycle, such as trains, aircraft, etc. The method is based on a two-level hybrid algorithm for simultaneous optimization of design and maintenance, using genetic algorithms. The first level is to select a design solution for a given system that considers the life cycle cost and the reliability. The second level consists of determining a dynamic and optimal maintenance plan to be deployed for a design solution. This level is based on the Maintenance Free Operating Period (MFOP) concept, which takes into account the decision criteria such as, total reliability, maintenance cost and maintenance time. Depending on the life cycle duration, the desired availability, and the desired business model (sales or rental), this tool provides visibility of overall costs and optimal product architecture.

Determination of the Thermophysical Characteristics of the Composite Material Clay Cement Paper

In Morocco, the building sector is largely responsible for the evolution of energy consumption. The control of energy in this sector remains a major issue despite the rise of renewable energies. The design of an environmentally friendly building requires mastery and knowledge of energy and bioclimatic aspects. This implies taking into consideration of all the elements making up the building and the way in which energy exchanges take place between these elements. In this context, thermal insulation seems to be an ideal starting point for reducing energy consumption and greenhouse gas emissions. In this context, thermal insulation seems to be an ideal starting point for reducing energy consumption and greenhouse gas emissions. The aim of this work is to provide some solutions to reduce energy consumption while maintaining thermal comfort in the building. The objective of our work is to present an experimental study on the characterization of local materials used in the thermal insulation of buildings. These are paper recycling stabilized with cement and clay. The thermal conductivity of these materials, which were constituted based on sand, clay, cement; water, as well as treated paper, was determined by the guarded-hot-plate method. It involves the design of two materials that will subsequently be subjected to thermal and mechanical tests to determine their thermophysical properties. The results show that the thermal conductivity decreases as well in the case of the paper-cement mixture as that of the paper-clay and seems to stabilize around 40%. Measurements of mechanical properties such as flexural strength have shown that the enrichment of the studied material with paper makes it possible to reduce the flexural strength by 20% while optimizing the conductivity.

Extending the Flipped Classroom Approach: Using Technology in Module Delivery to Students of English Language and Literature at the British University in Egypt

Technology-enhanced teaching has been in the limelight since the 90s when educators started investigating and experimenting with using computers in the classroom as a means of building 21st. century skills and motivating students. The concept of technology-enhanced strategies in education is kaleidoscopic! It has meant different things to different educators. For the purpose of this paper, however, it will be used to refer to the diverse technology-based strategies used to support and enrich the flipped learning process, in the classroom and outside. The paper will investigate how technology is put in the service of teaching and learning to improve the students’ learning experience as manifested in students’ attendance and engagement, achievement rates and finally, students’ projects at the end of the semester. The results will be supported by a student survey about relevant specific aspects of their learning experience in the modules in the study.

Measuring Banks’ Antifragility via Fuzzy Logic

Analysing the world banking sector, we realize that traditional risk measurement methodologies no longer reflect the actual scenario with uncertainty and leave out events that can change the dynamics of markets. Considering this, regulators and financial institutions began to search more realistic models. The aim is to include external influences and interdependencies between agents, to describe and measure the operationalization of these complex systems and their risks in a more coherent and credible way. Within this context, X-Events are more frequent than assumed and, with uncertainties and constant changes, the concept of antifragility starts to gain great prominence in comparison to others methodologies of risk management. It is very useful to analyse whether a system succumbs (fragile), resists (robust) or gets benefits (antifragile) from disorder and stress. Thus, this work proposes the creation of the Banking Antifragility Index (BAI), which is based on the calculation of a triangular fuzzy number – to "quantify" qualitative criteria linked to antifragility.

Analysis and Evaluation of the Public Responses to Traffic Congestion Pricing Schemes in Urban Streets

Traffic congestion pricing in urban streets is one of the most suitable options for solving the traffic problems and environment pollutions in the cities of the country. Unlike its acceptable outcomes, there are problems concerning the necessity to pay by the mass. Regarding the fact that public response in order to succeed in this strategy is so influential, studying their response and behavior to get the feedback and improve the strategies is of great importance. In this study, a questionnaire was used to examine the public reactions to the traffic congestion pricing schemes at the center of Tehran metropolis and the factors involved in people’s decision making in accepting or rejecting the congestion pricing schemes were assessed based on the data obtained from the questionnaire as well as the international experiences. Then, by analyzing and comparing the schemes, guidelines to reduce public objections to them are discussed. The results of reviewing and evaluating the public reactions show that all the pros and cons must be considered to guarantee the success of these projects. Consequently, with targeted public education and consciousness-raising advertisements, prior to initiating a scheme and ensuring the mechanism of the implementation after the start of the project, the initial opposition is reduced and, with the gradual emergence of the real and tangible benefits of its implementation, users’ satisfaction will increase.

Choice Experiment Approach on Evaluation of Non-Market Farming System Outputs: First Results from Lithuanian Case Study

Market and non-market outputs are produced jointly in agriculture. Their supply depends on the intensity and type of production. The role of agriculture as an economic activity and its effects are important for the Lithuanian case study, as agricultural land covers more than a half of country. Positive and negative externalities, created in agriculture are not considered in the market. Therefore, specific techniques such as stated preferences methods, in particular choice experiments (CE) are used for evaluation of non-market outputs in agriculture. The main aim of this paper is to present construction of the research path for evaluation of non-market farming system outputs in Lithuania. The conventional and organic farming, covering crops (including both cereal and industrial crops) and livestock (including dairy and cattle) production has been selected. The CE method and nested logit (NL) model were selected as appropriate for evaluation of non-market outputs of different farming systems in Lithuania. A pilot survey was implemented between October–November 2018, in order to test and improve the CE questionnaire. The results of the survey showed that the questionnaire is accepted and well understood by the respondents. The econometric modelling showed that the selected NL model could be used for the main survey. The understanding of the differences between organic and conventional farming by residents was identified. It was revealed that they are more willing to choose organic farming in comparison to conventional farming.