Assessment of Occupational Exposure and Individual Radio-Sensitivity in People Subjected to Ionizing Radiation

The estimation of accumulated radiation doses in people professionally exposed to ionizing radiation was performed using methods of biological (chromosomal aberrations frequency in lymphocytes) and physical (radionuclides analysis in urine, whole-body radiation meter, individual thermoluminescent dosimeters) dosimetry. A group of 84 "A" category employees after their work in the territory of former Semipalatinsk test site (Kazakhstan) was investigated. The dose rate in some funnels exceeds 40 μSv/h. After radionuclides determination in urine using radiochemical and WBC methods, it was shown that the total effective dose of personnel internal exposure did not exceed 0.2 mSv/year, while an acceptable dose limit for staff is 20 mSv/year. The range of external radiation doses measured with individual thermo-luminescent dosimeters was 0.3-1.406 µSv. The cytogenetic examination showed that chromosomal aberrations frequency in staff was 4.27±0.22%, which is significantly higher than at the people from non-polluting settlement Tausugur (0.87±0.1%) (р ≤ 0.01) and citizens of Almaty (1.6±0.12%) (р≤ 0.01). Chromosomal type aberrations accounted for 2.32±0.16%, 0.27±0.06% of which were dicentrics and centric rings. The cytogenetic analysis of different types group radiosensitivity among «professionals» (age, sex, ethnic group, epidemiological data) revealed no significant differences between the compared values. Using various techniques by frequency of dicentrics and centric rings, the average cumulative radiation dose for group was calculated, and that was 0.084-0.143 Gy. To perform comparative individual dosimetry using physical and biological methods of dose assessment, calibration curves (including own ones) and regression equations based on general frequency of chromosomal aberrations obtained after irradiation of blood samples by gamma-radiation with the dose rate of 0,1 Gy/min were used. Herewith, on the assumption of individual variation of chromosomal aberrations frequency (1–10%), the accumulated dose of radiation varied 0-0.3 Gy. The main problem in the interpretation of individual dosimetry results is reduced to different reaction of the objects to irradiation - radiosensitivity, which dictates the need of quantitative definition of this individual reaction and its consideration in the calculation of the received radiation dose. The entire examined contingent was assigned to a group based on the received dose and detected cytogenetic aberrations. Radiosensitive individuals, at the lowest received dose in a year, showed the highest frequency of chromosomal aberrations (5.72%). In opposite, radioresistant individuals showed the lowest frequency of chromosomal aberrations (2.8%). The cohort correlation according to the criterion of radio-sensitivity in our research was distributed as follows: radio-sensitive (26.2%) — medium radio-sensitivity (57.1%), radioresistant (16.7%). Herewith, the dispersion for radioresistant individuals is 2.3; for the group with medium radio-sensitivity — 3.3; and for radio-sensitive group — 9. These data indicate the highest variation of characteristic (reactions to radiation effect) in the group of radio-sensitive individuals. People with medium radio-sensitivity show significant long-term correlation (0.66; n=48, β ≥ 0.999) between the values of doses defined according to the results of cytogenetic analysis and dose of external radiation obtained with the help of thermoluminescent dosimeters. Mathematical models based on the type of violation of the radiation dose according to the professionals radiosensitivity level were offered.

Different Tillage Possibilities for Second Crop in Green Bean Farming

In this study, determining of reduced tillage techniques in green bean farming as a second crop after harvesting wheat was targeted. To this aim, four different soil tillage methods namely, heavy-duty disc harrow (HD), rotary tiller (ROT), heavy-duty disc harrow plus rotary tiller (HD+ROT) and no-tillage (NT) (seeding by direct drill) were examined. Experiments were arranged in a randomized block design with three replications. The highest green beans yields were obtained in HD+ROT and NT as 5,862.1 and 5,829.3 Mg/ha, respectively. The lowest green bean yield was found in HD as 3,076.7 Mg/ha. The highest fuel consumption was measured 30.60 L ha-1 for HD+ROT whereas the lowest value was found 7.50 L ha-1 for NT. No tillage method gave the best results for fuel consumption and effective power requirement. It is concluded that no-tillage method can be used in second crop green bean in the Thrace Region due to economic and erosion conditions.

Quantifying Key Factors Affecting Leagile Manufacturing System

In today’s market, striving hard has become necessary for the industries to survive due to the intense competition and globalization. In earlier days, there were few sellers and limited numbers of buyers, so customers were having fewer options to buy the product. But today, the market is highly competitive and volatile. Industries are focusing on robotics, advance manufacturing methods like AJM (Abrasive Jet Machining), EDM (Electric Discharge Machining), ECM (Electrochemical Machining) etc., CAD/CAM, CAE to make quality products and market them in shortest possible time. Leagile manufacturing system is ensuring best available solution at minimum cost to meet the market demand. This paper tries to assimilate the concept of Leagile manufacturing system in today’s scenario and evaluating key factors affecting Leagile manufacturing using digraph technique.

A Methodology of Testing Beam to Column Connection under Lateral Impact Load

Beam to column connection can be considered as the most important structural part that affects the response of buildings to progressive collapse. However, many studies were conducted to investigate the beam to column connection under accidental loads such as fire, blast and impact load to investigate the connection response. The study is a part of a PhD plan to investigate different types of connections under lateral impact load. The conventional test setups, such as cruciform setup, were designed to apply shear forces and bending moment on the connection, whilst, in the lateral impact case, the connection is subjected to combined tension and moment. Hence, a review is presented to introduce the previous test setup that is used to investigate the connection behaviour. Then, the design and fabrication of the novel test setup is presented. Finally, some trial test results to investigate the efficiency of the proposed setup are discussed. The final results indicate that the setup was efficient in terms of the simplicity and strength.

Quasi-Static Analysis of End Plate Beam-to-Column Connections

This paper presents a method for modelling and analysing end plate beam-to-column connections to obtain the quasi-static behaviour using non-linear dynamic explicit integration. In addition to its importance to study the static behaviour of a structural member, quasi-static behaviour is largely needed to be compared with the dynamic behaviour of such members in order to investigate the dynamic effect by proposing dynamic increase factors (DIFs). The beam-to-column bolted connections contain various contact surfaces at which the implicit procedure may have difficulties converging, resulting in a large number of iterations. Contrary, explicit procedure could deal effectively with complex contacts without converging problems. Hence, finite element modelling using ABAQUS/explicit is used in this study to address the dynamic effect may be produced using explicit procedure. Also, the effect of loading rate and mass scaling are discussed to investigate their effect on the time of analysis. The results show that the explicit procedure is valuable to model the end plate beam-to-column connections in terms of failure mode, load-displacement relationships. Also, it is concluded that loading rate and mass scaling should be carefully selected to avoid the dynamic effect in the solution.

Moving towards a General Definition of Public Happiness: A Grounded Theory Approach to the Recent Academic Research on Well-Being

Although there seems to be a growing interest in the study of the citizen’s happiness as an alternative measure of a country’s progress to GDP, happiness as a public concern is still an ambiguous concept, hard to define. Moreover, different notions are used indiscriminately to talk about the same thing. This investigation aims to determine the conceptions of happiness, well-being and quality of life that originate from the indexes that different governments and public institutions around the world have created to study them. Through the Scoping Review method, this study identifies the recent academic research in this field (a total of 267 documents between 2006 and 2016) from some of the most popular social sciences databases around the world, Web of Science, Scopus, JSTOR, Sage, EBSCO, IBSS and Google Scholar, and in Spain, ISOC and Dialnet. These 267 documents referenced 53 different indexes and researches. The Grounded Theory method has been applied to a sample of 13 indexes in order to identify the main categories they use to determine these three concepts. The results show that these are multi-dimensional concepts and similar indicators are used indistinctly to measure happiness, well-being and quality of life.

An Agent-Based Modelling Simulation Approach to Calculate Processing Delay of GEO Satellite Payload

The global coverage of broadband multimedia and internet-based services in terrestrial-satellite networks demand particular interests for satellite providers in order to enhance services with low latencies and high signal quality to diverse users. In particular, the delay of on-board processing is an inherent source of latency in a satellite communication that sometimes is discarded for the end-to-end delay of the satellite link. The frame work for this paper includes modelling of an on-orbit satellite payload using an agent model that can reproduce the properties of processing delays. In essence, a comparison of different spatial interpolation methods is carried out to evaluate physical data obtained by an GEO satellite in order to define a discretization function for determining that delay. Furthermore, the performance of the proposed agent and the development of a delay discretization function are together validated by simulating an hybrid satellite and terrestrial network. Simulation results show high accuracy according to the characteristics of initial data points of processing delay for Ku bands.

A Partially Accelerated Life Test Planning with Competing Risks and Linear Degradation Path under Tampered Failure Rate Model

In this paper, we propose a method to model the relationship between failure time and degradation for a simple step stress test where underlying degradation path is linear and different causes of failure are possible. It is assumed that the intensity function depends only on the degradation value. No assumptions are made about the distribution of the failure times. A simple step-stress test is used to shorten failure time of products and a tampered failure rate (TFR) model is proposed to describe the effect of the changing stress on the intensities. We assume that some of the products that fail during the test have a cause of failure that is only known to belong to a certain subset of all possible failures. This case is known as masking. In the presence of masking, the maximum likelihood estimates (MLEs) of the model parameters are obtained through an expectation-maximization (EM) algorithm by treating the causes of failure as missing values. The effect of incomplete information on the estimation of parameters is studied through a Monte-Carlo simulation. Finally, a real example is analyzed to illustrate the application of the proposed methods.

Hybrid Quasi-Steady Thermal Lattice Boltzmann Model for Studying the Behavior of Oil in Water Emulsions Used in Machining Tool Cooling and Lubrication

Oil in water (O/W) emulsions are utilized extensively for cooling and lubricating cutting tools during parts machining. A robust Lattice Boltzmann (LBM) thermal-surfactants model, which provides a useful platform for exploring complex emulsions’ characteristics under variety of flow conditions, is used here for the study of the fluid behavior during conventional tools cooling. The transient thermal capabilities of the model are employed for simulating the effects of the flow conditions of O/W emulsions on the cooling of cutting tools. The model results show that the temperature outcome is slightly affected by reversing the direction of upper plate (workpiece). On the other hand, an important increase in effective viscosity is seen which supports better lubrication during the work.

Numerical Modeling of Various Support Systems to Stabilize Deep Excavations

Urban development requires deep excavations near buildings and other structures. Deep excavation has become more a necessity for better utilization of space as the population of the world has dramatically increased. In Lebanon, some urban areas are very crowded and lack spaces for new buildings and underground projects, which makes the usage of underground space indispensable. In this paper, a numerical modeling is performed using the finite element method to study the deep excavation-diaphragm wall soil-structure interaction in the case of nonlinear soil behavior. The study is focused on a comparison of the results obtained using different support systems. Furthermore, a parametric study is performed according to the remoteness of the structure.

Sparse Unmixing of Hyperspectral Data by Exploiting Joint-Sparsity and Rank-Deficiency

In this work, we exploit two assumed properties of the abundances of the observed signatures (endmembers) in order to reconstruct the abundances from hyperspectral data. Joint-sparsity is the first property of the abundances, which assumes the adjacent pixels can be expressed as different linear combinations of same materials. The second property is rank-deficiency where the number of endmembers participating in hyperspectral data is very small compared with the dimensionality of spectral library, which means that the abundances matrix of the endmembers is a low-rank matrix. These assumptions lead to an optimization problem for the sparse unmixing model that requires minimizing a combined l2,p-norm and nuclear norm. We propose a variable splitting and augmented Lagrangian algorithm to solve the optimization problem. Experimental evaluation carried out on synthetic and real hyperspectral data shows that the proposed method outperforms the state-of-the-art algorithms with a better spectral unmixing accuracy.

FCNN-MR: A Parallel Instance Selection Method Based on Fast Condensed Nearest Neighbor Rule

Instance selection (IS) technique is used to reduce the data size to improve the performance of data mining methods. Recently, to process very large data set, several proposed methods divide the training set into some disjoint subsets and apply IS algorithms independently to each subset. In this paper, we analyze the limitation of these methods and give our viewpoint about how to divide and conquer in IS procedure. Then, based on fast condensed nearest neighbor (FCNN) rule, we propose a large data sets instance selection method with MapReduce framework. Besides ensuring the prediction accuracy and reduction rate, it has two desirable properties: First, it reduces the work load in the aggregation node; Second and most important, it produces the same result with the sequential version, which other parallel methods cannot achieve. We evaluate the performance of FCNN-MR on one small data set and two large data sets. The experimental results show that it is effective and practical.

An Improved Method on Static Binary Analysis to Enhance the Context-Sensitive CFI

Control Flow Integrity (CFI) is one of the most promising technique to defend Code-Reuse Attacks (CRAs). Traditional CFI Systems and recent Context-Sensitive CFI use coarse control flow graphs (CFGs) to analyze whether the control flow hijack occurs, left vast space for attackers at indirect call-sites. Coarse CFGs make it difficult to decide which target to execute at indirect control-flow transfers, and weaken the existing CFI systems actually. It is an unsolved problem to extract CFGs precisely and perfectly from binaries now. In this paper, we present an algorithm to get a more precise CFG from binaries. Parameters are analyzed at indirect call-sites and functions firstly. By comparing counts of parameters prepared before call-sites and consumed by functions, targets of indirect calls are reduced. Then the control flow would be more constrained at indirect call-sites in runtime. Combined with CCFI, we implement our policy. Experimental results on some popular programs show that our approach is efficient. Further analysis show that it can mitigate COOP and other advanced attacks.

Latent Factors of Severity in Truck-Involved and Non-Truck-Involved Crashes on Freeways

Truck-involved crashes have higher crash severity than non-truck-involved crashes. There have been many studies about the frequency of crashes and the development of severity models, but those studies only analyzed the relationship between observed variables. To identify why more people are injured or killed when trucks are involved in the crash, we must examine to quantify the complex causal relationship between severity of the crash and risk factors by adopting the latent factors of crashes. The aim of this study was to develop a structural equation or model based on truck-involved and non-truck-involved crashes, including five latent variables, i.e. a crash factor, environmental factor, road factor, driver’s factor, and severity factor. To clarify the unique characteristics of truck-involved crashes compared to non-truck-involved crashes, a confirmatory analysis method was used. To develop the model, we extracted crash data from 10,083 crashes on Korean freeways from 2008 through 2014. The results showed that the most significant variable affecting the severity of a crash is the crash factor, which can be expressed by the location, cause, and type of the crash. For non-truck-involved crashes, the crash and environment factors increase severity of the crash; conversely, the road and driver factors tend to reduce severity of the crash. For truck-involved crashes, the driver factor has a significant effect on severity of the crash although its effect is slightly less than the crash factor. The multiple group analysis employed to analyze the differences between the heterogeneous groups of drivers.

Investigation of Utilizing L-Band Horn Antenna in Landmine Detection

Landmine detection is an important and yet challenging problem remains to be solved. Ground Penetrating Radar (GPR) is a powerful and rapidly maturing technology for subsurface threat identification. The detection methodology of GPR depends mainly on the contrast of the dielectric properties of the searched target and its surrounding soil. This contrast produces a partial reflection of the electromagnetic pulses that are being transmitted into the soil and then being collected by the GPR.  One of the most critical hardware components for the performance of GPR is the antenna system. The current paper explores the design and simulation of a pyramidal horn antenna operating at L-band frequencies (1- 2 GHz) to detect a landmine. A prototype model of the GPR system setup is developed to simulate full wave analysis of the electromagnetic fields in different soil types. The contrast in the dielectric permittivity of the landmine and the sandy soil is the most important parameter to be considered for detecting the presence of landmine. L-band horn antenna is proved to be well-versed in the investigation of landmine detection.

Schrödinger Equation with Position-Dependent Mass: Staggered Mass Distributions

The Point canonical transformation method is applied for solving the Schrödinger equation with position-dependent mass. This class of problem has been solved for continuous mass distributions. In this work, a staggered mass distribution for the case of a free particle in an infinite square well potential has been proposed. The continuity conditions as well as normalization for the wave function are also considered. The proposal can be used for dealing with other kind of staggered mass distributions in the Schrödinger equation with different quantum potentials.

Analyzing the Perception of Social Networking Sites as a Learning Tool among University Students: Case Study of a Business School in India

Universities and higher education institutes are finding it increasingly difficult to engage students fruitfully through traditional pedagogic tools. Web 2.0 technologies comprising social networking sites (SNSs) offer a platform for students to collaborate and share information, thereby enhancing their learning experience. Despite the potential and reach of SNSs, its use has been limited in academic settings promoting higher education. The purpose of this paper is to assess the perception of social networking sites among business school students in India and analyze its role in enhancing quality of student experiences in a business school leading to the proposal of an agenda for future research. In this study, more than 300 students of a reputed business school were involved in a survey of their preferences of different social networking sites and their perceptions and attitudes towards these sites. A questionnaire with three major sections was designed, validated and distributed among  a sample of students, the research method being descriptive in nature. Crucial questions were addressed to the students concerning time commitment, reasons for usage, nature of interaction on these sites, and the propensity to share information leading to direct and indirect modes of learning. It was further supplemented with focus group discussion to analyze the findings. The paper notes the resistance in the adoption of new technology by a section of business school faculty, who are staunch supporters of the classical “face-to-face” instruction. In conclusion, social networking sites like Facebook and LinkedIn provide new avenues for students to express themselves and to interact with one another. Universities could take advantage of the new ways  in which students are communicating with one another. Although interactive educational options such as Moodle exist, social networking sites are rarely used for academic purposes. Using this medium opens new ways of academically-oriented interactions where faculty could discover more about students' interests, and students, in turn, might express and develop more intellectual facets of their lives. hitherto unknown intellectual facets.  This study also throws up the enormous potential of mobile phones as a tool for “blended learning” in business schools going forward.

Studying the Structural Behaviour of RC Beams with Circular Openings of Different Sizes and Locations Using FE Method

This paper aims to investigate the structural behaviour of RC beams with circular openings of different sizes and locations modelled using ABAQUS FEM software. Seven RC beams with the dimensions of 1200 mm×150 mm×150 mm were tested under three-point loading. Group A consists of three RC beams incorporating circular openings with diameters of 40 mm, 55 mm and 65 mm in the shear zone. However, Group B consists of three RC beams incorporating circular openings with diameters of 40 mm, 55 mm and 65 mm in the flexural zone. The final RC beam did not have any openings, to provide a control beam for comparison. The results show that increasing the diameter of the openings increases the maximum deflection and the ultimate failure load decreases relative to the control beam. In the shear zone, the presence of the openings caused an increase in the maximum deflection ranging between 4% and 22% and a decrease in the ultimate failure load of between 26% and 36% compared to the control beam. However, the presence of the openings in the flexural zone caused an increase in the maximum deflection of between 1.5% and 19.7% and a decrease in the ultimate failure load of between 6% and 13% relative to the control beam. In this study, the optimum location for placing circular openings was found to be in the flexural zone of the beam with a diameter of less than 30% of the depth of the beam.

Performance Analysis of Deterministic Stable Election Protocol Using Fuzzy Logic in Wireless Sensor Network

In Wireless Sensor Network (WSN), the sensor containing motes (nodes) incorporate batteries that can lament at some extent. To upgrade the energy utilization, clustering is one of the prototypical approaches for split sensor motes into a number of clusters where one mote (also called as node) proceeds as a Cluster Head (CH). CH selection is one of the optimization techniques for enlarging stability and network lifespan. Deterministic Stable Election Protocol (DSEP) is an effectual clustering protocol that makes use of three kinds of nodes with dissimilar residual energy for CH election. Fuzzy Logic technology is used to expand energy level of DSEP protocol by using fuzzy inference system. This paper presents protocol DSEP using Fuzzy Logic (DSEP-FL) CH by taking into account four linguistic variables such as energy, concentration, centrality and distance to base station. Simulation results show that our proposed method gives more effective results in term of a lifespan of network and stability as compared to the performance of other clustering protocols.

Assessment of Multi-Domain Energy Systems Modelling Methods

Emissions are a consequence of electricity generation. A major option for low carbon generation, local energy systems featuring Combined Heat and Power with solar PV (CHPV) has significant potential to increase energy performance, increase resilience, and offer greater control of local energy prices while complementing the UK’s emissions standards and targets. Recent advances in dynamic modelling and simulation of buildings and clusters of buildings using the IDEAS framework have successfully validated a novel multi-vector (simultaneous control of both heat and electricity) approach to integrating the wide range of primary and secondary plant typical of local energy systems designs including CHP, solar PV, gas boilers, absorption chillers and thermal energy storage, and associated electrical and hot water networks, all operating under a single unified control strategy. Results from this work indicate through simulation that integrated control of thermal storage can have a pivotal role in optimizing system performance well beyond the present expectations. Environmental impact analysis and reporting of all energy systems including CHPV LES presently employ a static annual average carbon emissions intensity for grid supplied electricity. This paper focuses on establishing and validating CHPV environmental performance against conventional emissions values and assessment benchmarks to analyze emissions performance without and with an active thermal store in a notional group of non-domestic buildings. Results of this analysis are presented and discussed in context of performance validation and quantifying the reduced environmental impact of CHPV systems with active energy storage in comparison with conventional LES designs.