Design and Fabrication of Micro-Bubble Oxygenator

This paper applies the MEMS technology to design and fabricate a micro-bubble generator by a piezoelectric actuator. Coupled with a nickel nozzle plate, an annular piezoelectric ceramic was utilized as the primary structure of the generator. In operations, the piezoelectric element deforms transversely under an electric field applied across the thickness of the generator. The surface of the nozzle plate can expand or contract because of the induction of radial strain, resulting in the whole structure to bend, and successively transport oxygen micro-bubbles into the blood flow for enhancing the oxygen content in blood. In the tests, a high magnification microscope and a high speed CCD camera were employed to photograph the time evolution of meniscus shape of gaseous bubbles dispensed from the micro-bubble generator for flow visualization. This investigation thus explored the bubble formation process including the influences of inlet gas pressure along with driving voltage and resonance frequency on the formed bubble extent.

User Survey on Food and Drinks in Japanese Public Libraries

Several decades ago, food and drinks were disallowed in most Japanese libraries. However, as discussions of “Library as a Place” have increased in recent years, the number of public and university libraries that have relaxed their policies to allow food and drinks have been increasing. This study focused on the opinions of library users on allowing food and drinks in public libraries and conducted a questionnaire survey among users of nine Japanese libraries. The results indicated that many users favored allowing food and drinks in libraries. Furthermore, it was found that users tend to frequently visit and stay longer in libraries where food and drinks are allowed.

Moving towards a General Definition of Public Happiness: A Grounded Theory Approach to the Recent Academic Research on Well-Being

Although there seems to be a growing interest in the study of the citizen’s happiness as an alternative measure of a country’s progress to GDP, happiness as a public concern is still an ambiguous concept, hard to define. Moreover, different notions are used indiscriminately to talk about the same thing. This investigation aims to determine the conceptions of happiness, well-being and quality of life that originate from the indexes that different governments and public institutions around the world have created to study them. Through the Scoping Review method, this study identifies the recent academic research in this field (a total of 267 documents between 2006 and 2016) from some of the most popular social sciences databases around the world, Web of Science, Scopus, JSTOR, Sage, EBSCO, IBSS and Google Scholar, and in Spain, ISOC and Dialnet. These 267 documents referenced 53 different indexes and researches. The Grounded Theory method has been applied to a sample of 13 indexes in order to identify the main categories they use to determine these three concepts. The results show that these are multi-dimensional concepts and similar indicators are used indistinctly to measure happiness, well-being and quality of life.

Number of Parametrization of Discrete-Time Systems without Unit-Delay Element: Single-Input Single-Output Case

In this paper, we consider the parametrization of the discrete-time systems without the unit-delay element within the framework of the factorization approach. In the parametrization, we investigate the number of required parameters. We consider single-input single-output systems in this paper. By the investigation, we find, on the discrete-time systems without the unit-delay element, three cases that are (1) there exist plants which require only one parameter and (2) two parameters, and (3) the number of parameters is at most three.

An Agent-Based Modelling Simulation Approach to Calculate Processing Delay of GEO Satellite Payload

The global coverage of broadband multimedia and internet-based services in terrestrial-satellite networks demand particular interests for satellite providers in order to enhance services with low latencies and high signal quality to diverse users. In particular, the delay of on-board processing is an inherent source of latency in a satellite communication that sometimes is discarded for the end-to-end delay of the satellite link. The frame work for this paper includes modelling of an on-orbit satellite payload using an agent model that can reproduce the properties of processing delays. In essence, a comparison of different spatial interpolation methods is carried out to evaluate physical data obtained by an GEO satellite in order to define a discretization function for determining that delay. Furthermore, the performance of the proposed agent and the development of a delay discretization function are together validated by simulating an hybrid satellite and terrestrial network. Simulation results show high accuracy according to the characteristics of initial data points of processing delay for Ku bands.

A Study on Strategy of Coordinative Symbiosis between New and Old Buildings: Case Study of Shanghai Citic Plaza and Surroundings

Along with the acceleration of Chinese urbanization, the expansion, renovation and demolition of old buildings is on the stage together with the design and construction of new buildings every day in downtown of the old city area. The coordinative symbiosis between new and old buildings is an important problem which needs to be solved in the process of urban development. By studying and analyzing the case of Shanghai Citic Plaza and surroundings, this paper contributes to discussing the concept, value and problems to be solved of the coordination of new and old buildings, meanwhile, striking the balance between new and old buildings from the aspects of architectural form, space, function and local context. As a result, the strategy of the coordinative symbiosis between new and old buildings is summarized, which can offer some guiding principles to urban development from now on.

The Sustainable Strategies Research for Renewal of “Villages in City”: A Case Study of Liuzhou in Southwestern China

Transformation under the reconfiguration of urban-rural relation in Liuzhou city has never been as radical and visible as it has been since the tremendous turn of the last century in China. Huanjiang village is located in Linhuashan Scenic Area in the middle east of Liuzhou city, with spectacular landscape and traditional features. Nowadays Huanjiang village has become a so-called "village in city", which is considered full of great potential for development because of the economic value of regional advantages during the urban sprawl. Communities of village found it difficult to acclimatize with the dramatic changes, which later led to numerous problems including ecological damage, unemployment of landless farmers and loss of traditional culture. Government has started up a series of renewal planings to resolve the problems, which are based on advanced technology and conform to sustainable and integrated strategies of city planning considering the original context and historical culture, superseding the traditional arrangements based on the guide of extensive economic growth. This paper aims to elaborate the context of Liuzhou city and Huanjiang village offered to both the traditional and sustainable planning approaches, in order to understand challenges and solutions of the rebuilding process. Through the analysis of the place relevant to architecture, society and culture, it will establish the corresponding systematic strategies. Considering the local features, it concludes with a comprehensive perspective on organic renewal in the case of Huanjiang village.

A Partially Accelerated Life Test Planning with Competing Risks and Linear Degradation Path under Tampered Failure Rate Model

In this paper, we propose a method to model the relationship between failure time and degradation for a simple step stress test where underlying degradation path is linear and different causes of failure are possible. It is assumed that the intensity function depends only on the degradation value. No assumptions are made about the distribution of the failure times. A simple step-stress test is used to shorten failure time of products and a tampered failure rate (TFR) model is proposed to describe the effect of the changing stress on the intensities. We assume that some of the products that fail during the test have a cause of failure that is only known to belong to a certain subset of all possible failures. This case is known as masking. In the presence of masking, the maximum likelihood estimates (MLEs) of the model parameters are obtained through an expectation-maximization (EM) algorithm by treating the causes of failure as missing values. The effect of incomplete information on the estimation of parameters is studied through a Monte-Carlo simulation. Finally, a real example is analyzed to illustrate the application of the proposed methods.

Hybrid Quasi-Steady Thermal Lattice Boltzmann Model for Studying the Behavior of Oil in Water Emulsions Used in Machining Tool Cooling and Lubrication

Oil in water (O/W) emulsions are utilized extensively for cooling and lubricating cutting tools during parts machining. A robust Lattice Boltzmann (LBM) thermal-surfactants model, which provides a useful platform for exploring complex emulsions’ characteristics under variety of flow conditions, is used here for the study of the fluid behavior during conventional tools cooling. The transient thermal capabilities of the model are employed for simulating the effects of the flow conditions of O/W emulsions on the cooling of cutting tools. The model results show that the temperature outcome is slightly affected by reversing the direction of upper plate (workpiece). On the other hand, an important increase in effective viscosity is seen which supports better lubrication during the work.

Developing Digital Competencies in Aboriginal Students through University-College Partnerships

This paper reports on a pilot project to develop a collaborative partnership between a community college in rural northern Ontario, Canada, and an urban university in the greater Toronto area in Oshawa, Canada. Partner institutions will collaborate to address learning needs of university applicants whose goals are to attain an undergraduate university BA in Educational Studies and Digital Technology degree, but who may not live in a geographical location that would facilitate this pathways process. The UOIT BA degree is attained through a 2+2 program, where students with a 2 year college diploma or equivalent can attain a four year undergraduate degree. The goals reported on the project are as: 1. Our aim is to expand the BA program to include an additional stream which includes serious educational games, simulations and virtual environments, 2. Develop fully (using both synchronous and asynchronous technologies) online learning modules for use by university applicants who otherwise are not geographically located close to a physical university site, 3. Assess the digital competencies of all students, including members of local, distance and Indigenous communities using a validated tool developed and tested by UOIT across numerous populations. This tool, the General Technical Competency Use and Scale (GTCU) will provide the collaborating institutions with data that will allow for analyzing how well students are prepared to succeed in fully online learning communities. Philosophically, the UOIT BA program is based on a fully online learning communities model (FOLC) that can be accessed from anywhere in the world through digital learning environments via audio video conferencing tools such as Adobe Connect. It also follows models of adult learning and mobile learning, and makes a university degree accessible to the increasing demographic of adult learners who may use mobile devices to learn anywhere anytime. The program is based on key principles of Problem Based Learning, allowing students to build their own understandings through the co-design of the learning environment in collaboration with the instructors and their peers. In this way, this degree allows students to personalize and individualize the learning based on their own culture, background and professional/personal experiences. Using modified flipped classroom strategies, students are able to interrogate video modules on their own time in preparation for one hour discussions occurring in video conferencing sessions. As a consequence of the program flexibility, students may continue to work full or part time. All of the partner institutions will co-develop four new modules, administer the GTCU and share data, while creating a new stream of the UOIT BA degree. This will increase accessibility for students to bridge from community colleges to university through a fully digital environment. We aim to work collaboratively with Indigenous elders, community members and distance education instructors to increase opportunities for more students to attain a university education.

Simulation of Utility Accrual Scheduling and Recovery Algorithm in Multiprocessor Environment

This paper presents the development of an event based Discrete Event Simulation (DES) for a recovery algorithm known Backward Recovery Global Preemptive Utility Accrual Scheduling (BR_GPUAS). This algorithm implements the Backward Recovery (BR) mechanism as a fault recovery solution under the existing Time/Utility Function/ Utility Accrual (TUF/UA) scheduling domain for multiprocessor environment. The BR mechanism attempts to take the faulty tasks back to its initial safe state and then proceeds to re-execute the affected section of the faulty tasks to enable recovery. Considering that faults may occur in the components of any system; a fault tolerance system that can nullify the erroneous effect is necessary to be developed. Current TUF/UA scheduling algorithm uses the abortion recovery mechanism and it simply aborts the erroneous task as their fault recovery solution. None of the existing algorithm in TUF/UA scheduling domain in multiprocessor scheduling environment have considered the transient fault and implement the BR mechanism as a fault recovery mechanism to nullify the erroneous effect and solve the recovery problem in this domain. The developed BR_GPUAS simulator has derived the set of parameter, events and performance metrics according to a detailed analysis of the base model. Simulation results revealed that BR_GPUAS algorithm can saved almost 20-30% of the accumulated utilities making it reliable and efficient for the real-time application in the multiprocessor scheduling environment.

Experimental Study on Thermomechanical Properties of New-Generation ODS Alloys

By using a combination of new technologies together with an unconventional use of different types of materials, specific mechanical properties and structures of the material can be achieved. Some possibilities are enabled by a combination of powder metallurgy in the preparation of a metal matrix with dispersed stable particles achieved by mechanical alloying and hot consolidation. This paper explains the thermomechanical properties of new generation of Oxide Dispersion Strengthened alloys (ODS) within three ranges of temperature with specified deformation profiles. The results show that the mechanical properties of new ODS alloys are significantly affected by the thermomechanical treatment.

Thermodynamic Analysis of an Ejector-Absorption Refrigeration Cycle with Using NH3-H2O

In this paper, the ejector-absorption refrigeration cycle is presented. This article deals with the thermodynamic simulation and the first and second law analysis of an ammonia-water. The effects of parameters such as condenser, absorber, generator, and evaporator temperatures have been investigated. The influence of the various operating parameters on the performance coefficient and exergy efficiency of this cycle has been studied. The results show that when the temperature of different parts increases, the performance coefficient and the exergy efficiency of the cycle decrease, except for evaporator and generator, that causes an increase in coefficient of performance (COP). According to the results, absorber and ejector have the highest exergy losses in the studied conditions.

Applying Different Working Fluids in a Combined Power and Ejector Refrigeration Cycle with Low Temperature Heat Sources

A power and cooling cycle, which combines the organic Rankine cycle and the ejector refrigeration cycle supplied by waste heat energy sources, is discussed in this paper. 13 working fluids including wet, dry, and isentropic fluids are studied in order to find their performances on the combined cycle. Various operating conditions’ effects on the proposed cycle are examined by fixing power/refrigeration ratio. According to the results, dry and isentropic fluids have better performance compared with wet fluids.

Numerical Modeling of Various Support Systems to Stabilize Deep Excavations

Urban development requires deep excavations near buildings and other structures. Deep excavation has become more a necessity for better utilization of space as the population of the world has dramatically increased. In Lebanon, some urban areas are very crowded and lack spaces for new buildings and underground projects, which makes the usage of underground space indispensable. In this paper, a numerical modeling is performed using the finite element method to study the deep excavation-diaphragm wall soil-structure interaction in the case of nonlinear soil behavior. The study is focused on a comparison of the results obtained using different support systems. Furthermore, a parametric study is performed according to the remoteness of the structure.

Generalization of Clustering Coefficient on Lattice Networks Applied to Criminal Networks

A lattice network is a special type of network in which all nodes have the same number of links, and its boundary conditions are periodic. The most basic lattice network is the ring, a one-dimensional network with periodic border conditions. In contrast, the Cartesian product of d rings forms a d-dimensional lattice network. An analytical expression currently exists for the clustering coefficient in this type of network, but the theoretical value is valid only up to certain connectivity value; in other words, the analytical expression is incomplete. Here we obtain analytically the clustering coefficient expression in d-dimensional lattice networks for any link density. Our analytical results show that the clustering coefficient for a lattice network with density of links that tend to 1, leads to the value of the clustering coefficient of a fully connected network. We developed a model on criminology in which the generalized clustering coefficient expression is applied. The model states that delinquents learn the know-how of crime business by sharing knowledge, directly or indirectly, with their friends of the gang. This generalization shed light on the network properties, which is important to develop new models in different fields where network structure plays an important role in the system dynamic, such as criminology, evolutionary game theory, econophysics, among others.

Sparse Unmixing of Hyperspectral Data by Exploiting Joint-Sparsity and Rank-Deficiency

In this work, we exploit two assumed properties of the abundances of the observed signatures (endmembers) in order to reconstruct the abundances from hyperspectral data. Joint-sparsity is the first property of the abundances, which assumes the adjacent pixels can be expressed as different linear combinations of same materials. The second property is rank-deficiency where the number of endmembers participating in hyperspectral data is very small compared with the dimensionality of spectral library, which means that the abundances matrix of the endmembers is a low-rank matrix. These assumptions lead to an optimization problem for the sparse unmixing model that requires minimizing a combined l2,p-norm and nuclear norm. We propose a variable splitting and augmented Lagrangian algorithm to solve the optimization problem. Experimental evaluation carried out on synthetic and real hyperspectral data shows that the proposed method outperforms the state-of-the-art algorithms with a better spectral unmixing accuracy.

FCNN-MR: A Parallel Instance Selection Method Based on Fast Condensed Nearest Neighbor Rule

Instance selection (IS) technique is used to reduce the data size to improve the performance of data mining methods. Recently, to process very large data set, several proposed methods divide the training set into some disjoint subsets and apply IS algorithms independently to each subset. In this paper, we analyze the limitation of these methods and give our viewpoint about how to divide and conquer in IS procedure. Then, based on fast condensed nearest neighbor (FCNN) rule, we propose a large data sets instance selection method with MapReduce framework. Besides ensuring the prediction accuracy and reduction rate, it has two desirable properties: First, it reduces the work load in the aggregation node; Second and most important, it produces the same result with the sequential version, which other parallel methods cannot achieve. We evaluate the performance of FCNN-MR on one small data set and two large data sets. The experimental results show that it is effective and practical.

An Improved Method on Static Binary Analysis to Enhance the Context-Sensitive CFI

Control Flow Integrity (CFI) is one of the most promising technique to defend Code-Reuse Attacks (CRAs). Traditional CFI Systems and recent Context-Sensitive CFI use coarse control flow graphs (CFGs) to analyze whether the control flow hijack occurs, left vast space for attackers at indirect call-sites. Coarse CFGs make it difficult to decide which target to execute at indirect control-flow transfers, and weaken the existing CFI systems actually. It is an unsolved problem to extract CFGs precisely and perfectly from binaries now. In this paper, we present an algorithm to get a more precise CFG from binaries. Parameters are analyzed at indirect call-sites and functions firstly. By comparing counts of parameters prepared before call-sites and consumed by functions, targets of indirect calls are reduced. Then the control flow would be more constrained at indirect call-sites in runtime. Combined with CCFI, we implement our policy. Experimental results on some popular programs show that our approach is efficient. Further analysis show that it can mitigate COOP and other advanced attacks.

Latent Factors of Severity in Truck-Involved and Non-Truck-Involved Crashes on Freeways

Truck-involved crashes have higher crash severity than non-truck-involved crashes. There have been many studies about the frequency of crashes and the development of severity models, but those studies only analyzed the relationship between observed variables. To identify why more people are injured or killed when trucks are involved in the crash, we must examine to quantify the complex causal relationship between severity of the crash and risk factors by adopting the latent factors of crashes. The aim of this study was to develop a structural equation or model based on truck-involved and non-truck-involved crashes, including five latent variables, i.e. a crash factor, environmental factor, road factor, driver’s factor, and severity factor. To clarify the unique characteristics of truck-involved crashes compared to non-truck-involved crashes, a confirmatory analysis method was used. To develop the model, we extracted crash data from 10,083 crashes on Korean freeways from 2008 through 2014. The results showed that the most significant variable affecting the severity of a crash is the crash factor, which can be expressed by the location, cause, and type of the crash. For non-truck-involved crashes, the crash and environment factors increase severity of the crash; conversely, the road and driver factors tend to reduce severity of the crash. For truck-involved crashes, the driver factor has a significant effect on severity of the crash although its effect is slightly less than the crash factor. The multiple group analysis employed to analyze the differences between the heterogeneous groups of drivers.