CO2 Emission and Cost Optimization of Reinforced Concrete Frame Designed by Performance Based Design Approach

As greenhouse effect has been recognized as serious environmental problem of the world, interests in carbon dioxide (CO2) emission which comprises major part of greenhouse gas (GHG) emissions have been increased recently. Since construction industry takes a relatively large portion of total CO2 emissions of the world, extensive studies about reducing CO2 emissions in construction and operation of building have been carried out after the 2000s. Also, performance based design (PBD) methodology based on nonlinear analysis has been robustly developed after Northridge Earthquake in 1994 to assure and assess seismic performance of building more exactly because structural engineers recognized that prescriptive code based design approach cannot address inelastic earthquake responses directly and assure performance of building exactly. Although CO2 emissions and PBD approach are recent rising issues on construction industry and structural engineering, there were few or no researches considering these two issues simultaneously. Thus, the objective of this study is to minimize the CO2 emissions and cost of building designed by PBD approach in structural design stage considering structural materials. 4 story and 4 span reinforced concrete building optimally designed to minimize CO2 emissions and cost of building and to satisfy specific seismic performance (collapse prevention in maximum considered earthquake) of building satisfying prescriptive code regulations using non-dominated sorting genetic algorithm-II (NSGA-II). Optimized design result showed that minimized CO2 emissions and cost of building were acquired satisfying specific seismic performance. Therefore, the methodology proposed in this paper can be used to reduce both CO2 emissions and cost of building designed by PBD approach.

Exploring Communities of Practice through Public Health Walks for Nurse Education

Introduction: Student nurses must develop skills in observation, communication and reflection as well as public health knowledge from their first year of training. This paper will explain a method developed for students to collect their own findings about public health in urban areas. These areas are both rich in the history of old public health that informs the content of many traditional public health walks, but are also locations where new public health concerns about chronic disease are concentrated. The learning method explained in this paper enables students to collect their own data and write original work as first year students. Examples of their findings will be given. Methodology: In small groups, health care students are instructed to walk in neighbourhoods near to the hospitals they will soon attend as apprentice nurses. On their walks, they wander slowly, engage in conversations, and enter places open to the public. As they drift, they observe with all five senses in the real three dimensional world to collect data for their reflective accounts of old and new public health. They are encouraged to stop for refreshments and taste, as well as look, hear, smell, and touch while on their walk. They reflect as a group and later develop an individual reflective account in which they write up their deep reflections about what they observed on their walk. In preparation for their walk, they are encouraged to look at studies of quality of Life and other neighbourhood statistics as well as undertaking a risk assessment for their walk. Findings: Reflecting on their walks, students apply theoretical concepts around social determinants of health and health inequalities to develop their understanding of communities in the neighbourhoods visited. They write about the treasured historical architecture made of stone, bronze and marble which have outlived those who built them; but also how the streets are used now. The students develop their observations into thematic analyses such as: what we drink as illustrated by the empty coke can tossed into a now disused drinking fountain; the shift in home-life balance illustrated by streets where families once lived over the shop which are now walked by commuters weaving around each other as they talk on their mobile phones; and security on the street, with CCTV cameras placed at regular intervals, signs warning trespasses and barbed wire; but little evidence of local people watching the street. Conclusion: In evaluations of their first year, students have reported the health walk as one of their best experiences. The innovative approach was commended by the UK governing body of nurse education and it received a quality award from the nurse education funding body. This approach to education allows students to develop skills in the real world and write original work.

Passive Non-Prehensile Manipulation on Helix Path Based on Mechanical Intelligence

Object manipulation techniques in robotics can be categorized in two major groups including manipulation with grasp and manipulation without grasp. The original aim of this paper is to develop an object manipulation method where in addition to being grasp-less, the manipulation task is done in a passive approach. In this method, linear and angular positions of the object are changed and its manipulation path is controlled. The manipulation path is a helix track with constant radius and incline. The method presented in this paper proposes a system which has not the actuator and the active controller. So this system requires a passive mechanical intelligence to convey the object from the status of the source along the specified path to the goal state. This intelligent is created based on utilizing the geometry of the system components. A general set up for the components of the system is considered to satisfy the required conditions. Then after kinematical analysis, detailed dimensions and geometry of the mechanism is obtained. The kinematical results are verified by simulation in ADAMS.

Design for Safety: Safety Consideration in Planning and Design of Airport Airsides

During airport planning and design stages, the major issues of capacity and safety in construction and operation of an airport need to be taken into consideration. The airside of an airport is a major and critical infrastructure that usually consists of runway(s), taxiway system, and apron(s) etc., which have to be designed according to the international standards and recommendations, and local limitations to accommodate the forecasted demands. However, in many cases, airport airsides are suffering from unexpected risks that occurred during airport operations. Therefore, safety risk assessment should be applied in the planning and design of airsides to cope with the probability of risks and their consequences, and to make decisions to reduce the risks to as low as reasonably practicable (ALARP) based on safety risk assessment. This paper presents a combination approach of Failure Modes, Effect, and Criticality Analysis (FMECA), Fuzzy Reasoning Approach (FRA), and Fuzzy Analytic Hierarchy Process (FAHP) to develop a risk analysis model for safety risk assessment. An illustrated example is used to the demonstrate risk assessment process on how the design of an airside in an airport can be analysed by using the proposed safety design risk assessment model.

Reducing Defects through Organizational Learning within a Housing Association Environment

Housing Associations (HAs) contribute circa 20% of the UK’s housing supply. HAs are however under increasing pressure as a result of funding cuts and rent reductions. Due to the increased pressure, a number of processes are currently being reviewed by HAs, especially how they manage and learn from defects. Learning from defects is considered a useful approach to achieving defect reduction within the UK housebuilding industry. This paper contributes to our understanding of how HAs learn from defects by undertaking an initial round table discussion with key HA stakeholders as part of an ongoing collaborative research project with the National House Building Council (NHBC) to better understand how house builders and HAs learn from defects to reduce their prevalence. The initial discussion shows that defect information runs through a number of groups, both internal and external of a HA during both the defects management process and organizational learning (OL) process. Furthermore, HAs are reliant on capturing and recording defect data as the foundation for the OL process. During the OL process defect data analysis is the primary enabler to recognizing a need for a change to organizational routines. When a need for change has been recognized, new options are typically pursued to design out defects via updates to a HAs Employer’s Requirements. Proposed solutions are selected by a review board and committed to organizational routine. After implementing a change, both structured and unstructured feedback is sought to establish the change’s success. The findings from the HA discussion demonstrates that OL can achieve defect reduction within the house building sector in the UK. The paper concludes by outlining a potential ‘learning from defects model’ for the housebuilding industry as well as describing future work.

Interactive Garments: Flexible Technologies for Textile Integration

Upon reviewing the literature and the pragmatic work done in the field of E- textiles, it is observed that the applications of wearable technologies have found a steady growth in the field of military, medical, industrial, sports; whereas fashion is at a loss to know how to treat this technology and bring it to market. The purpose of this paper is to understand the practical issues of integration of electronics in garments; cutting patterns for mass production, maintaining the basic properties of textiles and daily maintenance of garments that hinder the wide adoption of interactive fabric technology within Fashion and leisure wear. To understand the practical hindrances an experimental and laboratory approach is taken. “Techno Meets Fashion” has been an interactive fashion project where sensor technologies have been embedded with textiles that result in set of ensembles that are light emitting garments, sound sensing garments, proximity garments, shape memory garments etc. Smart textiles, especially in the form of textile interfaces, are drastically underused in fashion and other lifestyle product design. Clothing and some other textile products must be washable, which subjects to the interactive elements to water and chemical immersion, physical stress, and extreme temperature. The current state of the art tends to be too fragile for this treatment. The process for mass producing traditional textiles becomes difficult in interactive textiles. As cutting patterns from larger rolls of cloth and sewing them together to make garments breaks and reforms electronic connections in an uncontrolled manner. Because of this, interactive fabric elements are integrated by hand into textiles produced by standard methods. The Arduino has surely made embedding electronics into textiles much easier than before; even then electronics are not integral to the daily wear garments. Soft and flexible interfaces of MEMS (micro sensors and Micro actuators) can be an option to make this possible by blending electronics within E-textiles in a way that’s seamless and still retains functions of the circuits as well as the garment. Smart clothes, which offer simultaneously a challenging design and utility value, can be only mass produced if the demands of the body are taken care of i.e. protection, anthropometry, ergonomics of human movement, thermo- physiological regulation.

Investigation of Stoneley Waves in Multilayered Plates

Stoneley waves are interface waves that propagate at the interface between two solid media. In this study, the dispersion characteristics and wave structures of Stoneley waves in elastic multilayered plates are displayed and investigated. With a perspective of bulk wave, a reasonable assumption of the potential function forms of the expansion wave and shear wave in nth layer medium is adopted, and the characteristic equation of Stoneley waves in a three-layered plate is given in a determinant form. The dispersion curves and wave structures are solved and presented in both numerical and simulation results. It is observed that two Stoneley wave modes exist in a three-layered plate, that conspicuous dispersion occurs on low frequency band, that the velocity of each Stoneley wave mode approaches the corresponding Stoneley wave velocity at interface between two half infinite spaces. The wave structures reveal that the in-plane displacement of Stoneley waves are relatively high at interfaces, which shows great potential for interface defects detection.

Properties of Adipose Tissue Derived Mesenchymal Stem Cells with Long-Term Cryopreservation

Adult mesenchymal stem cells (MSCs) have been investigated using preclinical approaches for tissue regeneration. Porcine MSCs (pMSCs) are capable of growing and attaching to plastic with a fibroblast-like morphology and then differentiating into bone, adipose, and cartilage tissues in vitro. This study was conducted to investigate the proliferating abilities, differentiation potentials, and multipotency of miniature pig adipose tissue-derived MSCs (mpAD-MSCs) with or without long-term cryopreservation, considering that cryostorage has the potential for use in clinical applications. After confirming the characteristics of the mpAD-MSCs, we examined the effect of long-term cryopreservation (> 2 years) on expression of cell surface markers (CD34, CD90 and CD105), proliferating abilities (cumulative population doubling level, doubling time, colony-forming unit, and MTT assay) and differentiation potentials into mesodermal cell lineages. As a result, the expression of cell surface markers is similar between thawed and fresh mpAD-MSCs. However, long-term cryopreservation significantly lowered the differentiation potentials (adipogenic, chondrogenic, and osteogenic) of mpAD-MSCs. When compared with fresh mpAD-MSCs, thawed mpAD-MSCs exhibited lower expression of mesodermal cell lineage-related genes such as peroxisome proliferator-activated receptor-g2, lipoprotein lipase, collagen Type II alpha 1, osteonectin, and osteocalcin. Interestingly, long-term cryostoraged mpAD-MSCs exhibited significantly higher cell viability than the fresh mpAD-MSCs. Long-term cryopreservation induced a 30% increase in the cell viability of mpAD-MSCs when compared with the fresh mpAD-MSCs at 5 days after thawing. However, long-term cryopreservation significantly lowered expression of stemness markers such as Oct3/4, Sox2, and Nanog. Furthermore, long-term cryopreservation negatively affected expression of senescence-associated genes such as telomerase reverse transcriptase and heat shock protein 90 of mpAD-MSCs when compared with the fresh mpAD-MSCs. The results from this study might be important for the successful application of MSCs in clinical trials after long-term cryopreservation.

Assessing the Effect of Grid Connection of Large-Scale Wind Farms on Power System Small-Signal Angular Stability

Grid connection of a large-scale wind farm affects power system small-signal angular stability in two aspects. Firstly, connection of the wind farm brings about the change of load flow and configuration of a power system. Secondly, the dynamic interaction is introduced by the wind farm with the synchronous generators (SGs) in the power system. This paper proposes a method to assess the two aspects of the effect of the wind farm on power system small-signal angular stability. The effect of the change of load flow/system configuration brought about by the wind farm can be examined separately by displacing wind farms with constant power sources, then the effect of the dynamic interaction of the wind farm with the SGs can be also computed individually. Thus, a clearer picture and better understanding on the power system small-signal angular stability as affected by grid connection of the large-scale wind farm are provided. In the paper, an example power system with grid connection of a wind farm is presented to demonstrate the proposed approach.

Monitorization of Junction Temperature Using a Thermal-Test-Device

Due to the higher power loss levels in electronic components, the thermal design of PCBs (Printed Circuit Boards) of an assembled device becomes one of the most important quality factors in electronics. Nonetheless, some of leading causes of the microelectronic component failures are due to higher temperatures, the leakages or thermal-mechanical stress, which is a concern, is the reliability of microelectronic packages. This article presents an experimental approach to measure the junction temperature of exposed pad packages. The implemented solution is in a prototype phase, using a temperature-sensitive parameter (TSP) to measure temperature directly on the die, validating the numeric results provided by the Mechanical APDL (Ansys Parametric Design Language) under same conditions. The physical device-under-test is composed by a Thermal Test Chip (TTC-1002) and assembly in a QFN cavity, soldered to a test-board according to JEDEC Standards. Monitoring the voltage drop across a forward-biased diode, is an indirectly method but accurate to obtain the junction temperature of QFN component with an applied power range between 0,3W to 1.5W. The temperature distributions on the PCB test-board and QFN cavity surface were monitored by an infra-red thermal camera (Goby-384) controlled and images processed by the Xeneth software. The article provides a set-up to monitorize in real-time the junction temperature of ICs, namely devices with the exposed pad package (i.e. QFN). Presenting the PCB layout parameters that the designer should use to improve thermal performance, and evaluate the impact of voids in solder interface in the device junction temperature.

Ontologies for Social Media Digital Evidence

Online Social Networks (OSNs) are nowadays being used widely and intensively for crime investigation and prevention activities. As they provide a lot of information they are used by the law enforcement and intelligence. An extensive review on existing solutions and models for collecting intelligence from this source of information and making use of it for solving crimes has been presented in this article. The main focus is on smart solutions and models where ontologies have been used as the main approach for representing criminal domain knowledge. A framework for a prototype ontology named SC-Ont will be described. This defines terms of the criminal domain ontology and the relations between them. The terms and the relations are extracted during both this review and the discussions carried out with domain experts. The development of SC-Ont is still ongoing work, where in this paper, we report mainly on the motivation for using smart ontology models and the possible benefits of using them for solving crimes.

Learning Algorithms for Fuzzy Inference Systems Composed of Double- and Single-Input Rule Modules

Most of self-tuning fuzzy systems, which are automatically constructed from learning data, are based on the steepest descent method (SDM). However, this approach often requires a large convergence time and gets stuck into a shallow local minimum. One of its solutions is to use fuzzy rule modules with a small number of inputs such as DIRMs (Double-Input Rule Modules) and SIRMs (Single-Input Rule Modules). In this paper, we consider a (generalized) DIRMs model composed of double and single-input rule modules. Further, in order to reduce the redundant modules for the (generalized) DIRMs model, pruning and generative learning algorithms for the model are suggested. In order to show the effectiveness of them, numerical simulations for function approximation, Box-Jenkins and obstacle avoidance problems are performed.

Affine Projection Adaptive Filter with Variable Regularization

We propose two affine projection algorithms (APA) with variable regularization parameter. The proposed algorithms dynamically update the regularization parameter that is fixed in the conventional regularized APA (R-APA) using a gradient descent based approach. By introducing the normalized gradient, the proposed algorithms give birth to an efficient and a robust update scheme for the regularization parameter. Through experiments we demonstrate that the proposed algorithms outperform conventional R-APA in terms of the convergence rate and the misadjustment error.

Variable Regularization Parameter Normalized Least Mean Square Adaptive Filter

We present a normalized LMS (NLMS) algorithm with robust regularization. Unlike conventional NLMS with the fixed regularization parameter, the proposed approach dynamically updates the regularization parameter. By exploiting a gradient descent direction, we derive a computationally efficient and robust update scheme for the regularization parameter. In simulation, we demonstrate the proposed algorithm outperforms conventional NLMS algorithms in terms of convergence rate and misadjustment error.

An Approach to Noise Variance Estimation in Very Low Signal-to-Noise Ratio Stochastic Signals

This paper describes a method for AWGN (Additive White Gaussian Noise) variance estimation in noisy stochastic signals, referred to as Multiplicative-Noising Variance Estimation (MNVE). The aim was to develop an estimation algorithm with minimal number of assumptions on the original signal structure. The provided MATLAB simulation and results analysis of the method applied on speech signals showed more accuracy than standardized AR (autoregressive) modeling noise estimation technique. In addition, great performance was observed on very low signal-to-noise ratios, which in general represents the worst case scenario for signal denoising methods. High execution time appears to be the only disadvantage of MNVE. After close examination of all the observed features of the proposed algorithm, it was concluded it is worth of exploring and that with some further adjustments and improvements can be enviably powerful.

National Image in the Age of Mass Self-Communication: An Analysis of Internet Users' Perception of Portugal

Nowadays, massification of Internet access represents one of the major challenges to the traditional powers of the State, among which the power to control its external image. The virtual world has also sparked the interest of social sciences which consider it a new field of study, an immense open text where sense is expressed. In this paper, that immense text has been accessed to so as to understand the perception Internet users from all over the world have of Portugal. Ours is a quantitative and qualitative approach, as we have resorted to buzz, thematic and category analysis. The results confirm the predominance of sea stereotype in others' vision of the Portuguese people, and evidence that national image has adapted to network communication through processes of individuation and paganization.

A Model-Driven Approach of User Interface for MVP Rich Internet Application

This paper presents an approach for the model-driven generating of Rich Internet Application (RIA) focusing on the graphical aspect. We used well known Model-Driven Engineering (MDE) frameworks and technologies, such as Eclipse Modeling Framework (EMF), Graphical Modeling Framework (GMF), Query View Transformation (QVTo) and Acceleo to enable the design and the code automatic generation of the RIA. During the development of the approach, we focused on the graphical aspect of the application in terms of interfaces while opting for the Model View Presenter pattern that is designed for graphics interfaces. The paper describes the process followed to define the approach, the supporting tool and presents the results from a case study.

Hybrid Advanced Oxidative Pretreatment of Complex Industrial Effluent for Biodegradability Enhancement

The study explores the hybrid combination of Hydrodynamic Cavitation (HC) and Subcritical Wet Air Oxidation-based pretreatment of complex industrial effluent to enhance the biodegradability selectively (without major COD destruction) to facilitate subsequent enhanced downstream processing via anaerobic or aerobic biological treatment. Advanced oxidation based techniques can be less efficient as standalone options and a hybrid approach by combining Hydrodynamic Cavitation (HC), and Wet Air Oxidation (WAO) can lead to a synergistic effect since both the options are based on common free radical mechanism. The HC can be used for initial turbulence and generation of hotspots which can begin the free radical attack and this agitating mixture then can be subjected to less intense WAO since initial heat (to raise the activation energy) can be taken care by HC alone. Lab-scale venturi-based hydrodynamic cavitation and wet air oxidation reactor with biomethanated distillery wastewater (BMDWW) as a model effluent was examined for establishing the proof-of-concept. The results indicated that for a desirable biodegradability index (BOD: COD - BI) enhancement (up to 0.4), the Cavitation (standalone) pretreatment condition was: 5 bar and 88 min reaction time with a COD reduction of 36 % and BI enhancement of up to 0.27 (initial BI - 0.17). The optimum WAO condition (standalone) was: 150oC, 6 bar and 30 minutes with 31% COD reduction and 0.33 BI. The hybrid pretreatment (combined Cavitation + WAO) worked out to be 23.18 min HC (at 5 bar) followed by 30 min WAO at 150oC, 6 bar, at which around 50% COD was retained yielding a BI of 0.55. FTIR & NMR analysis of pretreated effluent indicated dissociation and/or reorientation of complex organic compounds in untreated effluent to simpler organic compounds post-pretreatment.

Numerical Analysis of the Turbulent Flow around DTMB 4119 Marine Propeller

This article presents a numerical analysis of a turbulent flow past DTMB 4119 marine propeller by the means of RANS approach; the propeller designed at David Taylor Model Basin in USA. The purpose of this study is to predict the hydrodynamic performance of the marine propeller, it aims also to compare the results obtained with the experiment carried out in open water tests; a periodical computational domain was created to reduce the unstructured mesh size generated. The standard kw turbulence model for the simulation is selected; the results were in a good agreement. Therefore, the errors were estimated respectively to 1.3% and 5.9% for KT and KQ.

Effects of Alternative Opportunities and Compensation on Turnover Intention of Singapore PMET

In Singapore, talent retention is one of the most persistent and real issue companies have to grapple with due to the tight labour market. Being resource-scarce, Singapore depends solely on its talented pool of high quality human resource to sustain its competitive advantage in the global economy. But the complex and multifaceted nature of turnover phenomenon makes the prescription of effective talent retention strategies in such a competitive labour market very challenging, especially when it comes to monetary incentives, companies struggle to answer the question of “How much is enough?” By examining the interactive effects of perceived alternative employment opportunities, annual salary and satisfaction with compensation on the turnover intention of 102 Singapore Professionals, Managers, Executives and Technicians (PMET) through correlation analyses and multiple regressions, important insights into the psyche of the Singapore talent pool can be drawn. It is found that annual salary influence turnover intention indirectly through mediation and moderation effects on PMET’s satisfaction on compensation. PMET are also found to be heavily swayed by better external opportunities. This implies that talent retention strategies should not adopt a purely monetary based blanket approach but rather a comprehensive and holistic one that considers the dynamics of prevailing market conditions.