Automatic Thresholding for Data Gap Detection for a Set of Sensors in Instrumented Buildings

Building systems are highly vulnerable to different kinds of faults and failures. In fact, various faults, failures and human behaviors could affect the building performance. This paper tackles the detection of unreliable sensors in buildings. Different literature surveys on diagnosis techniques for sensor grids in buildings have been published but all of them treat only bias and outliers. Occurences of data gaps have also not been given an adequate span of attention in the academia. The proposed methodology comprises the automatic thresholding for data gap detection for a set of heterogeneous sensors in instrumented buildings. Sensor measurements are considered to be regular time series. However, in reality, sensor values are not uniformly sampled. So, the issue to solve is from which delay each sensor become faulty? The use of time series is required for detection of abnormalities on the delays. The efficiency of the method is evaluated on measurements obtained from a real power plant: an office at Grenoble Institute of technology equipped by 30 sensors.

Determination of Surface Deformations with Global Navigation Satellite System Time Series

The development of Global Navigation Satellite System (GNSS) technology has led to increasingly widely and successful applications of GNSS surveys for monitoring crustal movements. Instead of the multi-period GNSS solutions, this study utilizes GNSS time series that are required to more precisely determine the vertical deformations in the study area. In recent years, the surface deformations that are parallel and semi-parallel to Bolvadin fault have occurred in Western Anatolia. These surface deformations have continued to occur in Bolvadin settlement area that is located mostly on alluvium ground. Due to these surface deformations, a number of cracks in the buildings located in the residential areas and breaks in underground water and sewage systems have been observed. In order to determine the amount of vertical surface deformations, two continuous GNSS stations have been established in the region. The stations have been operating since 2015 and 2017, respectively. In this study, GNSS observations from the mentioned two GNSS stations were processed with GAMIT/GLOBK (GNSS Analysis Massachusetts Institute of Technology/GLOBal Kalman) program package to create coordinate time series. With the time series analyses, the GNSS stations’ behaviour models (linear, periodical, etc.), the causes of these behaviours, and mathematical models were determined. The study results from the time series analysis of these two 2 GNSS stations show approximately 50-90 mm/yr vertical movement.

Influence of a High-Resolution Land Cover Classification on Air Quality Modelling

Poor air quality is one of the main environmental causes of premature deaths worldwide, and mainly in cities, where the majority of the population lives. It is a consequence of successive land cover (LC) and use changes, as a result of the intensification of human activities. Knowing these landscape modifications in a comprehensive spatiotemporal dimension is, therefore, essential for understanding variations in air pollutant concentrations. In this sense, the use of air quality models is very useful to simulate the physical and chemical processes that affect the dispersion and reaction of chemical species into the atmosphere. However, the modelling performance should always be evaluated since the resolution of the input datasets largely dictates the reliability of the air quality outcomes. Among these data, the updated LC is an important parameter to be considered in atmospheric models, since it takes into account the Earth’s surface changes due to natural and anthropic actions, and regulates the exchanges of fluxes (emissions, heat, moisture, etc.) between the soil and the air. This work aims to evaluate the performance of the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem), when different LC classifications are used as an input. The influence of two LC classifications was tested: i) the 24-classes USGS (United States Geological Survey) LC database included by default in the model, and the ii) CLC (Corine Land Cover) and specific high-resolution LC data for Portugal, reclassified according to the new USGS nomenclature (33-classes). Two distinct WRF-Chem simulations were carried out to assess the influence of the LC on air quality over Europe and Portugal, as a case study, for the year 2015, using the nesting technique over three simulation domains (25 km2, 5 km2 and 1 km2 horizontal resolution). Based on the 33-classes LC approach, particular emphasis was attributed to Portugal, given the detail and higher LC spatial resolution (100 m x 100 m) than the CLC data (5000 m x 5000 m). As regards to the air quality, only the LC impacts on tropospheric ozone concentrations were evaluated, because ozone pollution episodes typically occur in Portugal, in particular during the spring/summer, and there are few research works relating to this pollutant with LC changes. The WRF-Chem results were validated by season and station typology using background measurements from the Portuguese air quality monitoring network. As expected, a better model performance was achieved in rural stations: moderate correlation (0.4 – 0.7), BIAS (10 – 21µg.m-3) and RMSE (20 – 30 µg.m-3), and where higher average ozone concentrations were estimated. Comparing both simulations, small differences grounded on the Leaf Area Index and air temperature values were found, although the high-resolution LC approach shows a slight enhancement in the model evaluation. This highlights the role of the LC on the exchange of atmospheric fluxes, and stresses the need to consider a high-resolution LC characterization combined with other detailed model inputs, such as the emission inventory, to improve air quality assessment.

Evaluating Damage Spectra for Steel Braced Frames Due to Near-Field and Far-Field Earthquakes

Recent ground motion records demonstrate that the near-field earthquakes have various properties compared to far-field earthquakes. In general, most of these properties are affected by an important phenomenon called ‘forward directivity’ in near-fault earthquakes. Measuring structural damages are one of the common activities administered after an earthquake. Predicting the amount of damage caused by the earthquake as well as determining the vulnerability of the structure is extremely significant. In order to measure the amount of structural damages, instead of calculating the acceleration and velocity spectrum, it is possible to use the damage spectra of the structure. The damage spectrum is a kind of nonlinear spectrum that is drawn by setting the nonlinear parameters related to the single degree of freedom structures and its dynamic analysis under the specific record and measuring damage of any structure. In this study, the damage spectra of steel structures have been drawn. For this purpose, different kinds of concentric and eccentric braced structures with various ductility coefficients in hard and soft soil under near-field and far-field ground motion records have been considered using the Krawinkler and Zohrei damage index. The results indicate that, by increasing the structures' fundamental period, the amount of damage increases under the near-field earthquakes compared to far-field earthquakes. In addition, by increasing the structure ductility, the amount of damage based on near-field and far-field earthquakes decreases noticeably. Furthermore, in concentric braced structures, the amount of damage under the near-field earthquakes is almost two times more than the amount of damage in eccentrically braced structures especially for fundamental periods larger than 0.6 s.

CRYPTO COPYCAT: A Fashion Centric Blockchain Framework for Eliminating Fashion Infringement

The fashion industry represents a significant portion of the global gross domestic product, however, it is plagued by cheap imitators that infringe on the trademarks which destroys the fashion industry's hard work and investment. While eventually the copycats would be found and stopped, the damage has already been done, sales are missed and direct and indirect jobs are lost. The infringer thrives on two main facts: the time it takes to discover them and the lack of tracking technologies that can help the consumer distinguish them. Blockchain technology is a new emerging technology that provides a distributed encrypted immutable and fault resistant ledger. Blockchain presents a ripe technology to resolve the infringement epidemic facing the fashion industry. The significance of the study is that a new approach leveraging the state of the art blockchain technology coupled with artificial intelligence is used to create a framework addressing the fashion infringement problem. It transforms the current focus on legal enforcement, which is difficult at best, to consumer awareness that is far more effective. The framework, Crypto CopyCat, creates an immutable digital asset representing the actual product to empower the customer with a near real time query system. This combination emphasizes the consumer's awareness and appreciation of the product's authenticity, while provides real time feedback to the producer regarding the fake replicas. The main findings of this study are that implementing this approach can delay the fake product penetration of the original product market, thus allowing the original product the time to take advantage of the market. The shift in the fake adoption results in reduced returns, which impedes the copycat market and moves the emphasis to the original product innovation.

Performance Improvement of Information System of a Banking System Based on Integrated Resilience Engineering Design

Integrated resilience engineering (IRE) is capable of returning banking systems to the normal state in extensive economic circumstances. In this study, information system of a large bank (with several branches) is assessed and optimized under severe economic conditions. Data envelopment analysis (DEA) models are employed to achieve the objective of this study. Nine IRE factors are considered to be the outputs, and a dummy variable is defined as the input of the DEA models. A standard questionnaire is designed and distributed among executive managers to be considered as the decision-making units (DMUs). Reliability and validity of the questionnaire is examined based on Cronbach's alpha and t-test. The most appropriate DEA model is determined based on average efficiency and normality test. It is shown that the proposed integrated design provides higher efficiency than the conventional RE design. Results of sensitivity and perturbation analysis indicate that self-organization, fault tolerance, and reporting culture respectively compose about 50 percent of total weight.

The Social Reaction to the Wadi Salib Riots (1959) as Reflected in Contemporary Israeli Press

Social reactions to deviant groups with political goals follow two central patterns; one that associates personal characteristics with deviant behavior, and the other that claims that society is to be blamed for deviant behavior. The establishment usually tends towards the former notion and thus disclaims any responsibility for the distress of the underprivileged, while it is usually those who oppose government policies who believe that the fault lies with society. The purpose of the present research was to examine social reactions to the Wadi Salib riots that occurred in Haifa in 1959. These riots represented the first ethnic protest within Israeli society with its ideology of the ingathering of the exiles. The central question was whether this ideology contributed to the development of a different reaction when compared to reactions to similar events abroad. This question was examined by means of analyzing articles in the Israeli press of that period. The Israeli press representing the views of the establishment was at pains to point out that the rioters were criminals, their object being to obstruct the development of society. Opposition party leaders claimed that the rioters lived in poor circumstances, which constituted a direct result of government policies. An analysis of press reports on the Wadi Salib riots indicates a correspondence between the reaction to these events and similar events abroad. Nevertheless, the reaction to the Wadi Salib riots did not only express a conflict between different political camps, but also different symbolic universes. Each group exploited the events at Wadi Salib to prove that their ideology was the legitimate one.

States Estimation and Fault Detection of a Doubly Fed Induction Machine by Moving Horizon Estimation

This paper presents the estimation of the key parameters of a double fed induction machine (DFIM) by the use of the moving horizon estimator (MHE) for control and monitoring purpose. A study was conducted on the behavior of this observer in the presence of some faults which can occur during the operation of the machine. In the first case a stator phase has been suppressed. In the second case the rotor resistance has been multiplied by a factor. The results show a good estimation of different parameters such as rotor flux, rotor speed, stator current with a very small estimation error. The robustness of the observer was also tested in the practical case of DFIM by using another model different from the real one at a constant close. The very small estimation error makes the MHE a good software sensor candidate for monitoring purpose for the DFIM. 

Geology, Geomorphology and Genesis of Andarokh Karstic Cave, North-East Iran

Andarokh basin is one of the main karstic regions in Khorasan Razavi province NE Iran. This basin is part of Kopeh-Dagh mega zone extending from Caspian Sea in the east to northern Afghanistan in the west. This basin is covered by Mozdooran Formation, Ngr evaporative formation and quaternary alluvium deposits in descending order of age. Mozdooran carbonate formation is notably karstified. The main surface karstic features in Mozdooran formation are Groove karren, Cleft karren, Rain pit, Rill karren, Tritt karren, Kamintza, Domes, and Table karren. In addition to surface features, deep karstic feature Andarokh Cave also exists in the region. Studying Ca, Mg, Mn, Sr, Fe concentration and Sr/Mn ratio in Mozdooran formation samples with distance to main faults and joints system using PCA analyses demonstrates intense meteoric digenesis role in controlling carbonate rock geochemistry. The karst evaluation in Andarokh basin varies from early stages 'deep seated karst' in Mesozoic to mature karstic system 'Exhumed karst' in quaternary period. Andarokh cave (the main cave in Andarokh basin) is rudimentary branch work consists of three passages of A, B and C and two entrances Andarokh and Sky.

Evaluating 8D Reports Using Text-Mining

Increasing quality requirements make reliable and effective quality management indispensable. This includes the complaint handling in which the 8D method is widely used. The 8D report as a written documentation of the 8D method is one of the key quality documents as it internally secures the quality standards and acts as a communication medium to the customer. In practice, however, the 8D report is mostly faulty and of poor quality. There is no quality control of 8D reports today. This paper describes the use of natural language processing for the automated evaluation of 8D reports. Based on semantic analysis and text-mining algorithms the presented system is able to uncover content and formal quality deficiencies and thus increases the quality of the complaint processing in the long term.

CompleX-Machine: An Automated Testing Tool Using X-Machine Theory

This paper is aimed at creating an Automatic Java X-Machine testing tool for software development. The nature of software development is changing; thus, the type of software testing tools required is also changing. Software is growing increasingly complex and, in part due to commercial impetus for faster software releases with new features and value, increasingly in danger of containing faults. These faults can incur huge cost for software development organisations and users; Cambridge Judge Business School’s research estimated the cost of software bugs to the global economy is $312 billion. Beyond the cost, faster software development methodologies and increasing expectations on developers to become testers is driving demand for faster, automated, and effective tools to prevent potential faults as early as possible in the software development lifecycle. Using X-Machine theory, this paper will explore a new tool to address software complexity, changing expectations on developers, faster development pressures and methodologies, with a view to reducing the huge cost of fixing software bugs.

A Comparative Analysis of the Performance of COSMO and WRF Models in Quantitative Rainfall Prediction

The Numerical weather prediction (NWP) models are considered powerful tools for guiding quantitative rainfall prediction. A couple of NWP models exist and are used at many operational weather prediction centers. This study considers two models namely the Consortium for Small–scale Modeling (COSMO) model and the Weather Research and Forecasting (WRF) model. It compares the models’ ability to predict rainfall over Uganda for the period 21st April 2013 to 10th May 2013 using the root mean square (RMSE) and the mean error (ME). In comparing the performance of the models, this study assesses their ability to predict light rainfall events and extreme rainfall events. All the experiments used the default parameterization configurations and with same horizontal resolution (7 Km). The results show that COSMO model had a tendency of largely predicting no rain which explained its under–prediction. The COSMO model (RMSE: 14.16; ME: -5.91) presented a significantly (p = 0.014) higher magnitude of error compared to the WRF model (RMSE: 11.86; ME: -1.09). However the COSMO model (RMSE: 3.85; ME: 1.39) performed significantly (p = 0.003) better than the WRF model (RMSE: 8.14; ME: 5.30) in simulating light rainfall events. All the models under–predicted extreme rainfall events with the COSMO model (RMSE: 43.63; ME: -39.58) presenting significantly higher error magnitudes than the WRF model (RMSE: 35.14; ME: -26.95). This study recommends additional diagnosis of the models’ treatment of deep convection over the tropics.

Hybrid Approach for Software Defect Prediction Using Machine Learning with Optimization Technique

Software technology is developing rapidly which leads to the growth of various industries. Now-a-days, software-based applications have been adopted widely for business purposes. For any software industry, development of reliable software is becoming a challenging task because a faulty software module may be harmful for the growth of industry and business. Hence there is a need to develop techniques which can be used for early prediction of software defects. Due to complexities in manual prediction, automated software defect prediction techniques have been introduced. These techniques are based on the pattern learning from the previous software versions and finding the defects in the current version. These techniques have attracted researchers due to their significant impact on industrial growth by identifying the bugs in software. Based on this, several researches have been carried out but achieving desirable defect prediction performance is still a challenging task. To address this issue, here we present a machine learning based hybrid technique for software defect prediction. First of all, Genetic Algorithm (GA) is presented where an improved fitness function is used for better optimization of features in data sets. Later, these features are processed through Decision Tree (DT) classification model. Finally, an experimental study is presented where results from the proposed GA-DT based hybrid approach is compared with those from the DT classification technique. The results show that the proposed hybrid approach achieves better classification accuracy.

Transient Voltage Distribution on the Single Phase Transmission Line under Short Circuit Fault Effect

Single phase transmission lines are used to transfer data or energy between two users. Transient conditions such as switching operations and short circuit faults cause the generation of the fluctuation on the waveform to be transmitted. Spatial voltage distribution on the single phase transmission line may change owing to the position and duration of the short circuit fault in the system. In this paper, the state space representation of the single phase transmission line for short circuit fault and for various types of terminations is given. Since the transmission line is modeled in time domain using distributed parametric elements, the mathematical representation of the event is given in state space (time domain) differential equation form. It also makes easy to solve the problem because of the time and space dependent characteristics of the voltage variations on the distributed parametrically modeled transmission line.

Investigating the Behavior of Underground Structures in the Event of an Earthquake

The progress of technology and producing new machinery have made a big change in excavation operations and construction of underground structures. The limitations of space and some other economic, politic and military considerations gained the attention of most developed and developing countries towards the construction of these structures for mine, military, and development objectives. Underground highways, tunnels, subways, oil reservoir resources, fuels, nuclear wastes burying reservoir and underground stores are increasingly developing and being used in these countries. The existence and habitability of the cities depend on these underground installations or in other words these vital arteries. Stopping the flow of water, gas leakage and explosion, collapsing of sewage paths, etc., resulting from the earthquake are among the factors that can severely harm the environment and increase the casualty. Lack of sewage network and complete stoppage of the flow of water in Bam (Iran) is a good example of this kind. In this paper, we investigate the effect of wave orientation on structures and deformation of them and the effect of faulting on underground structures, and then, we study resistance of reinforced concrete against earthquake, simulate two different samples, analyze the result and point out the importance of paying attention to underground installations.

Crash and Injury Characteristics of Riders in Motorcycle-Passenger Vehicle Crashes

The motorcycle has become one of the most common type of vehicles used on the road, particularly in the Asia region, including Malaysia, due to its size-convenience and affordable price. This study focuses only on crashes involving motorcycles with passenger cars consisting 43 real world crashes obtained from in-depth crash investigation process from June 2016 till July 2017. The study collected and analyzed vehicle and site parameters obtained during crash investigation and injury information acquired from the patient-treating hospital. The investigation team, consisting of two personnel, is stationed at the Emergency Department of the treatment facility, and was dispatched to the crash scene once receiving notification of the related crashes. The injury information retrieved was coded according to the level of severity using the Abbreviated Injury Scale (AIS) and classified into different body regions. The data revealed that weekend crashes were significantly higher for the night time period and the crash occurrence was the highest during morning hours (commuting to work period) for weekdays. Bad weather conditions play a minimal effect towards the occurrence of motorcycle – passenger vehicle crashes and nearly 90% involved motorcycles with single riders. Riders up to 25 years old are heavily involved in crashes with passenger vehicles (60%), followed by 26-55 year age group with 35%. Male riders were dominant in each of the age segments. The majority of the crashes involved side impacts, followed by rear impacts and cars outnumbered the rest of the passenger vehicle types in terms of crash involvement with motorcycles. The investigation data also revealed that passenger vehicles were the most at-fault counterpart (62%) when involved in crashes with motorcycles and most of the crashes involved situations whereby both of the vehicles are travelling in the same direction and one of the vehicles is in a turning maneuver. More than 80% of the involved motorcycle riders had sustained yellow severity level during triage process. The study also found that nearly 30% of the riders sustained injuries to the lower extremities, while MAIS level 3 injuries were recorded for all body regions except for thorax region. The result showed that crashes in which the motorcycles were found to be at fault were more likely to occur during night and raining conditions. These types of crashes were also found to be more likely to involve other types of passenger vehicles rather than cars and possess higher likelihood in resulting higher ISS (>6) value to the involved rider. To reduce motorcycle fatalities, it first has to understand the characteristics concerned and focus may be given on crashes involving passenger vehicles as the most dominant crash partner on Malaysian roads.

Conditions for Fault Recovery of Interconnected Asynchronous Sequential Machines with State Feedback

In this paper, fault recovery for parallel interconnected asynchronous sequential machines is studied. An adversarial input can infiltrate into one of two submachines comprising parallel composition of the considered asynchronous sequential machine, causing an unauthorized state transition. The control objective is to elucidate the condition for the existence of a corrective controller that makes the closed-loop system immune against any occurrence of adversarial inputs. In particular, an efficient existence condition is presented that does not need the complete modeling of the interconnected asynchronous sequential machine.

Robust Coordinated Design of Multiple Power System Stabilizers Using Particle Swarm Optimization Technique

Power system stabilizers (PSS) are now routinely used in the industry to damp out power system oscillations. In this paper, particle swarm optimization (PSO) technique is applied to coordinately design multiple power system stabilizers (PSS) in a multi-machine power system. The design problem of the proposed controllers is formulated as an optimization problem and PSO is employed to search for optimal controller parameters. By minimizing the time-domain based objective function, in which the deviation in the oscillatory rotor speed of the generator is involved; stability performance of the system is improved. The non-linear simulation results are presented for various severe disturbances and small disturbance at different locations as well as for various fault clearing sequences to show the effectiveness and robustness of the proposed controller and their ability to provide efficient damping of low frequency oscillations.

Design and Implementation of 4 Bit Multiplier Using Fault Tolerant Hybrid Full Adder

The fault tolerant system plays a crucial role in the critical applications which are being used in the present scenario. A fault may change the functionality of circuits. Aim of this paper is to design multiplier using fault tolerant hybrid full adder. Fault tolerant hybrid full adder is designed to check and repair any fault in the circuit using self-checking circuit and the self-repairing circuit. Further, the use of conventional logic circuits may result in more area, delay as well as power consumption. In order to reduce these parameters of the circuit, GDI (Gate Diffusion Input) techniques with less number of transistors are used compared to conventional full adder circuit. This reduces the area, delay and power consumption. The proposed method solves the major problems occurring in the most crucial and critical applications.

Neural Network Based Approach of Software Maintenance Prediction for Laboratory Information System

Software maintenance phase is started once a software project has been developed and delivered. After that, any modification to it corresponds to maintenance. Software maintenance involves modifications to keep a software project usable in a changed or a changing environment, to correct discovered faults, and modifications, and to improve performance or maintainability. Software maintenance and management of software maintenance are recognized as two most important and most expensive processes in a life of a software product. This research is basing the prediction of maintenance, on risks and time evaluation, and using them as data sets for working with neural networks. The aim of this paper is to provide support to project maintenance managers. They will be able to pass the issues planned for the next software-service-patch to the experts, for risk and working time evaluation, and afterward to put all data to neural networks in order to get software maintenance prediction. This process will lead to the more accurate prediction of the working hours needed for the software-service-patch, which will eventually lead to better planning of budget for the software maintenance projects.