Retrospective Reconstruction of Time Series Data for Integrated Waste Management

The development, operation and maintenance of Integrated Waste Management Systems (IWMS) affects essentially the sustainable concern of every region. The features of such systems have great influence on all of the components of sustainability. In order to reach the optimal way of processes, a comprehensive mapping of the variables affecting the future efficiency of the system is needed such as analysis of the interconnections among the components and modeling of their interactions. The planning of a IWMS is based fundamentally on technical and economical opportunities and the legal framework. Modeling the sustainability and operation effectiveness of a certain IWMS is not in the scope of the present research. The complexity of the systems and the large number of the variables require the utilization of a complex approach to model the outcomes and future risks. This complex method should be able to evaluate the logical framework of the factors composing the system and the interconnections between them. The authors of this paper studied the usability of the Fuzzy Cognitive Map (FCM) approach modeling the future operation of IWMS’s. The approach requires two input data set. One is the connection matrix containing all the factors affecting the system in focus with all the interconnections. The other input data set is the time series, a retrospective reconstruction of the weights and roles of the factors. This paper introduces a novel method to develop time series by content analysis.

Scale Development for Measuring E-Service Quality in Banking

This study examines several critical dimensions of eservice quality overlooked in the existing literature and proposes a model and instrument framework for measuring customer perceived e-service quality in the banking sector. The initial design was derived from a pool of instrument dimensions and their items from the existing literature review by content analysis. Based on focused group discussion, nine dimensions were extracted. An exploratory factor analysis approach was applied to data from a survey of 323 respondents. The instrument has been designed specifically for the banking sector. Research data was collected from bank customers who use electronic banking in a developing economy. A nine-factor instrument has been proposed to measure the e-service quality. The instrument has been checked for reliability. The validity and sample place limited the applicability of the instrument across economies and service categories. Future research must be conducted to check the validity. This instrument can help bankers in developing economies like India to measure the e-service quality and make improvements. The present study offers a systematic procedure that provides insights on to the conceptual and empirical comprehension of customer perceived e-service quality and its constituents.

Non-Standard Monetary Policy Measures and Their Consequences

The study is a review of the literature concerning the consequences of non-standard monetary policy, which are used by central banks during unconventional periods, threatening banking sector instability. In particular, the attention was paid to the effects of non-standard monetary policy tools for financial markets. However, the empirical evidence about their effects and real consequences for financial markets is still not final. The main aim of the study is to survey consequences of standard and non-standard monetary policy instruments, implemented during the global financial crisis in the United States, United Kingdom and euro area, with particular attention to the results for the stabilization of global financial markets. The study consists mainly of the empirical review, indicating the impact of the implementation of these tools for financial markets. The following research methods were used in the study: literature studies, including domestic and foreign literature, cause and effect analysis and statistical analysis.

Predicting Application Layer DDoS Attacks Using Machine Learning Algorithms

A Distributed Denial of Service (DDoS) attack is a major threat to cyber security. It originates from the network layer or the application layer of compromised/attacker systems which are connected to the network. The impact of this attack ranges from the simple inconvenience to use a particular service to causing major failures at the targeted server. When there is heavy traffic flow to a target server, it is necessary to classify the legitimate access and attacks. In this paper, a novel method is proposed to detect DDoS attacks from the traces of traffic flow. An access matrix is created from the traces. As the access matrix is multi dimensional, Principle Component Analysis (PCA) is used to reduce the attributes used for detection. Two classifiers Naive Bayes and K-Nearest neighborhood are used to classify the traffic as normal or abnormal. The performance of the classifier with PCA selected attributes and actual attributes of access matrix is compared by the detection rate and False Positive Rate (FPR).

Fiber Braggs Grating Sensor Based Instrumentation to Evaluate Postural Balance and Stability on an Unstable Platform

This paper describes a novel application of Fiber Braggs Grating (FBG) sensors in the assessment of human postural stability and balance on an unstable platform. In this work, FBG sensor Stability Analyzing Device (FBGSAD) is developed for measurement of plantar strain to assess the postural stability of subjects on unstable platforms during different stances in eyes open and eyes closed conditions on a rocker board. The studies are validated by comparing the Centre of Gravity (CG) variations measured on the lumbar vertebra of subjects using a commercial accelerometer. The results obtained from the developed FBGSAD depict qualitative similarities with the data recorded by commercial accelerometer. The advantage of the FBGSAD is that it measures simultaneously plantar strain distribution and postural stability of the subject along with its inherent benefits like non-requirement of energizing voltage to the sensor, electromagnetic immunity and simple design which suits its applicability in biomechanical applications. The developed FBGSAD can serve as a tool/yardstick to mitigate space motion sickness, identify individuals who are susceptible to falls and to qualify subjects for balance and stability, which are important factors in the selection of certain unique professionals such as aircraft pilots, astronauts, cosmonauts etc.

The Effect of Discontinued Water Spray Cooling on the Heat Transfer Coefficient

Water spray cooling is a technique typically used in heat treatment and other metallurgical processes where controlled temperature regimes are required. Water spray cooling is used in static (without movement) or dynamic (with movement of the steel plate) regimes. The static regime is notable for the fixed position of the hot steel plate and fixed spray nozzle. This regime is typical for quenching systems focused on heat treatment of the steel plate. The second application of spray cooling is the dynamic regime. The dynamic regime is notable for its static section cooling system and moving steel plate. This regime is used in rolling and finishing mills. The fixed position of cooling sections with nozzles and the movement of the steel plate produce nonhomogeneous water distribution on the steel plate. The length of cooling sections and placement of water nozzles in combination with the nonhomogeneity of water distribution lead to discontinued or interrupted cooling conditions. The impact of static and dynamic regimes on cooling intensity and the heat transfer coefficient during the cooling process of steel plates is an important issue. Heat treatment of steel is accompanied by oxide scale growth. The oxide scale layers can significantly modify the cooling properties and intensity during the cooling. The combination of static and dynamic (section) regimes with the variable thickness of the oxide scale layer on the steel surface impact the final cooling intensity. The study of the influence of the oxide scale layers with different cooling regimes was carried out using experimental measurements and numerical analysis. The experimental measurements compared both types of cooling regimes and the cooling of scale-free surfaces and oxidized surfaces. A numerical analysis was prepared to simulate the cooling process with different conditions of the section and samples with different oxide scale layers.

Estimation of Thermal Conductivity of Nanofluids Using MD-Stochastic Simulation Based Approach

The thermal conductivity of a fluid can be significantly enhanced by dispersing nano-sized particles in it, and the resultant fluid is termed as "nanofluid". A theoretical model for estimating the thermal conductivity of a nanofluid has been proposed here. It is based on the mechanism that evenly dispersed nanoparticles within a nanofluid undergo Brownian motion in course of which the nanoparticles repeatedly collide with the heat source. During each collision a rapid heat transfer occurs owing to the solidsolid contact. Molecular dynamics (MD) simulation of the collision of nanoparticles with the heat source has shown that there is a pulselike pick up of heat by the nanoparticles within 20-100 ps, the extent of which depends not only on thermal conductivity of the nanoparticles, but also on the elastic and other physical properties of the nanoparticle. After the collision the nanoparticles undergo Brownian motion in the base fluid and release the excess heat to the surrounding base fluid within 2-10 ms. The Brownian motion and associated temperature variation of the nanoparticles have been modeled by stochastic analysis. Repeated occurrence of these events by the suspended nanoparticles significantly contributes to the characteristic thermal conductivity of the nanofluids, which has been estimated by the present model for a ethylene glycol based nanofluid containing Cu-nanoparticles of size ranging from 8 to 20 nm, with Gaussian size distribution. The prediction of the present model has shown a reasonable agreement with the experimental data available in literature.

The Efficiency of Mechanization in Weed Control in Artificial Regeneration of Oriental Beech (Fagus orientalis Lipsky.)

In this study which has been conducted in Akçasu Forest Range District of Devrek Forest Directorate; 3 methods (weed control with labourer power, cover removal with Hitachi F20 Excavator, and weed control with agricultural equipment mounted on a Ferguson 240S agriculture tractor) were utilized in weed control efforts in regeneration of degraded oriental beech forests have been compared. In this respect, 3 methods have been compared by determining certain work hours and standard durations of unit areas (1 hectare). For this purpose, evaluating the tasks made with human and machine force from the aspects of duration, productivity and costs, it has been aimed to determine the most productive method in accordance with the actual ecological conditions of research field. Within the scope of the study, the time studies have been conducted for 3 methods used in weed control efforts. While carrying out those studies, the performed implementations have been evaluated by dividing them into business stages. Also, the actual data have been used while calculating the cost accounts. In those calculations, the latest formulas and equations which are also used in developed countries have been utilized. The variance of analysis (ANOVA) was used in order to determine whether there is any statistically significant difference among obtained results, and the Duncan test was used for grouping if there is significant difference. According to the measurements and findings carried out within the scope of this study, it has been found during living cover removal efforts in regeneration efforts in demolished oriental beech forests that the removal of weed layer in 1 hectare of field has taken 920 hours with labourer force, 15.1 hours with excavator and 60 hours with an equipment mounted on a tractor. On the other hand, it has been determined that the cost of removal of living cover in unit area (1 hectare) was 3220.00 TL for labourer power, 1250 TL for excavator and 1825 TL for equipment mounted on a tractor. According to the obtained results, it has been found that the utilization of excavator in weed control effort in regeneration of degraded oriental beech regions under actual ecological conditions of research field has been found to be more productive from both of aspects of duration and costs. These determinations carried out should be repeated in weed control efforts in degraded forest fields with different ecological conditions, it is compulsory for finding the most efficient weed control method. These findings will light the way of technical staff of forestry directorate in determination of the most effective and economic weed control method. Thus, the more actual data will be used while preparing the weed control budgets, and there will be significant contributions to national economy. Also the results of this and similar studies are very important for developing the policies for our forestry in short and long term.

Examining the Perceived Usefulness of ICTs for Learning about Indigenous Foods

Science and technology has a major impact on many societal domains such as communication, medicine, food, transportation, etc. However, this dominance of modern technology can have a negative unintended impact on indigenous systems, and in particular on indigenous foods. This problem serves as a motivation to this study whose aim is to examine the perceptions of learners on the usefulness of Information and Communication Technologies (ICTs) for learning about indigenous foods. This aim will be subdivided into two types of research objectives. The design and identification of theories and models will be achieved using literature content analysis. The objective on the empirical testing of such theories and models will be achieved through the survey of Hospitality studies learners from different schools in the iLembe and Umgungundlovu Districts of the South African Kwazulu-Natal province. SPSS is used to quantitatively analyze the data collected by the questionnaire of this survey using descriptive statistics and Pearson correlations after the assessment of the validity and the reliability of the data. The main hypothesis behind this study is that there is a connection between the demographics of learners, their perceptions on the usefulness of ICTs for learning about indigenous foods, and the following personality and eLearning related theories constructs: Computer self-efficacy, Trust in ICT systems, and Conscientiousness; as suggested by existing studies on learning theories. This hypothesis was fully confirmed by the survey conducted by this study except for the demographic factors where gender and age were not found to be determinant factors of learners’ perceptions on the usefulness of ICTs for learning about indigenous foods.

Comparative Study of Line Voltage Stability Indices for Voltage Collapse Forecasting in Power Transmission System

At present, the evaluation of voltage stability assessment experiences sizeable anxiety in the safe operation of power systems. This is due to the complications of a strain power system. With the snowballing of power demand by the consumers and also the restricted amount of power sources, therefore, the system has to perform at its maximum proficiency. Consequently, the noteworthy to discover the maximum ability boundary prior to voltage collapse should be undertaken. A preliminary warning can be perceived to evade the interruption of power system’s capacity. The effectiveness of line voltage stability indices (LVSI) is differentiated in this paper. The main purpose of the indices used is to predict the proximity of voltage instability of the electric power system. On the other hand, the indices are also able to decide the weakest load buses which are close to voltage collapse in the power system. The line stability indices are assessed using the IEEE 14 bus test system to validate its practicability. Results demonstrated that the implemented indices are practically relevant in predicting the manifestation of voltage collapse in the system. Therefore, essential actions can be taken to dodge the incident from arising.

Stability Analysis of Three-Lobe Journal Bearing Lubricated with a Micropolar Fluids

In this paper, the dynamic characteristics of a threelobe journal bearing lubricated with micropolar fluids are determined by the linear stability theory. Lubricating oil containing additives and contaminants is modelled as micropolar fluid. The modified Reynolds equation is obtained using the micropolar lubrication theory .The finite difference technique has been used to determine the solution of the modified Reynolds equation. The dynamic characteristics in terms of stiffness, damping coefficients, the critical mass and whirl ratio are determined for various values of size of material characteristic length and the coupling number. The computed results show that the three-lobe bearing lubricated with micropolar fluid exhibits better stability compared with that lubricated with Newtonian fluid. According to the results obtained, the effect of the parameter micropolar fluid is remarkable on the dynamic characteristics and stability of the three-lobe bearing.

A Sequential Approach to Random-Effects Meta-Analysis

The objective of meta-analysis is to combine results from several independent studies in order to create generalization and provide evidence base for decision making. But recent studies show that the magnitude of effect size estimates reported in many areas of research significantly changed over time and this can impair the results and conclusions of meta-analysis. A number of sequential methods have been proposed for monitoring the effect size estimates in meta-analysis. However they are based on statistical theory applicable only to fixed effect model (FEM) of meta-analysis. For random-effects model (REM), the analysis incorporates the heterogeneity variance, τ 2 and its estimation create complications. In this paper we study the use of a truncated CUSUM-type test with asymptotically valid critical values for sequential monitoring in REM. Simulation results show that the test does not control the Type I error well, and is not recommended. Further work required to derive an appropriate test in this important area of applications.

Power Flow Analysis for Radial Distribution System Using Backward/Forward Sweep Method

This paper proposes a backward/forward sweep method to analyze the power flow in radial distribution systems. The distribution system has radial structure and high R/X ratios. So the newton-raphson and fast decoupled methods are failed with distribution system. The proposed method presents a load flow study using backward/forward sweep method, which is one of the most effective methods for the load-flow analysis of the radial distribution system. By using this method, power losses for each bus branch and voltage magnitudes for each bus node are determined. This method has been tested on IEEE 33-bus radial distribution system and effective results are obtained using MATLAB.

Theoretical Modal Analysis of Freely and Simply Supported RC Slabs

This paper focuses on the dynamic behavior of reinforced concrete (RC) slabs. Therefore, the theoretical modal analysis was performed using two different types of boundary conditions. Modal analysis method is the most important dynamic analyses. The analysis would be modal case when there is no external force on the structure. By using this method in this paper, the effects of freely and simply supported boundary conditions on the frequencies and mode shapes of RC square slabs are studied. ANSYS software was employed to derive the finite element model to determine the natural frequencies and mode shapes of the slabs. Then, the obtained results through numerical analysis (finite element analysis) would be compared with the exact solution. The main goal of the research study is to predict how the boundary conditions change the behavior of the slab structures prior to performing experimental modal analysis. Based on the results, it is concluded that simply support boundary condition has obvious influence to increase the natural frequencies and change the shape of the mode when it is compared with freely supported boundary condition of slabs. This means that such support conditions have the direct influence on the dynamic behavior of the slabs. Thus, it is suggested to use free-free boundary condition in experimental modal analysis to precisely reflect the properties of the structure. By using free-free boundary conditions, the influence of poorly defined supports is interrupted.

Prediction of Seismic Damage Using Scalar Intensity Measures Based On Integration of Spectral Values

A key issue in seismic risk analysis within the context of Performance-Based Earthquake Engineering is the evaluation of the expected seismic damage of structures under a specific earthquake ground motion. The assessment of the seismic performance strongly depends on the choice of the seismic Intensity Measure (IM), which quantifies the characteristics of a ground motion that are important to the nonlinear structural response. Several conventional IMs of ground motion have been used to estimate their damage potential to structures. Yet, none of them has been proved to be able to predict adequately the seismic damage. Therefore, alternative, scalar intensity measures, which take into account not only ground motion characteristics but also structural information have been proposed. Some of these IMs are based on integration of spectral values over a range of periods, in an attempt to account for the information that the shape of the acceleration, velocity or displacement spectrum provides. The adequacy of a number of these IMs in predicting the structural damage of 3D R/C buildings is investigated in the present paper. The investigated IMs, some of which are structure specific and some are non structure-specific, are defined via integration of spectral values. To achieve this purpose three symmetric in plan R/C buildings are studied. The buildings are subjected to 59 bidirectional earthquake ground motions. The two horizontal accelerograms of each ground motion are applied along the structural axes. The response is determined by nonlinear time history analysis. The structural damage is expressed in terms of the maximum interstory drift as well as the overall structural damage index. The values of the aforementioned seismic damage measures are correlated with seven scalar ground motion IMs. The comparative assessment of the results revealed that the structure-specific IMs present higher correlation with the seismic damage of the three buildings. However, the adequacy of the IMs for estimation of the structural damage depends on the response parameter adopted. Furthermore, it was confirmed that the widely used spectral acceleration at the fundamental period of the structure is a good indicator of the expected earthquake damage level.

Analysis of Combined Heat Transfer through the Core Materials of VIPs with Various Scattering Properties

Vacuum Insulation Panel (VIP) can achieve very low thermal conductivity by evacuating its inner space. Heat transfer in the core materials of highly-evacuated VIP occurs by conduction through the solid structure and radiation through the pore. The effect of various scattering modes in combined conduction-radiation in VIP is investigated through numerical analysis. The discrete ordinates interpolation method (DOIM) incorporated with the commercial code FLUENT® is employed. It is found that backward scattering is more effective in reducing the total heat transfer while isotropic scattering is almost identical with pure absorbing/emitting case of the same optical thickness. For a purely scattering medium, the results agrees well with additive solution with diffusion approximation, while a modified term is added in the effect of optical thickness to backward scattering is employed. For other scattering phase functions, it is also confirmed that backwardly scattering phase function gives a lower effective thermal conductivity. Thus the materials with backward scattering properties, with radiation shields are desirable to lower the thermal conductivity of VIPs.

Measurement of VIP Edge Conduction Using Vacuum Guarded Hot Plate

Vacuum insulation panel (VIP) is a promising thermal insulator for buildings, refrigerator, LNG carrier and so on. In general, it has the thermal conductivity of 2~4 mW/m·K. However, this thermal conductivity is that measured at the center of VIP. The total effective thermal conductivity of VIP is larger than this value due to the edge conduction through the envelope. In this paper, the edge conduction of VIP is examined theoretically, numerically and experimentally. To confirm the existence of the edge conduction, numerical analysis is performed for simple two-dimensional VIP model and a theoretical model is proposed to calculate the edge conductivity. Also, the edge conductivity is measured using the vacuum guarded hot plate and the experiment is validated against numerical analysis. The results show that the edge conductivity is dependent on the width of panel and thickness of Al-foil. To reduce the edge conduction, it is recommended that the VIP should be made as big as possible or made of thin Al film envelope.

Automatic Detection and Classification of Microcalcification, Mass, Architectural Distortion and Bilateral Asymmetry in Digital Mammogram

Mammography has been one of the most reliable methods for early detection of breast cancer. There are different lesions which are breast cancer characteristic such as microcalcifications, masses, architectural distortions and bilateral asymmetry. One of the major challenges of analysing digital mammogram is how to extract efficient features from it for accurate cancer classification. In this paper we proposed a hybrid feature extraction method to detect and classify all four signs of breast cancer. The proposed method is based on multiscale surrounding region dependence method, Gabor filters, multi fractal analysis, directional and morphological analysis. The extracted features are input to self adaptive resource allocation network (SRAN) classifier for classification. The validity of our approach is extensively demonstrated using the two benchmark data sets Mammographic Image Analysis Society (MIAS) and Digital Database for Screening Mammograph (DDSM) and the results have been proved to be progressive.

DNA Polymorphism Studies of β-Lactoglobulin Gene in Saudi Goats

Domestic goats (Capra hircus) are extremely diverse species and principal animal genetic resource of the developing world. These facilitate a persistent supply of meat, milk, fibre, and skin and are considered as important revenue generators in small pastoral environments. This study aimed to fingerprint β-LG gene at PCR-RFLP level in native Saudi goat breeds (Ardi, Habsi and Harri) in an attempt to have a preliminary image of β-LG genotypic patterns in Saudi breeds as compared to other foreign breeds such as Indian and Egyptian. Also, the Phylogenetic analysis was done to investigate evolutionary trends and similarities among the caprine β-LG gene with that of the other domestic specie, viz. cow, buffalo and sheep. Blood samples were collected from 300 animals (100 for each breed) and genomic DNA was extracted. A fragment of the β-LG gene (427bp) was amplified using specific primers. Subsequent digestion with Sac II restriction endonuclease revealed two alleles (A and B) and three different banding patterns or genotypes i.e. AA, AB and BB. The statistical analysis showed a general trend that β-LG AA genotype had higher milk yield than β-LG AB and β-LG BB genotypes. Nucleotide sequencing of the selected β-LG fragments was done and submitted to GenBank NCBI (Accession No. KJ544248, KJ588275, KJ588276, KJ783455, KJ783456 and KJ874959). Phylogenetic analysis on the basis of nucleotide sequences of native Saudi goats indicated evolutional similarity with the GenBank reference sequences of goat, Bubalus bubalis and Bos taurus. However, the origin of sheep which is the most closely related from the evolutionary point of view, was located some distance away.

A Holistic Workflow Modeling Method for Business Process Redesign

In a highly competitive environment, it becomes more important to shorten the whole business process while delivering or even enhancing the business value to the customers and suppliers. Although the workflow management systems receive much attention for its capacity to practically support the business process enactment, the effective workflow modeling method remain still challenging and the high degree of process complexity makes it more difficult to gain the short lead time. This paper presents a workflow structuring method in a holistic way that can reduce the process complexity using activity-needs and formal concept analysis, which eventually enhances the key performance such as quality, delivery, and cost in business process.