Influence of Non-Structural Elements on Dynamic Response of Multi-Storey Rc Building to Mining Shock

In the paper the results of calculations of the dynamic response of a multi-storey reinforced concrete building to a strong mining shock originated from the main region of mining activity in Poland (i.e. the Legnica-Glogow Copper District) are presented. The representative time histories of accelerations registered in three directions were used as ground motion data in calculations of the dynamic response of the structure. Two variants of a numerical model were applied: the model including only structural elements of the building and the model including both structural and non-structural elements (i.e. partition walls and ventilation ducts made of brick). It turned out that non-structural elements of multi-storey RC buildings have a small impact of about 10 % on natural frequencies of these structures. It was also proved that the dynamic response of building to mining shock obtained in case of inclusion of all non-structural elements in the numerical model is about 20 % smaller than in case of consideration of structural elements only. The principal stresses obtained in calculations of dynamic response of multi-storey building to strong mining shock are situated on the level of about 30% of values obtained from static analysis (dead load).

Values as a Predictor of Cyber-bullying Among Secondary School Students

The use of new technologies such internet (e-mail, chat rooms) and cell phones has steeply increased in recent years. Especially among children and young people, use of technological tools and equipments is widespread. Although many teachers and administrators now recognize the problem of school bullying, few are aware that students are being harassed through electronic communication. Referred to as electronic bullying, cyber bullying, or online social cruelty, this phenomenon includes bullying through email, instant messaging, in a chat room, on a website, or through digital messages or images sent to a cell phone. Cyber bullying is defined as causing deliberate/intentional harm to others using internet or other digital technologies. It has a quantitative research design nd uses relational survey as its method. The participants consisted of 300 secondary school students in the city of Konya, Turkey. 195 (64.8%) participants were female and 105 (35.2%) were male. 39 (13%) students were at grade 1, 187 (62.1%) were at grade 2 and 74 (24.6%) were at grade 3. The “Cyber Bullying Question List" developed by Ar─▒cak (2009) was given to students. Following questions about demographics, a functional definition of cyber bullying was provided. In order to specify students- human values, “Human Values Scale (HVS)" developed by Dilmaç (2007) for secondary school students was administered. The scale consists of 42 items in six dimensions. Data analysis was conducted by the primary investigator of the study using SPSS 14.00 statistical analysis software. Descriptive statistics were calculated for the analysis of students- cyber bullying behaviour and simple regression analysis was conducted in order to test whether each value in the scale could explain cyber bullying behaviour.

Problem Solving Techniques with Extensive Computational Network and Applying in an Educational Software

Knowledge bases are basic components of expert systems or intelligent computational programs. Knowledge bases provide knowledge, events that serve deduction activity, computation and control. Therefore, researching and developing of models for knowledge representation play an important role in computer science, especially in Artificial Intelligence Science and intelligent educational software. In this paper, the extensive deduction computational model is proposed to design knowledge bases whose attributes are able to be real values or functional values. The system can also solve problems based on knowledge bases. Moreover, the models and algorithms are applied to produce the educational software for solving alternating current problems or solving set of equations automatically.

Bioleaching of Spent Catalyst using Moderate Thermophiles with Different Pulp Densities and Varying Size Fractions without Fe Supplemented Growth Medium

Bioleaching of spent catalyst using moderate thermophilic chemolithotrophic acidophiles in growth medium without Fe source was investigated with two different pulp densities and three different size fractions. All the experiments were conducted on shake flasks at a temperature of 65 °C. The leaching yield of Ni and Al was found to be promising with very high leaching yield of 92-96% followed by Al as 41-76%, which means both Ni and Al leaching were favored by the moderate thermophilic bioleaching compared to the mesophilic bioleaching. The acid consumption was comparatively higher for the 10% pulp density experiments. Comparatively minimal difference in the leaching yield with different size fractions and different pulp densities show no requirement of grinding and using low pulp density less than 10%. This process would rather be economical as well as eco-friendly process for future optimization of the recovery of metal values from spent catalyst.

Physical and Mechanical Properties of Particleboard from Bamboo Waste

This research was to evaluate a technical feasibility of making single-layer experimental particleboard panels from bamboo waste (Dendrocalamus asper Backer) by converting bamboo into strips, which are used to make laminated bamboo furniture. Variable factors were density (600, 700 and 800 kg/m3) and temperature of condition (25, 40 and 55 °C). The experimental panels were tested for their physical and mechanical properties including modulus of elasticity (MOE), modulus of rupture (MOR), internal bonding strength (IB), screw holding strength (SH) and thickness swelling values according to the procedures defined by Japanese Industrial Standard (JIS). The test result of mechanical properties showed that the MOR, MOE and IB values were not in the set criteria, except the MOR values at the density of 700 kg/m3 at 25 °C and at the density of 800 kg/m3 at 25 and 40 °C, the IB values at the density of 600 kg/m3, at 40 °C, and at the density of 800 kg/m3 at 55 °C. The SH values had the test result according to the set standard, except with the density of 600 kg/m3, at 40 and 55 °C. Conclusively, a valuable renewable biomass, bamboo waste could be used to manufacture boards.

The Effect of Different Nozzle Configurations on Airflow Behaviour and Yarn Quality

Nozzle is the main part of various spinning systems such as air-jet and Murata air vortex systems. Recently, many researchers worked on the usage of the nozzle on different spinning systems such as conventional ring and compact spinning systems. In these applications, primary purpose is to improve the yarn quality. In present study, it was produced the yarns with two different nozzle types and determined the changes in yarn properties. In order to explain the effect of the nozzle, airflow structure in the nozzle was modelled and airflow variables were determined. In numerical simulation, ANSYS 12.1 package program and Fluid Flow (CFX) analysis method was used. As distinct from the literature, Shear Stress Turbulent (SST) model is preferred. And also air pressure at the nozzle inlet was measured by electronic mass flow meter and these values were used for the simulation of the airflow. At last, the yarn was modelled and the area from where the yarn is passing was included to the numerical analysis.

Low Air Velocity Measurement Characteristics- Variation Due to Flow Regime

The paper depicts air velocity values, reproduced by laser Doppler anemometer (LDA) and ultrasonic anemometer (UA), relations with calculated ones from flow rate measurements using the gas meter which calibration uncertainty is ± (0.15 – 0.30) %. Investigation had been performed in channel installed in aerodynamical facility used as a part of national standard of air velocity. Relations defined in a research let us confirm the LDA and UA for air velocity reproduction to be the most advantageous measures. The results affirm ultrasonic anemometer to be reliable and favourable instrument for measurement of mean velocity or control of velocity stability in the velocity range of 0.05 m/s – 10 (15) m/s when the LDA used. The main aim of this research is to investigate low velocity regularities, starting from 0.05 m/s, including region of turbulent, laminar and transitional air flows. Theoretical and experimental results and brief analysis of it are given in the paper. Maximum and mean velocity relations for transitional air flow having unique distribution are represented. Transitional flow having distinctive and different from laminar and turbulent flow characteristics experimentally have not yet been analysed.

A Renovated Cook's Distance Based On The Buckley-James Estimate In Censored Regression

There have been various methods created based on the regression ideas to resolve the problem of data set containing censored observations, i.e. the Buckley-James method, Miller-s method, Cox method, and Koul-Susarla-Van Ryzin estimators. Even though comparison studies show the Buckley-James method performs better than some other methods, it is still rarely used by researchers mainly because of the limited diagnostics analysis developed for the Buckley-James method thus far. Therefore, a diagnostic tool for the Buckley-James method is proposed in this paper. It is called the renovated Cook-s Distance, (RD* i ) and has been developed based on the Cook-s idea. The renovated Cook-s Distance (RD* i ) has advantages (depending on the analyst demand) over (i) the change in the fitted value for a single case, DFIT* i as it measures the influence of case i on all n fitted values Yˆ∗ (not just the fitted value for case i as DFIT* i) (ii) the change in the estimate of the coefficient when the ith case is deleted, DBETA* i since DBETA* i corresponds to the number of variables p so it is usually easier to look at a diagnostic measure such as RD* i since information from p variables can be considered simultaneously. Finally, an example using Stanford Heart Transplant data is provided to illustrate the proposed diagnostic tool.

Thermo-Sensitive Hydrogel: Control of Hydrophilic-Hydrophobic Transition

The study investigated the hydrophilic to hydrophobic transition of modified polyacrylamide hydrogel with the inclusion of N-isopropylacrylamide (NIAM). The modification was done by mimicking micellar polymerization, which resulted in better arrangement of NIAM chains in the polyacrylamide network. The degree of NIAM arrangement is described by NH number. The hydrophilic to hydrophobic transition was measured through the partition coefficient, K, of Orange II and Methylene Blue in hydrogel and in water. These dyes were chosen as a model for solutes with different degree of hydrophobicity. The study showed that the hydrogel with higher NH values resulted in better solubility of both dyes. Moreover, in temperature above the lower critical solution temperature (LCST) of Poly(N-isopropylacrylamide) (PNIAM)also caused the collapse of NIPAM chains which results in a more hydrophobic environment that increases the solubility of Methylene Blue and decreases the solubility of Orange II in the hydrogels with NIPAM present.

Geochemistry of Tektites from Maoming of Guandong Province, China

We measured the major and trace element contents and Rb-Sr isotopic compositions of 12 tektites from the Maoming area, Guandong province (south China). All the samples studied are splash-form tektites which show pitted or grooved surfaces with schlieren structures on some surfaces. The trace element ratios Ba/Rb (avg. 4.33), Th/Sm (avg. 2.31), Sm/Sc (avg. 0.44), Th/Sc (avg. 1.01) , La/Sc (avg. 2.86), Th/U (avg. 7.47), Zr/Hf (avg. 46.01) and the rare earth elements (REE) contents of tektites of this study are similar to the average upper continental crust. From the chemical composition, it is suggested that tektites in this study are derived from similar parental terrestrial sedimentary deposit which may be related to post-Archean upper crustal rocks. The tektites from the Maoming area have high positive εSr(0) values-ranging from 176.9~190.5 which indicate that the parental material for these tektites have similar Sr isotopic compositions to old terrestrial sedimentary rocks and they were not dominantly derived from recent young sediments (such as soil or loess). The Sr isotopic data obtained by the present study support the conclusion proposed by Blum et al. (1992)[1] that the depositional age of sedimentary target materials is close to 170Ma (Jurassic). Mixing calculations based on the model proposed by Ho and Chen (1996)[2] for various amounts and combinations of target rocks indicate that the best fit for tektites from the Maoming area is a mixture of 40% shale, 30% greywacke, 30% quartzite.

I-Vague Normal Groups

The notions of I-vague normal groups with membership and non-membership functions taking values in an involutary dually residuated lattice ordered semigroup are introduced which generalize the notions with truth values in a Boolean algebra as well as those usual vague sets whose membership and non-membership functions taking values in the unit interval [0, 1]. Various operations and properties are established.

Trimmed Mean as an Adaptive Robust Estimator of a Location Parameter for Weibull Distribution

One of the purposes of the robust method of estimation is to reduce the influence of outliers in the data, on the estimates. The outliers arise from gross errors or contamination from distributions with long tails. The trimmed mean is a robust estimate. This means that it is not sensitive to violation of distributional assumptions of the data. It is called an adaptive estimate when the trimming proportion is determined from the data rather than being fixed a “priori-. The main objective of this study is to find out the robustness properties of the adaptive trimmed means in terms of efficiency, high breakdown point and influence function. Specifically, it seeks to find out the magnitude of the trimming proportion of the adaptive trimmed mean which will yield efficient and robust estimates of the parameter for data which follow a modified Weibull distribution with parameter λ = 1/2 , where the trimming proportion is determined by a ratio of two trimmed means defined as the tail length. Secondly, the asymptotic properties of the tail length and the trimmed means are also investigated. Finally, a comparison is made on the efficiency of the adaptive trimmed means in terms of the standard deviation for the trimming proportions and when these were fixed a “priori". The asymptotic tail lengths defined as the ratio of two trimmed means and the asymptotic variances were computed by using the formulas derived. While the values of the standard deviations for the derived tail lengths for data of size 40 simulated from a Weibull distribution were computed for 100 iterations using a computer program written in Pascal language. The findings of the study revealed that the tail lengths of the Weibull distribution increase in magnitudes as the trimming proportions increase, the measure of the tail length and the adaptive trimmed mean are asymptotically independent as the number of observations n becomes very large or approaching infinity, the tail length is asymptotically distributed as the ratio of two independent normal random variables, and the asymptotic variances decrease as the trimming proportions increase. The simulation study revealed empirically that the standard error of the adaptive trimmed mean using the ratio of tail lengths is relatively smaller for different values of trimming proportions than its counterpart when the trimming proportions were fixed a 'priori'.

A File Splitting Technique for Reducing the Entropy of Text Files

A novel file splitting technique for the reduction of the nth-order entropy of text files is proposed. The technique is based on mapping the original text file into a non-ASCII binary file using a new codeword assignment method and then the resulting binary file is split into several subfiles each contains one or more bits from each codeword of the mapped binary file. The statistical properties of the subfiles are studied and it is found that they reflect the statistical properties of the original text file which is not the case when the ASCII code is used as a mapper. The nth-order entropy of these subfiles are determined and it is found that the sum of their entropies is less than that of the original text file for the same values of extensions. These interesting statistical properties of the resulting subfiles can be used to achieve better compression ratios when conventional compression techniques are applied to these subfiles individually and on a bit-wise basis rather than on character-wise basis.

Determination of Q and R Matrices for Optimal Pitch Aircraft Control

In this paper, the process of obtaining Q and R matrices for optimal pitch aircraft control system has been described. Since the innovation of optimal control method, the determination of Q and R matrices for such system has not been fully specified. The value of Q and R for optimal pitch aircraft control application, have been simulated and calculated. The suitable results for Q and R have been observed through the performance index (PI). If the PI is small “enough", we would say the Q & R values are suitable for that certain type of optimal control system. Moreover, for the same value of PI, we could have different Q and R sets. Due to the rule-free determination of Q and R matrices, a specific method is brought to find out the rough value of Q and R referring to rather small value of PI.

A New Performance Characterization of Transient Analysis Method

This paper proposes a new performance characterization for the test strategy intended for second order filters denominated Transient Analysis Method (TRAM). We evaluate the ability of the addressed test strategy for detecting deviation faults under simultaneous statistical fluctuation of the non-faulty parameters. For this purpose, we use Monte Carlo simulations and a fault model that considers as faulty only one component of the filter under test while the others components adopt random values (within their tolerance band) obtained from their statistical distributions. The new data reported here show (for the filters under study) the presence of hard-to-test components and relatively low fault coverage values for small deviation faults. These results suggest that the fault coverage value obtained using only nominal values for the non-faulty components (the traditional evaluation of TRAM) seem to be a poor predictor of the test performance.

The Removal of As(V) from Drinking Waters by Coagulation Process using Iron Salts

In this study arsenate [As(V)] removal from drinking water by coagulation process was investigated. Ferric chloride (FeCl3.6H2O) and ferrous sulfate (FeSO4.7H2O) were used as coagulant. The effects of major operating variables such as coagulant dose (1–30 mg/L) and pH (5.5–9.5) were investigated. Ferric chloride and ferrous sulfate were found as effective and reliable coagulant due to required dose, residual arsenate and coagulant concentration. Optimum pH values for maximum arsenate removal for ferrous sulfate and ferric chloride were found as 8 and 7.5. The arsenate removal efficiency decreased at neutral and acidic pH values for Fe(II) and at the high acidic and high alkaline pH for Fe(III). It was found that the increase of coagulant dose caused a substantial increase in the arsenate removal. But above a certain ferric chloride and ferrous sulfate dosage, the increase in arsenate removal was not significant. Ferric chloride and ferrous sulfate dose above 8 mg/L slightly increased arsenate removal.

Waste Lubricating Oil Treatment by Adsorption Process Using Different Adsorbents

Waste lubricating oil re-refining adsorption process by different adsorbent materials was investigated. Adsorbent materials such as oil adsorbent, egg shale powder, date palm kernel powder, and acid activated date palm kernel powder were used. The adsorption process over fixed amount of adsorbent at ambient conditions was investigated. The adsorption/extraction process was able to deposit the asphaltenic and metallic contaminants from the waste oil to lower values. It was found that the date palm kernel powder with contact time of 4 h was able to give the best conditions for treating the waste oil. The recovered solvent could be also reused. It was also found that the activated bentonite gave the best physical properties followed by the date palm kernel powder.

Using a Trust-Based Environment Key for Mobile Agent Code Protection

Human activities are increasingly based on the use of remote resources and services, and on the interaction between remotely located parties that may know little about each other. Mobile agents must be prepared to execute on different hosts with various environmental security conditions. The aim of this paper is to propose a trust based mechanism to improve the security of mobile agents and allow their execution in various environments. Thus, an adaptive trust mechanism is proposed. It is based on the dynamic interaction between the agent and the environment. Information collected during the interaction enables generation of an environment key. This key informs on the host-s trust degree and permits the mobile agent to adapt its execution. Trust estimation is based on concrete parameters values. Thus, in case of distrust, the source of problem can be located and a mobile agent appropriate behavior can be selected.

Dynamic Traffic Simulation for Traffic Congestion Problem Using an Enhanced Algorithm

Traffic congestion has become a major problem in many countries. One of the main causes of traffic congestion is due to road merges. Vehicles tend to move slower when they reach the merging point. In this paper, an enhanced algorithm for traffic simulation based on the fluid-dynamic algorithm and kinematic wave theory is proposed. The enhanced algorithm is used to study traffic congestion at a road merge. This paper also describes the development of a dynamic traffic simulation tool which is used as a scenario planning and to forecast traffic congestion level in a certain time based on defined parameter values. The tool incorporates the enhanced algorithm as well as the two original algorithms. Output from the three above mentioned algorithms are measured in terms of traffic queue length, travel time and the total number of vehicles passing through the merging point. This paper also suggests an efficient way of reducing traffic congestion at a road merge by analyzing the traffic queue length and travel time.

Experimental and Statistical Study of Nonlinear Effect of Carbon Nanotube on Mechanical Properties of Polypropylene Composites

In this study concept of experimental design is successfully applied for the determination of optimum condition to produce PP/SWCNT (Polypropylene/Single wall carbon nanotube) nanocomposite. Central composite design as one of experimental design techniques is employed for the optimization and statistical determination of the significant factors influencing on the tensile modulus and yield stress as mechanical properties of this nanocomposite. The significant factors are SWCNT weight fraction and acid treatment time for functionalizing the nanoparticles. Optimum conditions are in 0.7 % of SWCNT weight fraction and 210 min as acid treatment time for 1112.75 ± 28 MPa as maximum tensile modulus and in 216 min and 0.65 % as acid treatment time and SWCNT weight fraction respectively for 40.26 ± 0.3 MPa as maximum yield stress. Also after setting new experiments for test these optimum conditions, found excelent agreement with predicted values.