Determining Earthquake Performances of Existing Reinforced Concrete Buildings by Using ANN

In this study, an Artificial Neural Network (ANN) analytical method has been developed for analyzing earthquake performances of the Reinforced Concrete (RC) buildings. 66 RC buildings with four to ten storeys were subjected to performance analysis according to the parameters which are the existing material, loading and geometrical characteristics of the buildings. The selected parameters have been thought to be effective on the performance of RC buildings. In the performance analyses stage of the study, level of performance possible to be shown by these buildings in case of an earthquake was determined on the basis of the 4-grade performance levels specified in Turkish Earthquake Code-2007 (TEC-2007). After obtaining the 4-grade performance level, selected 23 parameters of each building have been matched with the performance level. In this stage, ANN-based fast evaluation algorithm mentioned above made an economic and rapid evaluation of four to ten storey RC buildings. According to the study, the prediction accuracy of ANN has been found about 74%.

Studying Frame-Resistant Steel Structures under near Field Ground Motion

This paper presents the influence of the vertical seismic component on the non-linear dynamics analysis of three different structures. The subject structures were analyzed and designed according to recent codes. This paper considers three types of buildings: 5-, 10-, and 15-story buildings. The non-linear dynamics analysis of the structures with assuming elastic-perfectly-plastic behavior was performed using RAM PERFORM-3D software; the horizontal component was taken into consideration with and without the incorporation of the corresponding vertical component. Dynamic responses obtained for the horizontal component acting alone were compared with those obtained from the simultaneous application of both seismic components. The results show that the effect of the vertical component of ground motion may increase the axial load significantly in the interior columns and, consequently, the stories. The plastic mechanisms would be changed. The P-Delta effect is expected to increase. The punching base plate shear of the columns should be considered. Moreover, the vertical component increases the input energy when the structures exhibit inelastic behavior and are taller.

A Preliminary Study for Design of Automatic Block Reallocation Algorithm with Genetic Algorithm Method in the Land Consolidation Projects

Land reallocation is one of the most important steps in land consolidation projects. Many different models were proposed for land reallocation in the literature such as Fuzzy Logic, block priority based land reallocation and Spatial Decision Support Systems. A model including four parts is considered for automatic block reallocation with genetic algorithm method in land consolidation projects. These stages are preparing data tables for a project land, determining conditions and constraints of land reallocation, designing command steps and logical flow chart of reallocation algorithm and finally writing program codes of Genetic Algorithm respectively. In this study, we designed the first three steps of the considered model comprising four steps.

Effects of ECCS on the Cold-Leg Fluid Temperature during SGTR Accidents

The LSTF experiment simulating the SGTR accident at the Mihama Unit-2 reactor is analyzed using the RELAP5/MOD3.3 code. In the accident, and thus in the experiment, the ECC water was injected not only into the cold legs but into the upper plenum. Overall transients during the experiment such as pressures and fluid temperatures are simulated well by the code. The cold-leg fluid temperatures are shown to decrease if the upper plenum injection system is connected to the cold leg. It is found that the cold-leg fluid temperatures also decrease if the upper-plenum injection is not used and the cold-leg injection alone is actuated.

Efficient Variable Modulation Scheme Based on Codebook in the MIMO-OFDM System

Because current wireless communication requires high reliability in a limited bandwidth environment, this paper proposes the variable modulation scheme based on the codebook. The variable modulation scheme adjusts transmission power using the codebook in accordance with channel state. Also, if the codebook is composed of many bits, the reliability is more improved by the proposed scheme. The simulation results show that the performance of proposed scheme has better reliability than the the performance of conventional scheme.

Cognitive Relaying in Interference Limited Spectrum Sharing Environment: Outage Probability and Outage Capacity

In this paper, we consider a cognitive relay network (CRN) in which the primary receiver (PR) is protected by peak transmit power ¯PST and/or peak interference power Q constraints. In addition, the interference effect from the primary transmitter (PT) is considered to show its impact on the performance of the CRN. We investigate the outage probability (OP) and outage capacity (OC) of the CRN by deriving closed-form expressions over Rayleigh fading channel. Results show that both the OP and OC improve by increasing the cooperative relay nodes as well as when the PT is far away from the SR.

Reduction of Multiple User Interference for Optical CDMA Systems Using Successive Interference Cancellation Scheme

Multiple User Interference (MUI) considers the primary problem in Optical Code-Division Multiple Access (OCDMA), which resulting from the overlapping among the users. In this article we aim to mitigate this problem by studying an interference cancellation scheme called successive interference cancellation (SIC) scheme. This scheme will be tested on two different detection schemes, spectral amplitude coding (SAC) and direct detection systems (DS), using partial modified prime (PMP) as the signature codes. It was found that SIC scheme based on both SAC and DS methods had a potential to suppress the intensity noise, that is to say it can mitigate MUI noise. Furthermore, SIC/DS scheme showed much lower bit error rate (BER) performance relative to SIC/SAC scheme for different magnitude of effective power. Hence, many more users can be supported by SIC/DS receiver system.

Psychopathic Disorders and Judges Sentencing: Can Neurosciences Change This Aggravating Factor in a Mitigating Factor?

Psychopathic disorders are taking an important part in judge sentencing, especially in Canada. First, we will see how this phenomenon can be illustrated by the high proportion of psychopath offenders incarcerated in North American prisons. Many decisions in Canadians courtrooms seem to point out that psychopathy is often used as a strong argument by the judges to preserve public safety. The fact that psychopathy is often associated with violence, recklessness and recidivism, could explain why many judges consider psychopathic disorders as an aggravating factor. Generally, the judge reasoning is based on Article 753 of Canadian Criminal Code related to dangerous offenders, which is used for individuals who show a pattern of repetitive and persistent aggressive behaviour. Then we will show how, with cognitive neurosciences, the psychopath’s situation in courtrooms would probably change. Cerebral imaging and news data provided by the neurosciences show that emotional and volitional functions in psychopath’s brains are impaired. Understanding these new issues could enable some judges to recognize psychopathic disorders as a mitigating factor. Finally, two important questions ought to be raised in this article: can exploring psychopaths ‘brains really change the judge sentencing in Canadian courtrooms? If yes, can judges consider psychopathy more as a mitigating factor than an aggravating factor?

Unsupervised Classification of DNA Barcodes Species Using Multi-Library Wavelet Networks

DNA Barcode provides good sources of needed information to classify living species. The classification problem has to be supported with reliable methods and algorithms. To analyze species regions or entire genomes, it becomes necessary to use the similarity sequence methods. A large set of sequences can be simultaneously compared using Multiple Sequence Alignment which is known to be NP-complete. However, all the used methods are still computationally very expensive and require significant computational infrastructure. Our goal is to build predictive models that are highly accurate and interpretable. In fact, our method permits to avoid the complex problem of form and structure in different classes of organisms. The empirical data and their classification performances are compared with other methods. Evenly, in this study, we present our system which is consisted of three phases. The first one, is called transformation, is composed of three sub steps; Electron-Ion Interaction Pseudopotential (EIIP) for the codification of DNA Barcodes, Fourier Transform and Power Spectrum Signal Processing. Moreover, the second phase step is an approximation; it is empowered by the use of Multi Library Wavelet Neural Networks (MLWNN). Finally, the third one, is called the classification of DNA Barcodes, is realized by applying the algorithm of hierarchical classification.

Influence of Flame-Holder on Existence Important Parameters in a Duct Combustion Simulator

The effects of flame-holder position, the ratio of flame holder diameter to combustion chamber diameter and injection angle on fuel propulsive droplets sizing and effective mass fraction have been studied by a cold flow. We named the mass of fuel vapor inside the flammability limit as the effective mass fraction. An empty cylinder as well as a flame-holder which are a simulator for duct combustion has been considered. The airflow comes into the cylinder from one side and injection operation will be done by four nozzles which are located on the entrance of cylinder. To fulfill the calculations a modified version of KIVA-3V code which is a transient, three-dimensional, multiphase, multi component code for the analysis of chemically reacting flows with sprays, is used.

Strategic Risk Issues for Film Distributors of Hindi Film Industry in Mumbai: A Grounded Theory Approach

The purpose of the paper is to address the strategic risk issues surrounding Hindi film distribution in Mumbai for a film distributor, who acts as an entrepreneur when launching a product (movie) in the market (film territory).The paper undertakes a fundamental review of films and risk in the Hindi film industry and applies Grounded Theory technique to understand the complex phenomena of risk taking behavior of the film distributors (both independent and studios) in Mumbai. Rich in-depth interviews with distributors are coded to develop core categories through constant comparison leading to conceptualization of the phenomena of interest. This paper is a first-of-its-kind-attempt to understand risk behavior of a distributor, which is akin to entrepreneurial risk behavior under conditions of uncertainty.

Comparative Safety Performance Evaluation of Profiled Deck Composite Slab from the Use of Slope-Intercept and Partial Shear Methods

The economic use and ease of construction of profiled deck composite slab is marred with the complex and un-economic strength verification required for the serviceability and general safety considerations. Beside these, albeit factors such as shear span length, deck geometries and mechanical frictions greatly influence the longitudinal shear strength, that determines the ultimate strength of profiled deck composite slab, and number of methods available for its determination; partial shear and slope-intercept are the two methods according to Euro-code 4 provision. However, the complexity associated with shear behavior of profiled deck composite slab, the use of these methods in determining the load carrying capacities of such slab yields different and conflicting values. This couple with the time and cost constraint associated with the strength verification is a source of concern that draws more attentions nowadays, the issue is critical. Treating some of these known shear strength influencing factors as random variables, the load carrying capacity violation of profiled deck composite slab from the use of the two-methods defined according to Euro-code 4 are determined using reliability approach, and comparatively studied. The study reveals safety values from the use of m-k method shows good standing compared with that from the partial shear method.

Influence of Flexural Reinforcement on the Shear Strength of RC Beams without Stirrups

Numerical investigations were conducted to study the influence of flexural reinforcement ratio on the diagonal cracking strength and ultimate shear strength of reinforced concrete (RC) beams without stirrups. Three-dimensional nonlinear finite element analyses (FEAs) of the beams with flexural reinforcement ratios ranging from 0.58% to 2.20% subjected to a mid-span concentrated load were carried out. It is observed that the load-deflection and loadstrain curves obtained from the numerical analyses agree with those obtained from the experiments. It is concluded that flexural reinforcement ratio has a significant effect on the shear strength and deflection capacity of RC beams without stirrups. The predictions of diagonal cracking strength and ultimate shear strength of beams obtained by using the equations defined by a number of codes and researchers are compared with each other and with the experimental values.

New Efficient Method for Coding Color Images

In this paper a novel color image compression technique for efficient storage and delivery of data is proposed. The proposed compression technique started by RGB to YCbCr color transformation process. Secondly, the canny edge detection method is used to classify the blocks into the edge and non-edge blocks. Each color component Y, Cb, and Cr compressed by discrete cosine transform (DCT) process, quantizing and coding step by step using adaptive arithmetic coding. Our technique is concerned with the compression ratio, bits per pixel and peak signal to noise ratio, and produce better results than JPEG and more recent published schemes (like CBDCT-CABS and MHC). The provided experimental results illustrate the proposed technique that is efficient and feasible in terms of compression ratio, bits per pixel and peak signal to noise ratio.

Improved Wi-Fi Backscatter System for Multi-to-Multi Communication

The conventional Wi-Fi backscatter system can only process one-to-one communication between the Wi-Fi reader and the Wi-Fi tag. For improvement of throughput of the conventional system, this paper proposes the multi-to-multi communication system. In the proposed system, the interference by the multi-to-multi communication is effectively cancelled by the orthogonal multiple access based on the identification code of the tag. Although the overhead is generated by the procedure for the multi-to-multi communication, because the procedure is processed by the Wi-Fi protocol, the overhead is insignificant for the entire communication procedure. From the numerical results, it is confirmed that the proposed system has nearly proportional increased throughput in according to the number of the tag that simultaneously participates in communication.

Learners’ Perceptions of Tertiary Level Teachers’ Code Switching: A Vietnamese Perspective

The literature on language teaching and second language acquisition has been largely driven by monolingual ideology with a common assumption that a second language (L2) is best taught and learned in the L2 only. The current study challenges this assumption by reporting learners' positive perceptions of tertiary level teachers' code switching practices in Vietnam. The findings of this study contribute to our understanding of code switching practices in language classrooms from a learners' perspective. Data were collected from student participants who were working towards a Bachelor degree in English within the English for Business Communication stream through the use of focus group interviews. The literature has documented that this method of interviewing has a number of distinct advantages over individual student interviews. For instance, group interactions generated by focus groups create a more natural environment than that of an individual interview because they include a range of communicative processes in which each individual may influence or be influenced by others - as they are in their real life. The process of interaction provides the opportunity to obtain the meanings and answers to a problem that are "socially constructed rather than individually created" leading to the capture of real-life data. The distinct feature of group interaction offered by this technique makes it a powerful means of obtaining deeper and richer data than those from individual interviews. The data generated through this study were analysed using a constant comparative approach. Overall, the students expressed positive views of this practice indicating that it is a useful teaching strategy. Teacher code switching was seen as a learning resource and a source supporting language output. This practice was perceived to promote student comprehension and to aid the learning of content and target language knowledge. This practice was also believed to scaffold the students' language production in different contexts. However, the students indicated their preference for teacher code switching to be constrained, as extensive use was believed to negatively impact on their L2 learning and trigger cognitive reliance on the L1 for L2 learning. The students also perceived that when the L1 was used to a great extent, their ability to develop as autonomous learners was negatively impacted. This study found that teacher code switching was supported in certain contexts by learners, thus suggesting that there is a need for the widespread assumption about the monolingual teaching approach to be re-considered.

Evaluation of Static Modulus of Elasticity Depending on Concrete Compressive Strength

The paper is focused on monitoring of dependencies of different composition concretes on elastic modulus values. To obtain a summary of elastic modulus development in dependence of concrete composition design variability was the objective of the experiment. Essential part of this work was initiated as a reaction to building practice when questions of elastic moduli arose at the same time and which mostly did not obtain the required and expected values from concrete constructions.

Combined Source and Channel Coding for Image Transmission Using Enhanced Turbo Codes in AWGN and Rayleigh Channel

Any signal transmitted over a channel is corrupted by noise and interference. A host of channel coding techniques has been proposed to alleviate the effect of such noise and interference. Among these Turbo codes are recommended, because of increased capacity at higher transmission rates and superior performance over convolutional codes. The multimedia elements which are associated with ample amount of data are best protected by Turbo codes. Turbo decoder employs Maximum A-posteriori Probability (MAP) and Soft Output Viterbi Decoding (SOVA) algorithms. Conventional Turbo coded systems employ Equal Error Protection (EEP) in which the protection of all the data in an information message is uniform. Some applications involve Unequal Error Protection (UEP) in which the level of protection is higher for important information bits than that of other bits. In this work, enhancement to the traditional Log MAP decoding algorithm is being done by using optimized scaling factors for both the decoders. The error correcting performance in presence of UEP in Additive White Gaussian Noise channel (AWGN) and Rayleigh fading are analyzed for the transmission of image with Discrete Cosine Transform (DCT) as source coding technique. This paper compares the performance of log MAP, Modified log MAP (MlogMAP) and Enhanced log MAP (ElogMAP) algorithms used for image transmission. The MlogMAP algorithm is found to be best for lower Eb/N0 values but for higher Eb/N0 ElogMAP performs better with optimized scaling factors. The performance comparison of AWGN with fading channel indicates the robustness of the proposed algorithm. According to the performance of three different message classes, class3 would be more protected than other two classes. From the performance analysis, it is observed that ElogMAP algorithm with UEP is best for transmission of an image compared to Log MAP and MlogMAP decoding algorithms.

Numerical Simulation of Free Surface Water Wave for the Flow around NACA 0012 Hydrofoil and Wigley Hull Using VOF Method

Steady three-dimensional and two free surface waves generated by moving bodies are presented, the flow problem to be simulated is rich in complexity and poses many modeling challenges because of the existence of breaking waves around the ship hull, and because of the interaction of the two-phase flow with the turbulent boundary layer. The results of several simulations are reported. The first study was performed for NACA0012 of hydrofoil with different meshes, this section is analyzed at h/c= 1, 0345 for 2D. In the second simulation a mathematically defined Wigley hull form is used to investigate the application of a commercial CFD code in prediction of the total resistance and its components from tangential and normal forces on the hull wetted surface. The computed resistance and wave profiles are used to estimate the coefficient of the total resistance for Wigley hull advancing in calm water under steady conditions. The commercial CFD software FLUENT version 12 is used for the computations in the present study. The calculated grid is established using the code computer GAMBIT 2.3.26. The shear stress k-ωSST model is used for turbulence modeling and the volume of fluid technique is employed to simulate the free-surface motion. The second order upwind scheme is used for discretizing the convection terms in the momentum transport equations, the Modified HRIC scheme for VOF discretization. The results obtained compare well with the experimental data.

Forecast Based on an Empirical Probability Function with an Adjusted Error Using Propagation of Error

This paper addresses a cutting edge method of business demand forecasting, based on an empirical probability function when the historical behavior of the data is random. Additionally, it presents error determination based on the numerical method technique ‘propagation of errors.’ The methodology was conducted characterization and process diagnostics demand planning as part of the production management, then new ways to predict its value through techniques of probability and to calculate their mistake investigated, it was tools used numerical methods. All this based on the behavior of the data. This analysis was determined considering the specific business circumstances of a company in the sector of communications, located in the city of Bogota, Colombia. In conclusion, using this application it was possible to obtain the adequate stock of the products required by the company to provide its services, helping the company reduce its service time, increase the client satisfaction rate, reduce stock which has not been in rotation for a long time, code its inventory, and plan reorder points for the replenishment of stock.