Experimental Inspection of Damage and Performance Evaluation after Repair and Strengthening of Jiamusi Highway Prestressed Concrete Bridge in China

The main objectives of this study are to inspect and identify any damage of jaimusi highway prestressed concrete bridge after repair and strengthening of damaged structural members and to evaluate the performance of the bridge structural members by adopting static load test. Inspection program after repair and strengthening includes identifying and evaluating the structural members of bridge such as T-shape cantilever structure, hanging beams, corbels, external tendons, anchor beams, sticking steel plate, and piers. The results of inspection show that the overall state of the bridge structural member after repair and strengthening is good. The results of rebound test of concrete strength show that the average strength of concrete is 46.31Mpa. Whereas, the average value of concrete strength of anchor beam is 49.82Mpa. According to the results of static load test, the experimental values are less than theoretical values of internal forces, deflection, and strain, indicating that the stiffness of the experimental structure, overall deformation and integrity satisfy the designed standard and the working performance is good, and the undertaking capacity has a certain surplus. There is not visible change in the length and width of cracks and there are not new cracks under experimental load.

Investigating the Fatigue Crack Initiation Location in Interference Fitted and/or bolt Clamped Al 2024-T3 Double Shear Lap Joints

In this paper the fatigue crack initiation location of double shear lap joints, treated by interference fit and bolt clamping, have been investigated both experimentally and numerically. To do so, using the fracture section of available fatigue tested specimens of interference fitted and torque tightened Aluminum 2024-T3 plates, the crack initiation location was determined. The stress distribution attained from the finite element analysis was used to help explain the results observed in the experimental tests. The results showed that the fatigue crack initiation location changes from top and mid plane at the hole edge to somewhere far from the hole edge (stress concentration region) in different combination of clamping force, interference fit size and applied cyclic load ranges. It is worth mentioning that the fatigue crack initiation location affects the fatigue life of the specimens too.

Effects of Adding Fibre on Strength and Permeability of Recycled Aggregate Concrete Containing Treated Coarse RCA

This paper presents the experiment results of investigating the effects of adding various types and proportions of fibre on mechanical strength and permeability characteristics of recycled aggregate concrete (RAC), which was produced with treated coarse recycled concrete aggregate (RCA). Two types of synthetic fibres (i.e., barchip and polypropylene fibre) with various volume fractions were added to the RAC, which was calculated by the weight of the cement. The hardened RAC properties such as compressive strength, flexural strength, ultrasonic pulse velocity, water absorption and total porosity at the curing ages of 7 and 28 days were evaluated and compared with the properties of the control specimens. Results indicate that the treated coarse RCA enhances the mechanical strength and permeability properties of RAC and adding barchip fibre further optimises the results. Adding 1.2% barchip fibre has the best effect on the mechanical strength performance of the RAC.

Elastic Strain-Concentration Factor of Cylindrical Bars with Circumferential Flat-Bottom Groove under Static Tension

Using finite element method (FEM), the elastic new strain-concentration factor (SNCF) of cylindrical bars with circumferential flat-bottom groove is studied. This new SNCF has been defined under triaxial stress state. The employed specimens have constant groove depth with net section and gross diameters of 10.0 and 16.7 mm, respectively. The length of flatness ao has been varied form 0.0 ~12.5 mm to study the elastic SNCF of this type of geometrical irregularities. The results that the elastic new SNCF rapidly drops from its elastic value of the groove with ao = 0.0, i.e. circumferential U-notch, and reaches minimum value at ao = 2 mm. After that the elastic new SNCF becomes nearly constant with increasing flatness length (ao). The value of tensile load at yielding at the groove root increases with increasing ao. The current results show that severity of the notch decreases with increasing flatness length ao.

Using Reverse Osmosis Membrane for Chromium Removal from Aqueous Solution

In this paper, removal of chromium(VI) from aqueous solution has been researched using reverse osmosis. The influence of transmembrane pressure and feed concentration on permeate flux, water recovery, permeate concentration, and salt rejection was studied. The results showed that according to the variation of transmembrane pressure and feed concentration, the permeate flux and salt rejection were in the range 19.17 to 58.75 l/m2.min and 99.51 to 99.8 %, respectively. The highest permeate flux, 58.75 l/m2.min, and water recovery, 42.47 %, were obtained in the highest pressure and the lowest feed concentration. On the other hand, the lowest permeate concentration, 0.01 mg/l, and the highest salt rejection, 99.8 %, were obtained in the highest pressure and the lowest feed concentration.

Information Support for Emergency Staff Processes and Effective Decisions

Managing the emergency situations at the Emergency Staff requires a high co-operation between its members and their fast decision making. For these purpose it is necessary to prepare Emergency Staff members adequately. The aim of this paper is to describe the development of information support that focuses to emergency staff processes and effective decisions. The information support is based on the principles of process management, and Process Framework for Emergency Management was used during the development. The output is the information system that allows users to simulate an emergency situation, including effective decision making. The system also evaluates the progress of the emergency processes solving by quantitative and qualitative indicators. By using the simulator, a higher quality education of specialists can be achieved. Therefore, negative impacts resulting from arising emergency situations can be directly reduced.

Financial Regulations in the Process of Global Financial Crisis and Macroeconomics Impact of Basel III

Basel III (or the Third Basel Accord) is a global regulatory standard on bank capital adequacy, stress testing and market liquidity risk agreed upon by the members of the Basel Committee on Banking Supervision in 2010-2011, and scheduled to be introduced from 2013 until 2018. Basel III is a comprehensive set of reform measures. These measures aim to; (1) improve the banking sector-s ability to absorb shocks arising from financial and economic stress, whatever the source, (2) improve risk management and governance, (3) strengthen banks- transparency and disclosures. Similarly the reform target; (1) bank level or micro-prudential, regulation, which will help raise the resilience of individual banking institutions to periods of stress. (2) Macro-prudential regulations, system wide risk that can build up across the banking sector as well as the pro-cyclical implication of these risks over time. These two approaches to supervision are complementary as greater resilience at the individual bank level reduces the risk system wide shocks. Macroeconomic impact of Basel III; OECD estimates that the medium-term impact of Basel III implementation on GDP growth is in the range -0,05 percent to -0,15 percent per year. On the other hand economic output is mainly affected by an increase in bank lending spreads as banks pass a rise in banking funding costs, due to higher capital requirements, to their customers. Consequently the estimated effects on GDP growth assume no active response from monetary policy. Basel III impact on economic output could be offset by a reduction (or delayed increase) in monetary policy rates by about 30 to 80 basis points. The aim of this paper is to create a framework based on the recent regulations in order to prevent financial crises. Thus the need to overcome the global financial crisis will contribute to financial crises that may occur in the future periods. In the first part of the paper, the effects of the global crisis on the banking system examine the concept of financial regulations. In the second part; especially in the financial regulations and Basel III are analyzed. The last section in this paper explored the possible consequences of the macroeconomic impacts of Basel III.

Generalization of SGIP Surface Tension Force Model in Three-Dimensional Flows and Compare to Other Models in Interfacial Flows

In this paper, the two-dimensional stagger grid interface pressure (SGIP) model has been generalized and presented into three-dimensional form. For this purpose, various models of surface tension force for interfacial flows have been investigated and compared with each other. The VOF method has been used for tracking the interface. To show the ability of the SGIP model for three-dimensional flows in comparison with other models, pressure contours, maximum spurious velocities, norm spurious flow velocities and pressure jump error for motionless drop of liquid and bubble of gas are calculated using different models. It has been pointed out that SGIP model in comparison with the CSF, CSS and PCIL models produces the least maximum and norm spurious velocities. Additionally, the new model produces more accurate results in calculating the pressure jumps across the interface for motionless drop of liquid and bubble of gas which is generated in surface tension force.

Combating Money Laundering in the Banking Industry: Malaysian Experience

Money laundering has been described by many as the lifeblood of crime and is a major threat to the economic and social well-being of societies. It has been recognized that the banking system has long been the central element of money laundering. This is in part due to the complexity and confidentiality of the banking system itself. It is generally accepted that effective anti-money laundering (AML) measures adopted by banks will make it tougher for criminals to get their "dirty money" into the financial system. In fact, for law enforcement agencies, banks are considered to be an important source of valuable information for the detection of money laundering. However, from the banks- perspective, the main reason for their existence is to make as much profits as possible. Hence their cultural and commercial interests are totally distinct from that of the law enforcement authorities. Undoubtedly, AML laws create a major dilemma for banks as they produce a significant shift in the way banks interact with their customers. Furthermore, the implementation of the laws not only creates significant compliance problems for banks, but also has the potential to adversely affect the operations of banks. As such, it is legitimate to ask whether these laws are effective in preventing money launderers from using banks, or whether they simply put an unreasonable burden on banks and their customers. This paper attempts to address these issues and analyze them against the background of the Malaysian AML laws. It must be said that effective coordination between AML regulator and the banking industry is vital to minimize problems faced by the banks and thereby to ensure effective implementation of the laws in combating money laundering.

Principal Component Analysis using Singular Value Decomposition of Microarray Data

A series of microarray experiments produces observations of differential expression for thousands of genes across multiple conditions. Principal component analysis(PCA) has been widely used in multivariate data analysis to reduce the dimensionality of the data in order to simplify subsequent analysis and allow for summarization of the data in a parsimonious manner. PCA, which can be implemented via a singular value decomposition(SVD), is useful for analysis of microarray data. For application of PCA using SVD we use the DNA microarray data for the small round blue cell tumors(SRBCT) of childhood by Khan et al.(2001). To decide the number of components which account for sufficient amount of information we draw scree plot. Biplot, a graphic display associated with PCA, reveals important features that exhibit relationship between variables and also the relationship of variables with observations.

Raman Scattering and PL Studies on AlGaN/GaN HEMT Layers on 200 mm Si(111)

The crystalline quality of the AlGaN/GaN high electron mobility transistor (HEMT) structure grown on a 200 mm silicon substrate has been investigated using UV-visible micro- Raman scattering and photoluminescence (PL). The visible Raman scattering probes the whole nitride stack with the Si substrate and shows the presence of a small component of residual in-plane stress in the thick GaN buffer resulting from a wafer bowing, while the UV micro-Raman indicates a tensile interfacial stress induced at the top GaN/AlGaN/AlN layers. PL shows a good crystal quality GaN channel where the yellow band intensity is very low compared to that of the near-band-edge transition. The uniformity of this sample is shown by measurements from several points across the epiwafer.

Performance Evaluation Standards and Innovation: An Empirical Investigation

In this empirical research, how marketing managers evaluate their firms- performances and decide to make innovation is examined. They use some standards which are past performance of the firm, target performance of the firm, competitor performance, and average performance of the industry to compare and evaluate the firms- performances. It is hypothesized that marketing managers and owners of the firm compare the firms- current performance with these four standards at the same time to decide when to make innovation relating to any aspects of the firm, either management style or products. Relationship between the comparison of the firm-s performance with these standards and innovation are searched in the same regression model. The results of the regression analysis are discussed and some recommendations are made for future studies and applicants.

Performance Evaluation of Neural Network Prediction for Data Prefetching in Embedded Applications

Embedded systems need to respect stringent real time constraints. Various hardware components included in such systems such as cache memories exhibit variability and therefore affect execution time. Indeed, a cache memory access from an embedded microprocessor might result in a cache hit where the data is available or a cache miss and the data need to be fetched with an additional delay from an external memory. It is therefore highly desirable to predict future memory accesses during execution in order to appropriately prefetch data without incurring delays. In this paper, we evaluate the potential of several artificial neural networks for the prediction of instruction memory addresses. Neural network have the potential to tackle the nonlinear behavior observed in memory accesses during program execution and their demonstrated numerous hardware implementation emphasize this choice over traditional forecasting techniques for their inclusion in embedded systems. However, embedded applications execute millions of instructions and therefore millions of addresses to be predicted. This very challenging problem of neural network based prediction of large time series is approached in this paper by evaluating various neural network architectures based on the recurrent neural network paradigm with pre-processing based on the Self Organizing Map (SOM) classification technique.

Analysis of Feature Space for a 2d/3d Vision based Emotion Recognition Method

In modern human computer interaction systems (HCI), emotion recognition is becoming an imperative characteristic. The quest for effective and reliable emotion recognition in HCI has resulted in a need for better face detection, feature extraction and classification. In this paper we present results of feature space analysis after briefly explaining our fully automatic vision based emotion recognition method. We demonstrate the compactness of the feature space and show how the 2d/3d based method achieves superior features for the purpose of emotion classification. Also it is exposed that through feature normalization a widely person independent feature space is created. As a consequence, the classifier architecture has only a minor influence on the classification result. This is particularly elucidated with the help of confusion matrices. For this purpose advanced classification algorithms, such as Support Vector Machines and Artificial Neural Networks are employed, as well as the simple k- Nearest Neighbor classifier.

Robust Minutiae Watermarking in Wavelet Domain for Fingerprint Security

In this manuscript, a wavelet-based blind watermarking scheme has been proposed as a means to provide security to authenticity of a fingerprint. The information used for identification or verification of a fingerprint mainly lies in its minutiae. By robust watermarking of the minutiae in the fingerprint image itself, the useful information can be extracted accurately even if the fingerprint is severely degraded. The minutiae are converted in a binary watermark and embedding these watermarks in the detail regions increases the robustness of watermarking, at little to no additional impact on image quality. It has been experimentally shown that when the minutiae is embedded into wavelet detail coefficients of a fingerprint image in spread spectrum fashion using a pseudorandom sequence, the robustness is observed to have a proportional response while perceptual invisibility has an inversely proportional response to amplification factor “K". The DWT-based technique has been found to be very robust against noises, geometrical distortions filtering and JPEG compression attacks and is also found to give remarkably better performance than DCT-based technique in terms of correlation coefficient and number of erroneous minutiae.

Therapeutic Product Preparation Bioprocess Modeling

An immunomodulator bioproduct is prepared in a batch bioprocess with a modified bacterium Pseudomonas aeruginosa. The bioprocess is performed in 100 L Bioengineering bioreactor with 42 L cultivation medium made of peptone, meat extract and sodium chloride. The optimal bioprocess parameters were determined: temperature – 37 0C, agitation speed - 300 rpm, aeration rate – 40 L/min, pressure – 0.5 bar, Dow Corning Antifoam M-max. 4 % of the medium volume, duration - 6 hours. This kind of bioprocesses are appreciated as difficult to control because their dynamic behavior is highly nonlinear and time varying. The aim of the paper is to present (by comparison) different models based on experimental data. The analysis criteria were modeling error and convergence rate. The estimated values and the modeling analysis were done by using the Table Curve 2D. The preliminary conclusions indicate Andrews-s model with a maximum specific growth rate of the bacterium in the range of 0.8 h-1.

NonStationary CMA for Decision Feedback Equalization of Markovian Time Varying Channels

In this paper, we propose a modified version of the Constant Modulus Algorithm (CMA) tailored for blind Decision Feedback Equalizer (DFE) of first order Markovian time varying channels. The proposed NonStationary CMA (NSCMA) is designed so that it explicitly takes into account the Markovian structure of the channel nonstationarity. Hence, unlike the classical CMA, the NSCMA is not blind with respect to the channel time variations. This greatly helps the equalizer in the case of realistic channels, and avoids frequent transmissions of training sequences. This paper develops a theoretical analysis of the steady state performance of the CMA and the NSCMA for DFEs within a time varying context. Therefore, approximate expressions of the mean square errors are derived. We prove that in the steady state, the NSCMA exhibits better performance than the classical CMA. These new results are confirmed by simulation. Through an experimental study, we demonstrate that the Bit Error Rate (BER) is reduced by the NSCMA-DFE, and the improvement of the BER achieved by the NSCMA-DFE is as significant as the channel time variations are severe.

Medical Image Segmentation Using Deformable Model and Local Fitting Binary: Thoracic Aorta

This paper presents an application of level sets for the segmentation of abdominal and thoracic aortic aneurysms in CTA datasets. An important challenge in reliably detecting aortic is the need to overcome problems associated with intensity inhomogeneities. Level sets are part of an important class of methods that utilize partial differential equations (PDEs) and have been extensively applied in image segmentation. A kernel function in the level set formulation aids the suppression of noise in the extracted regions of interest and then guides the motion of the evolving contour for the detection of weak boundaries. The speed of curve evolution has been significantly improved with a resulting decrease in segmentation time compared with previous implementations of level sets, and are shown to be more effective than other approaches in coping with intensity inhomogeneities. We have applied the Courant Friedrichs Levy (CFL) condition as stability criterion for our algorithm.

An Adaptive Memetic Algorithm With Dynamic Population Management for Designing HIV Multidrug Therapies

In this paper, a mathematical model of human immunodeficiency virus (HIV) is utilized and an optimization problem is proposed, with the final goal of implementing an optimal 900-day structured treatment interruption (STI) protocol. Two type of commonly used drugs in highly active antiretroviral therapy (HAART), reverse transcriptase inhibitors (RTI) and protease inhibitors (PI), are considered. In order to solving the proposed optimization problem an adaptive memetic algorithm with population management (AMAPM) is proposed. The AMAPM uses a distance measure to control the diversity of population in genotype space and thus preventing the stagnation and premature convergence. Moreover, the AMAPM uses diversity parameter in phenotype space to dynamically set the population size and the number of crossovers during the search process. Three crossover operators diversify the population, simultaneously. The progresses of crossover operators are utilized to set the number of each crossover per generation. In order to escaping the local optima and introducing the new search directions toward the global optima, two local searchers assist the evolutionary process. In contrast to traditional memetic algorithms, the activation of these local searchers is not random and depends on both the diversity parameters in genotype space and phenotype space. The capability of AMAPM in finding optimal solutions compared with three popular metaheurestics is introduced.

Investigation of Effective Parameters on Annealing and Hot Spotting Processes for Straightening of Bent Turbine Rotors

The most severe damage of the turbine rotor is its distortion. The rotor straightening process must lead, at the first stage, to removal of the stresses from the material by annealing and next, to straightening of the plastic distortion without leaving any stress by hot spotting. The straightening method does not produce stress accumulations and the heating technique, developed specifically for solid forged rotors and disks, enables to avoid local overheating and structural changes in the material. This process also does not leave stresses in the shaft material. An experimental study of hot spotting is carried out on a large turbine rotor and some of the most important effective parameters that must be considered on annealing and hot spotting processes are investigated in this paper.