Anti-Money Laundering Requirements – Perceived Effectiveness

Anti-money laundering is commonly recognized as a set of procedures, laws or regulations designed to reduce the practice of generating income through illegal actions. In Malaysia, the government and law enforcement agencies have stepped up their capacities and efforts to curb money laundering since 2001. One of these measures was the enactment of the Anti-Money Laundering Act (AMLA) in 2001. The implementation costs on anti-money laundering requirements (AMLR) can be burdensome to those who are involved in enforcing them. The objective of this paper is to explore the perceived effectiveness of AMLR from the enforcement agencies- perspective. This is a preliminary study whose findings will help to give direction for further AML research in Malaysia. In addition, the results of this study provide empirical evidences on the perceived effectiveness of AMLR prior to further investigations on barriers and improvements of the implementation of the anti-money laundering regime in Malaysia.

TACS : Thermo Acoustic Cooling System

Cooling with sound is a physical phenomenon allowed by Thermo-Acoustics in which acoustic energy is transformed into a negative heat transfer, in other words: into cooling! Without needing any harmful gas, the transformation is environmentally friendly and can respond to many needs in terms of air conditioning, food refrigeration for domestic use, and cooling medical samples for example. To explore the possibilities of this cooling solution on a small scale, the TACS prototype has been designed, consisting of a low cost thermoacoustic refrigerant “pipe” able to lower the temperature by a few degrees. The obtained results are providing an interesting element for possible future of thermo-acoustic refrigeration.

Control Improvement of a C Sugar Cane Crystallization Using an Auto-Tuning PID Controller Based on Linearization of a Neural Network

The industrial process of the sugar cane crystallization produces a residual that still contains a lot of soluble sucrose and the objective of the factory is to improve its extraction. Therefore, there are substantial losses justifying the search for the optimization of the process. Crystallization process studied on the industrial site is based on the “three massecuites process". The third step of this process constitutes the final stage of exhaustion of the sucrose dissolved in the mother liquor. During the process of the third step of crystallization (Ccrystallization), the phase that is studied and whose control is to be improved, is the growing phase (crystal growth phase). The study of this process on the industrial site is a problem in its own. A control scheme is proposed to improve the standard PID control law used in the factory. An auto-tuning PID controller based on instantaneous linearization of a neural network is then proposed.

Integration of CMOS Biosensor into a Polymeric Lab-on-a-Chip System

We present an integration approach of a CMOS biosensor into a polymer based microfluidic environment suitable for mass production. It consists of a wafer-level-package for the silicon die and laser bonding process promoted by an intermediate hot melt foil to attach the sensor package to the microfluidic chip, without the need for dispensing of glues or underfiller. A very good condition of the sensing area was obtained after introducing a protection layer during packaging. A microfluidic flow cell was fabricated and shown to withstand pressures up to Δp = 780 kPa without leakage. The employed biosensors were electrically characterized in a dry environment.

The Induction of Antioxidant Enzyme Activities in Cabbage Seedlings by Heavy Metal Stress

Cabbage seedlings grown in vitro were exposed to excess levels of heavy metals, including Cd, Mo, and Zn. High metal levels affected plant growth at cotyledonary stage. Seedlings under Cd, Mo, and Zn treatments could not produce root hairs and true leaves. Under stress conditions, seedlings accumulated a higher amount of anthocyanins in their cotyledons than those in the control. The pigments isolated from Cd and Zn stressed seedling cotyledons appeared as pink, while under Mo stress, was dark pink or purple. Moreover, excess Mo stress increased antioxidant enzyme activities of APX, CAT, SOD. These results suggest that, under excess Mo stress, the induced antioxidant enzyme activity of cabbage seedlings may function as a protective mechanism to shield the plants from toxicity and exacerbated growth.

Semi Classical Three-Valley Monte Carlo Simulation Analysis of Steady-State and Transient Electron Transport within Bulk Ga0.38In0.62P

to simulate the phenomenon of electronic transport in semiconductors, we try to adapt a numerical method, often and most frequently it’s that of Monte Carlo. In our work, we applied this method in the case of a ternary alloy semiconductor GaInP in its cubic form; The Calculations are made using a non-parabolic effective-mass energy band model. We consider a band of conduction to three valleys (ΓLX), major of the scattering mechanisms are taken into account in this modeling, as the interactions with the acoustic phonons (elastic collisions) and optics (inelastic collisions). The polar optical phonons cause anisotropic collisions, intra-valleys, very probable in the III-V semiconductors. Other optical phonons, no polar, allow transitions inter-valleys. Initially, we present the full results obtained by the simulation of Monte Carlo in GaInP in stationary regime. We consider thereafter the effects related to the application of an electric field varying according to time, we thus study the transient phenomenon which make their appearance in ternary material

Critical Assessment of Scoring Schemes for Protein-Protein Docking Predictions

Protein-protein interactions (PPI) play a crucial role in many biological processes such as cell signalling, transcription, translation, replication, signal transduction, and drug targeting, etc. Structural information about protein-protein interaction is essential for understanding the molecular mechanisms of these processes. Structures of protein-protein complexes are still difficult to obtain by biophysical methods such as NMR and X-ray crystallography, and therefore protein-protein docking computation is considered an important approach for understanding protein-protein interactions. However, reliable prediction of the protein-protein complexes is still under way. In the past decades, several grid-based docking algorithms based on the Katchalski-Katzir scoring scheme were developed, e.g., FTDock, ZDOCK, HADDOCK, RosettaDock, HEX, etc. However, the success rate of protein-protein docking prediction is still far from ideal. In this work, we first propose a more practical measure for evaluating the success of protein-protein docking predictions,the rate of first success (RFS), which is similar to the concept of mean first passage time (MFPT). Accordingly, we have assessed the ZDOCK bound and unbound benchmarks 2.0 and 3.0. We also createda new benchmark set for protein-protein docking predictions, in which the complexes have experimentally determined binding affinity data. We performed free energy calculation based on the solution of non-linear Poisson-Boltzmann equation (nlPBE) to improve the binding mode prediction. We used the well-studied thebarnase-barstarsystem to validate the parameters for free energy calculations. Besides,thenlPBE-based free energy calculations were conducted for the badly predicted cases by ZDOCK and ZRANK. We found that direct molecular mechanics energetics cannot be used to discriminate the native binding pose from the decoys.Our results indicate that nlPBE-based calculations appeared to be one of the promising approaches for improving the success rate of binding pose predictions.

Mathematical Modeling of Gas Turbine Blade Cooling

In contrast to existing methods which do not take into account multiconnectivity in a broad sense of this term, we develop mathematical models and highly effective combination (BIEM and FDM) numerical methods of calculation of stationary and quasistationary temperature field of a profile part of a blade with convective cooling (from the point of view of realization on PC). The theoretical substantiation of these methods is proved by appropriate theorems. For it, converging quadrature processes have been developed and the estimations of errors in the terms of A.Ziqmound continuity modules have been received. For visualization of profiles are used: the method of the least squares with automatic conjecture, device spline, smooth replenishment and neural nets. Boundary conditions of heat exchange are determined from the solution of the corresponding integral equations and empirical relationships. The reliability of designed methods is proved by calculation and experimental investigations heat and hydraulic characteristics of the gas turbine first stage nozzle blade.

An Empirical Analysis of the Influence of Application Experience on Working Methods of Process Modelers

In view of growing competition in the service sector, services are as much in need of modeling, analysis and improvement as business or working processes. Graphical process models are important means to capture process-related know-how for an effective management of the service process. In this contribution, a human performance analysis of process model development paying special attention to model development time and the working method was conducted. It was found that modelers with higher application experience need significantly less time for mental activities than modelers with lower application experience, spend more time on labeling graphical elements, and achieved higher process model quality in terms of activity label quality.

Neutronic Study of Two Reactor Cores Cooled with Light and Heavy Water Using Computation Method

Most HWRs currently use natural uranium fuel. Using enriched uranium fuel results in a significant improvement in fuel cycle costs and uranium utilization. On the other hand, reactivity changes of HWRs over the full range of operating conditions from cold shutdown to full power are small. This reduces the required reactivity worth of control devices and minimizes local flux distribution perturbations, minimizing potential problems due to transient local overheating of fuel. Analyzing heavy water effectiveness on neutronic parameters such as enrichment requirements, peaking factor and reactivity is important and should pay attention as primary concepts of a HWR core designing. Two nuclear nuclear reactors of CANDU-type and hexagonal-type reactor cores of 33 fuel assemblies and 19 assemblies in 1.04 P/D have been respectively simulated using MCNP-4C code. Using heavy water and light water as moderator have been compared for achieving less reactivity insertion and enrichment requirements. Two fuel matrixes of (232Th/235U)O2 and (238/235U)O2 have been compared to achieve more economical and safe design. Heavy water not only decreased enrichment needs, but it concluded in negative reactivity insertions during moderator density variations. Thorium oxide fuel assemblies of 2.3% enrichment loaded into the core of heavy water moderator resulted in 0.751 fission to absorption ratio and peaking factor of 1.7 using. Heavy water not only provides negative reactivity insertion during temperature raises which changes moderator density but concluded in 2 to 10 kg reduction of enrichment requirements, depend on geometry type.

The Effect of Loperamide and Fentanyl on the Distribution Kinetics of Verapamil in the Lung and Brain in Sprague Dawley Rats

Verapamil has been shown to inhibit fentanyl uptake in vitro and is a potent P-glycoprotein inhibitor. Tissue partitioning of loperamide, a commercially available opioid, is closely controlled by the P-gp efflux transporter. The following studies were designed to evaluate the effect of opioids on verapamil partitioning in the lung and brain, in vivo. Opioid (fentanyl or loperamide) was administered by intravenous infusion to Sprague Dawley rats alone or in combination with verapamil and plasma, with lung and brain tissues were collected at 1, 5, 6, 8, 10 and 60 minutes. Drug dispositions were modeled by recirculatory pharmacokinetic models. Fentanyl slightly increased the verapamil lung (PL) partition coefficient yet decreased the brain (PB) partition coefficient. Furthermore, loperamide significantly increased PLand PB. Fentanyl reduced the verapamil volume of distribution (V1) and verapamil elimination clearance (ClE). Fentanyl decreased verapamil brain partitioning, yet increased verapamil lung partitioning. Also, loperamide increased lung and brain partitioning in vivo. These results suggest that verapamil and fentanyl may be substrates of an unidentified inward transporter in brain tissue and confirm that verapamil and loperamide are substrates of the efflux transporter P-gp.

Dynamic Shock Bank Liquidity Analysis

Simulations are developed in this paper with usual DSGE model equations. The model is based on simplified version of Smets-Wouters equations in use at European Central Bank which implies 10 macro-economic variables: consumption, investment, wages, inflation, capital stock, interest rates, production, capital accumulation, labour and credit rate, and allows take into consideration the banking system. Throughout the simulations, this model will be used to evaluate the impact of rate shocks recounting the actions of the European Central Bank during 2008.

Efficient Pipelined Hardware Implementation of RIPEMD-160 Hash Function

In this paper an efficient implementation of Ripemd- 160 hash function is presented. Hash functions are a special family of cryptographic algorithms, which is used in technological applications with requirements for security, confidentiality and validity. Applications like PKI, IPSec, DSA, MAC-s incorporate hash functions and are used widely today. The Ripemd-160 is emanated from the necessity for existence of very strong algorithms in cryptanalysis. The proposed hardware implementation can be synthesized easily for a variety of FPGA and ASIC technologies. Simulation results, using commercial tools, verified the efficiency of the implementation in terms of performance and throughput. Special care has been taken so that the proposed implementation doesn-t introduce extra design complexity; while in parallel functionality was kept to the required levels.

JConqurr - A Multi-Core Programming Toolkit for Java

With the popularity of the multi-core and many-core architectures there is a great requirement for software frameworks which can support parallel programming methodologies. In this paper we introduce an Eclipse toolkit, JConqurr which is easy to use and provides robust support for flexible parallel progrmaming. JConqurr is a multi-core and many-core programming toolkit for Java which is capable of providing support for common parallel programming patterns which include task, data, divide and conquer and pipeline parallelism. The toolkit uses an annotation and a directive mechanism to convert the sequential code into parallel code. In addition to that we have proposed a novel mechanism to achieve the parallelism using graphical processing units (GPU). Experiments with common parallelizable algorithms have shown that our toolkit can be easily and efficiently used to convert sequential code to parallel code and significant performance gains can be achieved.

Designing Pictogram for Food Portion Size

The objective of this paper is to investigate a new approach based on the idea of pictograms for food portion size. This approach adopts the model of the United States Pharmacopeia- Drug Information (USP-DI). The representation of each food portion size composed of three parts: frame, the connotation of dietary portion sizes and layout. To investigate users- comprehension based on this approach, two experiments were conducted, included 122 Taiwanese people, 60 male and 62 female with ages between 16 and 64 (divided into age groups of 16-30, 31-45 and 46-64). In Experiment 1, the mean correcting rate of the understanding level of food items is 48.54% (S.D.= 95.08) and the mean response time 2.89sec (S.D.=2.14). The difference on the correct rates for different age groups is significant (P*=0.00

SELF-Cured Alkali Activated Slag Concrete Mixes- An Experimental Study

Alkali Activated Slag Concrete (AASC) mixes are manufactured by activating ground granulated blast furnace slag (GGBFS) using sodium hydroxide and sodium silicate solutions. The aim of the present experimental research was to investigate the effect of increasing the dosages of sodium oxide (Na2O, in the range of 4 to 8%) and the activator modulus (Ms) (i.e. the SiO2/Na2O ratio, in the range of 0.5 to 1.5) of the alkaline solutions, on the workability and strength characteristics of self-cured (air-cured) alkali activated Indian slag concrete mixes. Further the split tensile and flexure strengths for optimal mixes were studied for each dosage of Na2O.It is observed that increase in Na2O concentration increases the compressive, split-tensile and flexural strengths, both at the early and later-ages, while increase in Ms, decreases the workability of the mixes. An optimal Ms of 1.25 is found at various Na2O dosages. No significant differences in the strength performances were observed between AASCs manufactured with alkali solutions prepared using either of potable and de-ionized water.

Traces of Birdhouse Tradition in Anatolia

The birdhouses and dovecotes, which are the indicator of naturalness and human-animal relationship, are one of the traditional cultural values of Turkey. With their structures compatible with nature and respectful to humans the bird houses and dovecotes, which have an important position in local urbanization models as a representative of the civil architecture with their unique form and function are important subjects that should be evaluated in a wide frame comprising from architecture to urbanism, from ecologic agriculture to globalization. The traditional bird houses and dovecotes are disregarded due to the insensitivity affecting the city life and the change in the public sense of art. In this study, the characteristic properties of traditional dovecotes and birdhouses, started in 13th century and ended in 19th century in Anatolia, are tried to be defined for the sustainability of the tradition and for giving a new direction to the designers.

Qualitative Survey on Managing Building Maintenance Projects

Buildings are one of the valuable assets to provide people with shelters for work, leisure and rest. After years of attacks by weather, buildings will deteriorate which need proper maintenance in order to fulfill the requirements and satisfaction of the users. Poorly managed buildings not just give a negative image to the city itself, but also pose potential risk hazards to the health and safety of the general public. As a result, the management of maintenance projects has played an important role in cities like Hong Kong where the problem of urban decay has drawn much attention. However, most research has focused on managing new construction, and little research effort has been put on maintenance projects. Given the short duration and more diversified nature of work, repair and maintenance works are found to be more difficult to monitor and regulate when compared with new works. Project participants may face with problems in running maintenance projects which should be investigated so that proper strategies can be established. This paper aims to provide a thorough analysis on the problems of running maintenance projects. A review of literature on the characteristics of building maintenance projects was firstly conducted, which forms a solid basis for the empirical study. Results on the problems and difficulties of running maintenance projects from the viewpoints of industry practitioners will also be delivered with a view to formulating effective strategies for managing maintenance projects successfully.

Simulation Tools for Fixed Point DSP Algorithms and Architectures

This paper presents software tools that convert the C/Cµ floating point source code for a DSP algorithm into a fixedpoint simulation model that can be used to evaluate the numericalperformance of the algorithm on several different fixed pointplatforms including microprocessors, DSPs and FPGAs. The tools use a novel system for maintaining binary point informationso that the conversion from floating point to fixed point isautomated and the resulting fixed point algorithm achieves maximum possible precision. A configurable architecture is used during the simulation phase so that the algorithm can produce a bit-exact output for several different target devices.