The Evolution of Quality Improvement Methodology in Malaysia-s IT Industry: The Past, Current and Future

There are various approaches to implement quality improvements. Organizations aim for a management standard which is capable of providing customers with quality assurance on their product/service via continuous process improvement. Carefully planned steps are necessary to ensure the right quality improvement methodology (QIM) and business operations are consistent, reliable and truly meet the customers' needs. This paper traces the evolution of QIM in Malaysia-s Information Technology (IT) industry in the past, current and future; and highlights some of the thought of researchers who contributed to the science and practice of quality, and identifies leading methodologies in use today. Some of the misconceptions and mistakes leading to quality system failures will also be examined and discussed. This paper aims to provide a general overview of different types of QIMs available for IT businesses in maximizing business advantages, enhancing product quality, improving process routines and increasing performance earnings.

Energy Consumption Analysis of Design Patterns

The importance of low power consumption is widely acknowledged due to the increasing use of portable devices, which require minimizing the consumption of energy. Energy dissipation is heavily dependent on the software used in the system. Applying design patterns in object-oriented designs is a common practice nowadays. In this paper we analyze six design patterns and explore the effect of them on energy consumption and performance.

Exit Strategies from The Global Crisis

While the form of crises may change, their essence remains the same (such as a cycle of abundant liquidity, rapid credit growth, and a low-inflation environment followed by an asset-price bubble). The current market turbulence began in mid-2000s when the US economy shifted to imbalanced both internal and external macroeconomic positions. We see two key causes of these problems – loose US monetary policy in early 2000s and US government guarantees issued on the securities by government-sponsored enterprises what was further fueled by financial innovations such as structured credit products. We have discovered both negative and positive lessons deriving from this crisis and divided the negative lessons into three groups: financial products and valuation, processes and business models, and strategic issues. Moreover, we address key risk management lessons and exit strategies derived from the current crisis and recommend policies that should help diminish the negative impact of future potential crises.

Fracture Location Characterizations of Dissimilar Friction Stir Welds

This paper reports the tensile fracture location characterizations of dissimilar friction stir welds between 5754 aluminium alloy and C11000 copper. The welds were produced using three shoulder diameter tools; namely, 15, 18 and 25 mm by varying the process parameters. The rotational speeds considered were 600, 950 and 1200 rpm while the feed rates employed were 50, 150 and 300 mm/min to represent the low, medium and high settings respectively. The tensile fracture locations were evaluated using the optical microscope to identify the fracture locations and were characterized. It was observed that 70% of the tensile samples failed in the Thermo Mechanically Affected Zone (TMAZ) of copper at the weld joints. Further evaluation of the fracture surfaces of the pulled tensile samples revealed that welds with low Ultimate Tensile Strength either have defects or intermetallics present at their joint interfaces.

Molecular Epidemiology and Genotyping of Bovine Viral Diarrhea Virus in Xinjiang Uygur Autonomous Region of China

As part of national epidemiological survey on bovine viral diarrhea virus (BVDV), a total of 274 dejecta samples were collected from 14 cattle farms in 8 areas of Xinjiang Uygur Autonomous Region in northwestern China. Total RNA was extracted from each sample, and 5--untranslated region (UTR) of BVDV genome was amplified by using two-step reverse transcriptase-polymerase chain reaction (RT-PCR). The PCR products were subsequently sequenced to study the genetic variations of BVDV in these areas. Among the 274 samples, 33 samples were found virus-positive. According to sequence analysis of the PCR products, the 33 samples could be arranged into 16 groups. All the sequences, however, were highly conserved with BVDV Osloss strains. The virus possessed theses sequences belonged to BVDV-1b subtype by phylogenetic analysis. Based on these data, we established a typing tree for BVDV in these areas. Our results suggested that BVDV-1b was a predominant subgenotype in northwestern China and no correlation between the genetic and geographical distances could be observed above the farm level.

Neural Network Learning Based on Chaos

Chaos and fractals are novel fields of physics and mathematics showing up a new way of universe viewpoint and creating many ideas to solve several present problems. In this paper, a novel algorithm based on the chaotic sequence generator with the highest ability to adapt and reach the global optima is proposed. The adaptive ability of proposal algorithm is flexible in 2 steps. The first one is a breadth-first search and the second one is a depth-first search. The proposal algorithm is examined by 2 functions, the Camel function and the Schaffer function. Furthermore, the proposal algorithm is applied to optimize training Multilayer Neural Networks.

Error Correction Codes in Wireless Sensor Network: An Energy Aware Approach

Link reliability and transmitted power are two important design constraints in wireless network design. Error control coding (ECC) is a classic approach used to increase link reliability and to lower the required transmitted power. It provides coding gain, resulting in transmitter energy savings at the cost of added decoder power consumption. But the choice of ECC is very critical in the case of wireless sensor network (WSN). Since the WSNs are energy constraint in nature, both the BER and power consumption has to be taken into count. This paper develops a step by step approach in finding suitable error control codes for WSNs. Several simulations are taken considering different error control codes and the result shows that the RS(31,21) fits both in BER and power consumption criteria.

Multi-View Neural Network Based Gait Recognition

Human identification at a distance has recently gained growing interest from computer vision researchers. Gait recognition aims essentially to address this problem by identifying people based on the way they walk [1]. Gait recognition has 3 steps. The first step is preprocessing, the second step is feature extraction and the third one is classification. This paper focuses on the classification step that is essential to increase the CCR (Correct Classification Rate). Multilayer Perceptron (MLP) is used in this work. Neural Networks imitate the human brain to perform intelligent tasks [3].They can represent complicated relationships between input and output and acquire knowledge about these relationships directly from the data [2]. In this paper we apply MLP NN for 11 views in our database and compare the CCR values for these views. Experiments are performed with the NLPR databases, and the effectiveness of the proposed method for gait recognition is demonstrated.

Quadratic Pulse Inversion Ultrasonic Imaging(QPI): A Two-Step Procedure for Optimization of Contrast Sensitivity and Specificity

We have previously introduced an ultrasonic imaging approach that combines harmonic-sensitive pulse sequences with a post-beamforming quadratic kernel derived from a second-order Volterra filter (SOVF). This approach is designed to produce images with high sensitivity to nonlinear oscillations from microbubble ultrasound contrast agents (UCA) while maintaining high levels of noise rejection. In this paper, a two-step algorithm for computing the coefficients of the quadratic kernel leading to reduction of tissue component introduced by motion, maximizing the noise rejection and increases the specificity while optimizing the sensitivity to the UCA is presented. In the first step, quadratic kernels from individual singular modes of the PI data matrix are compared in terms of their ability of maximize the contrast to tissue ratio (CTR). In the second step, quadratic kernels resulting in the highest CTR values are convolved. The imaging results indicate that a signal processing approach to this clinical challenge is feasible.

The Contribution of Growth Rate to the Pathogenicity of Candida spp.

Fungal infections are becoming more common and the range of susceptible individuals has expanded. While Candida albicans remains the most common infective species, other Candida spp. are becoming increasingly significant. In a range of large-scale studies of candidaemia between 1999 and 2006, about 52% of 9717 cases involved C. albicans, about 30% involved either C. glabrata or C. parapsilosis and less than 15% involved C. tropicalis, C. krusei or C. guilliermondii. However, the probability of mortality within 30 days of infection with a particular species was at least 40% for C. tropicalis, C. albicans, C. glabrata and C. krusei and only 22% for C. parapsilopsis. Clinical isolates of Candida spp. grew at rates ranging from 1.65 h-1 to 4.9 h-1. Three species (C. krusei, C. albicans and C. glabrata) had relatively high growth rates (μm > 4 h-1), C. tropicalis and C. dubliniensis grew moderately quickly (Ôëê 3 h-1) and C. parapsilosis and C. guilliermondii grew slowly (< 2 h-1). Based on these data, the log of the odds of mortality within 30 days of diagnosis was linearly related to μm. From this the underlying probability of mortality is 0.13 (95% CI: 0.10-0.17) and it increases by about 0.09 ± 0.02 for each unit increase in μm. Given that the overall crude mortality is about 0.36, the growth of Candida spp. approximately doubles the rate, consistent with the results of larger case-matched studies of candidaemia.

Binary Classification Tree with Tuned Observation-based Clustering

There are several approaches for handling multiclass classification. Aside from one-against-one (OAO) and one-against-all (OAA), hierarchical classification technique is also commonly used. A binary classification tree is a hierarchical classification structure that breaks down a k-class problem into binary sub-problems, each solved by a binary classifier. In each node, a set of classes is divided into two subsets. A good class partition should be able to group similar classes together. Many algorithms measure similarity in term of distance between class centroids. Classes are grouped together by a clustering algorithm when distances between their centroids are small. In this paper, we present a binary classification tree with tuned observation-based clustering (BCT-TOB) that finds a class partition by performing clustering on observations instead of class centroids. A merging step is introduced to merge any insignificant class split. The experiment shows that performance of BCT-TOB is comparable to other algorithms.

Explorations in the Role of Emotion in Moral Judgment

Recent theorizations on the cognitive process of moral judgment have focused on the role of intuitions and emotions, marking a departure from previous emphasis on conscious, step-by-step reasoning. My study investigated how being in a disgusted mood state affects moral judgment. Participants were induced to enter a disgusted mood state through listening to disgusting sounds and reading disgusting descriptions. Results shows that they, when compared to control who have not been induced to feel disgust, are more likely to endorse actions that are emotionally aversive but maximizes utilitarian return The result is analyzed using the 'emotion-as-information' approach to decision making. The result is consistent with the view that emotions play an important role in determining moral judgment.

Detecting Interactions between Behavioral Requirements with OWL and SWRL

High quality requirements analysis is one of the most crucial activities to ensure the success of a software project, so that requirements verification for software system becomes more and more important in Requirements Engineering (RE) and it is one of the most helpful strategies for improving the quality of software system. Related works show that requirement elicitation and analysis can be facilitated by ontological approaches and semantic web technologies. In this paper, we proposed a hybrid method which aims to verify requirements with structural and formal semantics to detect interactions. The proposed method is twofold: one is for modeling requirements with the semantic web language OWL, to construct a semantic context; the other is a set of interaction detection rules which are derived from scenario-based analysis and represented with semantic web rule language (SWRL). SWRL based rules are working with rule engines like Jess to reason in semantic context for requirements thus to detect interactions. The benefits of the proposed method lie in three aspects: the method (i) provides systematic steps for modeling requirements with an ontological approach, (ii) offers synergy of requirements elicitation and domain engineering for knowledge sharing, and (3)the proposed rules can systematically assist in requirements interaction detection.

IAS 41 Implementation Challenges – The Case of Romania

Although agriculture is an important part of the world economy, accounting in agriculture still has many shortcomings. The adoption of IAS 41 “Agriculture” has tried to improve this situation and increase the comparability of financial statements of entities in the agricultural sector. Although controversial, IAS 41 is the first step of a consistent transition to fair value assessment in the agricultural sector. The objective of our work is the analysis of IAS 41 and current accounting agricultural situation in Romania. Accounting regulations in Romania are in accordance with European directives and, in many respects, converged with IFRS referential. Provisions of IAS 41, however, are not reflected directly in Romanian regulations. With the increase of forest land transactions, it is expected that recognition and measurement of biological assets under IAS 41 to become a necessity.

Methods of Estimating the Equilibrium Real Effective Exchange Rate (REER)

There are many debates now regarding undervalued and overvalued currencies currently traded on the world financial market. This paper contributes to these debates from a theoretical point of view. We present the three most commonly used methods of estimating the equilibrium real effective exchange rate (REER): macroeconomic balance approach, external sustainability approach and equilibrium real effective exchange rate approach in the reduced form. Moreover, we discuss key concepts of the calculation of the real exchange rate (RER) based on applied explanatory variables: nominal exchange rates, terms of trade and tradable and non-tradable goods. Last but not least, we discuss the three main driving forces behind real exchange rates movements which include terms of trade, relative productivity growth and the interest rate differential.

Color Lighting Efficiency of Light Emitting Diode Tube to Lure the Adult Coconut Hispine Beetle

The objective of this research was to investigate the efficiency of the light emitting diode (LED) tube in various color lights used to lure the adult coconut hispine beetle. The research was conducted by setting the forward bias on LED tubes, and the next step was to test luminous efficacy and quantity of electricity used to power each LED tube in different color lights. Finally, the researcher examined the efficiency of each color-light LED tube to lure the adult coconut hispine beetle. The results showed that the ultraviolet LED tubes had the most capacity to allure the adult coconut hispine beetles with the percentage of 82.92, followed by the blue LED tubes with the percentage of 59.76. Whereas the yellow, pink, red and warm white LED tubes had no influence to the adult coconut hispine beetles.

Teager-Huang Analysis Applied to Sonar Target Recognition

In this paper, a new approach for target recognition based on the Empirical mode decomposition (EMD) algorithm of Huang etal. [11] and the energy tracking operator of Teager [13]-[14] is introduced. The conjunction of these two methods is called Teager-Huang analysis. This approach is well suited for nonstationary signals analysis. The impulse response (IR) of target is first band pass filtered into subsignals (components) called Intrinsic mode functions (IMFs) with well defined Instantaneous frequency (IF) and Instantaneous amplitude (IA). Each IMF is a zero-mean AM-FM component. In second step, the energy of each IMF is tracked using the Teager energy operator (TEO). IF and IA, useful to describe the time-varying characteristics of the signal, are estimated using the Energy separation algorithm (ESA) algorithm of Maragos et al .[16]-[17]. In third step, a set of features such as skewness and kurtosis are extracted from the IF, IA and IMF energy functions. The Teager-Huang analysis is tested on set of synthetic IRs of Sonar targets with different physical characteristics (density, velocity, shape,? ). PCA is first applied to features to discriminate between manufactured and natural targets. The manufactured patterns are classified into spheres and cylinders. One hundred percent of correct recognition is achieved with twenty three echoes where sixteen IRs, used for training, are free noise and seven IRs, used for testing phase, are corrupted with white Gaussian noise.

Microbubbles Enhanced Synthetic Phorbol Ester Degradation by Ozonolysis

A phorbol-12-myristate-13-acetate (TPA) is a synthetic analogue of phorbol ester (PE), a natural toxic compound of Euphorbiaceae plant. The oil extracted from plants of this family is useful source for primarily biofuel. However this oil might also be used as a foodstuff due to its significant nutrition content. The limitations for utilizing the oil as a foodstuff are mainly due to a toxicity of PE. Currently, a majority of PE detoxification processes are expensive as include multi steps alcohol extraction sequence. Ozone is considered as a strong oxidative agent. It reacts with PE by attacking the carbon-carbon double bond of PE. This modification of PE molecular structure yields a non toxic ester with high lipid content. This report presents data on development of simple and cheap PE detoxification process with water application as a buffer and ozone as reactive component. The core of this new technique is an application for a new microscale plasma unit to ozone production and the technology permits ozone injection to the water-TPA mixture in form of microbubbles. The efficacy of a heterogeneous process depends on the diffusion coefficient which can be controlled by contact time and interfacial area. The low velocity of rising microbubbles and high surface to volume ratio allow efficient mass transfer to be achieved during the process. Direct injection of ozone is the most efficient way to process with such highly reactive and short lived chemical. Data on the plasma unit behavior are presented and the influence of gas oscillation technology on the microbubble production mechanism has been discussed. Data on overall process efficacy for TPA degradation is shown.

Optimal Solution of Constraint Satisfaction Problems

An optimal solution for a large number of constraint satisfaction problems can be found using the technique of substitution and elimination of variables analogous to the technique that is used to solve systems of equations. A decision function f(A)=max(A2) is used to determine which variables to eliminate. The algorithm can be expressed in six lines and is remarkable in both its simplicity and its ability to find an optimal solution. However it is inefficient in that it needs to square the updated A matrix after each variable elimination. To overcome this inefficiency the algorithm is analyzed and it is shown that the A matrix only needs to be squared once at the first step of the algorithm and then incrementally updated for subsequent steps, resulting in significant improvement and an algorithm complexity of O(n3).

A Study on the Attractiveness of Heavy Duty Motorcycle

The culture of riding heavy motorcycles originates from advanced countries and mainly comes from Europe, North America, and Japan. Heavy duty motorcycle riders are different from people who view motorcycles as a convenient mean of transportation. They regard riding them as a kind of enjoyment and high-level taste. The activities of riding heavy duty motorcycles have formes a distinctive landscape in domestic land in Taiwan. Previous studies which explored motorcycle culture in Taiwan still focused on the objects of motorcycle engine displacement under 50 cc.. The study aims to study the heavy duty motorcycles of engine displacement over 550 cc. and explores where their attractiveness is. For finding the attractiveness of heavy duty motorcycle, the study chooses Miryoku Engineering (Preference-Based Design) approach. Two steps are adopted to proceed the research. First, through arranging the letters obtained from interviewing experts, EGM (The Evaluation Grid Method) was applied to find out the structure of attractiveness. The attractive styles are eye-dazzling, leisure, classic, and racing competitive styles. Secondarily, Quantification Theory Type I analysis was adopted as a tool for analyzing the importance of attractiveness. The relationship between style and attractive parts was also discussed. The results could contribute to the design and research development of heavy duty motorcycle industry in Taiwan.