Evolution of Quality Function Deployment (QFD) via Fuzzy Concepts and Neural Networks

Quality Function Deployment (QFD) is an expounded, multi-step planning method for delivering commodity, services, and processes to customers, both external and internal to an organization. It is a way to convert between the diverse customer languages expressing demands (Voice of the Customer), and the organization-s languages expressing results that sate those demands. The policy is to establish one or more matrices that inter-relate producer and consumer reciprocal expectations. Due to its visual presence is called the “House of Quality" (HOQ). In this paper, we assumed HOQ in multi attribute decision making (MADM) pattern and through a proposed MADM method, rank technical specifications. Thereafter compute satisfaction degree of customer requirements and for it, we apply vagueness and uncertainty conditions in decision making by fuzzy set theory. This approach would propound supervised neural network (perceptron) for MADM problem solving.

Neural Network Based Determination of Splice Junctions by ROC Analysis

Gene, principal unit of inheritance, is an ordered sequence of nucleotides. The genes of eukaryotic organisms include alternating segments of exons and introns. The region of Deoxyribonucleic acid (DNA) within a gene containing instructions for coding a protein is called exon. On the other hand, non-coding regions called introns are another part of DNA that regulates gene expression by removing from the messenger Ribonucleic acid (RNA) in a splicing process. This paper proposes to determine splice junctions that are exon-intron boundaries by analyzing DNA sequences. A splice junction can be either exon-intron (EI) or intron exon (IE). Because of the popularity and compatibility of the artificial neural network (ANN) in genetic fields; various ANN models are applied in this research. Multi-layer Perceptron (MLP), Radial Basis Function (RBF) and Generalized Regression Neural Networks (GRNN) are used to analyze and detect the splice junctions of gene sequences. 10-fold cross validation is used to demonstrate the accuracy of networks. The real performances of these networks are found by applying Receiver Operating Characteristic (ROC) analysis.

Production of Carbon Nanotubes by Iron Catalyst

Carbon nanotubes (CNTs) with their high mechanical, electrical, thermal and chemical properties are regarded as promising materials for many different potential applications. Having unique properties they can be used in a wide range of fields such as electronic devices, electrodes, drug delivery systems, hydrogen storage, textile etc. Catalytic chemical vapor deposition (CCVD) is a common method for CNT production especially for mass production. Catalysts impregnated on a suitable substrate are important for production with chemical vapor deposition (CVD) method. Iron catalyst and MgO substrate is one of most common catalyst-substrate combination used for CNT. In this study, CNTs were produced by CCVD of acetylene (C2H2) on magnesium oxide (MgO) powder substrate impregnated by iron nitrate (Fe(NO3)3•9H2O) solution. The CNT synthesis conditions were as follows: at synthesis temperatures of 500 and 800°C multiwall and single wall CNTs were produced respectively. Iron (Fe) catalysts were prepared by with Fe:MgO ratio of 1:100, 5:100 and 10:100. The duration of syntheses were 30 and 60 minutes for all temperatures and catalyst percentages. The synthesized materials were characterized by thermal gravimetric analysis (TGA), transmission electron microscopy (TEM) and Raman spectroscopy.

Examining Corporate Tax Evaders: Evidence from the Finalized Audit Cases

This paper aims to (1) analyze the profiles of transgressors (detected evaders); (2) examine reason(s) that triggered a tax audit, causes of tax evasion, audit timeframe and tax penalty charged; and (3) to assess if tax auditors followed the guidelines as stated in the 'Tax Audit Framework' when conducting tax audits. In 2011, the Inland Revenue Board Malaysia (IRBM) had audited and finalized 557 company cases. With official permission, data of all the 557 cases were obtained from the IRBM. Of these, a total of 421 cases with complete information were analyzed. About 58.1% was small and medium corporations and from the construction industry (32.8%). The selection for tax audit was based on risk analysis (66.8%), information from third party (11.1%), and firm with low profitability or fluctuating profit pattern (7.8%). The three persistent causes of tax evasion by firms were over claimed expenses (46.8%), fraudulent reporting of income (38.5%) and overstating purchases (10.5%). These findings are consistent with past literature. Results showed that tax auditors took six to 18 months to close audit cases. More than half of tax evaders were fined 45% on additional tax raised during audit for the first offence. The study found tax auditors did follow the guidelines in the 'Tax Audit Framework' in audit selection, settlement and penalty imposition.

Analysis of FWM Penalties in DWDM Systems Based on G.652, G.653, and G.655 Optical Fibers

This paper presents an investigation of the power penalties imposed by four-wave mixing (FWM) on G.652 (Single- Mode Fiber - SMF), G.653 (Dispersion-Shifted Fiber - DSF), and G.655 (Non-Zero Dispersion-Shifted Fiber - NZDSF) compliant fibers, considering the DWDM grids suggested by the ITU-T Recommendations G.692, and G.694.1, with uniform channel spacing of 100, 50, 25, and 12.5 GHz. The mathematical/numerical model assumes undepleted pumping, and shows very clearly the deleterious effect of FWM on the performance of DWDM systems, measured by the signal-to-noise ratio (SNR). The results make it evident that non-uniform channel spacing is practically mandatory for WDM systems based on DSF fibers.

Effect of Geum Kokanicum Total Extract on Induced Nociception and Inflammation in Male Mice

The aim of this study is evaluating the antinociceptive and anti-inflamatory activity of Geum kokanicum. After determination total extract LD50, different doses of extract were chosen for intrapritoneal injections. In inflammation test, male NMRI mice were divided into 6 groups: control (normal saline), positive control (Dexamethasone 15mg/kg), and total extract (0.025, 0.05, 0.1, and 0.2 gr/kg). The inflammation was produced by xyleneinduced edema. In order to evaluate the antinociceptive effect of total extract, formalin test was used. Mice were divided into 6 groups: control, positive control (morphine 10mg/kg), and 4 groups which received total extract. Then they received Formalin. The animals were observed for the reaction to pain. Data were analyzed using One-way ANOVA followed by Tukey-Kramer multiple comparison test. LD50 was 1 gr/kg. Data indicated that 0.5,0.1 and 0.2 gr/kg doses of total extract have particular antinociceptive and antiinflammatory effects in a comparison with control (P

Modification of Anodized Mg Alloy Surface By Pulse Condition for Biodegradable Material

Magnesium is used implant material potentially for non-toxicity to the human body. Due to the excellent bio-compatibility, Mg alloys is applied to implants avoiding removal second surgery. However, it is found commercial magnesium alloys including aluminum has low corrosion resistance, resulting subcutaneous gas bubbles and consequently the approach as permanent bio-materials. Generally, Aluminum is known to pollution substance, and it raises toxicity to nervous system. Therefore especially Mg-35Zn-3Ca alloy is prepared for new biodegradable materials in this study. And the pulsed power is used in constant-current mode of DC power kinds of anodization. Based on the aforementioned study, it examines corrosion resistance and biocompatibility by effect of current and frequency variation. The surface properties and thickness were compared using scanning electronic microscopy. Corrosion resistance was assessed via potentiodynamic polarization and the effect of oxide layer on the body was assessed cell viability. Anodized Mg-35Zn-3Ca alloy has good biocompatibility in vitro by current and frequency variation.

Design of a DCT-based Image Compression with Efficient Enhancement Filter

The algorithm represents the DCT coefficients to concentrate signal energy and proposes combination and dictator to eliminate the correlation in the same level subband for encoding the DCT-based images. This work adopts DCT and modifies the SPIHT algorithm to encode DCT coefficients. The proposed algorithm also provides the enhancement function in low bit rate in order to improve the perceptual quality. Experimental results indicate that the proposed technique improves the quality of the reconstructed image in terms of both PSNR and the perceptual results close to JPEG2000 at the same bit rate.

Towards Development of Solution for Business Process-Oriented Data Analysis

This paper proposes a modeling methodology for the development of data analysis solution. The Author introduce the approach to address data warehousing issues at the at enterprise level. The methodology covers the process of the requirements eliciting and analysis stage as well as initial design of data warehouse. The paper reviews extended business process model, which satisfy the needs of data warehouse development. The Author considers that the use of business process models is necessary, as it reflects both enterprise information systems and business functions, which are important for data analysis. The Described approach divides development into three steps with different detailed elaboration of models. The Described approach gives possibility to gather requirements and display them to business users in easy manner.

Likelihood Estimation for Stochastic Epidemics with Heterogeneous Mixing Populations

We consider a heterogeneously mixing SIR stochastic epidemic process in populations described by a general graph. Likelihood theory is developed to facilitate statistic inference for the parameters of the model under complete observation. We show that these estimators are asymptotically Gaussian unbiased estimates by using a martingale central limit theorem.

Enhancing Cache Performance Based on Improved Average Access Time

A high performance computer includes a fast processor and millions bytes of memory. During the data processing, huge amount of information are shuffled between the memory and processor. Because of its small size and its effectiveness speed, cache has become a common feature of high performance computers. Enhancing cache performance proved to be essential in the speed up of cache-based computers. Most enhancement approaches can be classified as either software based or hardware controlled. The performance of the cache is quantified in terms of hit ratio or miss ratio. In this paper, we are optimizing the cache performance based on enhancing the cache hit ratio. The optimum cache performance is obtained by focusing on the cache hardware modification in the way to make a quick rejection to the missed line's tags from the hit-or miss comparison stage, and thus a low hit time for the wanted line in the cache is achieved. In the proposed technique which we called Even- Odd Tabulation (EOT), the cache lines come from the main memory into cache are classified in two types; even line's tags and odd line's tags depending on their Least Significant Bit (LSB). This division is exploited by EOT technique to reject the miss match line's tags in very low time compared to the time spent by the main comparator in the cache, giving an optimum hitting time for the wanted cache line. The high performance of EOT technique against the familiar mapping technique FAM is shown in the simulated results.

Promoting Mental and Spiritual Health among Postpartum Mothers to Extend Breastfeeding Period

The purpose of this study was to study postpartum breastfeeding mothers to determine the impact their psychosocial and spiritual dimensions play in promoting full-term (6 month duration) breastfeeding of their infants. Purposive and snowball sampling methods were used to identify and recruit the study's participants. A total of 23 postpartum mothers, who were breastfeeding within 6 weeks after giving birth, participated in this study. In-depth interviews combined with observations, participant focus groups, and ethnographic records were used for data collection. The Data were then analyzed using content analysis and typology. The results of this study illustrated that postpartum mothers experienced fear and worry that they would lack support from their spouse, family and peers, and that their infant would not get enough milk It was found that the main barrier mothers faced in breastfeeding to full-term was the difficulty of continuing to breastfeed when returning to work. 81.82% of the primiparous mothers and 91.67% of the non-primiparous mothers were able to breastfeed for the desired full-term of 6 months. Factors found to be related to breastfeeding for six months included 1) belief and faith in breastfeeding, 2) support from spouse and family members, 3) counseling from public health nurses and friends. The sample also provided evidence that religious principles such as tolerance, effort, love, and compassion to their infant, and positive thinking, were used in solving their physical, mental and spiritual problems.

An Interval-Based Multi-Attribute Decision Making Approach for Electric Utility Resource Planning

This paper presents an interval-based multi-attribute decision making (MADM) approach in support of the decision process with imprecise information. The proposed decision methodology is based on the model of linear additive utility function but extends the problem formulation with the measure of composite utility variance. A sample study concerning with the evaluation of electric generation expansion strategies is provided showing how the imprecise data may affect the choice toward the best solution and how a set of alternatives, acceptable to the decision maker (DM), may be identified with certain confidence.

Innovation Strategy in Slovak Businesses

The aim of the paper is based on detailed analysis of literary sources and carried out research to develop a model development and implementation of innovation strategy in the business. The paper brings the main results of the authors conducted research on a sample of 462 respondents that shows the current situation in the Slovak enterprises in the use of innovation strategy. Carried out research and analysis provided the base for a model development and implementation of innovation strategy in the business, which is in the paper in detail, step by step explained with emphasis on the implementation process. Implementing the innovation strategy is described a separate model. Paper contains recommendations for successful implementation of innovation strategy in the business. These recommendations should serve mainly business managers as valuable tool in implementing the innovation strategy.

An Effective Hybrid Genetic Algorithm for Job Shop Scheduling Problem

The job shop scheduling problem (JSSP) is well known as one of the most difficult combinatorial optimization problems. This paper presents a hybrid genetic algorithm for the JSSP with the objective of minimizing makespan. The efficiency of the genetic algorithm is enhanced by integrating it with a local search method. The chromosome representation of the problem is based on operations. Schedules are constructed using a procedure that generates full active schedules. In each generation, a local search heuristic based on Nowicki and Smutnicki-s neighborhood is applied to improve the solutions. The approach is tested on a set of standard instances taken from the literature and compared with other approaches. The computation results validate the effectiveness of the proposed algorithm.

CFD Analysis of Two Phase Flow in a Horizontal Pipe – Prediction of Pressure Drop

In designing of condensers, the prediction of pressure drop is as important as the prediction of heat transfer coefficient. Modeling of two phase flow, particularly liquid – vapor flow under diabatic conditions inside a horizontal tube using CFD analysis is difficult with the available two phase models in FLUENT due to continuously changing flow patterns. In the present analysis, CFD analysis of two phase flow of refrigerants inside a horizontal tube of inner diameter, 0.0085 m and 1.2 m length is carried out using homogeneous model under adiabatic conditions. The refrigerants considered are R22, R134a and R407C. The analysis is performed at different saturation temperatures and at different flow rates to evaluate the local frictional pressure drop. Using Homogeneous model, average properties are obtained for each of the refrigerants that is considered as single phase pseudo fluid. The so obtained pressure drop data is compared with the separated flow models available in literature.

Bandwidth Allocation for ABR Service in Cellular Networks

Available Bit Rate Service (ABR) is the lower priority service and the better service for the transmission of data. On wireline ATM networks ABR source is always getting the feedback from switches about increase or decrease of bandwidth according to the changing network conditions and minimum bandwidth is guaranteed. In wireless networks guaranteeing the minimum bandwidth is really a challenging task as the source is always in mobile and traveling from one cell to another cell. Re establishment of virtual circuits from start to end every time causes the delay in transmission. In our proposed solution we proposed the mechanism to provide more available bandwidth to the ABR source by re-usage of part of old Virtual Channels and establishing the new ones. We want the ABR source to transmit the data continuously (non-stop) inorderto avoid the delay. In worst case scenario at least minimum bandwidth is to be allocated. In order to keep the data flow continuously, priority is given to the handoff ABR call against new ABR call.

Image Segmentation Based on Graph Theoretical Approach to Improve the Quality of Image Segmentation

Graph based image segmentation techniques are considered to be one of the most efficient segmentation techniques which are mainly used as time & space efficient methods for real time applications. How ever, there is need to focus on improving the quality of segmented images obtained from the earlier graph based methods. This paper proposes an improvement to the graph based image segmentation methods already described in the literature. We contribute to the existing method by proposing the use of a weighted Euclidean distance to calculate the edge weight which is the key element in building the graph. We also propose a slight modification of the segmentation method already described in the literature, which results in selection of more prominent edges in the graph. The experimental results show the improvement in the segmentation quality as compared to the methods that already exist, with a slight compromise in efficiency.

Learning Style and Learner Satisfaction in a Course Delivery Context

This paper describes the results and implications of a correlational study of learning styles and learner satisfaction. The relationship of these empirical concepts was examined in the context of traditional versus e-blended modes of course delivery in an introductory graduate research course. Significant results indicated that the visual side of the visual-verbal dimension of students- learning style(s) was positively correlated to satisfaction with themselves as learners in an e-blended course delivery mode and negatively correlated to satisfaction with the classroom environment in the context of a traditional classroom course delivery mode.

Decomposition Method for Neural Multiclass Classification Problem

In this article we are going to discuss the improvement of the multi classes- classification problem using multi layer Perceptron. The considered approach consists in breaking down the n-class problem into two-classes- subproblems. The training of each two-class subproblem is made independently; as for the phase of test, we are going to confront a vector that we want to classify to all two classes- models, the elected class will be the strongest one that won-t lose any competition with the other classes. Rates of recognition gotten with the multi class-s approach by two-class-s decomposition are clearly better that those gotten by the simple multi class-s approach.