Realization of Design Features for Linear Flow Splitting in NX 6

Within the collaborative research center 666 a new product development approach and the innovative manufacturing method of linear flow splitting are being developed. So far the design process is supported by 3D-CAD models utilizing User Defined Features in standard CAD-Systems. This paper now presents new functions for generating 3D-models of integral sheet metal products with bifurcations using Siemens PLM NX 6. The emphasis is placed on design and semi-automated insertion of User Defined Features. Therefore User Defined Features for both, linear flow splitting and its derivative linear bend splitting, were developed. In order to facilitate the modeling process, an application was developed that guides through the insertion process. Its usability and dialog layout adapt known standard features. The work presented here has significant implications on the quality, accurateness and efficiency of the product generation process of sheet metal products with higher order bifurcations.

Enhanced-Delivery Overlay Multicasting Scheme by Optimizing Bandwidth and Latency Discrepancy Ratios

With optimized bandwidth and latency discrepancy ratios, Node Gain Scores (NGSs) are determined and used as a basis for shaping the max-heap overlay. The NGSs - determined as the respective bandwidth-latency-products - govern the construction of max-heap-form overlays. Each NGS is earned as a synergy of discrepancy ratio of the bandwidth requested with respect to the estimated available bandwidth, and latency discrepancy ratio between the nodes and the source node. The tree leads to enhanceddelivery overlay multicasting – increasing packet delivery which could, otherwise, be hindered by induced packet loss occurring in other schemes not considering the synergy of these parameters on placing the nodes on the overlays. The NGS is a function of four main parameters – estimated available bandwidth, Ba; individual node's requested bandwidth, Br; proposed node latency to its prospective parent (Lp); and suggested best latency as advised by source node (Lb). Bandwidth discrepancy ratio (BDR) and latency discrepancy ratio (LDR) carry weights of α and (1,000 - α ) , respectively, with arbitrary chosen α ranging between 0 and 1,000 to ensure that the NGS values, used as node IDs, maintain a good possibility of uniqueness and balance between the most critical factor between the BDR and the LDR. A max-heap-form tree is constructed with assumption that all nodes possess NGS less than the source node. To maintain a sense of load balance, children of each level's siblings are evenly distributed such that a node can not accept a second child, and so on, until all its siblings able to do so, have already acquired the same number of children. That is so logically done from left to right in a conceptual overlay tree. The records of the pair-wise approximate available bandwidths as measured by a pathChirp scheme at individual nodes are maintained. Evaluation measures as compared to other schemes – Bandwidth Aware multicaSt architecturE (BASE), Tree Building Control Protocol (TBCP), and Host Multicast Tree Protocol (HMTP) - have been conducted. This new scheme generally performs better in terms of trade-off between packet delivery ratio; link stress; control overhead; and end-to-end delays.

Software Development Processes Maturity versus Software Processes and Products Measurement

Unsatisfactory effectiveness of software systems development and enhancement projects is one of the main reasons why in software engineering there are attempts being made to use experiences coming from other engineering disciplines. In spite of specificity of software product and process a belief had come out that the execution of software could be more effective if these objects were subject to measurement – as it is true in other engineering disciplines for which measurement is an immanent feature. Thus objective and reliable approaches to the measurement of software processes and products have been sought in software engineering for several dozens of years already. This may be proved, among others, by the current version of CMMI for Development model. This paper is aimed at analyzing the approach to the software processes and products measurement proposed in the latest version of this very model, indicating growing acceptance for this issue in software engineering.

Using the Monte Carlo Simulation to Predict the Assembly Yield

Electronics Products that achieve high levels of integrated communications, computing and entertainment, multimedia features in small, stylish and robust new form factors are winning in the market place. Due to the high costs that an industry may undergo and how a high yield is directly proportional to high profits, IC (Integrated Circuit) manufacturers struggle to maximize yield, but today-s customers demand miniaturization, low costs, high performance and excellent reliability making the yield maximization a never ending research of an enhanced assembly process. With factors such as minimum tolerances, tighter parameter variations a systematic approach is needed in order to predict the assembly process. In order to evaluate the quality of upcoming circuits, yield models are used which not only predict manufacturing costs but also provide vital information in order to ease the process of correction when the yields fall below expectations. For an IC manufacturer to obtain higher assembly yields all factors such as boards, placement, components, the material from which the components are made of and processes must be taken into consideration. Effective placement yield depends heavily on machine accuracy and the vision of the system which needs the ability to recognize the features on the board and component to place the device accurately on the pads and bumps of the PCB. There are currently two methods for accurate positioning, using the edge of the package and using solder ball locations also called footprints. The only assumption that a yield model makes is that all boards and devices are completely functional. This paper will focus on the Monte Carlo method which consists in a class of computational algorithms (information processed algorithms) which depends on repeated random samplings in order to compute the results. This method utilized in order to recreate the simulation of placement and assembly processes within a production line.

Bioefficacy of Some Oil-Mixed Plant Derivatives against African Mud Catfish (Clarias gariepinus) Beetles, Dermestes maculatus and Necrobia rufipes

The efficacy of the separate mixing of four tropical spicy and medicinal plant products: Dennettia tripetala Baker (pepper fruit), Eugenia aromatica Hook (clove), Piper guineense (Schum and Thonn) (black pepper) and Monodora myristica (Dunal) (African nut-meg) with a household vegetable oil was evaluated under tropical storage conditions for the control and reproductive performance of Dermestes maculatus (De Geer) (hide beetle) and Necroba rufipes (De Geer) (copra beetle) on African catfish, Clarias gariepinus (Burchell). Each of the plant materials was pulverized into powder and applied as a mix of 1ml of oil and plant powder at 2.5, 5.0, 7.5 and 10.0g per 100g of dried fish, and allowed to dry for 6h. Each of the four oil-mixed powder treatments evoked significant (P < 05) mortalities of the two insects compared with the control (oil only) at 1, 3 and 7 days post treatment. The oil-powder mixture dosages did not prevent insect egg hatchability but while the emergent larvae on the treated samples died, the emergent larvae in the control survived into adults. The application of oil-mixed powders effectively suppressed the emergence of the larvae of the beetles. Similarly, each of the oil-powder mixtures significantly reduced weight loss in smoked fish that were exposed to D. maculatus and N. rufipes when compared to the control (P < 05). The results of this study suggest that the plant powders rather than the domestic oil demonstrated protective ability against the fish beetles and confirm the efficacy of the plant products as pest control agents.

Interoperable CNC System for Turning Operations

The changing economic climate has made global manufacturing a growing reality over the last decade, forcing companies from east and west and all over the world to collaborate beyond geographic boundaries in the design, manufacture and assemble of products. The ISO10303 and ISO14649 Standards (STEP and STEP-NC) have been developed to introduce interoperability into manufacturing enterprises so as to meet the challenge of responding to production on demand. This paper describes and illustrates a STEP compliant CAD/CAPP/CAM System for the manufacture of rotational parts on CNC turning centers. The information models to support the proposed system together with the data models defined in the ISO14649 standard used to create the NC programs are also described. A structured view of a STEP compliant CAD/CAPP/CAM system framework supporting the next generation of intelligent CNC controllers for turn/mill component manufacture is provided. Finally a proposed computational environment for a STEP-NC compliant system for turning operations (SCSTO) is described. SCSTO is the experimental part of the research supported by the specification of information models and constructed using a structured methodology and object-oriented methods. SCSTO was developed to generate a Part 21 file based on machining features to support the interactive generation of process plans utilizing feature extraction. A case study component has been developed to prove the concept for using the milling and turning parts of ISO14649 to provide a turn-mill CAD/CAPP/CAM environment.

Open Source Library Management System Software: A Review

Library management systems are commonly used in all educational related institutes. Many commercial products are available. However, many institutions may not be able to afford the cost of using commercial products. Therefore, an alternative solution in such situations would be open source software. This paper is focusing on reviewing open source library management system packages currently available. The review will focus on the abilities to perform four basic components which are traditional services, interlibrary load management, managing electronic materials and basic common management system such as security, alert system and statistical reports. In addition, environment, basic requirement and supporting aspects of each open source package are also mentioned.

A Hybrid Scheme for on-Line Diagnostic Decision Making Using Optimal Data Representation and Filtering Technique

The early diagnostic decision making in industrial processes is absolutely necessary to produce high quality final products. It helps to provide early warning for a special event in a process, and finding its assignable cause can be obtained. This work presents a hybrid diagnostic schmes for batch processes. Nonlinear representation of raw process data is combined with classification tree techniques. The nonlinear kernel-based dimension reduction is executed for nonlinear classification decision boundaries for fault classes. In order to enhance diagnosis performance for batch processes, filtering of the data is performed to get rid of the irrelevant information of the process data. For the diagnosis performance of several representation, filtering, and future observation estimation methods, four diagnostic schemes are evaluated. In this work, the performance of the presented diagnosis schemes is demonstrated using batch process data.

A Study of Replacement Policies for Warranty Products with Different Failure Rate

This paper provides a replacement policy for warranty products with different failure rate from the consumer-s viewpoint. Assume that the product is replaced once within a finite planning horizon, and the failure rate of the second product is lower than the failure rate of the first product. Within warranty period (WP), the failed product is corrected by minimal repair without any cost to the consumers. After WP, the failed product is repaired with a fixed repair cost to the consumers. However, each failure incurs a fixed downtime cost to the consumers over a finite planning horizon. In this paper, we derive the model of the expected total disbursement cost within a finite planning horizon and some properties of the optimal replacement policy under some reasonable conditions are obtained. Finally, numerical examples are given to illustrate the features of the optimal replacement policy under various maintenance costs.

A Refined Application of QFD in SCM, A New Approach

Due to the fact that in the new century customers tend to express globally increasing demands, networks of interconnected businesses have been established in societies and the management of such networks seems to be a major key through gaining competitive advantages. Supply chain management encompasses such managerial activities. Within a supply chain, a critical role is played by quality. QFD is a widely-utilized tool which serves the purpose of not only bringing quality to the ultimate provision of products or service packages required by the end customer or the retailer, but it can also initiate us into a satisfactory relationship with our initial customer; that is the wholesaler. However, the wholesalers- cooperation is considerably based on the capabilities that are heavily dependent on their locations and existing circumstances. Therefore, it is undeniable that for all companies each wholesaler possesses a specific importance ratio which can heavily influence the figures calculated in the House of Quality in QFD. Moreover, due to the competitiveness of the marketplace today, it-s been widely recognized that consumers- expression of demands has been highly volatile in periods of production. Apparently, such instability and proneness to change has been very tangibly noticed and taking it into account during the analysis of HOQ is widely influential and doubtlessly required. For a more reliable outcome in such matters, this article demonstrates the application viability of Analytic Network Process for considering the wholesalers- reputation and simultaneously introduces a mortality coefficient for the reliability and stability of the consumers- expressed demands in course of time. Following to this, the paper provides further elaboration on the relevant contributory factors and approaches through the calculation of such coefficients. In the end, the article concludes that an empirical application is needed to achieve broader validity.

Application of the Virtual Reality Modeling Language for Design of Automated Workplaces

Virtual Reality Modelling Language (VRML) is description language, which belongs to a field Window on World virtual reality system. The file, which is in VRML format, can be interpreted by VRML explorer in three-dimensional scene. VRML was created with aim to represent virtual reality on Internet easier. Development of 3D graphic is connected with Silicon Graphic Corporation. VRML 2.0 is the file format for describing interactive 3D scenes and objects. It can be used in collaboration with www, can be used for 3D complex representations creating of scenes, products or VR applications VRML 2.0 enables represent static and animated objects too. Interesting application of VRML is in area of manufacturing systems presentation.

A Multi-period Profit Maximization Policy for a Stochastic Demand Inventory System with Upward Substitution

This paper deals with a periodic-review substitutable inventory system for a finite and an infinite number of periods. Here an upward substitution structure, a substitution of a more costly item by a less costly one, is assumed, with two products. At the beginning of each period, a stochastic demand comes for the first item only, which is quality-wise better and hence costlier. Whenever an arriving demand finds zero inventory of this product, a fraction of unsatisfied customers goes for its substitutable second item. An optimal ordering policy has been derived for each period. The results are illustrated with numerical examples. A sensitivity analysis has been done to examine how sensitive the optimal solution and the maximum profit are to the values of the discount factor, when there is a large number of periods.

Six Sigma Process and its Impact on the Organizational Productivity

The six sigma method is a project-driven management approach to improve the organization-s products, services, and processes by continually reducing defects in the organization. Understanding the key features, obstacles, and shortcomings of the six sigma method allows organizations to better support their strategic directions, and increasing needs for coaching, mentoring, and training. It also provides opportunities to better implement six sigma projects. The purpose of this paper is the survey of six sigma process and its impact on the organizational productivity. So I have studied key concepts , problem solving process of six sigmaas well as the survey of important fields such as: DMAIC, six sigma and productivity applied programme, and other advantages of six sigma. In the end of this paper, present research conclusions. (direct and positive relation between six sigma and productivity)

Consumer Product Demand Forecasting based on Artificial Neural Network and Support Vector Machine

The nature of consumer products causes the difficulty in forecasting the future demands and the accuracy of the forecasts significantly affects the overall performance of the supply chain system. In this study, two data mining methods, artificial neural network (ANN) and support vector machine (SVM), were utilized to predict the demand of consumer products. The training data used was the actual demand of six different products from a consumer product company in Thailand. The results indicated that SVM had a better forecast quality (in term of MAPE) than ANN in every category of products. Moreover, another important finding was the margin difference of MAPE from these two methods was significantly high when the data was highly correlated.

Evaluation and Analysis of Lean-Based Manufacturing Equipment and Technology System for Jordanian Industries

International markets driven forces are changing continuously, therefore companies need to gain a competitive edge in such markets. Improving the company's products, processes and practices is no longer auxiliary. Lean production is a production management philosophy that consolidates work tasks with minimum waste resulting in improved productivity. Lean production practices can be mapped into many production areas. One of these is Manufacturing Equipment and Technology (MET). Many lean production practices can be implemented in MET, namely, specific equipment configurations, total preventive maintenance, visual control, new equipment/ technologies, production process reengineering and shared vision of perfection.The purpose of this paper is to investigate the implementation level of these six practices in Jordanian industries. To achieve that a questionnaire survey has been designed according to five-point Likert scale. The questionnaire is validated through pilot study and through experts review. A sample of 350 Jordanian companies were surveyed, the response rate was 83%. The respondents were asked to rate the extent of implementation for each of practices. A relationship conceptual model is developed, hypotheses are proposed, and consequently the essential statistical analyses are then performed. An assessment tool that enables management to monitor the progress and the effectiveness of lean practices implementation is designed and presented. Consequently, the results show that the average implementation level of lean practices in MET is 77%, Jordanian companies are implementing successfully the considered lean production practices, and the presented model has Cronbach-s alpha value of 0.87 which is good evidence on model consistency and results validation.

Exploiting Machine Learning Techniques for the Enhancement of Acceptance Sampling

This paper proposes an innovative methodology for Acceptance Sampling by Variables, which is a particular category of Statistical Quality Control dealing with the assurance of products quality. Our contribution lies in the exploitation of machine learning techniques to address the complexity and remedy the drawbacks of existing approaches. More specifically, the proposed methodology exploits Artificial Neural Networks (ANNs) to aid decision making about the acceptance or rejection of an inspected sample. For any type of inspection, ANNs are trained by data from corresponding tables of a standard-s sampling plan schemes. Once trained, ANNs can give closed-form solutions for any acceptance quality level and sample size, thus leading to an automation of the reading of the sampling plan tables, without any need of compromise with the values of the specific standard chosen each time. The proposed methodology provides enough flexibility to quality control engineers during the inspection of their samples, allowing the consideration of specific needs, while it also reduces the time and the cost required for these inspections. Its applicability and advantages are demonstrated through two numerical examples.

Joint Optimization of Pricing and Advertisement for Seasonal Branded Products

The goal of this paper is to develop a model to integrate “pricing" and “advertisement" for short life cycle products, such as branded fashion clothing products. To achieve this goal, we apply the concept of “Dynamic Pricing". There are two classes of advertisements, for the brand (regardless of product) and for a particular product. Advertising the brand affects the demand and price of all the products. Thus, the model considers all these products in relation with each other. We develop two different methods to integrate both types of advertisement and pricing. The first model is developed within the framework of dynamic programming. However, due to the complexity of the model, this method cannot be applicable for large size problems. Therefore, we develop another method, called hieratical approach, which is capable of handling the real world problems. Finally, we show the accuracy of this method, both theoretically and also by simulation.

The Micro Ecosystem Restoration Mechanism Applied for Feasible Research of Lakes Eutrophication Enhancement

The technique of inducing micro ecosystem restoration is one of aquatic ecology engineering methods used to retrieve the polluted water. Batch scale study, pilot plant study, and field study were carried out to observe the eutrophication using the Inducing Ecology Restorative Symbiosis Agent (IERSA) consisting mainly degraded products by using lactobacillus, saccharomycete, and phycomycete. The results obtained from the experiments of the batch scale and pilot plant study allowed us to development the parameters for the field study. A pond, 5 m to the outlet of a lake, with an area of 500 m2 and depth of 0.6-1.2 m containing about 500 tons of water was selected as a model. After the treatment with 10 mg IERSA/L water twice a week for 70 days, the micro restoration mechanisms consisted of three stages (i.e., restoration, impact maintenance, and ecology recovery experiment after impact). The COD, TN, TKN, and chlorophyll a were reduced significantly in the first week. Although the unexpected heavy rain and contaminate from sewage system might slow the ecology restoration. However, the self-cleaning function continued and the chlorophyll a reduced for 50% in one month. In the 4th week, amoeba, paramecium, rotifer, and red wriggle worm reappeared, and the number of fish flies appeared up to1000 fish fries/m3. Those results proved that inducing restorative mechanism can be applied to improve the eutrophication and to control the growth of algae in the lakes by gaining the selfcleaning through inducing and competition of microbes. The situation for growth of fishes also can reach an excellent result due to the improvement of water quality.

An Experimental Investigation of Factors Affecting Consumers' Reactions to Mobile APP-Based Promotions

The purpose of this study is to understand how consumers react to a company's promotional offers with mobile applications (APP) as premiums. This paper presents the results of an experimental study where five features of APP were involved: the cost (free/discounted) for earning APP, the relationship between APP and the promoted product, the perceived usefulness, the perceived ease of use, and the perceived playfulness of APP in the context of light foods purchase. The results support that the above features, except perceived ease of use, have substantial influences on consumers' intention to adopt the APP. Among the five features, the cost for earning APP has the most impact on the adopting intention of APP. The study also found a positive influence of adopting intention of APP on the consumer's purchase intention of the promoted product. Thus, APP-based premiums may enhance the consumer's purchase intention of a company's promoted products.

iDENTM Phones Automated Stress Testing

System testing is actually done to the entire system against the Functional Requirement Specification and/or the System Requirement Specification. Moreover, it is an investigatory testing phase, where the focus is to have almost a destructive attitude and test not only the design, but also the behavior and even the believed expectations of the customer. It is also intended to test up to and beyond the bounds defined in the software/hardware requirements specifications. In Motorola®, Automated Testing is one of the testing methodologies uses by GSG-iSGT (Global Software Group - iDEN TM Subcriber Group-Test) to increase the testing volume, productivity and reduce test cycle-time in iDEN TM phones testing. Testing is able to produce more robust products before release to the market. In this paper, iHopper is proposed as a tool to perform stress test on iDEN TM phonse. We will discuss the value that automation has brought to iDEN TM Phone testing such as improving software quality in the iDEN TM phone together with some metrics. We will also look into the advantages of the proposed system and some discussion of the future work as well.