Study of the Effect of Project Management on Manufacturing and Production Projects

In this article the accumulated results out of the effects and length of the manufacture and production projects in the university and research standard have been settled with the usefulness definition of the process of project management for the accessibility to the proportional pattern in the “time and action" stages. Studies show that many problems confronted by the researchers in these projects are connected to the non-profiting of: 1) autonomous timing for gathering the educational theme, 2) autonomous timing for planning and pattern, presenting before the construction, and 3) autonomous timing for manufacture and sample presentation from the output. The result of this study indicates the division of every manufacture and production projects into three smaller autonomous projects from its kind, budget and autonomous expenditure, shape and order of the stages for the management of these kinds of projects. In this case study real result are compared with theoretical results.

Ionic Liquid Pretreatment and Enzymatic Hydrolysis of Wood Biomass

Pretreatment of lignocellulosic biomass materials from poplar, acacia, oak, and fir with different ionic liquids (ILs) containing 1-alkyl-3-methyl-imidazolium cations and various anions has been carried out. The dissolved cellulose from biomass was precipitated by adding anti-solvents into the solution and vigorous stirring. Commercial cellulases Celluclast 1.5L and Accelerase 1000 have been used for hydrolysis of untreated and pretreated lignocellulosic biomass. Among the tested ILs, [Emim]COOCH3 showed the best efficiency, resulting in highest amount of liberated reducing sugars. Pretreatment of lignocellulosic biomass using glycerol-ionic liquids combined pretreatment and dilute acid-ionic liquids combined pretreatment were evaluated and compared with glycerol pretreatment, ionic liquids pretreatment and dilute acid pretreatment.

A Multi Objective Optimization Approach to Optimize Vehicle Ride and Handling Characteristics

Vehicle suspension design must fulfill some conflicting criteria. Among those is ride comfort which is attained by minimizing the acceleration transmitted to the sprung mass, via suspension spring and damper. Also good handling of a vehicle is a desirable property which requires stiff suspension and therefore is in contrast with a vehicle with good ride. Among the other desirable features of a suspension is the minimization of the maximum travel of suspension. This travel which is called suspension working space in vehicle dynamics literature is also a design constraint and it favors good ride. In this research a full car 8 degrees of freedom model has been developed and the three above mentioned criteria, namely: ride, handling and working space has been adopted as objective functions. The Multi Objective Programming (MOP) discipline has been used to find the Pareto Front and some reasoning used to chose a design point between these non dominated points of Pareto Front.

A New Approach for Prioritization of Failure Modes in Design FMEA using ANOVA

The traditional Failure Mode and Effects Analysis (FMEA) uses Risk Priority Number (RPN) to evaluate the risk level of a component or process. The RPN index is determined by calculating the product of severity, occurrence and detection indexes. The most critically debated disadvantage of this approach is that various sets of these three indexes may produce an identical value of RPN. This research paper seeks to address the drawbacks in traditional FMEA and to propose a new approach to overcome these shortcomings. The Risk Priority Code (RPC) is used to prioritize failure modes, when two or more failure modes have the same RPN. A new method is proposed to prioritize failure modes, when there is a disagreement in ranking scale for severity, occurrence and detection. An Analysis of Variance (ANOVA) is used to compare means of RPN values. SPSS (Statistical Package for the Social Sciences) statistical analysis package is used to analyze the data. The results presented are based on two case studies. It is found that the proposed new methodology/approach resolves the limitations of traditional FMEA approach.

Chaotic Response and Bifurcation Analysis of Gear-Bearing System with and without Porous Effect under Nonlinear Suspension

This study presents a systematic analysis of the dynamic behaviors of a gear-bearing system with porous squeeze film damper (PSFD) under nonlinear suspension, nonlinear oil-film force and nonlinear gear meshing force effect. It can be found that the system exhibits very rich forms of sub-harmonic and even the chaotic vibrations. The bifurcation diagrams also reveal that greater values of permeability may not only improve non-periodic motions effectively, but also suppress dynamic amplitudes of the system. Therefore, porous effect plays an important role to improve dynamic stability of gear-bearing systems or other mechanical systems. The results presented in this study provide some useful insights into the design and development of a gear-bearing system for rotating machinery that operates in highly rotational speed and highly nonlinear regimes.

Development of a Simple laser-based 2D Compensating System for the Contouring Accuracy of Machine Tools

The dynamical contouring error is a critical element for the accuracy of machine tools. The contouring error is defined as the difference between the processing actual path and commanded path, which is implemented by following the command curves from feeding driving system in machine tools. The contouring error is resulted from various factors, such as the external loads, friction, inertia moment, feed rate, speed control, servo control, and etc. Thus, the study proposes a 2D compensating system for the contouring accuracy of machine tools. Optical method is adopted by using stable frequency laser diode and the high precision position sensor detector (PSD) to performno-contact measurement. Results show the related accuracy of position sensor detector (PSD) of 2D contouring accuracy compensating system was ±1.5 μm for a calculated range of ±3 mm, and improvement accuracy is over 80% at high-speed feed rate.

Chemical Degradation of Dieldrin using Ferric Sulfide and Iron Powder

The chemical degradation of dieldrin in ferric sulfide and iron powder aqueous suspension was investigated in laboratory batch type experiments. To identify the reaction mechanism, reduced copper was used as reductant. More than 90% of dieldrin was degraded using both reaction systems after 29 days. Initial degradation rate of the pesticide using ferric sulfide was superior to that using iron powder. The reaction schemes were completely dissimilar even though the ferric ion plays an important role in both reaction systems. In the case of metallic iron powder, dieldrin undergoes partial dechlorination. This reaction proceeded by reductive hydrodechlorination with the generation of H+, which arise by oxidation of ferric iron. This reductive reaction was accelerated by reductant but mono-dechlorination intermediates were accumulated. On the other hand, oxidative degradation was observed in the reaction with ferric sulfide, and the stable chemical structure of dieldrin was decomposed into water-soluble intermediates. These reaction intermediates have no chemical structure of drin class. This dehalogenation reaction assumes to occur via the adsorbed hydroxyl radial generated on the surface of ferric sulfide.

Experiments and Modeling of Ion Exchange Resins for Nuclear Power Plants

Resins are used in nuclear power plants for water ultrapurification. Two approaches are considered in this work: column experiments and simulations. A software called OPTIPUR was developed, tested and used. The approach simulates the onedimensional reactive transport in porous medium with convectivedispersive transport between particles and diffusive transport within the boundary layer around the particles. The transfer limitation in the boundary layer is characterized by the mass transfer coefficient (MTC). The influences on MTC were measured experimentally. The variation of the inlet concentration does not influence the MTC; on the contrary of the Darcy velocity which influences. This is consistent with results obtained using the correlation of Dwivedi&Upadhyay. With the MTC, knowing the number of exchange site and the relative affinity, OPTIPUR can simulate the column outlet concentration versus time. Then, the duration of use of resins can be predicted in conditions of a binary exchange.

A New Preconditioned AOR Method for Z-matrices

In this paper, we present a preconditioned AOR-type iterative method for solving the linear systems Ax = b, where A is a Z-matrix. And give some comparison theorems to show that the rate of convergence of the preconditioned AOR-type iterative method is faster than the rate of convergence of the AOR-type iterative method.

Modeling of Material Removal on Machining of Ti-6Al-4V through EDM using Copper Tungsten Electrode and Positive Polarity

This paper deals optimized model to investigate the effects of peak current, pulse on time and pulse off time in EDM performance on material removal rate of titanium alloy utilizing copper tungsten as electrode and positive polarity of the electrode. The experiments are carried out on Ti6Al4V. Experiments were conducted by varying the peak current, pulse on time and pulse off time. A mathematical model is developed to correlate the influences of these variables and material removal rate of workpiece. Design of experiments (DOE) method and response surface methodology (RSM) techniques are implemented. The validity test of the fit and adequacy of the proposed models has been carried out through analysis of variance (ANOVA). The obtained results evidence that as the material removal rate increases as peak current and pulse on time increases. The effect of pulse off time on MRR changes with peak ampere. The optimum machining conditions in favor of material removal rate are verified and compared. The optimum machining conditions in favor of material removal rate are estimated and verified with proposed optimized results. It is observed that the developed model is within the limits of the agreeable error (about 4%) when compared to experimental results. This result leads to desirable material removal rate and economical industrial machining to optimize the input parameters.

Organizational De-Evolution; the Small Group or Single Actor Terrorist

Traditionally, terror groups have been formed by ideologically aligned actors who perceive a lack of options for achieving political or social change. However, terrorist attacks have been increasingly carried out by small groups of actors or lone individuals who may be only ideologically affiliated with larger, formal terrorist organizations. The formation of these groups represents the inverse of traditional organizational growth, whereby structural de-evolution within issue-based organizations leads to the formation of small, independent terror cells. Ideological franchising – the bypassing of formal affiliation to the “parent" organization – represents the de-evolution of traditional concepts of organizational structure in favor of an organic, independent, and focused unit. Traditional definitions of dark networks that are issue-based include focus on an identified goal, commitment to achieving this goal through unrestrained actions, and selection of symbolic targets. The next step in the de-evolution of small dark networks is the miniorganization, consisting of only a handful of actors working toward a common, violent goal. Information-sharing through social media platforms, coupled with civil liberties of democratic nations, provide the communication systems, access to information, and freedom of movement necessary for small dark networks to flourish without the aid of a parent organization. As attacks such as the 7/7 bombings demonstrate the effectiveness of small dark networks, terrorist actors will feel increasingly comfortable aligning with an ideology only, without formally organizing. The natural result of this de-evolving organization is the single actor event, where an individual seems to subscribe to a larger organization-s violent ideology with little or no formal ties.

Software Reliability Prediction Model Analysis

Software reliability prediction gives a great opportunity to measure the software failure rate at any point throughout system test. A software reliability prediction model provides with the technique for improving reliability. Software reliability is very important factor for estimating overall system reliability, which depends on the individual component reliabilities. It differs from hardware reliability in that it reflects the design perfection. Main reason of software reliability problems is high complexity of software. Various approaches can be used to improve the reliability of software. We focus on software reliability model in this article, assuming that there is a time redundancy, the value of which (the number of repeated transmission of basic blocks) can be an optimization parameter. We consider given mathematical model in the assumption that in the system may occur not only irreversible failures, but also a failure that can be taken as self-repairing failures that significantly affect the reliability and accuracy of information transfer. Main task of the given paper is to find a time distribution function (DF) of instructions sequence transmission, which consists of random number of basic blocks. We consider the system software unreliable; the time between adjacent failures has exponential distribution.

Analysis of Microalgae Lipids Isolated from Basin of Kazakhstan, to Assess the Prospects of Practical Use

It was analyzed of fatty acid composition of 16 strains of microalgae lipid fractions isolated from different basins of Kazakhstan and characterized by stable active growth in the laboratory. Three species of green microalgae (Oocystis rhomboideus, Chlorococcum infusionum, Dictyochlorella globosa) and three species of diatoms (Synedra sp., Nitzshia sp., Pleurosigma attenuatum) are characterized by a high content of lipids and are promising for further study as a source of polyunsaturated fatty acids.

Applying Complex Network Theory to Software Structure Analysis

Complex networks have been intensively studied across many fields, especially in Internet technology, biological engineering, and nonlinear science. Software is built up out of many interacting components at various levels of granularity, such as functions, classes, and packages, representing another important class of complex networks. It can also be studied using complex network theory. Over the last decade, many papers on the interdisciplinary research between software engineering and complex networks have been published. It provides a different dimension to our understanding of software and also is very useful for the design and development of software systems. This paper will explore how to use the complex network theory to analyze software structure, and briefly review the main advances in corresponding aspects.

Synergy in Vertical Transformations of Expert Designers

Existing literature ondesign reasoning seems to give either one sided accounts on expert design behaviour based on internal processing. In the same way ecological theoriesseem to focus one sidedly on external elementsthat result in a lack of unifying design cognition theory. Although current extended design cognition studies acknowledge the intellectual interaction between internal and external resources, there still seems to be insufficient understanding of the complexities involved in such interactive processes. As such,this paper proposes a novelmulti-directional model for design researchers tomap the complex and dynamic conduct controlling behaviour in which both the computational and ecological perspectives are integrated in a vertical manner. A clear distinction between identified intentional and emerging physical drivers, and relationships between them during the early phases of experts- design process, is demonstrated by presenting a case study in which the model was employed.

Improved Power Spectrum Estimation for RR-Interval Time Series

The RR interval series is non-stationary and unevenly spaced in time. For estimating its power spectral density (PSD) using traditional techniques like FFT, require resampling at uniform intervals. The researchers have used different interpolation techniques as resampling methods. All these resampling methods introduce the low pass filtering effect in the power spectrum. The lomb transform is a means of obtaining PSD estimates directly from irregularly sampled RR interval series, thus avoiding resampling. In this work, the superiority of Lomb transform method has been established over FFT based approach, after applying linear and cubicspline interpolation as resampling methods, in terms of reproduction of exact frequency locations as well as the relative magnitudes of each spectral component.

Treatment of Recycled Concrete Aggregates by Si-Based Polymers

The recycling of concrete, bricks and masonry rubble as concrete aggregates is an important way to contribute to a sustainable material flow. However, there are still various uncertainties limiting the widespread use of Recycled Concrete Aggregates (RCA). The fluctuations in the composition of grade recycled aggregates and their influence on the properties of fresh and hardened concrete are of particular concern regarding the use of RCA. Most of problems occurring while using recycled concrete aggregates as aggregates are due to higher porosity and hence higher water absorption, lower mechanical strengths, residual impurities on the surface of the RCA forming weaker bond between cement paste and aggregate. So, the reuse of RCA is still limited. Efficient polymer based treatment is proposed in order to reuse RCA easier. The silicon-based polymer treatments of RCA were carried out and were compared. This kind of treatment can improve the properties of RCA such as the rate of water absorption on treated RCA is significantly reduced.

A Study on the Differential Diagnostic Model for Newborn Hearing Loss Screening

According to the statistics, the prevalence of congenital hearing loss in Taiwan is approximately six thousandths; furthermore, one thousandths of infants have severe hearing impairment. Hearing ability during infancy has significant impact in the development of children-s oral expressions, language maturity, cognitive performance, education ability and social behaviors in the future. Although most children born with hearing impairment have sensorineural hearing loss, almost every child more or less still retains some residual hearing. If provided with a hearing aid or cochlear implant (a bionic ear) timely in addition to hearing speech training, even severely hearing-impaired children can still learn to talk. On the other hand, those who failed to be diagnosed and thus unable to begin hearing and speech rehabilitations on a timely manner might lose an important opportunity to live a complete and healthy life. Eventually, the lack of hearing and speaking ability will affect the development of both mental and physical functions, intelligence, and social adaptability. Not only will this problem result in an irreparable regret to the hearing-impaired child for the life time, but also create a heavy burden for the family and society. Therefore, it is necessary to establish a set of computer-assisted predictive model that can accurately detect and help diagnose newborn hearing loss so that early interventions can be provided timely to eliminate waste of medical resources. This study uses information from the neonatal database of the case hospital as the subjects, adopting two different analysis methods of using support vector machine (SVM) for model predictions and using logistic regression to conduct factor screening prior to model predictions in SVM to examine the results. The results indicate that prediction accuracy is as high as 96.43% when the factors are screened and selected through logistic regression. Hence, the model constructed in this study will have real help in clinical diagnosis for the physicians and actually beneficial to the early interventions of newborn hearing impairment.

Device for 3D Analysis of Basic Movements of the Lower Extremity

This document details the process of developing a wireless device that captures the basic movements of the foot (plantar flexion, dorsal flexion, abduction, adduction.), and the knee movement (flexion). It implements a motion capture system by using a hardware based on optical fiber sensors, due to the advantages in terms of scope, noise immunity and speed of data transmission and reception. The operating principle used by this system is the detection and transmission of joint movement by mechanical elements and their respective measurement by optical ones (in this case infrared). Likewise, Visual Basic software is used for reception, analysis and signal processing of data acquired by the device, generating a 3D graphical representation in real time of each movement. The result is a boot in charge of capturing the movement, a transmission module (Implementing Xbee Technology) and a receiver module for receiving information and sending it to the PC for their respective processing. The main idea with this device is to help on topics such as bioengineering and medicine, by helping to improve the quality of life and movement analysis.

A Fast Sign Localization System Using Discriminative Color Invariant Segmentation

Building intelligent traffic guide systems has been an interesting subject recently. A good system should be able to observe all important visual information to be able to analyze the context of the scene. To do so, signs in general, and traffic signs in particular, are usually taken into account as they contain rich information to these systems. Therefore, many researchers have put an effort on sign recognition field. Sign localization or sign detection is the most important step in the sign recognition process. This step filters out non informative area in the scene, and locates candidates in later steps. In this paper, we apply a new approach in detecting sign locations using a new color invariant model. Experiments are carried out with different datasets introduced in other works where authors claimed the difficulty in detecting signs under unfavorable imaging conditions. Our method is simple, fast and most importantly it gives a high detection rate in locating signs.