An Approximate Solution of the Classical Van der Pol Oscillator Coupled Gyroscopically to a Linear Oscillator Using Parameter-Expansion Method

In this article, we are dealing with a model consisting of a classical Van der Pol oscillator coupled gyroscopically to a linear oscillator. The major problem is analyzed. The regular dynamics of the system is considered using analytical methods. In this case, we provide an approximate solution for this system using parameter-expansion method. Also, we find approximate values for frequencies of the system. In parameter-expansion method the solution and unknown frequency of oscillation are expanded in a series by a bookkeeping parameter. By imposing the non-secularity condition at each order in the expansion the method provides different approximations to both the solution and the frequency of oscillation. One iteration step provides an approximate solution which is valid for the whole solution domain.

Sperm Identification Using Elliptic Model and Tail Detection

The conventional assessment of human semen is a highly subjective assessment, with considerable intra- and interlaboratory variability. Computer-Assisted Sperm Analysis (CASA) systems provide a rapid and automated assessment of the sperm characteristics, together with improved standardization and quality control. However, the outcome of CASA systems is sensitive to the method of experimentation. While conventional CASA systems use digital microscopes with phase-contrast accessories, producing higher contrast images, we have used raw semen samples (no staining materials) and a regular light microscope, with a digital camera directly attached to its eyepiece, to insure cost benefits and simple assembling of the system. However, since the accurate finding of sperms in the semen image is the first step in the examination and analysis of the semen, any error in this step can affect the outcome of the analysis. This article introduces and explains an algorithm for finding sperms in low contrast images: First, an image enhancement algorithm is applied to remove extra particles from the image. Then, the foreground particles (including sperms and round cells) are segmented form the background. Finally, based on certain features and criteria, sperms are separated from other cells.

Efficient Method for ECG Compression Using Two Dimensional Multiwavelet Transform

In this paper we introduce an effective ECG compression algorithm based on two dimensional multiwavelet transform. Multiwavelets offer simultaneous orthogonality, symmetry and short support, which is not possible with scalar two-channel wavelet systems. These features are known to be important in signal processing. Thus multiwavelet offers the possibility of superior performance for image processing applications. The SPIHT algorithm has achieved notable success in still image coding. We suggested applying SPIHT algorithm to 2-D multiwavelet transform of2-D arranged ECG signals. Experiments on selected records of ECG from MIT-BIH arrhythmia database revealed that the proposed algorithm is significantly more efficient in comparison with previously proposed ECG compression schemes.

Evolutionary Feature Selection for Text Documents using the SVM

Text categorization is the problem of classifying text documents into a set of predefined classes. After a preprocessing step, the documents are typically represented as large sparse vectors. When training classifiers on large collections of documents, both the time and memory restrictions can be quite prohibitive. This justifies the application of feature selection methods to reduce the dimensionality of the document-representation vector. In this paper, we present three feature selection methods: Information Gain, Support Vector Machine feature selection called (SVM_FS) and Genetic Algorithm with SVM (called GA_SVM). We show that the best results were obtained with GA_SVM method for a relatively small dimension of the feature vector.

Testing Loaded Programs Using Fault Injection Technique

Fault tolerance is critical in many of today's large computer systems. This paper focuses on improving fault tolerance through testing. Moreover, it concentrates on the memory faults: how to access the editable part of a process memory space and how this part is affected. A special Software Fault Injection Technique (SFIT) is proposed for this purpose. This is done by sequentially scanning the memory of the target process, and trying to edit maximum number of bytes inside that memory. The technique was implemented and tested on a group of programs in software packages such as jet-audio, Notepad, Microsoft Word, Microsoft Excel, and Microsoft Outlook. The results from the test sample process indicate that the size of the scanned area depends on several factors. These factors are: process size, process type, and virtual memory size of the machine under test. The results show that increasing the process size will increase the scanned memory space. They also show that input-output processes have more scanned area size than other processes. Increasing the virtual memory size will also affect the size of the scanned area but to a certain limit.

The Shaping of a Triangle Steel Plate into an Equilateral Vertical Steel by Finite-Element Modeling

The orthogonal processes to shape the triangle steel plate into a equilateral vertical steel are examined by an incremental elasto-plastic finite-element method based on an updated Lagrangian formulation. The highly non-linear problems due to the geometric changes, the inelastic constitutive behavior and the boundary conditions varied with deformation are taken into account in an incremental manner. On the contact boundary, a modified Coulomb friction mode is specially considered. A weighting factor r-minimum is employed to limit the step size of loading increment to linear relation. In particular, selective reduced integration was adopted to formulate the stiffness matrix. The simulated geometries of verticality could clearly demonstrate the vertical processes until unloading. A series of experiments and simulations were performed to validate the formulation in the theory, leading to the development of the computer codes. The whole deformation history and the distribution of stress, strain and thickness during the forming process were obtained by carefully considering the moving boundary condition in the finite-element method. Therefore, this modeling can be used for judging whether a equilateral vertical steel can be shaped successfully. The present work may be expected to improve the understanding of the formation of the equilateral vertical steel.

Thermal and Morphological Evaluation of Chemically Pretreated Sugarcane Bagasse

Enzymatic hydrolysis is one of the major steps involved in the conversion from sugarcane bagasse to yield ethanol. This process offers potential for yields and selectivity higher, lower energy costs and milder operating conditions than chemical processes. However, the presence of some factors such as lignin content, crystallinity degree of the cellulose, and particle sizes, limits the digestibility of the cellulose present in the lignocellulosic biomasses. Pretreatment aims to improve the access of the enzyme to the substrate. In this study sugarcane bagasse was submitted chemical pretreatment that consisted of two consecutive steps, the first with dilute sulfuric acid (1 % (v/v) H2SO4), and the second with alkaline solutions with different concentrations of NaOH (1, 2, 3 and 4 % (w/v)). Thermal Analysis (TG/ DTG and DTA) was used to evaluate hemicellulose, cellulose and lignin contents in the samples. Scanning Electron Microscopy (SEM) was used to evaluate the morphological structures of the in natura and chemically treated samples. Results showed that pretreatments were effective in chemical degradation of lignocellulosic materials of the samples, and also was possible to observe the morphological changes occurring in the biomasses after pretreatments.

Controller Design of Discrete Systems by Order Reduction Technique Employing Differential Evolution Optimization Algorithm

One of the main objectives of order reduction is to design a controller of lower order which can effectively control the original high order system so that the overall system is of lower order and easy to understand. In this paper, a simple method is presented for controller design of a higher order discrete system. First the original higher order discrete system in reduced to a lower order model. Then a Proportional Integral Derivative (PID) controller is designed for lower order model. An error minimization technique is employed for both order reduction and controller design. For the error minimization purpose, Differential Evolution (DE) optimization algorithm has been employed. DE method is based on the minimization of the Integral Squared Error (ISE) between the desired response and actual response pertaining to a unit step input. Finally the designed PID controller is connected to the original higher order discrete system to get the desired specification. The validity of the proposed method is illustrated through a numerical example.

Phase Error Accumulation Methodology for On-Chip Cell Characterization

This paper describes the design of new method of propagation delay measurement in micro and nanostructures during characterization of ASIC standard library cell. Providing more accuracy timing information about library cell to the design team we can improve a quality of timing analysis inside of ASIC design flow process. Also, this information could be very useful for semiconductor foundry team to make correction in technology process. By comparison of the propagation delay in the CMOS element and result of analog SPICE simulation. It was implemented as digital IP core for semiconductor manufacturing process. Specialized method helps to observe the propagation time delay in one element of the standard-cell library with up-to picoseconds accuracy and less. Thus, the special useful solutions for VLSI schematic to parameters extraction, basic cell layout verification, design simulation and verification are announced.

Design for Manufacturability and Concurrent Engineering for Product Development

In the 1980s, companies began to feel the effect of three major influences on their product development: newer and innovative technologies, increasing product complexity and larger organizations. And therefore companies were forced to look for new product development methods. This paper tries to focus on the two of new product development methods (DFM and CE). The aim of this paper is to see and analyze different product development methods specifically on Design for Manufacturability and Concurrent Engineering. Companies can achieve and be benefited by minimizing product life cycle, cost and meeting delivery schedule. This paper also presents simplified models that can be modified and used by different companies based on the companies- objective and requirements. Methodologies that are followed to do this research are case studies. Two companies were taken and analysed on the product development process. Historical data, interview were conducted on these companies in addition to that, Survey of literatures and previous research works on similar topics has been done during this research. This paper also tries to show the implementation cost benefit analysis and tries to calculate the implementation time. From this research, it has been found that the two companies did not achieve the delivery time to the customer. Some of most frequently coming products are analyzed and 50% to 80 % of their products are not delivered on time to the customers. The companies are following the traditional way of product development that is sequentially design and production method, which highly affect time to market. In the case study it is found that by implementing these new methods and by forming multi disciplinary team in designing and quality inspection; the company can reduce the workflow steps from 40 to 30.

Feature Based Unsupervised Intrusion Detection

The goal of a network-based intrusion detection system is to classify activities of network traffics into two major categories: normal and attack (intrusive) activities. Nowadays, data mining and machine learning plays an important role in many sciences; including intrusion detection system (IDS) using both supervised and unsupervised techniques. However, one of the essential steps of data mining is feature selection that helps in improving the efficiency, performance and prediction rate of proposed approach. This paper applies unsupervised K-means clustering algorithm with information gain (IG) for feature selection and reduction to build a network intrusion detection system. For our experimental analysis, we have used the new NSL-KDD dataset, which is a modified dataset for KDDCup 1999 intrusion detection benchmark dataset. With a split of 60.0% for the training set and the remainder for the testing set, a 2 class classifications have been implemented (Normal, Attack). Weka framework which is a java based open source software consists of a collection of machine learning algorithms for data mining tasks has been used in the testing process. The experimental results show that the proposed approach is very accurate with low false positive rate and high true positive rate and it takes less learning time in comparison with using the full features of the dataset with the same algorithm.

Accounting Research from the Globalization Perspective

This paper explores the idea of globalisation and considers accounting-s role in that process in order to develop new spaces for accounting research. That-s why in this paper we are looking for questions not necessary for answers. Adopting an 'alternative' view of accounting it-s related to the fact that we sees accounting as social and evolutionist process, that pays heed to those voices arguing for greater social and environmental justice, and that draws attention to the role of accounting researchers in the process of globalisation. The paper defines globalisation and expands the globalisation and accounting research agenda introducing in this context the harmonization process in accounting. There are the two main systems which are disputing the first stage of being the benchmark: GAAP and IFRS. Each of them has his pluses and minuses on being the selected one. Due to this fact a convergence of the two, joining the advantages and disadvantages of the two should be the solution for an unique international accounting solution. Is this idea realizable, what steps has been made until now, what should be done in the future. The paper is emphasising the role of the cultural differences in the process of imposing of an unique international accounting system by the global organizations..

Three-Level Tracking Method for Animating a 3D Humanoid Character

With a rapid growth in 3D graphics technology over the last few years, people are desired to see more flexible reacting motions of a biped in animations. In particular, it is impossible to anticipate all reacting motions of a biped while facing a perturbation. In this paper, we propose a three-level tracking method for animating a 3D humanoid character. First, we take the laws of physics into account to attach physical attributes, such as mass, gravity, friction, collision, contact, and torque, to bones and joints of a character. The next step is to employ PD controller to follow a reference motion as closely as possible. Once the character cannot tolerate a strong perturbation to prevent itself from falling down, we are capable of tracking a desirable falling-down action to avoid any falling condition inaccuracy. From the experimental results, we demonstrate the effectiveness and flexibility of the proposed method in comparison with conventional data-driven approaches.

New Simultaneous High Performance Liquid Chromatographic Method for Determination of NSAIDs and Opioid Analgesics in Advanced Drug Delivery Systems and Human Plasma

A new and cost effective RP-HPLC method was developed and validated for simultaneous analysis of non steroidal anti inflammatory dugs Diclofenac sodium (DFS), Flurbiprofen (FLP) and an opioid analgesic Tramadol (TMD) in advanced drug delivery systems (Liposome and Microcapsules), marketed brands and human plasma. Isocratic system was employed for the flow of mobile phase consisting of 10 mM sodium dihydrogen phosphate buffer and acetonitrile in molar ratio of 67: 33 with adjusted pH of 3.2. The stationary phase was hypersil ODS column (C18, 250×4.6 mm i.d., 5 μm) with controlled temperature of 30 C°. DFS in liposomes, microcapsules and marketed drug products was determined in range of 99.76-99.84%. FLP and TMD in microcapsules and brands formulation were 99.78 - 99.94 % and 99.80 - 99.82 %, respectively. Single step liquid-liquid extraction procedure using combination of acetonitrile and trichloroacetic acid (TCA) as protein precipitating agent was employed. The detection limits (at S/N ratio 3) of quality control solutions and plasma samples were 10, 20, and 20 ng/ml for DFS, FLP and TMD, respectively. The Assay was acceptable in linear dynamic range. All other validation parameters were found in limits of FDA and ICH method validation guidelines. The proposed method is sensitive, accurate and precise and could be applicable for routine analysis in pharmaceutical industry as well as in human plasma samples for bioequivalence and pharmacokinetics studies.

Synthesis and Characterization of Gallosilicate Sodalite Containing NO2- Ions

Pure phase gallosilicate nitrite sodalite has been synthesized in a single step by low temperature (373 oK) hydrothermal technique. The product obtained was characterized using a combination of techniques including X-ray powder diffraction, IR, Raman spectroscopy, SEM, MAS NMR spectroscopy as well as thermogravimetry. Sodalite with an ideal composition was obtained after synthesis at 3730K and seven days duration using alkaline medium. The structural features of the Na8[GaSiO4]6(NO2)2 sodalite were investigated by IR, MAS NMR spectroscopy of 29Si and 23Na nuclei and by Reitveld refinement of X-ray powder diffraction data. The crystal structure of this sodalite has been refined in the space group P 4 3n; with a cell parameter 8.98386Å, V= 726.9 Å, (Rwp= 0.077 and Rp=0.0537) and Si-O-Ga angle is found to be 132.920 . MAS NMR study confirms complete ordering of Si and Ga in the gallosilicate framework. The surface area of single entity with stoichiometry Na8[GaSiO4]6(NO2)2 was found to be 8.083 x10-15 cm2/g.

Pedometer Development Utilizing an Accelerometer Sensor

This paper develops a pedometer with a three-axis acceleration sensor that can be placed with any angle. The proposed pedometer measures the number of steps while users walk, jog or run. It can be worn on users’ waistband or placed within pocket or backpack. The work address to improve on the general pedometers, which can only be used in a single direction or can only count of steps without the continuous exercise judgment mechanism. Finally, experimental results confirm the superior performance of the proposed pedometer.

Impact of the Decoder Connection Schemes on Iterative Decoding of GPCB Codes

In this paper we present a study of the impact of connection schemes on the performance of iterative decoding of Generalized Parallel Concatenated block (GPCB) constructed from one step majority logic decodable (OSMLD) codes and we propose a new connection scheme for decoding them. All iterative decoding connection schemes use a soft-input soft-output threshold decoding algorithm as a component decoder. Numerical result for GPCB codes transmitted over Additive White Gaussian Noise (AWGN) channel are provided. It will show that the proposed scheme is better than Hagenauer-s scheme and Lucas-s scheme [1] and slightly better than the Pyndiah-s scheme.

Comparison of the Existing Methods in Determination of the Characteristic Polynomial

This paper presents comparison among methods of determination of the characteristic polynomial coefficients. First, the resultant systems from the methods are compared based on frequency criteria such as the closed loop bandwidth, gain and phase margins. Then the step responses of the resultant systems are compared on the basis of the transient behavior criteria including overshoot, rise time, settling time and error (via IAE, ITAE, ISE and ITSE integral indices). Also relative stability of the systems is compared together. Finally the best choices in regards to the above diverse criteria are presented.

Development of a New CFD Multi-Coupling Tool Based on Immersed Boundary Method: toward SRM Analysis

The ongoing effort to develop an in-house compressible solver with multi-disciplinary physics is presented in this paper. Basic compressible solver combined with IBM technique provides us an effective numerical tool able to tackle the physics phenomena and especially physic phenomena involved in Solid Rocket Motors (SRMs). Main principles are introduced step by step describing its implementation. This paper sheds light on the whole potentiality of our proposed numerical model and we strongly believe a way to introduce multi-physics mechanisms strongly coupled is opened to ablation in nozzle, fluid/structure interaction and burning propellant surface with time.

Integrating Low and High Level Object Recognition Steps by Probabilistic Networks

In pattern recognition applications the low level segmentation and the high level object recognition are generally considered as two separate steps. The paper presents a method that bridges the gap between the low and the high level object recognition. It is based on a Bayesian network representation and network propagation algorithm. At the low level it uses hierarchical structure of quadratic spline wavelet image bases. The method is demonstrated for a simple circuit diagram component identification problem.