On Bounds For The Zeros of Univariate Polynomial

Problems on algebraical polynomials appear in many fields of mathematics and computer science. Especially the task of determining the roots of polynomials has been frequently investigated.Nonetheless, the task of locating the zeros of complex polynomials is still challenging. In this paper we deal with the location of zeros of univariate complex polynomials. We prove some novel upper bounds for the moduli of the zeros of complex polynomials. That means, we provide disks in the complex plane where all zeros of a complex polynomial are situated. Such bounds are extremely useful for obtaining a priori assertations regarding the location of zeros of polynomials. Based on the proven bounds and a test set of polynomials, we present an experimental study to examine which bound is optimal.

Utilization of 3-N-trimethylamino-1-propanol by Rhodococcus sp. strain A4 isolated from Natural Soil

The aim of this study was to screen for microorganism that able to utilize 3-N-trimethylamino-1-propanol (homocholine) as a sole source of carbon and nitrogen. The aerobic degradation of homocholine has been found by a gram-positive Rhodococcus sp. bacterium isolated from soil. The isolate was identified as Rhodococcus sp. strain A4 based on the phenotypic features, physiologic and biochemical characteristics, and phylogenetic analysis. The cells of the isolated strain grown on both basal-TMAP and nutrient agar medium displayed elementary branching mycelia fragmented into irregular rod and coccoid elements. Comparative 16S rDNA sequencing studies indicated that the strain A4 falls into the Rhodococcus erythropolis subclade and forms a monophyletic group with the type-strains of R. opacus, and R. wratislaviensis. Metabolites analysis by capillary electrophoresis, fast atom bombardment-mass spectrometry, and gas chromatography- mass spectrometry, showed trimethylamine (TMA) as the major metabolite beside β-alanine betaine and trimethylaminopropionaldehyde. Therefore, the possible degradation pathway of trimethylamino propanol in the isolated strain is through consequence oxidation of alcohol group (-OH) to aldehyde (-CHO) and acid (-COOH), and thereafter the cleavage of β-alanine betaine C-N bonds yielded trimethylamine and alkyl chain.

Design Optimization of a Double Stator Cup- Rotor Machine

This paper presents the optimum design for a double stator, cup rotor machine; a novel type of BLDC PM Machine. The optimization approach is divided into two stages: the first stage is calculating the machine configuration using Matlab, and the second stage is the optimization of the machine using Finite Element Modeling (FEM). Under the design specifications, the machine model will be selected from three pole numbers, namely, 8, 10 and 12 with an appropriate slot number. A double stator brushless DC permanent magnet machine is designed to achieve low cogging torque; high electromagnetic torque and low ripple torque.

The Effect of Ownership Structure on Stock Prices after Crisis: A Study on Ise 100 Index

Using Turkish data, in this study it is investigated that whether a firm’s ownership structure has an impact on its stock prices after the crisis. A linear regression model is conducted on the data of non-financial firms that are trading in Istanbul Stock Exchange 100 Index (ISE 100) index. The findings show that, all explanatory variables such as inside ownership, largest ownership, concentrated ownership, foreign shareholders, family controlled and dispersed ownership are not very important to explain stock prices after the crisis. Family controlled firms and concentrated ownership is positively related to stock price, dispersed ownership, largest ownership, foreign shareholders, and inside ownership structures have negative interaction between stock prices, but because of the p value is not under the value of 0.05 this relation is not significant. In addition, the analysis shows that, the shares of firms that have inside, largest and dispersed ownership structure are outperform comparing with the other firms. Furthermore, ownership concentrated firms outperform to family controlled firms.

Automatic Distance Compensation for Robust Voice-based Human-Computer Interaction

Distant-talking voice-based HCI system suffers from performance degradation due to mismatch between the acoustic speech (runtime) and the acoustic model (training). Mismatch is caused by the change in the power of the speech signal as observed at the microphones. This change is greatly influenced by the change in distance, affecting speech dynamics inside the room before reaching the microphones. Moreover, as the speech signal is reflected, its acoustical characteristic is also altered by the room properties. In general, power mismatch due to distance is a complex problem. This paper presents a novel approach in dealing with distance-induced mismatch by intelligently sensing instantaneous voice power variation and compensating model parameters. First, the distant-talking speech signal is processed through microphone array processing, and the corresponding distance information is extracted. Distance-sensitive Gaussian Mixture Models (GMMs), pre-trained to capture both speech power and room property are used to predict the optimal distance of the speech source. Consequently, pre-computed statistic priors corresponding to the optimal distance is selected to correct the statistics of the generic model which was frozen during training. Thus, model combinatorics are post-conditioned to match the power of instantaneous speech acoustics at runtime. This results to an improved likelihood in predicting the correct speech command at farther distances. We experiment using real data recorded inside two rooms. Experimental evaluation shows voice recognition performance using our method is more robust to the change in distance compared to the conventional approach. In our experiment, under the most acoustically challenging environment (i.e., Room 2: 2.5 meters), our method achieved 24.2% improvement in recognition performance against the best-performing conventional method.

Iran’s Gas Flare Recovery Options Using MCDM

In this paper, five options of Iran’s gas flare recovery have been compared via MCDM method. For developing the model, the weighing factor of each indicator an AHP method is used via the Expert-choice software. Several cases were considered in this analysis. They are defined where the priorities were defined always keeping one criterion in first position, while the priorities of the other criteria were defined by ordinal information defining the mutual relations of the criteria and the respective indicators. The results, show that amongst these cases, priority is obtained for CHP usage where availability indicator is highly weighted while the pipeline usage is obtained where environmental indicator highly weighted and the injection priority is obtained where economic indicator is highly weighted and also when the weighing factor of all the criteria are the same the Injection priority is obtained.

Monitoring Patents Using the Statistical Process Control

The statistical process control (SPC) is one of the most powerful tools developed to assist ineffective control of quality, involves collecting, organizing and interpreting data during production. This article aims to show how the use of CEP industries can control and continuously improve product quality through monitoring of production that can detect deviations of parameters representing the process by reducing the amount of off-specification products and thus the costs of production. This study aimed to conduct a technological forecasting in order to characterize the research being done related to the CEP. The survey was conducted in the databases Spacenet, WIPO and the National Institute of Industrial Property (INPI). Among the largest are the United States depositors and deposits via PCT, the classification section that was presented in greater abundance to F.

A Novel Frequency Offset Estimation Scheme for OFDM Systems

In this paper, we propose a novel frequency offset estimation scheme for orthogonal frequency division multiplexing (OFDM) systems. By correlating the OFDM signals within the coherence phase bandwidth and employing a threshold in the frequency offset estimation process, the proposed scheme is not only robust to the timing offset but also has a reduced complexity compared with that of the conventional scheme. Moreover, a timing offset estimation scheme is also proposed as the next stage of the proposed frequency offset estimation. Numerical results show that the proposed scheme can estimate frequency offset with lower computational complexity and does not require additional memory while maintaining the same level of estimation performance.

Effect of Scanning Speed on Material Efficiency of Laser Metal Deposited Ti6Al4V

The study of effect of laser scanning speed on material efficiency in Ti6Al4V application is very important because unspent powder is not reusable because of high temperature oxygen pick-up and contamination. This study carried out an extensive study on the effect of scanning speed on material efficiency by varying the speed between 0.01 to 0.1m/sec. The samples are wire brushed and cleaned with acetone after each deposition to remove un-melted particles from the surface of the deposit. The substrate is weighed before and after deposition. A formula was developed to calculate the material efficiency and the scanning speed was compared with the powder efficiency obtained. The results are presented and discussed. The study revealed that the optimum scanning speed exists for this study at 0.01m/sec, above and below which the powder efficiency will drop

Constitutive Equations for Human Saphenous Vein Coronary Artery Bypass Graft

Coronary artery bypass grafts (CABG) are widely studied with respect to hemodynamic conditions which play important role in presence of a restenosis. However, papers which concern with constitutive modeling of CABG are lacking in the literature. The purpose of this study is to find a constitutive model for CABG tissue. A sample of the CABG obtained within an autopsy underwent an inflation–extension test. Displacements were recoredered by CCD cameras and subsequently evaluated by digital image correlation. Pressure – radius and axial force – elongation data were used to fit material model. The tissue was modeled as onelayered composite reinforced by two families of helical fibers. The material is assumed to be locally orthotropic, nonlinear, incompressible and hyperelastic. Material parameters are estimated for two strain energy functions (SEF). The first is classical exponential. The second SEF is logarithmic which allows interpretation by means of limiting (finite) strain extensibility. Presented material parameters are estimated by optimization based on radial and axial equilibrium equation in a thick-walled tube. Both material models fit experimental data successfully. The exponential model fits significantly better relationship between axial force and axial strain than logarithmic one.

Drag Analysis of an Aircraft Wing Model withand without Bird Feather like Winglet

This work describes the aerodynamic characteristic for aircraft wing model with and without bird feather like winglet. The aerofoil used to construct the whole structure is NACA 653-218 Rectangular wing and this aerofoil has been used to compare the result with previous research using winglet. The model of the rectangular wing with bird feather like winglet has been fabricated using polystyrene before design using CATIA P3 V5R13 software and finally fabricated in wood. The experimental analysis for the aerodynamic characteristic for rectangular wing without winglet, wing with horizontal winglet and wing with 60 degree inclination winglet for Reynolds number 1.66×105, 2.08×105 and 2.50×105 have been carried out in open loop low speed wind tunnel at the Aerodynamics laboratory in Universiti Putra Malaysia. The experimental result shows 25-30 % reduction in drag coefficient and 10-20 % increase in lift coefficient by using bird feather like winglet for angle of attack of 8 degree.

Two Dimensionnal Model for Extraction Packed Column Simulation using Finite Element Method

Modeling transfer phenomena in several chemical engineering operations leads to the resolution of partial differential equations systems. According to the complexity of the operations mechanisms, the equations present a nonlinear form and analytical solution became difficult, we have then to use numerical methods which are based on approximations in order to transform a differential system to an algebraic one.Finite element method is one of numerical methods which can be used to obtain an accurate solution in many complex cases of chemical engineering.The packed columns find a large application like contactor for liquid-liquid systems such solvent extraction. In the literature, the modeling of this type of equipment received less attention in comparison with the plate columns.A mathematical bidimensionnal model with radial and axial dispersion, simulating packed tower extraction behavior was developed and a partial differential equation was solved using the finite element method by adopting the Galerkine model. We developed a Mathcad program, which can be used for a similar equations and concentration profiles are obtained along the column. The influence of radial dispersion was prooved and it can-t be neglected, the results were compared with experimental concentration at the top of the column in the extraction system: acetone/toluene/water.

Efficacy of Anti-phishing Measures and Strategies - A Research Analysis

Statistics indicate that more than 1000 phishing attacks are launched every month. With 57 million people hit by the fraud so far in America alone, how do we combat phishing?This publication aims to discuss strategies in the war against Phishing. This study is an examination of the analysis and critique found in the ways adopted at various levels to counter the crescendo of phishing attacks and new techniques being adopted for the same. An analysis of the measures taken up by the varied popular Mail servers and popular browsers is done under this study. This work intends to increase the understanding and awareness of the internet user across the globe and even discusses plausible countermeasures at the users as well as the developers end. This conceptual paper will contribute to future research on similar topics.

A Comparative Study of Main Memory Databases and Disk-Resident Databases

Main Memory Database systems (MMDB) store their data in main physical memory and provide very high-speed access. Conventional database systems are optimized for the particular characteristics of disk storage mechanisms. Memory resident systems, on the other hand, use different optimizations to structure and organize data, as well as to make it reliable. This paper provides a brief overview on MMDBs and one of the memory resident systems named FastDB and compares the processing time of this system with a typical disc resident database based on the results of the implementation of TPC benchmarks environment on both.

Use of Heliox during Spontaneous Ventilation: Model Study

The study deals with the modelling of the gas flow during heliox therapy. A special model has been developed to study the effect of the helium upon the gas flow in the airways during the spontaneous breathing. Lower density of helium compared with air decreases the Reynolds number and it allows improving the flow during the spontaneous breathing. In the cases, where the flow becomes turbulent while the patient inspires air the flow is still laminar when the patient inspires heliox. The use of heliox decreases the work of breathing and improves ventilation. It allows in some cases to prevent the intubation of the patients.

64 bit Computer Architectures for Space Applications – A study

The more recent satellite projects/programs makes extensive usage of real – time embedded systems. 16 bit processors which meet the Mil-Std-1750 standard architecture have been used in on-board systems. Most of the Space Applications have been written in ADA. From a futuristic point of view, 32 bit/ 64 bit processors are needed in the area of spacecraft computing and therefore an effort is desirable in the study and survey of 64 bit architectures for space applications. This will also result in significant technology development in terms of VLSI and software tools for ADA (as the legacy code is in ADA). There are several basic requirements for a special processor for this purpose. They include Radiation Hardened (RadHard) devices, very low power dissipation, compatibility with existing operational systems, scalable architectures for higher computational needs, reliability, higher memory and I/O bandwidth, predictability, realtime operating system and manufacturability of such processors. Further on, these may include selection of FPGA devices, selection of EDA tool chains, design flow, partitioning of the design, pin count, performance evaluation, timing analysis etc. This project deals with a brief study of 32 and 64 bit processors readily available in the market and designing/ fabricating a 64 bit RISC processor named RISC MicroProcessor with added functionalities of an extended double precision floating point unit and a 32 bit signal processing unit acting as co-processors. In this paper, we emphasize the ease and importance of using Open Core (OpenSparc T1 Verilog RTL) and Open “Source" EDA tools such as Icarus to develop FPGA based prototypes quickly. Commercial tools such as Xilinx ISE for Synthesis are also used when appropriate.

EGCL: An Extended G-Code Language with Flow Control, Functions and Mnemonic Variables

In the context of computer numerical control (CNC) and computer aided manufacturing (CAM), the capabilities of programming languages such as symbolic and intuitive programming, program portability and geometrical portfolio have special importance. They allow to save time and to avoid errors during part programming and permit code re-usage. Our updated literature review indicates that the current state of art presents voids in parametric programming, program portability and programming flexibility. In response to this situation, this article presents a compiler implementation for EGCL (Extended G-code Language), a new, enriched CNC programming language which allows the use of descriptive variable names, geometrical functions and flow-control statements (if-then-else, while). Our compiler produces low-level generic, elementary ISO-compliant Gcode, thus allowing for flexibility in the choice of the executing CNC machine and in portability. Our results show that readable variable names and flow control statements allow a simplified and intuitive part programming and permit re-usage of the programs. Future work includes allowing the programmer to define own functions in terms of EGCL, in contrast to the current status of having them as library built-in functions.

Investigation of Short Time Scale Variation of Solar Radiation Spectrum in UV, PAR, and NIR Bands due to Atmospheric Aerosol and Water Vapor

Long terms variation of solar insolation had been widely studied. However, its parallel observations in short time scale is rather lacking. This paper aims to investigate the short time scale evolution of solar radiation spectrum (UV, PAR, and NIR bands) due to atmospheric aerosols and water vapors. A total of 25 days of global and diffused solar spectrum ranges from air mass 2 to 6 were collected using ground-based spectrometer with shadowband technique. The result shows that variation of solar radiation is the least in UV fraction, followed by PAR and the most in NIR. Broader variations in PAR and NIR are associated with the short time scale fluctuations of aerosol and water vapors. The corresponding daily evolution of UV, PAR, and NIR fractions implies that aerosol and water vapors variation could also be responsible for the deviation pattern in the Langley-plot analysis.

Higher-Dimensional Quantum Cryptography

We report on a high-speed quantum cryptography system that utilizes simultaneous entanglement in polarization and in “time-bins". With multiple degrees of freedom contributing to the secret key, we can achieve over ten bits of random entropy per detected coincidence. In addition, we collect from multiple spots o the downconversion cone to further amplify the data rate, allowing usto achieve over 10 Mbits of secure key per second.

Surrogate based Evolutionary Algorithm for Design Optimization

Optimization is often a critical issue for most system design problems. Evolutionary Algorithms are population-based, stochastic search techniques, widely used as efficient global optimizers. However, finding optimal solution to complex high dimensional, multimodal problems often require highly computationally expensive function evaluations and hence are practically prohibitive. The Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model presented in our earlier work [14] reduced computation time by controlled use of meta-models to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the meta-model are generated from a single uniform model. Situations like model formation involving variable input dimensions and noisy data certainly can not be covered by this assumption. In this paper we present an enhanced version of DAFHEA that incorporates a multiple-model based learning approach for the SVM approximator. DAFHEA-II (the enhanced version of the DAFHEA framework) also overcomes the high computational expense involved with additional clustering requirements of the original DAFHEA framework. The proposed framework has been tested on several benchmark functions and the empirical results illustrate the advantages of the proposed technique.