Properties of Biodiesel Produced by Enzymatic Transesterification of Lipids Extracted from Microalgae in Supercritical Carbon Dioxide Medium

Biodiesel, as an alternative renewable fuel, has been receiving increasing attention due to the limited supply of fossil fuels and the increasing need for energy. Microalgae are promising source for lipids, which can be converted to biodiesel. The biodiesel production from microalgae lipids using lipase catalyzed reaction in supercritical CO2 medium has several advantages over conventional production processes. However, identifying the optimum microalgae lipid extraction and transesterification conditions is still a challenge. In this study, the quality of biodiesel produced from lipids extracted from Scenedesmus sp. and their enzymatic transesterification using supercritical carbon dioxide have been investigated. At the optimum conditions, the highest biodiesel production yield was found to be 82%. The fuel properties of the produced biodiesel, without any separation step, at optimum reaction condition, were determined and compared to ASTM standards. The properties were found to comply with the limits, and showed a low glycerol content, without any separation step.

Studies of Rule Induction by STRIM from the Decision Table with Contaminated Attribute Values from Missing Data and Noise — In the Case of Critical Dataset Size —

STRIM (Statistical Test Rule Induction Method) has been proposed as a method to effectively induct if-then rules from the decision table which is considered as a sample set obtained from the population of interest. Its usefulness has been confirmed by simulation experiments specifying rules in advance, and by comparison with conventional methods. However, scope for future development remains before STRIM can be applied to the analysis of real-world data sets. The first requirement is to determine the size of the dataset needed for inducting true rules, since finding statistically significant rules is the core of the method. The second is to examine the capacity of rule induction from datasets with contaminated attribute values created by missing data and noise, since real-world datasets usually contain such contaminated data. This paper examines the first problem theoretically, in connection with the rule length. The second problem is then examined in a simulation experiment, utilizing the critical size of dataset derived from the first step. The experimental results show that STRIM is highly robust in the analysis of datasets with contaminated attribute values, and hence is applicable to real-world data

The Application of FSI Techniques in Modeling of Realist Pulmonary Systems

The modeling lung respiratory system that has complex anatomy and biophysics presents several challenges including tissue-driven flow patterns and wall motion. Also, the pulmonary lung system because of that they stretch and recoil with each breath, has not static walls and structures. The direct relationship between air flow and tissue motion in the lung structures naturally prefers an FSI simulation technique. Therefore, in order to toward the realistic simulation of pulmonary breathing mechanics the development of a coupled FSI computational model is an important step. A simple but physiologically relevant three-dimensional deep long geometry is designed and fluid-structure interaction (FSI) coupling technique is utilized for simulating the deformation of the lung parenchyma tissue that produces airflow fields. The real understanding of respiratory tissue system as a complex phenomenon have been investigated with respect to respiratory patterns, fluid dynamics and tissue viscoelasticity and tidal breathing period. 

Cell Biomass and Lipid Productivities of Meyerella planktonica under Autotrophic and Heterotrophic Growth Conditions

Microalgae Meyerella planktonica is a potential biofuel source because it can grow in bulk in either autotrophic or heterotrophic condition. However, the quantitative growth of this algal type is still low as it tends to precipitates on the bottom. Besides, the lipid concentration is still low when grown in autotrophic condition. In contrast, heterotrophic condition can enhance the lipid concentration. The combination of autotrophic condition and agitation treatment was conducted to increase the density of the culture. On the other hand, a heterotrophic condition was set up to raise the lipid production. A two-stage experiment was applied to increase the density at the first step and to increase the lipid concentration in the next step. The autotrophic condition resulted higher density but lower lipid concentration compared to heterotrophic one. The agitation treatment produced higher density in both autotrophic and heterotrophic conditions. The two-stage experiment managed to enhance the density during the autotrophic stage and the lipid concentration during the heterotrophic stage. The highest yield was performed by using 0.4% v/v glycerol as a carbon source (2.9±0.016 x 10^6 cells w/w) attained 7 days after the heterotrophic stage began. The lipid concentration was stable starting from day 7.

Various Modifications of Electrochemical Barrier Layer Thinning of Anodic Aluminum Oxide

In this paper, two options of anodic alumina barrier layer thinning have been demonstrated. The approaches varied with the duration of the voltage step. It was found that too long step of the barrier layer thinning process leads to chemical etching of the nanopores on their top. At the bottoms pores are not fully opened what is disadvantageous for further applications in nanofabrication. On the other hand, while the duration of the voltage step is controlled by the current density (value of the current density cannot exceed 75% of the value recorded during previous voltage step) the pores are fully opened. However, pores at the bottom obtained with this procedure have smaller diameter, nevertheless this procedure provides electric contact between the bare aluminum (substrate) and electrolyte, what is suitable for template assisted electrodeposition, one of the most cost-efficient synthesis method in nanotechnology.

Selecting the Best Sub-Region Indexing the Images in the Case of Weak Segmentation Based On Local Color Histograms

Color Histogram is considered as the oldest method used by CBIR systems for indexing images. In turn, the global histograms do not include the spatial information; this is why the other techniques coming later have attempted to encounter this limitation by involving the segmentation task as a preprocessing step. The weak segmentation is employed by the local histograms while other methods as CCV (Color Coherent Vector) are based on strong segmentation. The indexation based on local histograms consists of splitting the image into N overlapping blocks or sub-regions, and then the histogram of each block is computed. The dissimilarity between two images is reduced, as consequence, to compute the distance between the N local histograms of the both images resulting then in N*N values; generally, the lowest value is taken into account to rank images, that means that the lowest value is that which helps to designate which sub-region utilized to index images of the collection being asked. In this paper, we make under light the local histogram indexation method in the hope to compare the results obtained against those given by the global histogram. We address also another noteworthy issue when Relying on local histograms namely which value, among N*N values, to trust on when comparing images, in other words, which sub-region among the N*N sub-regions on which we base to index images. Based on the results achieved here, it seems that relying on the local histograms, which needs to pose an extra overhead on the system by involving another preprocessing step naming segmentation, does not necessary mean that it produces better results. In addition to that, we have proposed here some ideas to select the local histogram on which we rely on to encode the image rather than relying on the local histogram having lowest distance with the query histograms.

Effect of Different Microbial Strains on Biological Pretreatment of Sugarcane Bagasse for Enzymatic Hydrolysis

Among agricultural residues, sugarcane bagasse is one of the most convincing raw materials for the production of bioethanol due to its availability, and low cost through enzymatic hydrolysis and yeast fermentation. A pretreatment step is needed to enhance the enzymatic step. In this study, sugarcane bagasse (SCB), one of the most abundant agricultural residues in Thailand, was pretreated biologically with various microorganisms of white-rot fungus—Phanerochaete sordid (SK 7), Cellulomonas sp. (TISTR 784), and strain A 002 (Bacillus subtilis isolated from Thai higher termites). All samples with various microbial pretreatments were further hydrolyzed enzymatically by a commercial enzyme obtained from Aspergillus niger. The results showed that the pretreatment with the white-rot fungus gave the highest glucose concentration around two-fold higher when compared with the others.

A Method for Improving the Embedded Runge Kutta Fehlberg 4(5)

In this paper, we introduce a method for improving the embedded Runge-Kutta-Fehlberg4(5) method. At each integration step, the proposed method is comprised of two equations for the solution and the error, respectively. These solution and error are obtained by solving an initial value problem whose solution has the information of the error at each integration step. The constructed algorithm controls both the error and the time step size simultaneously and possesses a good performance in the computational cost compared to the original method. For the assessment of the effectiveness, EULR problem is numerically solved.

Determination of Vitamin C (Ascorbic Acid) in Orange Juices Product

This research describes a voltammetric approach to determine amounts of vitamin C (Ascorbic acid) in orange juice sample, using three screen printed electrode. The anodic currents of vitamin C were proportional to vitamin C concentration in the range of 0 – 10.0 mM with the limit of detection of 1.36 mM. The method was successfully employed with 2 µL of the working solution dropped on the electrode surface. The proposed method was applied for the analysis of vitamin C in packed orange juice without sample purification or complexion of sample preparation step.

Level Set and Morphological Operation Techniques in Application of Dental Image Segmentation

Medical image analysis is one of the great effects of computer image processing. There are several processes to analysis the medical images which the segmentation process is one of the challenging and most important step. In this paper the segmentation method proposed in order to segment the dental radiograph images. Thresholding method has been applied to simplify the images and to morphologically open binary image technique performed to eliminate the unnecessary regions on images. Furthermore, horizontal and vertical integral projection techniques used to extract the each individual tooth from radiograph images. Segmentation process has been done by applying the level set method on each extracted images. Nevertheless, the experiments results by 90% accuracy demonstrate that proposed method achieves high accuracy and promising result.

Effect of Progressive Type-I Right Censoring on Bayesian Statistical Inference of Simple Step–Stress Acceleration Life Testing Plan under Weibull Life Distribution

This paper discusses the effects of using progressive Type-I right censoring on the design of the Simple Step Accelerated Life testing using Bayesian approach for Weibull life products under the assumption of cumulative exposure model. The optimization criterion used in this paper is to minimize the expected pre-posterior variance of the Pth percentile time of failures. The model variables are the stress changing time and the stress value for the first step. A comparison between the conventional and the progressive Type-I right censoring is provided. The results have shown that the progressive Type-I right censoring reduces the cost of testing on the expense of the test precision when the sample size is small. Moreover, the results have shown that using strong priors or large sample size reduces the sensitivity of the test precision to the censoring proportion. Hence, the progressive Type-I right censoring is recommended in these cases as progressive Type-I right censoring reduces the cost of the test and doesn't affect the precision of the test a lot. Moreover, the results have shown that using direct or indirect priors affects the precision of the test.

Low Cost IMU \ GPS Integration Using Kalman Filtering for Land Vehicle Navigation Application

Land vehicle navigation system technology is a subject of great interest today. Global Positioning System (GPS) is a common choice for positioning in such systems. However, GPS alone is incapable of providing continuous and reliable positioning, because of its inherent dependency on external electromagnetic signals. Inertial Navigation is the implementation of inertial sensors to determine the position and orientation of a vehicle. As such, inertial navigation has unbounded error growth since the error accumulates at each step. Thus in order to contain these errors some form of external aiding is required. The availability of low cost Micro-Electro-Mechanical-System (MEMS) inertial sensors is now making it feasible to develop Inertial Navigation System (INS) using an inertial measurement unit (IMU), in conjunction with GPS to fulfill the demands of such systems. Typically IMU’s are very expensive systems; however this INS will use “low cost” components. Unfortunately with low cost also comes low performance and is the main reason for the inclusion of GPS and Kalman filtering into the system. The aim of this paper is to develop a GPS/MEMS INS integrated system, which is able to provide a navigation solution with accuracy levels appropriate for land vehicle navigation. The primary piece of equipment used was a MEMS-based Crista IMU (from Cloud Cap Technology Inc.) and a Garmin GPS 18 PC (which is both a receiver and antenna). The integration of GPS with INS can be implemented using a Kalman filter in loosely coupled mode. In this integration mode the INS error states, together with any navigation state (position, velocity, and attitude) and other unknown parameters of interest, are estimated using GPS measurements. All important equations regarding navigation are presented along with discussion.

Backcalculation of HMA Stiffness Based On Finite Element Model

Stiffness of Hot Mix Asphalt (HMA) in flexible pavement is largely dependent of temperature, mode of testing and age of pavement. Accurate measurement of HMA stiffness is thus quite challenging. This study determines HMA stiffness based on Finite Element Model (FEM) and validates the results using field data. As a first step, stiffnesses of different layers of a pavement section on Interstate 40 (I-40) in New Mexico were determined by Falling Weight Deflectometer (FWD) test. Pavement temperature was not measured at that time due to lack of temperature probe. Secondly, a FE model is developed in ABAQUS. Stiffness of the base, subbase and subgrade were taken from the FWD test output obtained from the first step. As HMA stiffness largely varies with temperature it was assigned trial and error approach. Thirdly, horizontal strain and vertical stress at the bottom of the HMA and temperature at different depths of the pavement were measured with installed sensors on the whole day on December 25th, 2012. Fourthly, outputs of FEM were correlated with measured stress-strain responses. After a number of trials a relationship was developed between the trial stiffness of HMA and measured mid-depth HMA temperature. At last, the obtained relationship between stiffness and temperature is verified by further FWD test when pavement temperature was recorded. A promising agreement between them is observed. Therefore, conclusion can be drawn that linear elastic FEM can accurately predict the stiffness and the structural response of flexible pavement.

Heat Transfer to Laminar Flow over a Double Backward-Facing Step

Heat transfer and laminar air flow over a double backward-facing step numerically studied in this paper. The simulations was performed by using ANSYS ICEM for meshing process and using ANSYS fluent 14 (CFD) for solving. The k-ɛ standard model adopted with Reynolds number varied between 98.5 to 512 and three step height at constant heat flux (q=2000 W/m2). The top of wall and bottom of upstream are insulated with bottom of downstream is heated. The results show increase in Nusselt number with increases of Reynolds number for all cases and the maximum of Nusselt number happens at the first step in compared to the second step. Due to increase of cross section area of downstream to generate sudden expansion then Nusselt number decrease but the profile of Nusselt number keep same trend for all cases where increase after the first and second steps. Recirculation region after the first and second steps are denoted by contour of streamline velocity. The higher augmentation of heat transfer rate observed for case 1 at Reynolds number of 512 and heat flux q=2000 W/m2.

An Impulse-Momentum Approach to Swing-Up Control of Double Inverted Pendulum on a Cart

The challenge in the swing-up problem of double inverted pendulum on a cart (DIPC) is to design a controller that bring all DIPC's states, especially the joint angles of the two links, into the region of attraction of the desired equilibrium. This paper proposes a new method to swing-up DIPC based on a series of restto- rest maneuvers of the first link about its vertically upright configuration while holding the cart fixed at the origin. The rest-torest maneuvers are designed such that each one results in a net gain in energy of the second link. This results in swing-up of DIPC-s configuration to the region of attraction of the desired equilibrium. A three-step algorithm is provided for swing-up control followed by the stabilization step. Simulation results with a comparison to an experimental work done in the literature are presented to demonstrate the efficacy of the approach.

Despeckling of Synthetic Aperture Radar Images Using Inner Product Spaces in Undecimated Wavelet Domain

This paper introduces the effective speckle reduction of synthetic aperture radar (SAR) images using inner product spaces in undecimated wavelet domain. There are two major areas in projection onto span algorithm where improvement can be made. First is the use of undecimated wavelet transformation instead of discrete wavelet transformation. And second area is the use of smoothing filter namely directional smoothing filter which is an additional step. Proposed method does not need any noise estimation and thresholding technique. More over proposed method gives good results on both single polarimetric and fully polarimetric SAR images.

GA Based Optimal Feature Extraction Method for Functional Data Classification

Classification is an interesting problem in functional data analysis (FDA), because many science and application problems end up with classification problems, such as recognition, prediction, control, decision making, management, etc. As the high dimension and high correlation in functional data (FD), it is a key problem to extract features from FD whereas keeping its global characters, which relates to the classification efficiency and precision to heavens. In this paper, a novel automatic method which combined Genetic Algorithm (GA) and classification algorithm to extract classification features is proposed. In this method, the optimal features and classification model are approached via evolutional study step by step. It is proved by theory analysis and experiment test that this method has advantages in improving classification efficiency, precision and robustness whereas using less features and the dimension of extracted classification features can be controlled.

Parallelization of Ensemble Kalman Filter (EnKF) for Oil Reservoirs with Time-lapse Seismic Data

In this paper we describe the design and implementation of a parallel algorithm for data assimilation with ensemble Kalman filter (EnKF) for oil reservoir history matching problem. The use of large number of observations from time-lapse seismic leads to a large turnaround time for the analysis step, in addition to the time consuming simulations of the realizations. For efficient parallelization it is important to consider parallel computation at the analysis step. Our experiments show that parallelization of the analysis step in addition to the forecast step has good scalability, exploiting the same set of resources with some additional efforts.

Design and Microfabrication of a High Throughput Thermal Cycling Platform with Various Annealing Temperatures

This study describes a micro device integrated with multi-chamber for polymerase chain reaction (PCR) with different annealing temperatures. The device consists of the reaction polydimethylsiloxane (PDMS) chip, a cover glass chip, and is equipped with cartridge heaters, fans, and thermocouples for temperature control. In this prototype, commercial software is utilized to determine the geometric and operational parameters those are responsible for creating the denaturation, annealing, and extension temperatures within the chip. Two cartridge heaters are placed at two sides of the chip and maintained at two different temperatures to achieve a thermal gradient on the chip during the annealing step. The temperatures on the chip surface are measured via an infrared imager. Some thermocouples inserted into the reaction chambers are used to obtain the transient temperature profiles of the reaction chambers during several thermal cycles. The experimental temperatures compared to the simulated results show a similar trend. This work should be interesting to persons involved in the high-temperature based reactions and genomics or cell analysis.