Design Optimization of Aerocapture with Aerodynamic-Environment-Adaptive Variable Geometry Flexible Aeroshell

This paper proposes the concept of aerocapture with aerodynamic-environment-adaptive variable geometry flexible aeroshell that vehicle deploys. The flexible membrane is composed of thin-layer film or textile as its aeroshell in order to solve some problems obstructing realization of aerocapture technique. Multi-objective optimization study is conducted to investigate solutions and derive design guidelines. As a result, solutions which can avoid aerodynamic heating and enlarge the corridor width up to 10% are obtained successfully, so that the effectiveness of this concept can be demonstrated. The deformation-use optimum solution changes its drag coefficient from 1.6 to 1.1, along with the change in dynamic pressure. Moreover, optimization results show that deformation-use solution requires the membrane for which upper temperature limit and strain limit are more than 700 K and 120%, respectively, and elasticity (Young-s modulus) is of order of 106 Pa.

A Perceptually Optimized Wavelet Embedded Zero Tree Image Coder

In this paper, we propose a Perceptually Optimized Embedded ZeroTree Image Coder (POEZIC) that introduces a perceptual weighting to wavelet transform coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to the coding quality obtained using the SPIHT algorithm only. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEZIC quality assessment. Our POEZIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) luminance masking and Contrast masking, 2) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting, 3) the Wavelet Error Sensitivity WES used to reduce the perceptual quantization errors. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.

Robust Steam Temperature Regulation for Distillation of Essential Oil Extraction Process using Hybrid Fuzzy-PD plus PID Controller

This paper presents a hybrid fuzzy-PD plus PID (HFPP) controller and its application to steam distillation process for essential oil extraction system. Steam temperature is one of the most significant parameters that can influence the composition of essential oil yield. Due to parameter variations and changes in operation conditions during distillation, a robust steam temperature controller becomes nontrivial to avoid the degradation of essential oil quality. Initially, the PRBS input is triggered to the system and output of steam temperature is modeled using ARX model structure. The parameter estimation and tuning method is adopted by simulation using HFPP controller scheme. The effectiveness and robustness of proposed controller technique is validated by real time implementation to the system. The performance of HFPP using 25 and 49 fuzzy rules is compared. The experimental result demonstrates the proposed HFPP using 49 fuzzy rules achieves a better, consistent and robust controller compared to PID when considering the test on tracking the set point and the effects due to disturbance.

An Experimental Study on Development of the Connection System of Concrete Barriers Applicable to Modular Bridge

Although many studies on the assembly technology of the bridge construction have dealt mostly with on the pier, girder or the deck of the bridge, studies on the prefabricated barrier have rarely been performed. For understanding structural characteristics and application of the concrete barrier in the modular bridge, which is an assembly of structure members, static loading test was performed. Structural performances as a road barrier of the three methods, conventional cast-in-place(ST), vertical bolt connection(BVC) and horizontal bolt connection(BHC) were evaluated and compared through the analyses of load-displacement curves, strain curves of the steel, concrete strain curves and the visual appearances of crack patterns. The vertical bolt connection(BVC) method demonstrated comparable performance as an alternative to conventional cast-in-place(ST) while providing all the advantages of prefabricated technology. Necessities for the future improvement in nuts enforcement as well as legal standard and regulation are also addressed.

Dengue Disease Mapping with Standardized Morbidity Ratio and Poisson-gamma Model: An Analysis of Dengue Disease in Perak, Malaysia

Dengue disease is an infectious vector-borne viral disease that is commonly found in tropical and sub-tropical regions, especially in urban and semi-urban areas, around the world and including Malaysia. There is no currently available vaccine or chemotherapy for the prevention or treatment of dengue disease. Therefore prevention and treatment of the disease depend on vector surveillance and control measures. Disease risk mapping has been recognized as an important tool in the prevention and control strategies for diseases. The choice of statistical model used for relative risk estimation is important as a good model will subsequently produce a good disease risk map. Therefore, the aim of this study is to estimate the relative risk for dengue disease based initially on the most common statistic used in disease mapping called Standardized Morbidity Ratio (SMR) and one of the earliest applications of Bayesian methodology called Poisson-gamma model. This paper begins by providing a review of the SMR method, which we then apply to dengue data of Perak, Malaysia. We then fit an extension of the SMR method, which is the Poisson-gamma model. Both results are displayed and compared using graph, tables and maps. Results of the analysis shows that the latter method gives a better relative risk estimates compared with using the SMR. The Poisson-gamma model has been demonstrated can overcome the problem of SMR when there is no observed dengue cases in certain regions. However, covariate adjustment in this model is difficult and there is no possibility for allowing spatial correlation between risks in adjacent areas. The drawbacks of this model have motivated many researchers to propose other alternative methods for estimating the risk.

Mobile to Server Face Recognition: A System Overview

This paper presents a system overview of Mobile to Server Face Recognition, which is a face recognition application developed specifically for mobile phones. Images taken from mobile phone cameras lack of quality due to the low resolution of the cameras. Thus, a prototype is developed to experiment the chosen method. However, this paper shows a result of system backbone without the face recognition functionality. The result demonstrated in this paper indicates that the interaction between mobile phones and server is successfully working. The result shown before the database is completely ready. The system testing is currently going on using real images and a mock-up database to test the functionality of the face recognition algorithm used in this system. An overview of the whole system including screenshots and system flow-chart are presented in this paper. This paper also presents the inspiration or motivation and the justification in developing this system.

Enhancing the Error-Correcting Performance of LDPC Codes through an Efficient Use of Decoding Iterations

The decoding of Low-Density Parity-Check (LDPC) codes is operated over a redundant structure known as the bipartite graph, meaning that the full set of bit nodes is not absolutely necessary for decoder convergence. In 2008, Soyjaudah and Catherine designed a recovery algorithm for LDPC codes based on this assumption and showed that the error-correcting performance of their codes outperformed conventional LDPC Codes. In this work, the use of the recovery algorithm is further explored to test the performance of LDPC codes while the number of iterations is progressively increased. For experiments conducted with small blocklengths of up to 800 bits and number of iterations of up to 2000, the results interestingly demonstrate that contrary to conventional wisdom, the error-correcting performance keeps increasing with increasing number of iterations.

Mining Image Features in an Automatic Two-Dimensional Shape Recognition System

The number of features required to represent an image can be very huge. Using all available features to recognize objects can suffer from curse dimensionality. Feature selection and extraction is the pre-processing step of image mining. Main issues in analyzing images is the effective identification of features and another one is extracting them. The mining problem that has been focused is the grouping of features for different shapes. Experiments have been conducted by using shape outline as the features. Shape outline readings are put through normalization and dimensionality reduction process using an eigenvector based method to produce a new set of readings. After this pre-processing step data will be grouped through their shapes. Through statistical analysis, these readings together with peak measures a robust classification and recognition process is achieved. Tests showed that the suggested methods are able to automatically recognize objects through their shapes. Finally, experiments also demonstrate the system invariance to rotation, translation, scale, reflection and to a small degree of distortion.

Providing On-Demand Path and Arrival Time Information Considering Realtime Delays of Buses

This paper demonstrates the bus location system for the route bus through the experiment in the real environment. A bus location system is a system that provides information such as the bus delay and positions. This system uses actual services and positions data of buses, and those information should match data on the database. The system has two possible problems. One, the system could cost high in preparing devices to get bus positions. Two, it could be difficult to match services data of buses. To avoid these problems, we have developed this system at low cost and short time by using the smart phone with GPS and the bus route system. This system realizes the path planning considering bus delay and displaying position of buses on the map. The bus location system was demonstrated on route buses with smart phones for two months.

Green Building Materials: Hemp Oil Based Biocomposites

Novel acrylated epoxidized hemp oil (AEHO) based bioresins were successfully synthesised, characterized and applied to biocomposites reinforced with woven jute fibre. Characterisation of the synthesised AEHO consisted of acid number titrations and FTIR spectroscopy to assess the success of the acrylation reaction. Three different matrices were produced (vinylester (VE), 50/50 blend of AEHO/VE and 100% AEHO) and reinforced with jute fibre to form three different types of biocomposite samples. Mechanical properties in the form of flexural and interlaminar shear strength (ILSS) were investigated and compared for the different samples. Results from the mechanical tests showed that AEHO and 50/50 based neat bioresins displayed lower flexural properties compared with the VE samples. However when applied to biocomposites and compared with VE based samples, AEHO biocomposites demonstrated comparable flexural performance and improved ILSS. These results are attributed to improved fibre-matrix interfacial adhesion due to surface-chemical compatibility between the natural fibres and bioresin.

Multidimensional and Data Mining Analysis for Property Investment Risk Analysis

Property investment in the real estate industry has a high risk due to the uncertainty factors that will affect the decisions made and high cost. Analytic hierarchy process has existed for some time in which referred to an expert-s opinion to measure the uncertainty of the risk factors for the risk analysis. Therefore, different level of experts- experiences will create different opinion and lead to the conflict among the experts in the field. The objective of this paper is to propose a new technique to measure the uncertainty of the risk factors based on multidimensional data model and data mining techniques as deterministic approach. The propose technique consist of a basic framework which includes four modules: user, technology, end-user access tools and applications. The property investment risk analysis defines as a micro level analysis as the features of the property will be considered in the analysis in this paper.

Visual Attention Analysis on Mutated Brand Name using Eye-Tracking: A Case Study

Brand name plays a vital role for in-shop buying behavior of consumers and mutated brand name may affect the selling of leading branded products. In Indian market, there are many products with mutated brand names which are either orthographically or phonologically similar. Due to presence of such products, Indian consumers very often fall under confusion when buying some regularly used stuff. Authors of the present paper have attempted to demonstrate relationship between less attention and false recognition of mutated brand names during a product selection process. To achieve this goal, visual attention study was conducted on 15 male college students using eye-tracker against a mutated brand name and errors in recognition were noted using questionnaire. Statistical analysis of the acquired data revealed that there was more false recognition of mutated brand name when less attention was paid during selection of favorite product. Moreover, it was perceived that eye tracking is an effective tool for analyzing false recognition of brand name mutation.

Profile of Viral Hepatitis in Saudi Arabia

The study was conducted to investigate the profile of hepatitis in Kingdom of Saudi Arabia, and to determine which age group hepatitis viruses most commonly infect. The epidemiology of viral hepatitis in Saudi Arabia has undergone major changes, concurrent with major socioeconomic developments over the last two to three decades. This disease represents a major public health problem in Saudi Arabia resulting in the need for considerable healthcare resources. A retrospective cross sectional analysis of the reported cases of viral hepatitis was conducted based on the reports of The Ministry of Health in Saudi Arabia about Hepatitis A, B and C infections in all regions from the period of January 2006 to December 2010. The study demonstrated that incidence of viral Hepatitis is decreasing, except for Hepatitis B that showed minimal increase. Of hepatitis A, B, and C, Hepatitis B virus (HBV) was the most predominant type, accounting for (53%) of the cases, followed by Hepatitis C virus (HCV) (30%) and HAV (17%). HAV infection predominates in children (5–14 years) with 60% of viral hepatitis cases, HBV in young adults (15–44 years) with 69% of viral hepatitis cases, and HCV in older adults (>45 years) with 59% of viral hepatitis cases. Despite significant changes in the prevalence of viral hepatitis A, B and C, it remains a major public health problem in Saudi Arabia; however, it showed a significant decline in the last two decades that could be attributed to the vaccination programs and the improved health facilities. Further researches are needed to identify the risk factors making a specific age group or a specific region in Saudi Arabia targeted for a specific type of hepatitis viruses.

Rigorous Electromagnetic Model of Fourier Transform Infrared (FT-IR) Spectroscopic Imaging Applied to Automated Histology of Prostate Tissue Specimens

Fourier transform infrared (FT-IR) spectroscopic imaging is an emerging technique that provides both chemically and spatially resolved information. The rich chemical content of data may be utilized for computer-aided determinations of structure and pathologic state (cancer diagnosis) in histological tissue sections for prostate cancer. FT-IR spectroscopic imaging of prostate tissue has shown that tissue type (histological) classification can be performed to a high degree of accuracy [1] and cancer diagnosis can be performed with an accuracy of about 80% [2] on a microscopic (≈ 6μm) length scale. In performing these analyses, it has been observed that there is large variability (more than 60%) between spectra from different points on tissue that is expected to consist of the same essential chemical constituents. Spectra at the edges of tissues are characteristically and consistently different from chemically similar tissue in the middle of the same sample. Here, we explain these differences using a rigorous electromagnetic model for light-sample interaction. Spectra from FT-IR spectroscopic imaging of chemically heterogeneous samples are different from bulk spectra of individual chemical constituents of the sample. This is because spectra not only depend on chemistry, but also on the shape of the sample. Using coupled wave analysis, we characterize and quantify the nature of spectral distortions at the edges of tissues. Furthermore, we present a method of performing histological classification of tissue samples. Since the mid-infrared spectrum is typically assumed to be a quantitative measure of chemical composition, classification results can vary widely due to spectral distortions. However, we demonstrate that the selection of localized metrics based on chemical information can make our data robust to the spectral distortions caused by scattering at the tissue boundary.

An Improved Data Mining Method Applied to the Search of Relationship between Metabolic Syndrome and Lifestyles

A data cutting and sorting method (DCSM) is proposed to optimize the performance of data mining. DCSM reduces the calculation time by getting rid of redundant data during the data mining process. In addition, DCSM minimizes the computational units by splitting the database and by sorting data with support counts. In the process of searching for the relationship between metabolic syndrome and lifestyles with the health examination database of an electronics manufacturing company, DCSM demonstrates higher search efficiency than the traditional Apriori algorithm in tests with different support counts.

Contributory Factors to Diabetes Dietary Regimen Non Adherence in Adults with Diabetes

A cross sectional survey design was used to collect data from 370 diabetic patients. Two instruments were used in obtaining data; in-depth interview guide and researchers- developed questionnaire. Fisher's exact test was used to investigate association between the identified factors and nonadherence. Factors identified were: socio-demographic factors such as: gender, age, marital status, educational level and occupation; psychosocial obstacles such as: non-affordability of prescribed diet, frustration due to the restriction, limited spousal support, feelings of deprivation, feeling that temptation is inevitable, difficulty in adhering in social gatherings and difficulty in revealing to host that one is diabetic; health care providers obstacles were: poor attitude of health workers, irregular diabetes education in clinics , limited number of nutrition education sessions/ inability of the patients to estimate the desired quantity of food, no reminder post cards or phone calls about upcoming patient appointments and delayed start of appointment / time wasting in clinics.

Correlation of Viscosity in Nanofluids using Genetic Algorithm-neural Network (GA-NN)

An accurate and proficient artificial neural network (ANN) based genetic algorithm (GA) is developed for predicting of nanofluids viscosity. A genetic algorithm (GA) is used to optimize the neural network parameters for minimizing the error between the predictive viscosity and the experimental one. The experimental viscosity in two nanofluids Al2O3-H2O and CuO-H2O from 278.15 to 343.15 K and volume fraction up to 15% were used from literature. The result of this study reveals that GA-NN model is outperform to the conventional neural nets in predicting the viscosity of nanofluids with mean absolute relative error of 1.22% and 1.77% for Al2O3-H2O and CuO-H2O, respectively. Furthermore, the results of this work have also been compared with others models. The findings of this work demonstrate that the GA-NN model is an effective method for prediction viscosity of nanofluids and have better accuracy and simplicity compared with the others models.

Periodic Oscillations in a Delay Population Model

In this paper, a nonlinear delay population model is investigated. Choosing the delay as a bifurcation parameter, we demonstrate that Hopf bifurcation will occur when the delay exceeds a critical value. Global existence of bifurcating periodic solutions is established. Numerical simulations supporting the theoretical findings are included.

Library Aware Power Conscious Realization of Complementary Boolean Functions

In this paper, we consider the problem of logic simplification for a special class of logic functions, namely complementary Boolean functions (CBF), targeting low power implementation using static CMOS logic style. The functions are uniquely characterized by the presence of terms, where for a canonical binary 2-tuple, D(mj) ∪ D(mk) = { } and therefore, we have | D(mj) ∪ D(mk) | = 0 [19]. Similarly, D(Mj) ∪ D(Mk) = { } and hence | D(Mj) ∪ D(Mk) | = 0. Here, 'mk' and 'Mk' represent a minterm and maxterm respectively. We compare the circuits minimized with our proposed method with those corresponding to factored Reed-Muller (f-RM) form, factored Pseudo Kronecker Reed-Muller (f-PKRM) form, and factored Generalized Reed-Muller (f-GRM) form. We have opted for algebraic factorization of the Reed-Muller (RM) form and its different variants, using the factorization rules of [1], as it is simple and requires much less CPU execution time compared to Boolean factorization operations. This technique has enabled us to greatly reduce the literal count as well as the gate count needed for such RM realizations, which are generally prone to consuming more cells and subsequently more power consumption. However, this leads to a drawback in terms of the design-for-test attribute associated with the various RM forms. Though we still preserve the definition of those forms viz. realizing such functionality with only select types of logic gates (AND gate and XOR gate), the structural integrity of the logic levels is not preserved. This would consequently alter the testability properties of such circuits i.e. it may increase/decrease/maintain the same number of test input vectors needed for their exhaustive testability, subsequently affecting their generalized test vector computation. We do not consider the issue of design-for-testability here, but, instead focus on the power consumption of the final logic implementation, after realization with a conventional CMOS process technology (0.35 micron TSMC process). The quality of the resulting circuits evaluated on the basis of an established cost metric viz., power consumption, demonstrate average savings by 26.79% for the samples considered in this work, besides reduction in number of gates and input literals by 39.66% and 12.98% respectively, in comparison with other factored RM forms.

Energy Efficient Clustering and Data Aggregation in Wireless Sensor Networks

Wireless Sensor Networks (WSNs) are wireless networks consisting of number of tiny, low cost and low power sensor nodes to monitor various physical phenomena like temperature, pressure, vibration, landslide detection, presence of any object, etc. The major limitation in these networks is the use of nonrechargeable battery having limited power supply. The main cause of energy consumption WSN is communication subsystem. This paper presents an efficient grid formation/clustering strategy known as Grid based level Clustering and Aggregation of Data (GCAD). The proposed clustering strategy is simple and scalable that uses low duty cycle approach to keep non-CH nodes into sleep mode thus reducing energy consumption. Simulation results demonstrate that our proposed GCAD protocol performs better in various performance metrics.