Supercritical Fluid Extraction of Lutein Esters from Marigold Flowers and their Hydrolysis by Improved Saponification and Enzyme Biocatalysis

Lutein is a dietary oxycarotenoid which is found to reduce the risks of Age-related Macular Degeneration (AMD). Supercritical fluid extraction of lutein esters from marigold petals was carried out and was found to be much effective than conventional solvent extraction. The saponification of pre-concentrated lutein esters to produce free lutein was studied which showed a composition of about 88% total carotenoids (UV-VIS spectrophotometry) and 90.7% lutein (HPLC). The lipase catalyzed hydrolysis of lutein esters in conventional medium was investigated. The optimal temperature, pH, enzyme concentration and water activity were found to be 50°C, 7, 15% and 0.33 respectively and the activity loss of lipase was about 25% after 8 times re-use in at 50°C for 12 days. However, the lipase catalyzed hydrolysis of lutein esters in conventional media resulted in poor conversions (16.4%).

An Algorithm for Autonomous Aerial Navigation using MATLAB® Mapping Tool Box

In the present era of aviation technology, autonomous navigation and control have emerged as a prime area of active research. Owing to the tremendous developments in the field, autonomous controls have led today’s engineers to claim that future of aerospace vehicle is unmanned. Development of guidance and navigation algorithms for an unmanned aerial vehicle (UAV) is an extremely challenging task, which requires efforts to meet strict, and at times, conflicting goals of guidance and control. In this paper, aircraft altitude and heading controllers and an efficient algorithm for self-governing navigation using MATLAB® mapping toolbox is presented which also enables loitering of a fixed wing UAV over a specified area. For this purpose, a nonlinear mathematical model of a UAV is used. The nonlinear model is linearized around a stable trim point and decoupled for controller design. The linear controllers are tested on the nonlinear aircraft model and navigation algorithm is subsequently developed for for autonomous flight of the UAV. The results are presented for trajectory controllers and waypoint based navigation. Our investigation reveals that MATLAB® mapping toolbox can be exploited to successfully deliver an efficient algorithm for autonomous aerial navigation for a UAV.

Financial Regulations in the Process of Global Financial Crisis and Macroeconomics Impact of Basel III

Basel III (or the Third Basel Accord) is a global regulatory standard on bank capital adequacy, stress testing and market liquidity risk agreed upon by the members of the Basel Committee on Banking Supervision in 2010-2011, and scheduled to be introduced from 2013 until 2018. Basel III is a comprehensive set of reform measures. These measures aim to; (1) improve the banking sector-s ability to absorb shocks arising from financial and economic stress, whatever the source, (2) improve risk management and governance, (3) strengthen banks- transparency and disclosures. Similarly the reform target; (1) bank level or micro-prudential, regulation, which will help raise the resilience of individual banking institutions to periods of stress. (2) Macro-prudential regulations, system wide risk that can build up across the banking sector as well as the pro-cyclical implication of these risks over time. These two approaches to supervision are complementary as greater resilience at the individual bank level reduces the risk system wide shocks. Macroeconomic impact of Basel III; OECD estimates that the medium-term impact of Basel III implementation on GDP growth is in the range -0,05 percent to -0,15 percent per year. On the other hand economic output is mainly affected by an increase in bank lending spreads as banks pass a rise in banking funding costs, due to higher capital requirements, to their customers. Consequently the estimated effects on GDP growth assume no active response from monetary policy. Basel III impact on economic output could be offset by a reduction (or delayed increase) in monetary policy rates by about 30 to 80 basis points. The aim of this paper is to create a framework based on the recent regulations in order to prevent financial crises. Thus the need to overcome the global financial crisis will contribute to financial crises that may occur in the future periods. In the first part of the paper, the effects of the global crisis on the banking system examine the concept of financial regulations. In the second part; especially in the financial regulations and Basel III are analyzed. The last section in this paper explored the possible consequences of the macroeconomic impacts of Basel III.

Turkic - Indian Lexical Parallels in the Framework of the Nostratic Language's Macrofamily

From ancient times Turkic languages have been in contact with numerous representatives of different language families. The article discusses the Turkic - Indian language contact and were shown promise and necessity of this trend for the Turkic linguistics, were given Turkic - Indian lexical parallels in the framework of the nostratic language's macro family. The research work has done on the base of lexical parallels (LP) -of Turkic (which belong to the Altaic family of languages) and Indian (including Dravidian and Indo-Aryan languages).

Towards for Admission Control in WIMAX Relay Station Mesh Network for Mobile Stations out of Coverage Using Ad-Hoc

WIMAX relay station mesh network has been approved by IEEE 802.16j as a standard to provide a highly data rate transmission, the RS was implemented to extend the coverage zone of the BS, for instance the MSs previously were out of the coverage of the BS they become in the coverage of the RS, therefore these MSs can have Admission control from the BS through the RS. This paper describe a problem in the mesh network Relay station, for instance the problem of how to serve the mobile stations (MSs) which are out of the Relay station coverage. This paper also proposed a solution for mobile stations out of the coverage of the WIMAX Relay stations mesh Network. Therefore Ad-hoc network defined as a solution by using its admission control schema and apply it on the mobiles inside and outside the Relay station coverage.

Generalization of SGIP Surface Tension Force Model in Three-Dimensional Flows and Compare to Other Models in Interfacial Flows

In this paper, the two-dimensional stagger grid interface pressure (SGIP) model has been generalized and presented into three-dimensional form. For this purpose, various models of surface tension force for interfacial flows have been investigated and compared with each other. The VOF method has been used for tracking the interface. To show the ability of the SGIP model for three-dimensional flows in comparison with other models, pressure contours, maximum spurious velocities, norm spurious flow velocities and pressure jump error for motionless drop of liquid and bubble of gas are calculated using different models. It has been pointed out that SGIP model in comparison with the CSF, CSS and PCIL models produces the least maximum and norm spurious velocities. Additionally, the new model produces more accurate results in calculating the pressure jumps across the interface for motionless drop of liquid and bubble of gas which is generated in surface tension force.

Towards Better Understanding of the Concept of Tacit Knowledge – A Cognitive Approach

Tacit knowledge has been one of the most discussed and contradictory concepts in the field of knowledge management since the mid 1990s. The concept is used relatively vaguely to refer to any type of information that is difficult to articulate, which has led to discussions about the original meaning of the concept (adopted from Polanyi-s philosophy) and the nature of tacit knowing. It is proposed that the subject should be approached from the perspective of cognitive science in order to connect tacit knowledge to empirically studied cognitive phenomena. Some of the most important examples of tacit knowing presented by Polanyi are analyzed in order to trace the cognitive mechanisms of tacit knowing and to promote better understanding of the nature of tacit knowledge. The cognitive approach to Polanyi-s theory reveals that the tacit/explicit typology of knowledge often presented in the knowledge management literature is not only artificial but totally opposite approach compared to Polanyi-s thinking.

Enhancing Multi-Frame Images Using Self-Delaying Dynamic Networks

This paper presents the use of a newly created network structure known as a Self-Delaying Dynamic Network (SDN) to create a high resolution image from a set of time stepped input frames. These SDNs are non-recurrent temporal neural networks which can process time sampled data. SDNs can store input data for a lifecycle and feature dynamic logic based connections between layers. Several low resolution images and one high resolution image of a scene were presented to the SDN during training by a Genetic Algorithm. The SDN was trained to process the input frames in order to recreate the high resolution image. The trained SDN was then used to enhance a number of unseen noisy image sets. The quality of high resolution images produced by the SDN is compared to that of high resolution images generated using Bi-Cubic interpolation. The SDN produced images are superior in several ways to the images produced using Bi-Cubic interpolation.

Design and Operation of a Multicarrier Energy System Based On Multi Objective Optimization Approach

Multi-energy systems will enhance the system reliability and power quality. This paper presents an integrated approach for the design and operation of distributed energy resources (DER) systems, based on energy hub modeling. A multi-objective optimization model is developed by considering an integrated view of electricity and natural gas network to analyze the optimal design and operating condition of DER systems, by considering two conflicting objectives, namely, minimization of total cost and the minimization of environmental impact which is assessed in terms of CO2 emissions. The mathematical model considers energy demands of the site, local climate data, and utility tariff structure, as well as technical and financial characteristics of the candidate DER technologies. To provide energy demands, energy systems including photovoltaic, and co-generation systems, boiler, central power grid are considered. As an illustrative example, a hotel in Iran demonstrates potential applications of the proposed method. The results prove that increasing the satisfaction degree of environmental objective leads to increased total cost.

Principal Component Analysis using Singular Value Decomposition of Microarray Data

A series of microarray experiments produces observations of differential expression for thousands of genes across multiple conditions. Principal component analysis(PCA) has been widely used in multivariate data analysis to reduce the dimensionality of the data in order to simplify subsequent analysis and allow for summarization of the data in a parsimonious manner. PCA, which can be implemented via a singular value decomposition(SVD), is useful for analysis of microarray data. For application of PCA using SVD we use the DNA microarray data for the small round blue cell tumors(SRBCT) of childhood by Khan et al.(2001). To decide the number of components which account for sufficient amount of information we draw scree plot. Biplot, a graphic display associated with PCA, reveals important features that exhibit relationship between variables and also the relationship of variables with observations.

Evaluating Sinusoidal Functions by a Low Complexity Cubic Spline Interpolator with Error Optimization

We present a novel scheme to evaluate sinusoidal functions with low complexity and high precision using cubic spline interpolation. To this end, two different approaches are proposed to find the interpolating polynomial of sin(x) within the range [- π , π]. The first one deals with only a single data point while the other with two to keep the realization cost as low as possible. An approximation error optimization technique for cubic spline interpolation is introduced next and is shown to increase the interpolator accuracy without increasing complexity of the associated hardware. The architectures for the proposed approaches are also developed, which exhibit flexibility of implementation with low power requirement.

Raman Scattering and PL Studies on AlGaN/GaN HEMT Layers on 200 mm Si(111)

The crystalline quality of the AlGaN/GaN high electron mobility transistor (HEMT) structure grown on a 200 mm silicon substrate has been investigated using UV-visible micro- Raman scattering and photoluminescence (PL). The visible Raman scattering probes the whole nitride stack with the Si substrate and shows the presence of a small component of residual in-plane stress in the thick GaN buffer resulting from a wafer bowing, while the UV micro-Raman indicates a tensile interfacial stress induced at the top GaN/AlGaN/AlN layers. PL shows a good crystal quality GaN channel where the yellow band intensity is very low compared to that of the near-band-edge transition. The uniformity of this sample is shown by measurements from several points across the epiwafer.

Performance Evaluation Standards and Innovation: An Empirical Investigation

In this empirical research, how marketing managers evaluate their firms- performances and decide to make innovation is examined. They use some standards which are past performance of the firm, target performance of the firm, competitor performance, and average performance of the industry to compare and evaluate the firms- performances. It is hypothesized that marketing managers and owners of the firm compare the firms- current performance with these four standards at the same time to decide when to make innovation relating to any aspects of the firm, either management style or products. Relationship between the comparison of the firm-s performance with these standards and innovation are searched in the same regression model. The results of the regression analysis are discussed and some recommendations are made for future studies and applicants.

The Analysis of Printing Quality of Offset - Printing Ink with Coconut Oil Base

The objectives of this research are to produce prototype coconut oil based solvent offset printing inks and to analyze a basic quality of printing work derived from coconut oil based solvent offset printing inks, by mean of bringing coconut oil for producing varnish and bringing such varnish to produce black offset printing inks. Then, analysis of qualities i.e. CIELAB value, density value, and dot gain value of printing work from coconut oil based solvent offset printing inks which printed on gloss-coated woodfree paper weighs 130 grams were done. The research result of coconut oil based solvent offset printing inks indicated that the suitable varnish formulation is using 51% of coconut oil, 36% of phenolic resin, and 14% of solvent oil 14%, while the result of producing black offset ink displayed that the suitable formula of printing ink is using varnish mixed with 20% of coconut oil, and the analyzing printing work of coconut oil based solvent offset printing inks which printed on paper, the results were as follows: CIELAB value of black offset printing ink is at L* = 31.90, a* = 0.27, and b* = 1.86, density value is at 1.27 and dot gain value was high at mid tone area of image area.

How Prior Knowledge Affects User's Understanding of System Requirements?

Requirements are critical to system validation as they guide all subsequent stages of systems development. Inadequately specified requirements generate systems that require major revisions or cause system failure entirely. Use Cases have become the main vehicle for requirements capture in many current Object Oriented (OO) development methodologies, and a means for developers to communicate with different stakeholders. In this paper we present the results of a laboratory experiment that explored whether different types of use case format are equally effective in facilitating high knowledge user-s understanding. Results showed that the provision of diagrams along with the textual use case descriptions significantly improved user comprehension of system requirements in both familiar and unfamiliar application domains. However, when comparing groups that received models of textual description accompanied with diagrams of different level of details (simple and detailed) we found no significant difference in performance.

Effect Comparison of Speckle Noise Reduction Filters on 2D-Echocardigraphic Images

Echocardiography imaging is one of the most common diagnostic tests that are widely used for assessing the abnormalities of the regional heart ventricle function. The main goal of the image enhancement task in 2D-echocardiography (2DE) is to solve two major anatomical structure problems; speckle noise and low quality. Therefore, speckle noise reduction is one of the important steps that used as a pre-processing to reduce the distortion effects in 2DE image segmentation. In this paper, we present the common filters that based on some form of low-pass spatial smoothing filters such as Mean, Gaussian, and Median. The Laplacian filter was used as a high-pass sharpening filter. A comparative analysis was presented to test the effectiveness of these filters after being applied to original 2DE images of 4-chamber and 2-chamber views. Three statistical quantity measures: root mean square error (RMSE), peak signal-to-ratio (PSNR) and signal-tonoise ratio (SNR) are used to evaluate the filter performance quantitatively on the output enhanced image.

Conditions on Blind Source Separability of Linear FIR-MIMO Systems with Binary Inputs

In this note, we investigate the blind source separability of linear FIR-MIMO systems. The concept of semi-reversibility of a system is presented. It is shown that for a semi-reversible system, if the input signals belong to a binary alphabet, then the source data can be blindly separated. One sufficient condition for a system to be semi-reversible is obtained. It is also shown that the proposed criteria is weaker than that in the literature which requires that the channel matrix is irreducible/invertible or reversible.

An FPGA Implementation of Intelligent Visual Based Fall Detection

Falling has been one of the major concerns and threats to the independence of the elderly in their daily lives. With the worldwide significant growth of the aging population, it is essential to have a promising solution of fall detection which is able to operate at high accuracy in real-time and supports large scale implementation using multiple cameras. Field Programmable Gate Array (FPGA) is a highly promising tool to be used as a hardware accelerator in many emerging embedded vision based system. Thus, it is the main objective of this paper to present an FPGA-based solution of visual based fall detection to meet stringent real-time requirements with high accuracy. The hardware architecture of visual based fall detection which utilizes the pixel locality to reduce memory accesses is proposed. By exploiting the parallel and pipeline architecture of FPGA, our hardware implementation of visual based fall detection using FGPA is able to achieve a performance of 60fps for a series of video analytical functions at VGA resolutions (640x480). The results of this work show that FPGA has great potentials and impacts in enabling large scale vision system in the future healthcare industry due to its flexibility and scalability.

A Programmer’s Survey of the Quantum Computing Paradigm

Research in quantum computation is looking for the consequences of having information encoding, processing and communication exploit the laws of quantum physics, i.e. the laws which govern the ultimate knowledge that we have, today, of the foreign world of elementary particles, as described by quantum mechanics. This paper starts with a short survey of the principles which underlie quantum computing, and of some of the major breakthroughs brought by the first ten to fifteen years of research in this domain; quantum algorithms and quantum teleportation are very biefly presented. The next sections are devoted to one among the many directions of current research in the quantum computation paradigm, namely quantum programming languages and their semantics. A few other hot topics and open problems in quantum information processing and communication are mentionned in few words in the concluding remarks, the most difficult of them being the physical implementation of a quantum computer. The interested reader will find a list of useful references at the end of the paper.

Baking Quality of Hulled Wheat Species in Organic Farming

The organic farmers use wider range of crop varieties than the conventional farming. Bread wheat is the most favorite and the most common food crop. The organic bread wheat is usually of worse technological quality. Therefore, it is supposed to be an attractive alternative to the hulled wheat species (einkorn, emmer wheat and spelt). Twenty-five hulled bread wheat varieties and control bread wheat ones were grown on the certified organic parcel in České Budějovice (the Czech Republic) between 2009 and 2012. Their baking quality was measured and evaluated with standard methods, and in accordance with ICC. The results have shown that the grain of hulled wheat varieties contain a lot of proteins in grains (up to 18 percent); even the organic hulled bread wheat varieties are characterized by such good baking quality. Einkorn and emmer wheat are of worse technological quality of proteins (low values of gluten index and Zeleny test), which is a disadvantage of these two wheat species. On the other hand, spelt wheat is of better technological quality and is similar to the control bread wheat varieties. Mixtures consisting of bread wheat, among others, are considered good alternatives; they may contribute to wider range of use of the hulled wheat species. It is one of the possibilities which may increase the proportion of proteins in bread wheat grains; the nutrition-rich hulled wheat grains may be also used in such way at the same time.