Integral Operators Related to Problems of Interface Dynamics

This research work is concerned with the eigenvalue problem for the integral operators which are obtained by linearization of a nonlocal evolution equation. The purpose of section II.A is to describe the nature of the problem and the objective of the project. The problem is related to the “stable solution" of the evolution equation which is the so-called “instanton" that describe the interface between two stable phases. The analysis of the instanton and its asymptotic behavior are described in section II.C by imposing the Green function and making use of a probability kernel. As a result , a classical Theorem which is important for an instanton is proved. Section III devoted to a study of the integral operators related to interface dynamics which concern the analysis of the Cauchy problem for the evolution equation with initial data close to different phases and different regions of space.

Liquid-Liquid Equilibria for Ternary Mixtures of (Water + Carboxylic Acid+ MIBK), Experimental, Simulation, and Optimization

In this work, Experimental tie-line results and solubility (binodal) curves were obtained for the ternary systems (water + acetic acid + methyl isobutyl ketone (MIBK)), (water + lactic acid+ methyl isobutyl ketone) at T = 294.15K and atmospheric pressure. The consistency of the values of the experimental tie-lines was determined through the Othmer-Tobias and Hands correlations. For the extraction effectiveness of solvents, the distribution and selectivity curves were plotted. In addition, these experimental tieline data were also correlated with NRTL model. The interaction parameters for the NRTL model were retrieved from the obtained experimental results by means of a combination of the homotopy method and the genetic algorithms.

DAMQ-Based Approach for Efficiently Using the Buffer Spaces of a NoC Router

In this paper we present high performance dynamically allocated multi-queue (DAMQ) buffer schemes for fault tolerance systems on chip applications that require an interconnection network. Two virtual channels shared the same buffer space. Fault tolerant mechanisms for interconnection networks are becoming a critical design issue for large massively parallel computers. It is also important to high performance SoCs as the system complexity keeps increasing rapidly. On the message switching layer, we make improvement to boost system performance when there are faults involved in the components communication. The proposed scheme is when a node or a physical channel is deemed as faulty, the previous hop node will terminate the buffer occupancy of messages destined to the failed link. The buffer usage decisions are made at switching layer without interactions with higher abstract layer, thus buffer space will be released to messages destined to other healthy nodes quickly. Therefore, the buffer space will be efficiently used in case fault occurs at some nodes.

Evaluating and Selecting Optimization Software Packages: A Framework for Business Applications

Owing the fact that optimization of business process is a crucial requirement to navigate, survive and even thrive in today-s volatile business environment, this paper presents a framework for selecting a best-fit optimization package for solving complex business problems. Complexity level of the problem and/or using incorrect optimization software can lead to biased solutions of the optimization problem. Accordingly, the proposed framework identifies a number of relevant factors (e.g. decision variables, objective functions, and modeling approach) to be considered during the evaluation and selection process. Application domain, problem specifications, and available accredited optimization approaches are also to be regarded. A recommendation of one or two optimization software is the output of the framework which is believed to provide the best results of the underlying problem. In addition to a set of guidelines and recommendations on how managers can conduct an effective optimization exercise is discussed.

Study on the Effect of Sulphur, Glucose, Nitrogen and Plant Residues on the Immobilization of Sulphate-S in Soil

In order to evaluate the relationship between the sulphur (S), glucose (G), nitrogen (N) and plant residues (st), sulphur immobilization and microbial transformation were monitored in five soil samples from 0-30 cm of Bastam farmers fields of Shahrood area following 11 treatments with different levels of Sulphur (S), glucose (G), N and plant residues (wheat straw) in a randomized block design with three replications and incubated over 20, 45 and 60 days, the immobilization of SO4 -2-S presented as a percentage of that added, was inversely related to its addition rate. Additions of glucose and plant residues increased with the C-to-S ratio of the added amendments, irrespective of their origins (glucose and plant residues). In the presence of C sources (glucose or plant residues). N significantly increased the immobilization of SO4 -2-S, whilst the effect of N was insignificant in the absence of a C amendment. In first few days the amounts of added SO4 -2-S immobilized were linearly correlated with the amounts of added S recovered in the soil microbial biomass. With further incubation the proportions of immobilized SO4 -2-S remaining as biomass-S decreased. Decrease in biomass-S was thought to be due to the conversion of biomass-S into soil organic-S. Glucose addition increased the immobilization (microbial utilization and incorporation into the soil organic matter) of native soil SO4 -2-S. However, N addition enhance the mineralization of soil organic-S, increasing the concentration of SO4 - 2-S in soil.

Probe Selection for Pathway-Specific Microarray Probe Design Minimizing Melting Temperature Variance

In molecular biology, microarray technology is widely and successfully utilized to efficiently measure gene activity. If working with less studied organisms, methods to design custom-made microarray probes are available. One design criterion is to select probes with minimal melting temperature variances thus ensuring similar hybridization properties. If the microarray application focuses on the investigation of metabolic pathways, it is not necessary to cover the whole genome. It is more efficient to cover each metabolic pathway with a limited number of genes. Firstly, an approach is presented which minimizes the overall melting temperature variance of selected probes for all genes of interest. Secondly, the approach is extended to include the additional constraints of covering all pathways with a limited number of genes while minimizing the overall variance. The new optimization problem is solved by a bottom-up programming approach which reduces the complexity to make it computationally feasible. The new method is exemplary applied for the selection of microarray probes in order to cover all fungal secondary metabolite gene clusters for Aspergillus terreus.

Comparison of Different Techniques for Processing and Preserving fish Rastrineobola argentea from Lake Victoria, Kenya

This study was set to determine the antimicrobial activities of brine salting, chlorinated solution, and oil frying treatments on enteric bacteria and fungi in Rastrineobola argentea fish from fish landing beaches within L. Victoria basin of western Kenya. Statistical differences in effectiveness of the different treatment methods was determined by single factor ANOVA, and paired two-tail t-Test was performed to compare the differences in moisture contents before and after storage. Oil fried fish recorded the lowest microbial loads, sodium chloride at 10% concentration was the second most effective and chlorinated solution even at 150ppm was the least effective against the bacteria and fungi in fish. Moisture contents of the control and treated fish were significantly lower after storage. These results show that oil frying of fish should be adopted for processing and preserving Rastrineobola argentea which is the most abundant and affordable fish species from Lake Victoria.

The Grey Relational Analysis of the Influence Factors of Profit in Cartoon-s Character Merchandising Rights

This paper constructs a four factors theoretical model of Chinese small and medium enterprises based on the “cartoon characters- reputation - enterprise marketing and management capabilities – protection of the cartoon image - institutional environment" by literature research, case studies and investigation. The empirical study show that the greatest impact on current merchandising rights income is the institutional environment friendliness, followed by marketing and management capabilities, input of character image protection and Cartoon characters- reputation through the real-time grey relational analysis, and the greatest impact on post-merchandising rights profit is Cartoon characters reputation, followed by the institutional environment friendliness, then marketing and management ability and input of character image protection through the time-delay grey relational analysis.

mCRM-s New Opportunities of Customer Satisfaction

This paper aims at a new challenge of customer satisfaction on mobile customer relationship management. In this paper presents a conceptualization of mCRM on its unique characteristics of customer satisfaction. Also, this paper develops an empirical framework in conception of customer satisfaction in mCRM. A single-case study is applied as the methodology. In order to gain an overall view of the empirical case, this paper accesses to invisible and important information of company in this investigation. Interview is the key data source form the main informants of the company through which the issues are identified and the proposed framework is built. It supports the development of customer satisfaction in mCRM; links this theoretical framework into practice; and provides the direction for future research. Therefore, this paper is very useful for the industries as it helps them to understand how customer satisfaction changes the mCRM structure and increase the business competitive advantage. Finally, this paper provides a contribution in practice by linking a theoretical framework in conception of customer satisfaction in mCRM for companies to a practical real case.

Heuristic Continuous-time Associative Memories

In this paper, a novel associative memory model will be proposed and applied to memory retrievals based on the conventional continuous time model. The conventional model presents memory capacity is very low and retrieval process easily converges to an equilibrium state which is very different from the stored patterns. Genetic Algorithms is well-known with the capability of global optimal search escaping local optimum on progress to reach a global optimum. Based on the well-known idea of Genetic Algorithms, this work proposes a heuristic rule to make a mutation when the state of the network is trapped in a spurious memory. The proposal heuristic associative memory show the stored capacity does not depend on the number of stored patterns and the retrieval ability is up to ~ 1.

A Reproduction of Boundary Conditions in Three-Dimensional Continuous Casting Problem

The paper discusses a 3D numerical solution of the inverse boundary problem for a continuous casting process of alloy. The main goal of the analysis presented within the paper was to estimate heat fluxes along the external surface of the ingot. The verified information on these fluxes was crucial for a good design of a mould, effective cooling system and generally the whole caster. In the study an enthalpy-porosity technique implemented in Fluent package was used for modeling the solidification process. In this method, the phase change interface was determined on the basis of the liquid fraction approach. In inverse procedure the sensitivity analysis was applied for retrieving boundary conditions. A comparison of the measured and retrieved values showed a high accuracy of the computations. Additionally, the influence of the accuracy of measurements on the estimated heat fluxes was also investigated.

Estimation of Forest Fire Emission in Thailand by Using Remote Sensing Information

The forest fires in Thailand are annual occurrence which is the cause of air pollutions. This study intended to estimate the emission from forest fire during 2005-2009 using MODerateresolution Imaging Spectro-radiometer (MODIS) sensor aboard the Terra and Aqua satellites, experimental data, and statistical data. The forest fire emission is estimated using equation established by Seiler and Crutzen in 1982. The spatial and temporal variation of forest fire emission is analyzed and displayed in the form of grid density map. From the satellite data analysis suggested between 2005 and 2009, the number of fire hotspots occurred 86,877 fire hotspots with a significant highest (more than 80% of fire hotspots) in the deciduous forest. The peak period of the forest fire is in January to May. The estimation on the emissions from forest fires during 2005 to 2009 indicated that the amount of CO, CO2, CH4, and N2O was about 3,133,845 tons, 47,610.337 tons, 204,905 tons, and 6,027 tons, respectively, or about 6,171,264 tons of CO2eq. They also emitted 256,132 tons of PM10. The year 2007 was found to be the year when the emissions were the largest. Annually, March is the period that has the maximum amount of forest fire emissions. The areas with high density of forest fire emission were the forests situated in the northern, the western, and the upper northeastern parts of the country.

Promoting Electric Vehicles for Sustainable Urban Transport: How to Do It This Time Right

In recent years various types of electric vehicles has gained again increasing attention as an environmentally benign technology in transport. Especially for urban areas with high local pollution this Zero-emission technology (at the point of use) is considered to provide proper solutions. Yet, the bad economics and the limited driving ranges are still major barriers for a broader market penetration of battery electric vehicles (BEV) and of fuel cell vehicles (FCV). The major result of our analyses is that the most important precondition for a further dissemination of BEV in urban areas are emission-free zones. This is an instrument which allows the promotion of BEV without providing excessive subsidies. In addition, it is important to note that the full benefits of EV can only be harvested if the electricity used is produced from renewable energy sources. That is to say, it has to be ensured that the use of BEV in urban areas is clearly linked to a green electricity purchase model. And moreover, the introduction of a CO2- emission-based tax system would support this requirement.

Development and in vitro Characterization of Self-nanoemulsifying Drug Delivery Systems of Valsartan

The present study is aim to prepare and evaluate the selfnanoemulsifying drug delivery (SNEDDS) system of a poorly water soluble drug valsartan in order to achieve a better dissolution rate which would further help in enhancing oral bioavailability. The present research work describes a SNEDDS of valsartan using labrafil M 1944 CS, Tween 80 and Transcutol HP. The pseudoternary phase diagrams with presence and absence of drug were plotted to check for the emulsification range and also to evaluate the effect of valsartan on the emulsification behavior of the phases. The mixtures consisting of oil (labrafil M 1944 CS) with surfactant (tween 80), co-surfactant (Transcutol HP) were found to be optimum formulations. Prepared formulations were evaluated for its particle size distribution, nanoemulsifying properties, robustness to dilution, self emulsication time, turbidity measurement, drug content and invitro dissolution. The optimized formulations are further evaluated for heating cooling cycle, centrifugation studies, freeze thaw cycling, particle size distribution and zeta potential were carried out to confirm the stability of the formed SNEDDS formulations. The prepared formulation revealed t a significant improvement in terms of the drug solubility as compared with marketed tablet and pure drug.

Cognitive Radio Networks (CRN): Resource Allocation Techniques Based On DNA-inspired Computing

Spectrum is a scarce commodity, and considering the spectrum scarcity faced by the wireless-based service providers led to high congestion levels. Technical inefficiencies from pooled, since all networks share a common pool of channels, exhausting the available channels will force networks to block the services. Researchers found that cognitive radio (CR) technology may resolve the spectrum scarcity. A CR is a self-configuring entity in a wireless networking that senses its environment, tracks changes, and frequently exchanges information with their networks. However, CRN facing challenges and condition become worst while tracks changes i.e. reallocation of another under-utilized channels while primary network user arrives. In this paper, channels or resource reallocation technique based on DNA-inspired computing algorithm for CRN has been proposed.

Estimating Frequency, Amplitude and Phase of Two Sinusoids with Very Close Frequencies

This paper presents an algorithm to estimate the parameters of two closely spaced sinusoids, providing a frequency resolution that is more than 800 times greater than that obtained by using the Discrete Fourier Transform (DFT). The strategy uses a highly optimized grid search approach to accurately estimate frequency, amplitude and phase of both sinusoids, keeping at the same time the computational effort at reasonable levels. The proposed method has three main characteristics: 1) a high frequency resolution; 2) frequency, amplitude and phase are all estimated at once using one single package; 3) it does not rely on any statistical assumption or constraint. Potential applications to this strategy include the difficult task of resolving coincident partials of instruments in musical signals.

Error Effects on SAR Image Resolution using Range Doppler Imaging Algorithm

Synthetic Aperture Radar (SAR) is an imaging radar form by taking full advantage of the relative movement of the antenna with respect to the target. Through the simultaneous processing of the radar reflections over the movement of the antenna via the Range Doppler Algorithm (RDA), the superior resolution of a theoretical wider antenna, termed synthetic aperture, is obtained. Therefore, SAR can achieve high resolution two dimensional imagery of the ground surface. In addition, two filtering steps in range and azimuth direction provide accurate enough result. This paper develops a simulation in which realistic SAR images can be generated. Also, the effect of velocity errors in the resulting image has also been investigated. Taking some velocity errors into account, the simulation results on the image resolution would be presented. Most of the times, algorithms need to be adjusted for particular datasets, or particular applications.

Context Modeling and Reasoning Approach in Context-Aware Middleware for URC System

To realize the vision of ubiquitous computing, it is important to develop a context-aware infrastructure which can help ubiquitous agents, services, and devices become aware of their contexts because such computational entities need to adapt themselves to changing situations. A context-aware infrastructure manages the context model representing contextual information and provides appropriate information. In this paper, we introduce Context-Aware Middleware for URC System (hereafter CAMUS) as a context-aware infrastructure for a network-based intelligent robot system and discuss the ontology-based context modeling and reasoning approach which is used in that infrastructure.

New Hybrid Algorithm for Task Scheduling in Grid Computing to Decrease missed Task

The purpose of Grid computing is to utilize computational power of idle resources which are distributed in different areas. Given the grid dynamism and its decentralize resources, there is a need for an efficient scheduler for scheduling applications. Since task scheduling includes in the NP-hard problems various researches have focused on invented algorithms especially the genetic ones. But since genetic is an inherent algorithm which searches the problem space globally and does not have the efficiency required for local searching, therefore, its combination with local searching algorithms can compensate for this shortcomings. The aim of this paper is to combine the genetic algorithm and GELS (GAGELS) as a method to solve scheduling problem by which simultaneously pay attention to two factors of time and number of missed tasks. Results show that the proposed algorithm can decrease makespan while minimizing the number of missed tasks compared with the traditional methods.

On Positive Definite Solutions of Quaternionic Matrix Equations

The real representation of the quaternionic matrix is definited and studied. The relations between the positive (semi)define quaternionic matrix and its real representation matrix are presented. By means of the real representation, the relation between the positive (semi)definite solutions of quaternionic matrix equations and those of corresponding real matrix equations is established.