Kaolin for Production of Souvenirs

Ranong province has the best kaolin, and it is the most useful of all the clay types used in ceramic making. Until recently, there has been only one community business making ceramics in Ranong province. And this business could not build the mix of body and glaze from their raw material without assistance. Considering these problems, this research is aimed to test the composition of ceramic body and glaze which suit. Kaolin from Ranong is the raw material which these search focuses on. All other raw materials use in the investigation will come from southern Thailand, kaolin and limestone from Ranong province, ball clay from Surat Thani province, white sand from Songkhla province, and feldspar from Nakhon Si Thammarat province. Results can be used to develop the efficiency of industrial production which in return will enhance the business process.

Mathematical Model of Smoking Time Temperature Effect on Ribbed Smoked Sheets Quality

The quality of Ribbed Smoked Sheets (RSS) primarily based on color, dryness, and the presence or absence of fungus and bubbles. This quality is strongly influenced by the drying and fumigation process namely smoking process. Smoking that is held in high temperature long time will result scorched dark brown sheets, whereas if the temperature is too low or slow drying rate would resulted in less mature sheets and growth of fungus. Therefore need to find the time and temperature for optimum quality of sheets. Enhance, unmonitored heat and mass transfer during smoking process lead to high losses of energy balance. This research aims to generate simple empirical mathematical model describing the effect of smoking time and temperature to RSS quality of color, water content, fungus and bubbles. The second goal of study was to analyze energy balance during smoking process. Experimental study was conducted by measuring temperature, residence time and quality parameters of 16 sheets sample in smoking rooms. Data for energy consumption balance such as mass of fuel wood, mass of sheets being smoked, construction temperature, ambient temperature and relative humidity were taken directly along the smoking process. It was found that mathematical model correlating smoking temperature and time with color is Color = -169 - 0.184 T4 - 0.193 T3 - 0.160 0.405 T1 + T2 + 0.388 t1 +3.11 t2 + 3.92t3 + 0.215 t4 with R square 50.8% and with moisture is Moisture = -1.40-0.00123 T4 + 0.00032 T3 + 0.00260 T2 - 0.00292 T1 - 0.0105 t1 + 0.0290 t2 + 0.0452 t3 + 0.00061 t4 with R square of 49.9%. Smoking room energy analysis found useful energy was 27.8%. The energy stored in the material construction 7.3%. Lost of energy in conversion of wood combustion, ventilation and others were 16.6%. The energy flowed out through the contact of material construction with the ambient air was found to be the highest contribution to energy losses, it reached 48.3%.

A Power-Gating Scheme to Reduce Leakage Power for P-type Adiabatic Logic Circuits

With rapid technology scaling, the proportion of the static power consumption catches up with dynamic power consumption gradually. To decrease leakage consumption is becoming more and more important in low-power design. This paper presents a power-gating scheme for P-DTGAL (p-type dual transmission gate adiabatic logic) circuits to reduce leakage power dissipations under deep submicron process. The energy dissipations of P-DTGAL circuits with power-gating scheme are investigated in different processes, frequencies and active ratios. BSIM4 model is adopted to reflect the characteristics of the leakage currents. HSPICE simulations show that the leakage loss is greatly reduced by using the P-DTGAL with power-gating techniques.

The Spiral_OWL Model – Towards Spiral Knowledge Engineering

The Spiral development model has been used successfully in many commercial systems and in a good number of defense systems. This is due to the fact that cost-effective incremental commitment of funds, via an analogy of the spiral model to stud poker and also can be used to develop hardware or integrate software, hardware, and systems. To support adaptive, semantic collaboration between domain experts and knowledge engineers, a new knowledge engineering process, called Spiral_OWL is proposed. This model is based on the idea of iterative refinement, annotation and structuring of knowledge base. The Spiral_OWL model is generated base on spiral model and knowledge engineering methodology. A central paradigm for Spiral_OWL model is the concentration on risk-driven determination of knowledge engineering process. The collaboration aspect comes into play during knowledge acquisition and knowledge validation phase. Design rationales for the Spiral_OWL model are to be easy-to-implement, well-organized, and iterative development cycle as an expanding spiral.

Shape Restoration of the Left Ventricle

This paper describes an automatic algorithm to restore the shape of three-dimensional (3D) left ventricle (LV) models created from magnetic resonance imaging (MRI) data using a geometry-driven optimization approach. Our basic premise is to restore the LV shape such that the LV epicardial surface is smooth after the restoration. A geometrical measure known as the Minimum Principle Curvature (κ2) is used to assess the smoothness of the LV. This measure is used to construct the objective function of a two-step optimization process. The objective of the optimization is to achieve a smooth epicardial shape by iterative in-plane translation of the MRI slices. Quantitatively, this yields a minimum sum in terms of the magnitude of κ 2, when κ2 is negative. A limited memory quasi-Newton algorithm, L-BFGS-B, is used to solve the optimization problem. We tested our algorithm on an in vitro theoretical LV model and 10 in vivo patient-specific models which contain significant motion artifacts. The results show that our method is able to automatically restore the shape of LV models back to smoothness without altering the general shape of the model. The magnitudes of in-plane translations are also consistent with existing registration techniques and experimental findings.

Revisiting the Concept of Risk Analysis within the Context of Geospatial Database Design: A Collaborative Framework

The aim of this research is to design a collaborative framework that integrates risk analysis activities into the geospatial database design (GDD) process. Risk analysis is rarely undertaken iteratively as part of the present GDD methods in conformance to requirement engineering (RE) guidelines and risk standards. Accordingly, when risk analysis is performed during the GDD, some foreseeable risks may be overlooked and not reach the output specifications especially when user intentions are not systematically collected. This may lead to ill-defined requirements and ultimately in higher risks of geospatial data misuse. The adopted approach consists of 1) reviewing risk analysis process within the scope of RE and GDD, 2) analyzing the challenges of risk analysis within the context of GDD, and 3) presenting the components of a risk-based collaborative framework that improves the collection of the intended/forbidden usages of the data and helps geo-IT experts to discover implicit requirements and risks.

Using Target Costing to Investigates Competitive Price

This paper has presented research in progress concerning the contribution of target costing approach to achievement competitive price in the Iraqi firm. The title of the paper is one of the subjects that get large concerns in the finance and business world in the present time. That is because many competitive firms have appeared in the regional and global markets and the rapid changes that covered all fields of life. On the other hand, this paper concentrated on lack knowledge of the industrial firms, regarding the significant role of target cost for achieving the competitive prices. The paper depends on the main supposition, using the competitive price to get the target cost in the industrial firms. In order to achieve competitive advantage in business world the firms should rely on modern methods to manage cost and profit. From strategic perspective the target cost achieves a so powerful competitive advantage represented in cost reduction. Nevertheless the target cost does not exclude the calculation and survey of costs during the production process. Products- estimated costs are calculated and compared with the target costs.

Adaptive Neuro-Fuzzy Inference System for Financial Trading using Intraday Seasonality Observation Model

The prediction of financial time series is a very complicated process. If the efficient market hypothesis holds, then the predictability of most financial time series would be a rather controversial issue, due to the fact that the current price contains already all available information in the market. This paper extends the Adaptive Neuro Fuzzy Inference System for High Frequency Trading which is an expert system that is capable of using fuzzy reasoning combined with the pattern recognition capability of neural networks to be used in financial forecasting and trading in high frequency. However, in order to eliminate unnecessary input in the training phase a new event based volatility model was proposed. Taking volatility and the scaling laws of financial time series into consideration has brought about the development of the Intraday Seasonality Observation Model. This new model allows the observation of specific events and seasonalities in data and subsequently removes any unnecessary data. This new event based volatility model provides the ANFIS system with more accurate input and has increased the overall performance of the system.

Model of Continuous Cheese Whey Fermentation by Candida Pseudotropicalis

The utilization of cheese whey as a fermentation substrate to produce bio-ethanol is an effort to supply bio-ethanol demand as a renewable energy. Like other process systems, modeling is also required for fermentation process design, optimization and plant operation. This research aims to study the fermentation process of cheese whey by applying mathematics and fundamental concept in chemical engineering, and to investigate the characteristic of the cheese whey fermentation process. Steady state simulation results for inlet substrate concentration of 50, 100 and 150 g/l, and various values of hydraulic retention time, showed that the ethanol productivity maximum values were 0.1091, 0.3163 and 0.5639 g/l.h respectively. Those values were achieved at hydraulic retention time of 20 hours, which was the minimum value used in this modeling. This showed that operating reactor at low hydraulic retention time was favorable. Model of bio-ethanol production from cheese whey will enhance the understanding of what really happen in the fermentation process.

Decolourization of Melanoidin Containing Wastewater Using South African Coal Fly Ash

Batch adsorption of recalcitrant melanoidin using the abundantly available coal fly ash was carried out. It had low specific surface area (SBET) of 1.7287 m2/g and pore volume of 0.002245 cm3/g while qualitative evaluation of the predominant phases in it was done by XRD analysis. Colour removal efficiency was found to be dependent on various factors studied. Maximum colour removal was achieved around pH 6, whereas increasing sorbent mass from 10g/L to 200 g/L enhanced colour reduction from 25% to 86% at 298 K. Spontaneity of the process was suggested by negative Gibbs free energy while positive values for enthalpy change showed endothermic nature of the process. Non-linear optimization of error functions resulted in Freundlich and Redlich-Peterson isotherms describing sorption equilibrium data best. The coal fly ash had maximum sorption capacity of 53 mg/g and could thus be used as a low cost adsorbent in melanoidin removal.

Modeling Oxygen-transfer by Multiple Plunging Jets using Support Vector Machines and Gaussian Process Regression Techniques

The paper investigates the potential of support vector machines and Gaussian process based regression approaches to model the oxygen–transfer capacity from experimental data of multiple plunging jets oxygenation systems. The results suggest the utility of both the modeling techniques in the prediction of the overall volumetric oxygen transfer coefficient (KLa) from operational parameters of multiple plunging jets oxygenation system. The correlation coefficient root mean square error and coefficient of determination values of 0.971, 0.002 and 0.945 respectively were achieved by support vector machine in comparison to values of 0.960, 0.002 and 0.920 respectively achieved by Gaussian process regression. Further, the performances of both these regression approaches in predicting the overall volumetric oxygen transfer coefficient was compared with the empirical relationship for multiple plunging jets. A comparison of results suggests that support vector machines approach works well in comparison to both empirical relationship and Gaussian process approaches, and could successfully be employed in modeling oxygen-transfer.

The Data Mining usage in Production System Management

The paper gives the pilot results of the project that is oriented on the use of data mining techniques and knowledge discoveries from production systems through them. They have been used in the management of these systems. The simulation models of manufacturing systems have been developed to obtain the necessary data about production. The authors have developed the way of storing data obtained from the simulation models in the data warehouse. Data mining model has been created by using specific methods and selected techniques for defined problems of production system management. The new knowledge has been applied to production management system. Gained knowledge has been tested on simulation models of the production system. An important benefit of the project has been proposal of the new methodology. This methodology is focused on data mining from the databases that store operational data about the production process.

A Programmable FSK-Modulator in 350nm CMOS Technology

This paper describes the design of a programmable FSK-modulator based on VCO and its implementation in 0.35m CMOS process. The circuit is used to transmit digital data at 100Kbps rate in the frequency range of 400-600MHz. The design and operation of the modulator is discussed briefly. Further the characteristics of PLL, frequency synthesizer, VCO and the whole design are elaborated. The variation among the proposed and tested specifications is presented. Finally, the layout of sub-modules, pin configurations, final chip and test results are presented.

A Visual Educational Modeling Language to Help Teachers in Learning Scenario Design

The success of an e-learning system is highly dependent on the quality of its educational content and how effective, complete, and simple the design tool can be for teachers. Educational modeling languages (EMLs) are proposed as design languages intended to teachers for modeling diverse teaching-learning experiences, independently of the pedagogical approach and in different contexts. However, most existing EMLs are criticized for being too abstract and too complex to be understood and manipulated by teachers. In this paper, we present a visual EML that simplifies the process of designing learning scenarios for teachers with no programming background. Based on the conceptual framework of the activity theory, our resulting visual EML focuses on using Domainspecific modeling techniques to provide a pedagogical level of abstraction in the design process.

Adaptive Total Variation Based on Feature Scale

The widely used Total Variation de-noising algorithm can preserve sharp edge, while removing noise. However, since fixed regularization parameter over entire image, small details and textures are often lost in the process. In this paper, we propose a modified Total Variation algorithm to better preserve smaller-scaled features. This is done by allowing an adaptive regularization parameter to control the amount of de-noising in any region of image, according to relative information of local feature scale. Experimental results demonstrate the efficient of the proposed algorithm. Compared with standard Total Variation, our algorithm can better preserve smaller-scaled features and show better performance.

The Statistical Properties of Filtered Signals

In this paper, the statistical properties of filtered or convolved signals are considered by deriving the resulting density functions as well as the exact mean and variance expressions given a prior knowledge about the statistics of the individual signals in the filtering or convolution process. It is shown that the density function after linear convolution is a mixture density, where the number of density components is equal to the number of observations of the shortest signal. For circular convolution, the observed samples are characterized by a single density function, which is a sum of products.

Neural Network Based Determination of Splice Junctions by ROC Analysis

Gene, principal unit of inheritance, is an ordered sequence of nucleotides. The genes of eukaryotic organisms include alternating segments of exons and introns. The region of Deoxyribonucleic acid (DNA) within a gene containing instructions for coding a protein is called exon. On the other hand, non-coding regions called introns are another part of DNA that regulates gene expression by removing from the messenger Ribonucleic acid (RNA) in a splicing process. This paper proposes to determine splice junctions that are exon-intron boundaries by analyzing DNA sequences. A splice junction can be either exon-intron (EI) or intron exon (IE). Because of the popularity and compatibility of the artificial neural network (ANN) in genetic fields; various ANN models are applied in this research. Multi-layer Perceptron (MLP), Radial Basis Function (RBF) and Generalized Regression Neural Networks (GRNN) are used to analyze and detect the splice junctions of gene sequences. 10-fold cross validation is used to demonstrate the accuracy of networks. The real performances of these networks are found by applying Receiver Operating Characteristic (ROC) analysis.

Innovation Strategy in Slovak Businesses

The aim of the paper is based on detailed analysis of literary sources and carried out research to develop a model development and implementation of innovation strategy in the business. The paper brings the main results of the authors conducted research on a sample of 462 respondents that shows the current situation in the Slovak enterprises in the use of innovation strategy. Carried out research and analysis provided the base for a model development and implementation of innovation strategy in the business, which is in the paper in detail, step by step explained with emphasis on the implementation process. Implementing the innovation strategy is described a separate model. Paper contains recommendations for successful implementation of innovation strategy in the business. These recommendations should serve mainly business managers as valuable tool in implementing the innovation strategy.

The PARADIGMA Approach for Cooperative Work in the Medical Domain

PARADIGMA (PARticipative Approach to DIsease Global Management) is a pilot project which aims to develop and demonstrate an Internet based reference framework to share scientific resources and findings in the treatment of major diseases. PARADIGMA defines and disseminates a common methodology and optimised protocols (Clinical Pathways) to support service functions directed to patients and individuals on matters like prevention, posthospitalisation support and awareness. PARADIGMA will provide a platform of information services - user oriented and optimised against social, cultural and technological constraints - supporting the Health Care Global System of the Euro-Mediterranean Community in a continuous improvement process.

Lessons Learned from Observing User Behavior through Repeated Usability Evaluations

Academic research information service is a must for surveying previous studies in research and development process. OntoFrame is an academic research information service under Semantic Web framework different from simple keyword-based services such as CiteSeer and Google Scholar. The first purpose of this study is for revealing user behavior in their surveys, the objects of using academic research information services, and their needs. The second is for applying lessons learned from the results to OntoFrame.