Multiwavelet and Biological Signal Processing

In this paper we are to find the optimum multiwavelet for compression of electrocardiogram (ECG) signals and then, selecting it for using with SPIHT codec. At present, it is not well known which multiwavelet is the best choice for optimum compression of ECG. In this work, we examine different multiwavelets on 24 sets of ECG data with entirely different characteristics, selected from MIT-BIH database. For assessing the functionality of the different multiwavelets in compressing ECG signals, in addition to known factors such as Compression Ratio (CR), Percent Root Difference (PRD), Distortion (D), Root Mean Square Error (RMSE) in compression literature, we also employed the Cross Correlation (CC) criterion for studying the morphological relations between the reconstructed and the original ECG signal and Signal to reconstruction Noise Ratio (SNR). The simulation results show that the Cardinal Balanced Multiwavelet (cardbal2) by the means of identity (Id) prefiltering method to be the best effective transformation. After finding the most efficient multiwavelet, we apply SPIHT coding algorithm on the transformed signal by this multiwavelet.

Combustion Characteristics of Syngas and Natural Gas in Micro-pilot Ignited Dual-fuel Engine

The objective of this study is to investigate the combustion in a pilot-ignited supercharged dual-fuel engine, fueled with different types of gaseous fuels under various equivalence ratios. It is found that if certain operating conditions are maintained, conventional dual-fuel engine combustion mode can be transformed to the combustion mode with the two-stage heat release. This mode of combustion was called the PREMIER (PREmixed Mixture Ignition in the End-gas Region) combustion. During PREMIER combustion, initially, the combustion progresses as the premixed flame propagation and then, due to the mixture autoignition in the end-gas region, ahead of the propagating flame front, the transition occurs with the rapid increase in the heat release rate.

Process and Supply-Chain Optimization for Testing and Verification of Formation Tester/Pressure-While- Drilling Tools

Applying a rigorous process to optimize the elements of a supply-chain network resulted in reduction of the waiting time for a service provider and customer. Different sources of downtime of hydraulic pressure controller/calibrator (HPC) were causing interruptions in the operations. The process examined all the issues to drive greater efficiencies. The issues included inherent design issues with HPC pump, contamination of the HPC with impurities, and the lead time required for annual calibration in the USA. HPC is used for mandatory testing/verification of formation tester/pressure measurement/logging-while drilling tools by oilfield service providers, including Halliburton. After market study andanalysis, it was concluded that the current HPC model is best suited in the oilfield industry. To use theexisting HPC model effectively, design andcontamination issues were addressed through design and process improvements. An optimum network is proposed after comparing different supply-chain models for calibration lead-time reduction.

Stress Relaxation of Date at Different Temperature and Moisture Content of Product: A New Approach

Iran is one of the greatest producers of date in the world. However due to lack of information about its viscoelastic properties, much of the production downgraded during harvesting and postharvesting processes. In this study the effect of temperature and moisture content of product were investigated on stress relaxation characteristics. Therefore, the freshly harvested date (kabkab) at tamar stage were put in controlled environment chamber to obtain different temperature levels (25, 35, 45, and 55 0C) and moisture contents (8.5, 8.7, 9.2, 15.3, 20, 32.2 %d.b.). A texture analyzer TAXT2 (Stable Microsystems, UK) was used to apply uniaxial compression tests. A chamber capable to control temperature was designed and fabricated around the plunger of texture analyzer to control the temperature during the experiment. As a new approach a CCD camera (A4tech, 30 fps) was mounted on a cylindrical glass probe to scan and record contact area between date and disk. Afterwards, pictures were analyzed using image processing toolbox of Matlab software. Individual date fruit was uniaxially compressed at speed of 1 mm/s. The constant strain of 30% of thickness of date was applied to the horizontally oriented fruit. To select a suitable model for describing stress relaxation of date, experimental data were fitted with three famous stress relaxation models including the generalized Maxwell, Nussinovitch, and Pelege. The constant in mentioned model were determined and correlated with temperature and moisture content of product using non-linear regression analysis. It was found that Generalized Maxwell and Nussinovitch models appropriately describe viscoelastic characteristics of date fruits as compared to Peleg mode.

Leaf Chlorophyll of Corn, Sweet basil and Borage under Intercropping System in Weed Interference

Intercropping is one of the sustainable agricultural factors. The SPAD meter can be used to predict nitrogen index reliably, it may also be a useful tool for assessing the relative impact of weeds on crops. In order to study the effect of weeds on SPAD in corn (Zea mays L.), sweet basil (Ocimum basilicum L.) and borage (Borago officinalis L.) in intercropping system, a factorial experiment was conducted in three replications in 2011. Experimental factors were included intercropping of corn with sweet basil and borage in different ratios (100:0, 75:25, 50:50, 25:75 and 0:100 corn: borage or sweet basil) and weed infestation (weed control and weed interference). The results showed that intercropping of corn with sweet basil and borage increased the SPAD value of corn compare to monoculture in weed interference condition. Sweet basil SPAD value in weed control treatments (43.66) was more than weed interference treatments (40.17). Corn could increase the borage SPAD value compare to monoculture in weed interference treatments.

Directors- Islamic Code of Ethics

This paper discusses a new model of Islamic code of ethics for directors. Several corporate scandals and local (example Transmile and Megan Media) and overseas corporate (example Parmalat and Enron) collapses show that the current corporate governance and regulatory reform are unable to prevent these events from recurring. Arguably, the code of ethics for directors is under research and the current code of ethics only concentrates on binding the work of the employee of the organization as a whole, without specifically putting direct attention to the directors, the group of people responsible for the performance of the company. This study used a semi-structured interview survey of well-known Islamic scholars such as the Mufti to develop the model. It is expected that the outcome of the research is a comprehensive model of code of ethics based on the Islamic principles that can be applied and used by the company to construct a code of ethics for their directors.

The Analysis of the Software Industry in Thailand

The software industry has been considered a critical infrastructure for any nation. Several studies have indicated that national competitiveness increasingly depends upon Information and Communication Technology (ICT), and software is one of the major components of ICT, important for both large and small enterprises. Even though there has been strong growth in the software industry in Thailand, the industry has faced many challenges and problems that need to be resolved. For example, the amount of pirated software has been rising, and Thailand still has a large gap in the digital divide. Additionally, the adoption among SMEs has been slow. This paper investigates various issues in the software industry in Thailand, using information acquired through analysis of secondary sources, observation, and focus groups. The results of this study can be used as “lessons learned" for the development of the software industry in any developing country.

Probabilistic Model Development for Project Performance Forecasting

In this paper, based on the past project cost and time performance, a model for forecasting project cost performance is developed. This study presents a probabilistic project control concept to assure an acceptable forecast of project cost performance. In this concept project activities are classified into sub-groups entitled control accounts. Then obtain the Stochastic S-Curve (SS-Curve), for each sub-group and the project SS-Curve is obtained by summing sub-groups- SS-Curves. In this model, project cost uncertainties are considered through Beta distribution functions of the project activities costs required to complete the project at every selected time sections through project accomplishment, which are extracted from a variety of sources. Based on this model, after a percentage of the project progress, the project performance is measured via Earned Value Management to adjust the primary cost probability distribution functions. Then, accordingly the future project cost performance is predicted by using the Monte-Carlo simulation method.

Non-negative Principal Component Analysis for Face Recognition

Principle component analysis is often combined with the state-of-art classification algorithms to recognize human faces. However, principle component analysis can only capture these features contributing to the global characteristics of data because it is a global feature selection algorithm. It misses those features contributing to the local characteristics of data because each principal component only contains some levels of global characteristics of data. In this study, we present a novel face recognition approach using non-negative principal component analysis which is added with the constraint of non-negative to improve data locality and contribute to elucidating latent data structures. Experiments are performed on the Cambridge ORL face database. We demonstrate the strong performances of the algorithm in recognizing human faces in comparison with PCA and NREMF approaches.

Designing and Implementing an Innovative Course about World Wide Web, Based on the Conceptual Representations of Students

Internet is nowadays included to all National Curriculums of the elementary school. A comparative study of their goals leads to the conclusion that a complete curriculum should aim to student-s acquisition of the abilities to navigate and search for information and additionally to emphasize on the evaluation of the information provided by the World Wide Web. In a constructivistic knowledge framework the design of a course has to take under consideration the conceptual representations of students. The following paper presents the conceptual representation of students of eleven years old, attending the Sixth Grade of Greek Elementary School about World Wide Web and their use in the design and implementation of an innovative course.

A Microscopic Simulation Model for Earthmoving Operations

Earthmoving operations are a major part of many construction projects. Because of the complexity and fast-changing environment of such operations, the planning and estimating are crucial on both planning and operational levels. This paper presents the framework ofa microscopic discrete-event simulation system for modeling earthmoving operations and conducting productivity estimations on an operational level.A prototype has been developed to demonstrate the applicability of the proposed framework, and this simulation system is presented via a case study based on an actual earthmoving project. The case study shows that the proposed simulation model is capable of evaluating alternative operating strategies and resource utilization at a very detailed level.

Identifying Significant Factors of Brick Laying Process through Design of Experiment and Computer Simulation: A Case Study

Improving performance measures in the construction processes has been a major concern for managers and decision makers in the industry. They seek for ways to recognize the key factors which have the largest effect on the process. Identifying such factors can guide them to focus on the right parts of the process in order to gain the best possible result. In the present study design of experiment (DOE) has been applied to a computer simulation model of brick laying process to determine significant factors while productivity has been chosen as the response of the experiment. To this end, four controllable factors and their interaction have been experimented and the best factor level has been calculated for each one. The results indicate that three factors, namely, labor of brick, labor of mortar and inter arrival time of mortar along with interaction of labor of brick and labor of mortar are significant.

Modeling of Dielectric Heating in Radio- Frequency Applicator Optimized for Uniform Temperature by Means of Genetic Algorithms

The paper presents an optimization study based on genetic algorithms (GA-s) for a radio-frequency applicator used in heating dielectric band products. The weakly coupled electro-thermal problem is analyzed using 2D-FEM. The design variables in the optimization process are: the voltage of a supplementary “guard" electrode and six geometric parameters of the applicator. Two objective functions are used: temperature uniformity and total active power absorbed by the dielectric. Both mono-objective and multiobjective formulations are implemented in GA optimization.

Hydrogen Integration in Petrochemical Complexes, Using Modified Automated Targeting Method

Owing to extensive use of hydrogen in refining or petrochemical units, it is essential to manage hydrogen network in order to make the most efficient utilization of hydrogen. On the other hand, hydrogen is an important byproduct not properly used through petrochemical complexes and mostly sent to the fuel system. A few works have been reported in literature to improve hydrogen network for petrochemical complexes. In this study a comprehensive analysis is carried out on petrochemical units using a modified automated targeting technique which is applied to determine the minimum hydrogen consumption. Having applied the modified targeting method in two petrochemical cases, the results showed a significant reduction in required fresh hydrogen.

Characterization and Development of Anthropomorphic Phantoms Liver for Use in Nuclear Medicine

The objective this study was to characterize and develop anthropomorphic liver phantoms in tomography hepatic procedures for quality control and improvement professionals in nuclear medicine. For the conformation of the anthropomorphic phantom was used in plaster and acrylic. We constructed three phantoms representing processes with liver cirrhosis. The phantoms were filled with 99mTc diluted with water to obtain the scintigraphic images. Tomography images were analyzed anterior and posterior phantom representing a body with a greater degree cirrhotic. It was noted that the phantoms allow the acquisition of images similar to real liver with cirrhosis. Simulations of hemangiomas may contribute to continued professional education of nuclear medicine, on the question of image acquisition, allowing of the study parameters such of the matrix, energy window and count statistics.

The Willingness of Business Students on T Innovative Behavior within the Theory of Planned Behavior

Classes on creativity, innovation, and entrepreneurship are becoming quite popular at universities throughout the world. However, it is not easy for business students to get involved to innovative activities, especially patent application. The present study investigated how to enhance business students- intention to participate in innovative activities and which incentives universities should consider. A 22-item research scale was used, and confirmatory factor analysis was conducted to verify its reliability and validity. Multiple regression and discriminant analyses were also conducted. The results demonstrate the effect of growth-need strength on innovative behavior and indicate that the theory of planned behavior can explain and predict business students- intention to participate in innovative activities. Additionally, the results suggest that applying our proposed model in practice would effectively strengthen business students- intentions to engage in innovative activities.

Modeling of Reinforcement in Concrete Beams Using Machine Learning Tools

The paper discusses the results obtained to predict reinforcement in singly reinforced beam using Neural Net (NN), Support Vector Machines (SVM-s) and Tree Based Models. Major advantage of SVM-s over NN is of minimizing a bound on the generalization error of model rather than minimizing a bound on mean square error over the data set as done in NN. Tree Based approach divides the problem into a small number of sub problems to reach at a conclusion. Number of data was created for different parameters of beam to calculate the reinforcement using limit state method for creation of models and validation. The results from this study suggest a remarkably good performance of tree based and SVM-s models. Further, this study found that these two techniques work well and even better than Neural Network methods. A comparison of predicted values with actual values suggests a very good correlation coefficient with all four techniques.

Effect of COD Loading Rate on Hydrogen Production from Alcohol Wastewater

The objective of this study was to investigate hydrogen production from alcohol wastewater by anaerobic sequencing batch reactor (ASBR) under thermophillic operation. The ASBR unit used in this study had a liquid holding volume of 4 L and was operated at 6 cycles per day. The seed sludge taken from an upflow anaerobic sludge blanket unit treating the same wastewater was boiled at 95 °C for 15 min before being fed to the ASBR unit. The ASBR system was operated at different COD loading rates at a thermophillic temperature (55 °C), and controlled pH of 5.5. When the system was operated under optimum conditions (providing maximum hydrogen production performance) at a feed COD of 60 000 mg/l, and a COD loading rate of 68 kg/m3 d, the produced gas contained 43 % H2 content in the produced gas. Moreover, the hydrogen yield and the specific hydrogen production rate (SHPR) were 130 ml H2/g COD removed and 2100 ml H2/l d, respectively.

Thermal Treatment Influence on the Quality of Rye Bread Packaged in Different Polymer Films

this study was carried out to investigate the changes in quality parameters of rye bread packaged in different polymer films during convection air-flow thermal treatment process. Whole loafs of bread were placed in polymer pouches, which were sealed in reduced pressure air ambiance, bread was thermally treated in at temperature +(130; 140; and 150) ± 5 ºC within 40min, as long as the core temperature of the samples have reached accordingly +80±1 ºC. For bread packaging pouches were used: anti-fog Mylar®OL12AF and thermo resistant combined polymer material. Main quality parameters was analysed using standard methods: temperature in bread core, bread crumb and crust firmness value, starch granules volume and microflora. In the current research it was proved, that polymer films significantly influence rye bread quality parameters changes during thermal treatment. Thermo resistant combined polymer material film could be recommendable for packaged rye bread pasteurization, for maximal bread quality parameter keeping.

Novelty as a Measure of Interestingness in Knowledge Discovery

Rule Discovery is an important technique for mining knowledge from large databases. Use of objective measures for discovering interesting rules leads to another data mining problem, although of reduced complexity. Data mining researchers have studied subjective measures of interestingness to reduce the volume of discovered rules to ultimately improve the overall efficiency of KDD process. In this paper we study novelty of the discovered rules as a subjective measure of interestingness. We propose a hybrid approach based on both objective and subjective measures to quantify novelty of the discovered rules in terms of their deviations from the known rules (knowledge). We analyze the types of deviation that can arise between two rules and categorize the discovered rules according to the user specified threshold. We implement the proposed framework and experiment with some public datasets. The experimental results are promising.