The Possibility-Probability Relationship for Bloodstream Concentrations of Physiologically Active Substances

If a possibility distribution and a probability distribution are describing values x of one and the same system or process x(t), can they relate to each other? Though in general the possibility and probability distributions might be not connected at all, we can assume that in some particular cases there is an association linked them. In the presented paper, we consider distributions of bloodstream concentrations of physiologically active substances and propose that the probability to observe a concentration x of a substance X can be produced from the possibility of the event X = x . The proposed assumptions and resulted theoretical distributions are tested against the data obtained from various panel studies of the bloodstream concentrations of the different physiologically active substances in patients and healthy adults as well.

Effect of Humic Acid on Physical and Engineering Properties of Lime-Treated Organic Clay

The present work deals with the stabilisation of organic clay using hydrated lime. Artificial organic clays were prepared by adding kaolin and different humic acid contents. Results given by physical testing show that the presence of humic acid has a drawback effect on the untreated organic clay. The decrease in specific gravity value was accompanied by a decrease in dry density and plasticity of clay at higher humic acid contents. Significant increase in shear strength at 7 days of curing period is observed in the lime-treated samples up to 5% lime content. However shear strength of lime-treated organic clay decreases at longer curing periods. The results given by laboratory testing is further verified by microstructure analysis. Based on the results obtained in this study, it can be concluded that the presence of more than 1.5% humic acid reduces significantly the efficiency of lime stabilization in organic clays.

Gait Recognition System: Bundle Rectangle Approach

Biometrics methods include recognition techniques such as fingerprint, iris, hand geometry, voice, face, ears and gait. The gait recognition approach has some advantages, for example it does not need the prior concern of the observed subject and it can record many biometric features in order to make deeper analysis, but most of the research proposals use high computational cost. This paper shows a gait recognition system with feature subtraction on a bundle rectangle drawn over the observed person. Statistical results within a database of 500 videos are shown.

Effect of a Gravel Bed Flocculator on the Efficiency of a Low Cost Water Treatment Plants

The principal objective of a water treatment plant is to produce water that satisfies a set of drinking water quality standards at a reasonable price to the consumers. The gravel-bed flocculator provide a simple and inexpensive design for flocculation in small water treatment plants (less than 5000 m3/day capacity). The packed bed of gravel provides ideal conditions for the formation of compact settleable flocs because of continuous recontact provided by the sinuous flow of water through the interstices formed by the gravel. The field data which were obtained from the operation of the water supply treatment unit cover the physical, chemical and biological water qualities of the raw and settled water as obtained by the operation of the treatment unit. The experiments were carried out with the aim of assessing the efficiency of the gravel filter in removing the turbidity, pathogenic bacteria, from the raw water. The water treatment plant, which was constructed for the treatment of river water, was in principle a rapid sand filter. The results show that the average value of the turbidity level of the settled water was 4.83 NTU with a standard deviation of turbidity 2.893 NTU. This indicated that the removal efficiency of the sedimentation tank (gravel filter) was about 67.8 %. for pH values fluctuated between 7.75 and 8.15, indicating the alkaline nature of the raw water of the river Shatt Al-Hilla, as expected. Raw water pH is depressed slightly following alum coagulation. The pH of the settled water ranged from 7.75 to a maximum of 8.05. The bacteriological tests which were carried out on the water samples were: total coliform test, E-coli test, and the plate count test. In each test the procedure used was as outlined in the Standard Methods for the Examination of Water and Wastewater (APHA, AWWA, and WPCF, 1985). The gravel filter exhibit a low performance in removing bacterial load. The percentage bacterial removal, which is maximum for total plate count (19%) and minimum for total coliform (16.82%).

A Software Framework for Predicting Oil-Palm Yield from Climate Data

Intelligent systems based on machine learning techniques, such as classification, clustering, are gaining wide spread popularity in real world applications. This paper presents work on developing a software system for predicting crop yield, for example oil-palm yield, from climate and plantation data. At the core of our system is a method for unsupervised partitioning of data for finding spatio-temporal patterns in climate data using kernel methods which offer strength to deal with complex data. This work gets inspiration from the notion that a non-linear data transformation into some high dimensional feature space increases the possibility of linear separability of the patterns in the transformed space. Therefore, it simplifies exploration of the associated structure in the data. Kernel methods implicitly perform a non-linear mapping of the input data into a high dimensional feature space by replacing the inner products with an appropriate positive definite function. In this paper we present a robust weighted kernel k-means algorithm incorporating spatial constraints for clustering the data. The proposed algorithm can effectively handle noise, outliers and auto-correlation in the spatial data, for effective and efficient data analysis by exploring patterns and structures in the data, and thus can be used for predicting oil-palm yield by analyzing various factors affecting the yield.

The Applicability of the Zipper Strut to Seismic Rehabilitation of Steel Structures

Chevron frames (Inverted-V-braced frames or Vbraced frames) have seismic disadvantages, such as not good exhibit force redistribution capability and compression brace buckles immediately. Researchers developed new design provisions on increasing both the ductility and lateral resistance of these structures in seismic areas. One of these new methods is adding zipper columns, as proposed by Khatib et al. (1988) [2]. Zipper columns are vertical members connecting the intersection points of the braces above the first floor. In this paper applicability of the suspended zipper system to Seismic Rehabilitation of Steel Structures is investigated. The models are 3-, 6-, 9-, and 12-story Inverted-V-braced frames. In this case, it is assumed that the structures must be rehabilitated. For rehabilitation of structures, zipper column is used. The result of researches showed that the suspended zipper system is effective in case of 3-, 6-, and 9-story Inverted-V-braced frames and it would increase lateral resistance of structure up to life safety level. But in case of high-rise buildings (such as 12 story frame), it doesn-t show good performance. For solving this problem, the braced bay can consist of small “units" over the height of the entire structure, which each of them is a zipper-braced bay with a few stories. By using this method the lateral resistance of 12 story Inverted-V-braced frames is increased up to safety life level.

Downlink Scheduling and Radio Resource Allocation in Adaptive OFDMA Wireless Communication Systems for User-Individual QoS

In this paper, we address the problem of adaptive radio resource allocation (RRA) and packet scheduling in the downlink of a cellular OFDMA system, and propose a downlink multi-carrier proportional fair (MPF) scheduler and its joint with adaptive RRA algorithm to distribute radio resources among multiple users according to their individual QoS requirements. The allocation and scheduling objective is to maximize the total throughput, while at the same time maintaining the fairness among users. The simulation results demonstrate that the methods presented provide for user more explicit fairness relative to RRA algorithm, but the joint scheme achieves the higher sum-rate capacity with flexible parameters setting compared with MPF scheduler.

Retail Inventory Management for Perishable Products with Two Bins Strategy

Perishable goods constitute a large portion of retailer inventory and lose value with time due to deterioration and/or obsolescence. Retailers dealing with such goods required considering the factors of short shelf life and the dependency of sales on inventory displayed in determining optimal procurement policy. Many retailers follow the practice of using two bins - primary bin sales fresh items at a list price and secondary bin sales unsold items at a discount price transferred from primary bin on attaining certain age. In this paper, mathematical models are developed for primary bin and for secondary bin that maximizes profit with decision variables of order quantities, optimal review period and optimal selling price at secondary bin. The demand rates in two bins are assumed to be deterministic and dependent on displayed inventory level, price and age but independent of each other. The validity of the model is shown by solving an example and the sensitivity analysis of the model is also reported.

Effects of Sodium Bicarbonate Content and Vulcanization Method on Properties of NBR/PVC Thermal Insulator Foam

In this research sodium bicarbonate (NaHCO3) was introduced to generate carbon dioxide gas (CO2) to the existing nitrogen gas (N2) of elastomeric foam, to lower thermal conductivity (K). Various loadings of NaHCO3 (0 to 60 phr) were added into the azodicarbonamide (AZC)-containing compound and its properties were then determined. Two vulcanization methods, i.e., hot air and infrared (IR), were employed and compared in this study. Results revealed that compound viscosity tended to increase slightly with increasing NaHCO3 content but cure time was delayed. The effect of NaHCO3 content on thermal conductivity depended on the vulcanization method. For hot air method, the thermal conductivity was insignificantly changed with increasing NaHCO3 up to 40 phr whereas it tended to decrease gradually for IR method. At higher NaHCO3 content (60 phr), unexpected increase of thermal conductivity was observed. The water absorption was also determined and foam structures were then used to explain the results.

Chemical Analysis of PM2.5 during Dry Deforestation Season in Southeast Asia

In Southeast Asia, during the dry season (August to October) forest fires in Indonesia emit pollutants into the atmosphere. For two years during this period, a total of 67 samples of 2.5 μm particulate matters were collected and analyzed for total mass and elemental composition with ICP - MS after microwave digestion. A study of 60 elements measured during these periods suggest that the concentration of most of elements, even those usually related to crustal source, are extremely high and unpredictable during the haze period. In By contrast, trace element concentration in non - haze months is more stable and covers a lower range. Other unexpected events and their effects on the findings are discussed.

Non-Smooth Economic Dispatch Solution by Using Enhanced Bat-Inspired Optimization Algorithm

Economic dispatch (ED) has been considered to be one of the key functions in electric power system operation which can help to build up effective generating management plans. The practical ED problem has non-smooth cost function with nonlinear constraints which make it difficult to be effectively solved. This paper presents a novel heuristic and efficient optimization approach based on the new Bat algorithm (BA) to solve the practical non-smooth economic dispatch problem. The proposed algorithm easily takes care of different constraints. In addition, two newly introduced modifications method is developed to improve the variety of the bat population when increasing the convergence speed simultaneously. The simulation results obtained by the proposed algorithms are compared with the results obtained using other recently develop methods available in the literature.

Initialization Method of Reference Vectors for Improvement of Recognition Accuracy in LVQ

Initial values of reference vectors have significant influence on recognition accuracy in LVQ. There are several existing techniques, such as SOM and k-means, for setting initial values of reference vectors, each of which has provided some positive results. However, those results are not sufficient for the improvement of recognition accuracy. This study proposes an ACO-used method for initializing reference vectors with an aim to achieve recognition accuracy higher than those obtained through conventional methods. Moreover, we will demonstrate the effectiveness of the proposed method by applying it to the wine data and English vowel data and comparing its results with those of conventional methods.

Comparison of Different Gas Turbine Inlet Air Cooling Methods

Gas turbine air inlet cooling is a useful method for increasing output for regions where significant power demand and highest electricity prices occur during the warm months. Inlet air cooling increases the power output by taking advantage of the gas turbine-s feature of higher mass flow rate when the compressor inlet temperature decreases. Different methods are available for reducing gas turbine inlet temperature. There are two basic systems currently available for inlet cooling. The first and most cost-effective system is evaporative cooling. Evaporative coolers make use of the evaporation of water to reduce the gas turbine-s inlet air temperature. The second system employs various ways to chill the inlet air. In this method, the cooling medium flows through a heat exchanger located in the inlet duct to remove heat from the inlet air. However, the evaporative cooling is limited by wet-bulb temperature while the chilling can cool the inlet air to temperatures that are lower than the wet bulb temperature. In the present work, a thermodynamic model of a gas turbine is built to calculate heat rate, power output and thermal efficiency at different inlet air temperature conditions. Computational results are compared with ISO conditions herein called "base-case". Therefore, the two cooling methods are implemented and solved for different inlet conditions (inlet temperature and relative humidity). Evaporative cooler and absorption chiller systems results show that when the ambient temperature is extremely high with low relative humidity (requiring a large temperature reduction) the chiller is the more suitable cooling solution. The net increment in the power output as a function of the temperature decrease for each cooling method is also obtained.

Multi-Criteria Decision-Making Selection Model with Application to Chemical Engineering Management Decisions

Chemical industry project management involves complex decision making situations that require discerning abilities and methods to make sound decisions. Project managers are faced with decision environments and problems in projects that are complex. In this work, case study is Research and Development (R&D) project selection. R&D is an ongoing process for forward thinking technology-based chemical industries. R&D project selection is an important task for organizations with R&D project management. It is a multi-criteria problem which includes both tangible and intangible factors. The ability to make sound decisions is very important to success of R&D projects. Multiple-criteria decision making (MCDM) approaches are major parts of decision theory and analysis. This paper presents all of MCDM approaches for use in R&D project selection. It is hoped that this work will provide a ready reference on MCDM and this will encourage the application of the MCDM by chemical engineering management.

Combined Simulated Annealing and Genetic Algorithm to Solve Optimization Problems

Combinatorial optimization problems arise in many scientific and practical applications. Therefore many researchers try to find or improve different methods to solve these problems with high quality results and in less time. Genetic Algorithm (GA) and Simulated Annealing (SA) have been used to solve optimization problems. Both GA and SA search a solution space throughout a sequence of iterative states. However, there are also significant differences between them. The GA mechanism is parallel on a set of solutions and exchanges information using the crossover operation. SA works on a single solution at a time. In this work SA and GA are combined using new technique in order to overcome the disadvantages' of both algorithms.

A New Iterative Method for Solving Nonlinear Equations

In this study, a new root-finding method for solving nonlinear equations is proposed. This method requires two starting values that do not necessarily bracketing a root. However, when the starting values are selected to be close to a root, the proposed method converges to the root quicker than the secant method. Another advantage over all iterative methods is that; the proposed method usually converges to two distinct roots when the given function has more than one root, that is, the odd iterations of this new technique converge to a root and the even iterations converge to another root. Some numerical examples, including a sine-polynomial equation, are solved by using the proposed method and compared with results obtained by the secant method; perfect agreements are found.

A New Algorithm to Stereo Correspondence Using Rank Transform and Morphology Based On Genetic Algorithm

This paper presents a novel algorithm of stereo correspondence with rank transform. In this algorithm we used the genetic algorithm to achieve the accurate disparity map. Genetic algorithms are efficient search methods based on principles of population genetic, i.e. mating, chromosome crossover, gene mutation, and natural selection. Finally morphology is employed to remove the errors and discontinuities.

A Weighted Sum Technique for the Joint Optimization of Performance and Power Consumption in Data Centers

With data centers, end-users can realize the pervasiveness of services that will be one day the cornerstone of our lives. However, data centers are often classified as computing systems that consume the most amounts of power. To circumvent such a problem, we propose a self-adaptive weighted sum methodology that jointly optimizes the performance and power consumption of any given data center. Compared to traditional methodologies for multi-objective optimization problems, the proposed self-adaptive weighted sum technique does not rely on a systematical change of weights during the optimization procedure. The proposed technique is compared with the greedy and LR heuristics for large-scale problems, and the optimal solution for small-scale problems implemented in LINDO. the experimental results revealed that the proposed selfadaptive weighted sum technique outperforms both of the heuristics and projects a competitive performance compared to the optimal solution.

Mining Genes Relations in Microarray Data Combined with Ontology in Colon Cancer Automated Diagnosis System

MATCH project [1] entitle the development of an automatic diagnosis system that aims to support treatment of colon cancer diseases by discovering mutations that occurs to tumour suppressor genes (TSGs) and contributes to the development of cancerous tumours. The constitution of the system is based on a) colon cancer clinical data and b) biological information that will be derived by data mining techniques from genomic and proteomic sources The core mining module will consist of the popular, well tested hybrid feature extraction methods, and new combined algorithms, designed especially for the project. Elements of rough sets, evolutionary computing, cluster analysis, self-organization maps and association rules will be used to discover the annotations between genes, and their influence on tumours [2]-[11]. The methods used to process the data have to address their high complexity, potential inconsistency and problems of dealing with the missing values. They must integrate all the useful information necessary to solve the expert's question. For this purpose, the system has to learn from data, or be able to interactively specify by a domain specialist, the part of the knowledge structure it needs to answer a given query. The program should also take into account the importance/rank of the particular parts of data it analyses, and adjusts the used algorithms accordingly.