Microbial Leaching Process to Recover Valuable Metals from Spent Petroleum Catalyst Using Iron Oxidizing Bacteria

Spent petroleum catalyst from Korean petrochemical industry contains trace amount of metals such as Ni, V and Mo. Therefore an attempt was made to recover those trace metal using bioleaching process. Different leaching parameters such as Fe(II) concentration, pulp density, pH, temperature and particle size of spent catalyst particle were studied to evaluate their effects on the leaching efficiency. All the three metal ions like Ni, V and Mo followed dual kinetics, i.e., initial faster followed by slower rate. The percentage of leaching efficiency of Ni and V were higher than Mo. The leaching process followed a diffusion controlled model and the product layer was observed to be impervious due to formation of ammonium jarosite (NH4)Fe3(SO4)2(OH)6. In addition, the lower leaching efficiency of Mo was observed due to a hydrophobic coating of elemental sulfur over Mo matrix in the spent catalyst.

Denoising by Spatial Domain Averaging for Wireless Local Area Network Terminal Localization

Terminal localization for indoor Wireless Local Area Networks (WLANs) is critical for the deployment of location-aware computing inside of buildings. A major challenge is obtaining high localization accuracy in presence of fluctuations of the received signal strength (RSS) measurements caused by multipath fading. This paper focuses on reducing the effect of the distance-varying noise by spatial filtering of the measured RSS. Two different survey point geometries are tested with the noise reduction technique: survey points arranged in sets of clusters and survey points uniformly distributed over the network area. The results show that the location accuracy improves by 16% when the filter is used and by 18% when the filter is applied to a clustered survey set as opposed to a straight-line survey set. The estimated locations are within 2 m of the true location, which indicates that clustering the survey points provides better localization accuracy due to superior noise removal.

Prototype for Enhancing Information Security Awareness in Industry

Human-related information security breaches within organizations are primarily caused by employees who have not been made aware of the importance of protecting the information they work with. Information security awareness is accordingly attracting more attention from industry, because stakeholders are held accountable for the information with which they work. The authors developed an Information Security Retrieval and Awareness model – entitled “ISRA" – that is tailored specifically towards enhancing information security awareness in industry amongst all users of information, to address shortcomings in existing information security awareness models. This paper is principally aimed at expounding a prototype for the ISRA model to highlight the advantages of utilizing the model. The prototype will focus on the non-technical, humanrelated information security issues in industry. The prototype will ensure that all stakeholders in an organization are part of an information security awareness process, and that these stakeholders are able to retrieve specific information related to information security issues relevant to their job category, preventing them from being overburdened with redundant information.

Minimization of Power Loss in Distribution Networks by Different Techniques

Accurate loss minimization is the critical component for efficient electrical distribution power flow .The contribution of this work presents loss minimization in power distribution system through feeder restructuring, incorporating DG and placement of capacitor. The study of this work was conducted on IEEE distribution network and India Electricity Board benchmark distribution system. The executed experimental result of Indian system is recommended to board and implement practically for regulated stable output.

Optimization Approaches for a Complex Dairy Farm Simulation Model

This paper describes the optimization of a complex dairy farm simulation model using two quite different methods of optimization, the Genetic algorithm (GA) and the Lipschitz Branch-and-Bound (LBB) algorithm. These techniques have been used to improve an agricultural system model developed by Dexcel Limited, New Zealand, which describes a detailed representation of pastoral dairying scenarios and contains an 8-dimensional parameter space. The model incorporates the sub-models of pasture growth and animal metabolism, which are themselves complex in many cases. Each evaluation of the objective function, a composite 'Farm Performance Index (FPI)', requires simulation of at least a one-year period of farm operation with a daily time-step, and is therefore computationally expensive. The problem of visualization of the objective function (response surface) in high-dimensional spaces is also considered in the context of the farm optimization problem. Adaptations of the sammon mapping and parallel coordinates visualization are described which help visualize some important properties of the model-s output topography. From this study, it is found that GA requires fewer function evaluations in optimization than the LBB algorithm.

A Low Power SRAM Base on Novel Word-Line Decoding

This paper proposes a low power SRAM based on five transistor SRAM cell. Proposed SRAM uses novel word-line decoding such that, during read/write operation, only selected cell connected to bit-line whereas, in conventional SRAM (CV-SRAM), all cells in selected row connected to their bit-lines, which in turn develops differential voltages across all bit-lines, and this makes energy consumption on unselected bit-lines. In proposed SRAM memory array divided into two halves and this causes data-line capacitance to reduce. Also proposed SRAM uses one bit-line and thus has lower bit-line leakage compared to CV-SRAM. Furthermore, the proposed SRAM incurs no area overhead, and has comparable read/write performance versus the CV-SRAM. Simulation results in standard 0.25μm CMOS technology shows in worst case proposed SRAM has 80% smaller dynamic energy consumption in each cycle compared to CV-SRAM. Besides, energy consumption in each cycle of proposed SRAM and CV-SRAM investigated analytically, the results of which are in good agreement with the simulation results.

Applying 5S Lean Technology: An Infrastructure for Continuous Process Improvement

This paper presents an application of 5S lean technology to a production facility. Due to increased demand, high product variety, and a push production system, the plant has suffered from excessive wastes, unorganized workstations, and unhealthy work environment. This has translated into increased production cost, frequent delays, and low workers morale. Under such conditions, it has become difficult, if not impossible, to implement effective continuous improvement studies. Hence, the lean project is aimed at diagnosing the production process, streamlining the workflow, removing/reducing process waste, cleaning the production environment, improving plant layout, and organizing workstations. 5S lean technology is utilized for achieving project objectives. The work was a combination of both culture changes and tangible/physical changes on the shop floor. The project has drastically changed the plant and developed the infrastructure for a successful implementation of continuous improvement as well as other best practices and quality initiatives.

Simulation of Enhanced Biomass Gasification for Hydrogen Production using iCON

Due to the environmental and price issues of current energy crisis, scientists and technologists around the globe are intensively searching for new environmentally less-impact form of clean energy that will reduce the high dependency on fossil fuel. Particularly hydrogen can be produced from biomass via thermochemical processes including pyrolysis and gasification due to the economic advantage and can be further enhanced through in-situ carbon dioxide removal using calcium oxide. This work focuses on the synthesis and development of the flowsheet for the enhanced biomass gasification process in PETRONAS-s iCON process simulation software. This hydrogen prediction model is conducted at operating temperature between 600 to 1000oC at atmospheric pressure. Effects of temperature, steam-to-biomass ratio and adsorbent-to-biomass ratio were studied and 0.85 mol fraction of hydrogen is predicted in the product gas. Comparisons of the results are also made with experimental data from literature. The preliminary economic potential of developed system is RM 12.57 x 106 which equivalent to USD 3.77 x 106 annually shows economic viability of this process.

Investigation of Behavior on the Contact Surface of the Tire and Ground by CFD Simulation

Tread design has evolved over the years to achieve the common tread pattern used in current vehicles. However, to meet safety and comfort requirements, tread design considers more than one design factor. Tread design must consider the grip and drainage, and the manner in which to reduce rolling noise, which is one of the main factors considered by manufacturers. The main objective of this study was the application the computational fluid dynamics (CFD) technique to simulate the contact surface of the tire and ground. The results demonstrated an air-Pumping and large pressure drop effect in the process of contact surface. The results also revealed that the pressure can be used to analyze sound pressure level (SPL).

Software Model for a Computer Based Training for an HVDC Control Desk Simulator

With major technological advances and to reduce the cost of training apprentices for real-time critical systems, it was necessary the development of Intelligent Tutoring Systems for training apprentices in these systems. These systems, in general, have interactive features so that the learning is actually more efficient, making the learner more familiar with the mechanism in question. In the home stage of learning, tests are performed to obtain the student's income, a measure on their use. The aim of this paper is to present a framework to model an Intelligent Tutoring Systems using the UML language. The various steps of the analysis are considered the diagrams required to build a general model, whose purpose is to present the different perspectives of its development.

An Integrated Model of Urban Conservation and Revitalization from the Point of Immigration and Its Effects on Reyhan Urban Site in Turkey as a Case Study

This paper presents the effects of migration at the urban sites with an integrated model under the sustainable local development policies for the conservation and revitalization of the site areas as a case at Reyhan heritage site in Bursa. It is known as the “City of immigrants" because of its richness of cultural plurality. The city has always regarded the dynamic impact of immigration as a positive contribution. As a result of this situation, the city created the earliest urbanization practices: being the first capital city of the Ottoman Empire. Bursa created the first modern movement practices and set the first Organized Industrial Zone. The most important aim of the study is to be offer a model for the similar areas with the context of conservation and revitalization of the historical areas, subjected to the local integrated sustainable development policies of local goverments.

The Potential Use of Nanofilters to Supply Potable Water in Persian Gulf and Oman Sea Watershed Basin

In a world worried about water resources with the shadow of drought and famine looming all around, the quality of water is as important as its quantity. The source of all concerns is the constant reduction of per capita quality water for different uses. Iran With an average annual precipitation of 250 mm compared to the 800 mm world average, Iran is considered a water scarce country and the disparity in the rainfall distribution, the limitations of renewable resources and the population concentration in the margins of desert and water scarce areas have intensified the problem. The shortage of per capita renewable freshwater and its poor quality in large areas of the country, which have saline, brackish or hard water resources, and the profusion of natural and artificial pollutant have caused the deterioration of water quality. Among methods of treatment and use of these waters one can refer to the application of membrane technologies, which have come into focus in recent years due to their great advantages. This process is quite efficient in eliminating multi-capacity ions; and due to the possibilities of production at different capacities, application as treatment process in points of use, and the need for less energy in comparison to Reverse Osmosis processes, it can revolutionize the water and wastewater sector in years to come. The article studied the different capacities of water resources in the Persian Gulf and Oman Sea watershed basins, and processes the possibility of using nanofiltration process to treat brackish and non-conventional waters in these basins.

A Trainable Neural Network Ensemble for ECG Beat Classification

This paper illustrates the use of a combined neural network model for classification of electrocardiogram (ECG) beats. We present a trainable neural network ensemble approach to develop customized electrocardiogram beat classifier in an effort to further improve the performance of ECG processing and to offer individualized health care. We process a three stage technique for detection of premature ventricular contraction (PVC) from normal beats and other heart diseases. This method includes a denoising, a feature extraction and a classification. At first we investigate the application of stationary wavelet transform (SWT) for noise reduction of the electrocardiogram (ECG) signals. Then feature extraction module extracts 10 ECG morphological features and one timing interval feature. Then a number of multilayer perceptrons (MLPs) neural networks with different topologies are designed. The performance of the different combination methods as well as the efficiency of the whole system is presented. Among them, Stacked Generalization as a proposed trainable combined neural network model possesses the highest recognition rate of around 95%. Therefore, this network proves to be a suitable candidate in ECG signal diagnosis systems. ECG samples attributing to the different ECG beat types were extracted from the MIT-BIH arrhythmia database for the study.

Modeling of Session Initiation Protocol Invite Transaction using Colored Petri Nets

Wireless mobile communications have experienced the phenomenal growth through last decades. The advances in wireless mobile technologies have brought about a demand for high quality multimedia applications and services. For such applications and services to work, signaling protocol is required for establishing, maintaining and tearing down multimedia sessions. The Session Initiation Protocol (SIP) is an application layer signaling protocols, based on request/response transaction model. This paper considers SIP INVITE transaction over an unreliable medium, since it has been recently modified in Request for Comments (RFC) 6026. In order to help in assuring that the functional correctness of this modification is achieved, the SIP INVITE transaction is modeled and analyzed using Colored Petri Nets (CPNs). Based on the model analysis, it is concluded that the SIP INVITE transaction is free of livelocks and dead codes, and in the same time it has both desirable and undesirable deadlocks. Therefore, SIP INVITE transaction should be subjected for additional updates in order to eliminate undesirable deadlocks. In order to reduce the cost of implementation and maintenance of SIP, additional remodeling of the SIP INVITE transaction is recommended.

Complex Energy Signal Model for Digital Human Fingerprint Matching

This paper describes a complex energy signal model that is isomorphic with digital human fingerprint images. By using signal models, the problem of fingerprint matching is transformed into the signal processing problem of finding a correlation between two complex signals that differ by phase-rotation and time-scaling. A technique for minutiae matching that is independent of image translation, rotation and linear-scaling, and is resistant to missing minutiae is proposed. The method was tested using random data points. The results show that for matching prints the scaling and rotation angles are closely estimated and a stronger match will have a higher correlation.

Research on IBR-Driven Distributed Collaborative Visualization System

Image-based Rendering(IBR) techniques recently reached in broad fields which leads to a critical challenge to build up IBR-Driven visualization platform where meets requirement of high performance, large bounds of distributed visualization resource aggregation and concentration, multiple operators deploying and CSCW design employing. This paper presents an unique IBR-based visualization dataflow model refer to specific characters of IBR techniques and then discusses prominent feature of IBR-Driven distributed collaborative visualization (DCV) system before finally proposing an novel prototype. The prototype provides a well-defined three level modules especially work as Central Visualization Server, Local Proxy Server and Visualization Aid Environment, by which data and control for collaboration move through them followed the previous dataflow model. With aid of this triple hierarchy architecture of that, IBR oriented application construction turns to be easy. The employed augmented collaboration strategy not only achieve convenient multiple users synchronous control and stable processing management, but also is extendable and scalable.

A Survey: Clustering Ensembles Techniques

The clustering ensembles combine multiple partitions generated by different clustering algorithms into a single clustering solution. Clustering ensembles have emerged as a prominent method for improving robustness, stability and accuracy of unsupervised classification solutions. So far, many contributions have been done to find consensus clustering. One of the major problems in clustering ensembles is the consensus function. In this paper, firstly, we introduce clustering ensembles, representation of multiple partitions, its challenges and present taxonomy of combination algorithms. Secondly, we describe consensus functions in clustering ensembles including Hypergraph partitioning, Voting approach, Mutual information, Co-association based functions and Finite mixture model, and next explain their advantages, disadvantages and computational complexity. Finally, we compare the characteristics of clustering ensembles algorithms such as computational complexity, robustness, simplicity and accuracy on different datasets in previous techniques.

SLM Using Riemann Sequence Combined with DCT Transform for PAPR Reduction in OFDM Communication Systems

Orthogonal Frequency Division Multiplexing (OFDM) is an efficient method of data transmission for high speed communication systems. However, the main drawback of OFDM systems is that, it suffers from the problem of high Peak-to-Average Power Ratio (PAPR) which causes inefficient use of the High Power Amplifier and could limit transmission efficiency. OFDM consist of large number of independent subcarriers, as a result of which the amplitude of such a signal can have high peak values. In this paper, we propose an effective reduction scheme that combines DCT and SLM techniques. The scheme is composed of the DCT followed by the SLM using the Riemann matrix to obtain phase sequences for the SLM technique. The simulation results show PAPR can be greatly reduced by applying the proposed scheme. In comparison with OFDM, while OFDM had high values of PAPR –about 10.4dB our proposed method achieved about 4.7dB reduction of the PAPR with low complexities computation. This approach also avoids randomness in phase sequence selection, which makes it simpler to decode at the receiver. As an added benefit, the matrices can be generated at the receiver end to obtain the data signal and hence it is not required to transmit side information (SI).

Wind-tunnel Measurement of the Drag-reducing Effect of Compliant Coating

A specially designed flat plate was mounted vertically over the axial line in the wind tunnel of the Aerospace Department of the Pusan National University. The plate is 2 m long, 0.8 m high and 8 cm thick. The measurements were performed in velocity range from 15 to 60 m/s. A sand paper turbulizer was placed close to the plate nose to provide fully developed turbulent boundary layer over the most part of the plate. Strain balances were mounted in the trailing part of the plate to measure the skin friction drag over removable insertions of 0.55×0.25m2 size. A set of the insertions was designed and manufactured: 3mm thick polished metal surface and three compliant surfaces. The compliant surfaces were manufactured of a silicone rubber Silastic® S2 (Dow Corning company). To modify the viscoelastic properties of the rubber, its composition was varied: 90% of the rubber + 10% catalyst (standard), 92.5% + 7.5% (weak), 85% + 15% (strong). Modulus of elasticity and the loss tangent were measured accurately for these materials in the frequency range from 40 Hz to 3 KHz using the unique proposed technique.

A New Evolutionary Algorithm for Cluster Analysis

Clustering is a very well known technique in data mining. One of the most widely used clustering techniques is the kmeans algorithm. Solutions obtained from this technique depend on the initialization of cluster centers and the final solution converges to local minima. In order to overcome K-means algorithm shortcomings, this paper proposes a hybrid evolutionary algorithm based on the combination of PSO, SA and K-means algorithms, called PSO-SA-K, which can find better cluster partition. The performance is evaluated through several benchmark data sets. The simulation results show that the proposed algorithm outperforms previous approaches, such as PSO, SA and K-means for partitional clustering problem.