Detection and Quantification of Ozone in Screen Printing Facilities

Most often the contaminants are not taken seriously into consideration, and this behavior comes out directly from the lack of monitoring and professional reporting about pollution in the printing facilities in Serbia. The goal of planned and systematic ozone measurements in ambient air of the screen printing facilities in Novi Sad is to examine of its impact on the employees health, and to track trends in concentration. In this study, ozone concentrations were determined by using discontinuous and continuous method during the automatic and manual screen printing process. Obtained results indicates that the average concentrations of ozone measured during the automatic process were almost 3 to 28 times higher for discontinuous and 10 times higher for continuous method (1.028 ppm) compared to the values prescribed by OSHA. In the manual process, average concentrations of ozone were within prescribed values for discontinuous and almost 3 times higher for continuous method (0.299 ppm).

Analysis of Linear Equalizers for Cooperative Multi-User MIMO Based Reporting System

In this paper, we consider a multi user multiple input multiple output (MU-MIMO) based cooperative reporting system for cognitive radio network. In the reporting network, the secondary users forward the primary user data to the common fusion center (FC). The FC is equipped with linear equalizers and an energy detector to make the decision about the spectrum. The primary user data are considered to be a digital video broadcasting - terrestrial (DVB-T) signal. The sensing channel and the reporting channel are assumed to be an additive white Gaussian noise and an independent identically distributed Raleigh fading respectively. We analyzed the detection probability of MU-MIMO system with linear equalizers and arrived at the closed form expression for average detection probability. Also the system performance is investigated under various MIMO scenarios through Monte Carlo simulations.

Extraction of Semantic Digital Signatures from MRI Photos for Image-Identification Purposes

This paper makes an attempt to solve the problem of searching and retrieving of similar MRI photos via Internet services using morphological features which are sourced via the original image. This study is aiming to be considered as an additional tool of searching and retrieve methods. Until now the main way of the searching mechanism is based on the syntactic way using keywords. The technique it proposes aims to serve the new requirements of libraries. One of these is the development of computational tools for the control and preservation of the intellectual property of digital objects, and especially of digital images. For this purpose, this paper proposes the use of a serial number extracted by using a previously tested semantic properties method. This method, with its center being the multi-layers of a set of arithmetic points, assures the following two properties: the uniqueness of the final extracted number and the semantic dependence of this number on the image used as the method-s input. The major advantage of this method is that it can control the authentication of a published image or its partial modification to a reliable degree. Also, it acquires the better of the known Hash functions that the digital signature schemes use and produces alphanumeric strings for cases of authentication checking, and the degree of similarity between an unknown image and an original image.

Speckle Reducing Contourlet Transform for Medical Ultrasound Images

Speckle noise affects all coherent imaging systems including medical ultrasound. In medical images, noise suppression is a particularly delicate and difficult task. A tradeoff between noise reduction and the preservation of actual image features has to be made in a way that enhances the diagnostically relevant image content. Even though wavelets have been extensively used for denoising speckle images, we have found that denoising using contourlets gives much better performance in terms of SNR, PSNR, MSE, variance and correlation coefficient. The objective of the paper is to determine the number of levels of Laplacian pyramidal decomposition, the number of directional decompositions to perform on each pyramidal level and thresholding schemes which yields optimal despeckling of medical ultrasound images, in particular. The proposed method consists of the log transformed original ultrasound image being subjected to contourlet transform, to obtain contourlet coefficients. The transformed image is denoised by applying thresholding techniques on individual band pass sub bands using a Bayes shrinkage rule. We quantify the achieved performance improvement.

Analysis and Design of a Novel Active Soft Switched Phase-Shifted Full Bridge Converter

This paper proposes an active soft-switching circuit for bridge converters aiming to improve the power conversion efficiency. The proposed circuit achieves loss-less switching for both main and auxiliary switches without increasing the main switch current/voltage rating. A winding coupled to the primary of power transformer ensures ZCS for the auxiliary switches during their turn-off. A 350 W, 100 kHz phase shifted full bridge (PSFB) converter is built to validate the analysis and design. Theoretical loss calculations for proposed circuit is presented. The proposed circuit is compared with passive soft switched PSFB in terms of efficiency and loss in duty cycle.

Characterization of Corn Cobs from Microwave and Potassium Hydroxide Pretreatment

The complexity of lignocellulosic biomass requires a pretreatment step to improve the yield of fermentable sugars. The efficient pretreatment of corn cobs using microwave and potassium hydroxide and enzymatic hydrolysis was investigated. The objective of this work was to characterize the optimal condition of pretreatment of corn cobs using microwave and potassium hydroxide enhance enzymatic hydrolysis. Corn cobs were submerged in different potassium hydroxide concentration at varies temperature and resident time. The pretreated corn cobs were hydrolyzed to produce the reducing sugar for analysis. The morphology and microstructure of samples were investigated by Thermal gravimetric analysis (TGA, scanning electron microscope (SEM), X-ray diffraction (XRD). The results showed that lignin and hemicellulose were removed by microwave/potassium hydroxide pretreatment. The crystallinity of the pretreated corn cobs was higher than the untreated. This method was compared with autoclave and conventional heating method. The results indicated that microwave-alkali treatment was an efficient way to improve the enzymatic hydrolysis rate by increasing its accessibility hydrolysis enzymes.

Hierarchies Based On the Number of Cooperating Systems of Finite Automata on Four-Dimensional Input Tapes

In theoretical computer science, the Turing machine has played a number of important roles in understanding and exploiting basic concepts and mechanisms in computing and information processing [20]. It is a simple mathematical model of computers [9]. After that, M.Blum and C.Hewitt first proposed two-dimensional automata as a computational model of two-dimensional pattern processing, and investigated their pattern recognition abilities in 1967 [7]. Since then, a lot of researchers in this field have been investigating many properties about automata on a two- or three-dimensional tape. On the other hand, the question of whether processing fourdimensional digital patterns is much more difficult than two- or threedimensional ones is of great interest from the theoretical and practical standpoints. Thus, the study of four-dimensional automata as a computasional model of four-dimensional pattern processing has been meaningful [8]-[19],[21]. This paper introduces a cooperating system of four-dimensional finite automata as one model of four-dimensional automata. A cooperating system of four-dimensional finite automata consists of a finite number of four-dimensional finite automata and a four-dimensional input tape where these finite automata work independently (in parallel). Those finite automata whose input heads scan the same cell of the input tape can communicate with each other, that is, every finite automaton is allowed to know the internal states of other finite automata on the same cell it is scanning at the moment. In this paper, we mainly investigate some accepting powers of a cooperating system of eight- or seven-way four-dimensional finite automata. The seven-way four-dimensional finite automaton is an eight-way four-dimensional finite automaton whose input head can move east, west, south, north, up, down, or in the fu-ture, but not in the past on a four-dimensional input tape.

Improved IDR(s) Method for Gaining Very Accurate Solutions

The IDR(s) method based on an extended IDR theorem was proposed by Sonneveld and van Gijzen. The original IDR(s) method has excellent property compared with the conventional iterative methods in terms of efficiency and small amount of memory. IDR(s) method, however, has unexpected property that relative residual 2-norm stagnates at the level of less than 10-12. In this paper, an effective strategy for stagnation detection, stagnation avoidance using adaptively information of parameter s and improvement of convergence rate itself of IDR(s) method are proposed in order to gain high accuracy of the approximated solution of IDR(s) method. Through numerical experiments, effectiveness of adaptive tuning IDR(s) method is verified and demonstrated.

Bottom Up Text Mining through Hierarchical Document Representation

Most of the existing text mining approaches are proposed, keeping in mind, transaction databases model. Thus, the mined dataset is structured using just one concept: the “transaction", whereas the whole dataset is modeled using the “set" abstract type. In such cases, the structure of the whole dataset and the relationships among the transactions themselves are not modeled and consequently, not considered in the mining process. We believe that taking into account structure properties of hierarchically structured information (e.g. textual document, etc ...) in the mining process, can leads to best results. For this purpose, an hierarchical associations rule mining approach for textual documents is proposed in this paper and the classical set-oriented mining approach is reconsidered profits to a Direct Acyclic Graph (DAG) oriented approach. Natural languages processing techniques are used in order to obtain the DAG structure. Based on this graph model, an hierarchical bottom up algorithm is proposed. The main idea is that each node is mined with its parent node.

Effect of Superplasticizer and NaOH Molarity on Workability, Compressive Strength and Microstructure Properties of Self-Compacting Geopolymer Concrete

The research investigates the effects of super plasticizer and molarity of sodium hydroxide alkaline solution on the workability, microstructure and compressive strength of self compacting geopolymer concrete (SCGC). SCGC is an improved way of concreting execution that does not require compaction and is made by complete elimination of ordinary Portland cement content. The parameters studied were superplasticizer (SP) dosage and molarity of NaOH solution. SCGC were synthesized from low calcium fly ash, activated by combinations of sodium hydroxide and sodium silicate solutions, and by incorporation of superplasticizer for self compactability. The workability properties such as filling ability, passing ability and resistance to segregation were assessed using slump flow, T-50, V-funnel, L-Box and J-ring test methods. It was found that the essential workability requirements for self compactability according to EFNARC were satisfied. Results showed that the workability and compressive strength improved with the increase in superplasticizer dosage. An increase in strength and a decrease in workability of these concrete samples were observed with the increase in molarity of NaOH solution from 8M to 14M. Improvement of interfacial transition zone (ITZ) and micro structure with the increase of SP and increase of concentration from 8M to 12M were also identified.

The Effects of Sodium Chloride in the Formation of Size and Shape of Gold (Au)Nanoparticles by Microwave-Polyol Method for Mercury Adsorption

Mercury is a natural occurring element and present in various concentrations in the environment. Due to its toxic effects, it is desirable to research mercury sensitive materials to adsorb mercury. This paper describes the preparation of Au nanoparticles for mercury adsorption by using a microwave (MW)-polyol method in the presence of three different Sodium Chloride (NaCl) concentrations (10, 20 and 30 mM). Mixtures of spherical, triangular, octahedral, decahedral particles and 1-D product were obtained using this rapid method. Sizes and shapes was found strongly depend on the concentrations of NaCl. Without NaCl concentration, spherical, triangular plates, octahedral, decahedral nanoparticles and 1D product were produced. At the lower NaCl concentration (10 mM), spherical, octahedral and decahedral nanoparticles were present, while spherical and decahedral nanoparticles were preferentially form by using 20 mM of NaCl concentration. Spherical, triangular plates, octahedral and decahedral nanoparticles were obtained at the highest NaCl concentration (30 mM). The amount of mercury adsorbed using 20 ppm mercury solution is the highest (67.5 %) for NaCl concentration of 30 mM. The high yield of polygonal particles will increase the mercury adsorption. In addition, the adsorption of mercury is also due to the sizes of the particles. The sizes of particles become smaller with increasing NaCl concentrations (size ranges, 5- 16 nm) than those synthesized without addition of NaCl (size ranges 11-32 nm). It is concluded that NaCl concentrations affects the formation of sizes and shapes of Au nanoparticles thus affects the mercury adsorption.

Minimizing of Target Localization Error using Multi-robot System and Particle Filters

In recent years a number of applications with multirobot systems (MRS) is growing in various areas. But their design is in practice often difficult and algorithms are proposed for the theoretical background and do not consider errors and noise in real conditions, so they are not usable in real environment. These errors are visible also in task of target localization enough, when robots try to find and estimate the position of the target by the sensors. Localization of target is possible also with one robot but as it was examined target finding and localization with group of mobile robots can estimate the target position more accurately and faster. The accuracy of target position estimation is made by cooperation of MRS and particle filtering. Advantage of usage the MRS with particle filtering was tested on task of fixed target localization by group of mobile robots.

Optimal Policy for a Deteriorating Inventory Model with Finite Replenishment Rate and with Price Dependant Demand Rate and Cycle Length Dependant Price

In this paper, an inventory model with finite and constant replenishment rate, price dependant demand rate, time value of money and inflation, finite time horizon, lead time and exponential deterioration rate and with the objective of maximizing the present worth of the total system profit is developed. Using a dynamic programming based solution algorithm, the optimal sequence of the cycles can be found and also different optimal selling prices, optimal order quantities and optimal maximum inventories can be obtained for the cycles with unequal lengths, which have never been done before for this model. Also, a numerical example is used to show accuracy of the solution procedure.

Deterministic Random Number Generators for Online Applications

Cryptography, Image watermarking and E-banking are filled with apparent oxymora and paradoxes. Random sequences are used as keys to encrypt information to be used as watermark during embedding the watermark and also to extract the watermark during detection. Also, the keys are very much utilized for 24x7x365 banking operations. Therefore a deterministic random sequence is very much useful for online applications. In order to obtain the same random sequence, we need to supply the same seed to the generator. Many researchers have used Deterministic Random Number Generators (DRNGs) for cryptographic applications and Pseudo Noise Random sequences (PNs) for watermarking. Even though, there are some weaknesses in PN due to attacks, the research community used it mostly in digital watermarking. On the other hand, DRNGs have not been widely used in online watermarking due to its computational complexity and non-robustness. Therefore, we have invented a new design of generating DRNG using Pi-series to make it useful for online Cryptographic, Digital watermarking and Banking applications.

The Solar Wall in the Italian Climates

Passive systems were born with the purpose of the greatest exploitation of solar energy in cold climates and high altitudes. They spread themselves until the 80-s all over the world without any attention to the specific climate and the summer behavior; this caused the deactivation of the systems due to a series of problems connected to the summer overheating, the complex management and the rising of the dust. Until today the European regulation limits only the winter consumptions without any attention to the summer behavior but, the recent European EN 15251 underlines the relevance of the indoor comfort, and the necessity of the analytic studies validation by monitoring case studies. In the porpose paper we demonstrate that the solar wall is an efficient system both from thermal comfort and energy saving point of view and it is the most suitable for our temperate climates because it can be used as a passive cooling sistem too. In particular the paper present an experimental and numerical analisys carried out on a case study with nine different solar passive systems in Ancona, Italy. We carried out a detailed study of the lodging provided by the solar wall by the monitoring and the evaluation of the indoor conditions. Analyzing the monitored data, on the base of recognized models of comfort (ISO, ASHRAE, Givoni-s BBCC), is emerged that the solar wall has an optimal behavior in the middle seasons. In winter phase this passive system gives more advantages in terms of energy consumptions than the other systems, because it gives greater heat gain and therefore smaller consumptions. In summer, when outside air temperature return in the mean seasonal value, the indoor comfort is optimal thanks to an efficient transversal ventilation activated from the same wall.

Scrum as the Method Supporting the Implementation of Knowledge Management in an Organization

Many companies have switched their processes to project-oriented in the last years. This brings new possibilities and effectiveness not only in the field of external processes connected with the product delivery but also the internal processes as well. However centralized project organization which is based on the role of project manager in the team has proved insufficient in some cases. Agile methods of project organization are trying to solve this problem by bringing new view on the project organization, roles, processes and competences. Scrum is one of these methods which builds on the principles of knowledge management to drive the project to effectiveness from all view angles. Using this method to organize internal and delivery projects helps the organization to create and share knowledge throughout the company. It also supports forming unique competences of individuals and project teams and drives innovations in the company.

Mining Frequent Patterns with Functional Programming

Frequent patterns are patterns such as sets of features or items that appear in data frequently. Finding such frequent patterns has become an important data mining task because it reveals associations, correlations, and many other interesting relationships hidden in a dataset. Most of the proposed frequent pattern mining algorithms have been implemented with imperative programming languages such as C, Cµ, Java. The imperative paradigm is significantly inefficient when itemset is large and the frequent pattern is long. We suggest a high-level declarative style of programming using a functional language. Our supposition is that the problem of frequent pattern discovery can be efficiently and concisely implemented via a functional paradigm since pattern matching is a fundamental feature supported by most functional languages. Our frequent pattern mining implementation using the Haskell language confirms our hypothesis about conciseness of the program. The performance studies on speed and memory usage support our intuition on efficiency of functional language.

Preliminary Analysis of Energy Efficiency in Data Center: Case Study

As the data-driven economy is growing faster than ever and the demand for energy is being spurred, we are facing unprecedented challenges of improving energy efficiency in data centers. Effectively maximizing energy efficiency or minimising the cooling energy demand is becoming pervasive for data centers. This paper investigates overall energy consumption and the energy efficiency of cooling system for a data center in Finland as a case study. The power, cooling and energy consumption characteristics and operation condition of facilities are examined and analysed. Potential energy and cooling saving opportunities are identified and further suggestions for improving the performance of cooling system are put forward. Results are presented as a comprehensive evaluation of both the energy performance and good practices of energy efficient cooling operations for the data center. Utilization of an energy recovery concept for cooling system is proposed. The conclusion we can draw is that even though the analysed data center demonstrated relatively high energy efficiency, based on its power usage effectiveness value, there is still a significant potential for energy saving from its cooling systems.

Enhancing Operational Effectiveness in the Norwegian Army through Simulation-Based Training

The Norwegian Military Academy (Army) has initiated a project with the main ambition to explore possible avenues to enhancing operational effectiveness through an increased use of simulation-based training and exercises. Within a cost/benefit framework, we discuss opportunities and limitations of vertical and horizontal integration of the existing tactical training system. Vertical integration implies expanding the existing training system to span the full range of training from tactical level (platoon, company) to command and staff level (battalion, brigade). Horizontal integration means including other domains than army tactics and staff procedures in the training, such as military ethics, foreign languages, leadership and decision making. We discuss each of the integration options with respect to purpose and content of training, "best practice" for organising and conducting simulation-based training, and suggest how to evaluate training procedures and measure learning outcomes. We conclude by giving guidelines towards further explorative work and possible implementation.

Primer Design with Specific PCR Product using Particle Swarm Optimization

Before performing polymerase chain reactions (PCR), a feasible primer set is required. Many primer design methods have been proposed for design a feasible primer set. However, the majority of these methods require a relatively long time to obtain an optimal solution since large quantities of template DNA need to be analyzed. Furthermore, the designed primer sets usually do not provide a specific PCR product. In recent years, evolutionary computation has been applied to PCR primer design and yielded promising results. In this paper, a particle swarm optimization (PSO) algorithm is proposed to solve primer design problems associated with providing a specific product for PCR experiments. A test set of the gene CYP1A1, associated with a heightened lung cancer risk was analyzed and the comparison of accuracy and running time with the genetic algorithm (GA) and memetic algorithm (MA) was performed. A comparison of results indicated that the proposed PSO method for primer design finds optimal or near-optimal primer sets and effective PCR products in a relatively short time.