Kohonen Self-Organizing Maps as a New Method for Determination of Salt Composition of Multi-Component Solutions

The paper presents the results of clusterization by Kohonen self-organizing maps (SOM) applied for analysis of array of Raman spectra of multi-component solutions of inorganic salts, for determination of types of salts present in the solution. It is demonstrated that use of SOM is a promising method for solution of clusterization and classification problems in spectroscopy of multicomponent objects, as attributing a pattern to some cluster may be used for recognition of component composition of the object.

Application of Adaptive Neural Network Algorithms for Determination of Salt Composition of Waters Using Laser Spectroscopy

In this study, a comparative analysis of the approaches associated with the use of neural network algorithms for effective solution of a complex inverse problem – the problem of identifying and determining the individual concentrations of inorganic salts in multicomponent aqueous solutions by the spectra of Raman scattering of light – is performed. It is shown that application of artificial neural networks provides the average accuracy of determination of concentration of each salt no worse than 0.025 M. The results of comparative analysis of input data compression methods are presented. It is demonstrated that use of uniform aggregation of input features allows decreasing the error of determination of individual concentrations of components by 16-18% on the average.

Using Artificial Neural Networks for Optical Imaging of Fluorescent Biomarkers

The article presents the results of the application of artificial neural networks to separate the fluorescent contribution of nanodiamonds used as biomarkers, adsorbents and carriers of drugs in biomedicine, from a fluorescent background of own biological fluorophores. The principal possibility of solving this problem is shown. Use of neural network architecture let to detect fluorescence of nanodiamonds against the background autofluorescence of egg white with high accuracy - better than 3 ug/ml.

Selecting the Best Sub-Region Indexing the Images in the Case of Weak Segmentation Based On Local Color Histograms

Color Histogram is considered as the oldest method used by CBIR systems for indexing images. In turn, the global histograms do not include the spatial information; this is why the other techniques coming later have attempted to encounter this limitation by involving the segmentation task as a preprocessing step. The weak segmentation is employed by the local histograms while other methods as CCV (Color Coherent Vector) are based on strong segmentation. The indexation based on local histograms consists of splitting the image into N overlapping blocks or sub-regions, and then the histogram of each block is computed. The dissimilarity between two images is reduced, as consequence, to compute the distance between the N local histograms of the both images resulting then in N*N values; generally, the lowest value is taken into account to rank images, that means that the lowest value is that which helps to designate which sub-region utilized to index images of the collection being asked. In this paper, we make under light the local histogram indexation method in the hope to compare the results obtained against those given by the global histogram. We address also another noteworthy issue when Relying on local histograms namely which value, among N*N values, to trust on when comparing images, in other words, which sub-region among the N*N sub-regions on which we base to index images. Based on the results achieved here, it seems that relying on the local histograms, which needs to pose an extra overhead on the system by involving another preprocessing step naming segmentation, does not necessary mean that it produces better results. In addition to that, we have proposed here some ideas to select the local histogram on which we rely on to encode the image rather than relying on the local histogram having lowest distance with the query histograms.

Frequent Itemset Mining Using Rough-Sets

Frequent pattern mining is the process of finding a pattern (a set of items, subsequences, substructures, etc.) that occurs frequently in a data set. It was proposed in the context of frequent itemsets and association rule mining. Frequent pattern mining is used to find inherent regularities in data. What products were often purchased together? Its applications include basket data analysis, cross-marketing, catalog design, sale campaign analysis, Web log (click stream) analysis, and DNA sequence analysis. However, one of the bottlenecks of frequent itemset mining is that as the data increase the amount of time and resources required to mining the data increases at an exponential rate. In this investigation a new algorithm is proposed which can be uses as a pre-processor for frequent itemset mining. FASTER (FeAture SelecTion using Entropy and Rough sets) is a hybrid pre-processor algorithm which utilizes entropy and roughsets to carry out record reduction and feature (attribute) selection respectively. FASTER for frequent itemset mining can produce a speed up of 3.1 times when compared to original algorithm while maintaining an accuracy of 71%.

A Contribution to the Polynomial Eigen Problem

The relationship between eigenstructure (eigenvalues and eigenvectors) and latent structure (latent roots and latent vectors) is established. In control theory eigenstructure is associated with the state space description of a dynamic multi-variable system and a latent structure is associated with its matrix fraction description. Beginning with block controller and block observer state space forms and moving on to any general state space form, we develop the identities that relate eigenvectors and latent vectors in either direction. Numerical examples illustrate this result. A brief discussion of the potential of these identities in linear control system design follows. Additionally, we present a consequent result: a quick and easy method to solve the polynomial eigenvalue problem for regular matrix polynomials.

Study of Temperature Difference and Current Distribution in Parallel-Connected Cells at Low Temperature

Two types of commercial cylindrical lithium ion batteries (Panasonic 3.4 Ah NCR-18650B and Samsung 2.9 Ah INR-18650), were investigated experimentally. The capacities of these samples were individually measured using constant current-constant voltage (CC-CV) method at different ambient temperatures (-10°C, 0°C, 25°C). Their internal resistance was determined by electrochemical impedance spectroscopy (EIS) and pulse discharge methods. The cells with different configurations of parallel connection NCR-NCR, INR-INR and NCR-INR were charged/discharged at the aforementioned ambient temperatures. The results showed that the difference of internal resistance between cells much more evident at low temperatures. Furthermore, the parallel connection of NCR-NCR exhibits the most uniform temperature distribution in cells at -10°C, this feature is quite favorable for the safety of the battery pack.

Temperature Distribution in Friction Stir Welding Using Finite Element Method

During welding, the amount of heat present in weld zones determines the quality of weldment produced. Thus, the heat distribution characteristics and its magnitude in weld zones with respect to process variables such as tool pin-shoulder rotational and traveling speed during welding is analyzed using thermal finite element analyses method. For this purpose, transient thermal finite element analyses are performed to model the temperatures distribution and its quantities in weld-zones with respect to process variables such as rotational speed and traveling speed during welding. Commercially available software Altair HyperWork is used to model three-dimensional tool pin-shoulder vs. workpieces and to simulate the friction stir process. The results show that increasing tool rotational speed, at a constant traveling speed, will increase the amount of heat generated in weld-zones. In contrary, increasing traveling speed, at constant tool pin-shoulder rotational speeds, will reduce the amount of heat generated in weld zones.

The Effect of Solution Density on the Synthesis of Magnesium Borate from Boron-Gypsum

Boron-gypsum is a waste which occurs in the boric acid production process. In this study, the boron content of this waste is evaluated for the use in synthesis of magnesium borates and such evaluation of this kind of waste is useful more than storage or disposal. Magnesium borates, which are a sub-class of boron minerals, are useful additive materials for the industries due to their remarkable thermal and mechanical properties. Magnesium borates were obtained hydrothermally at different temperatures. Novelty of this study is the search of the solution density effects to magnesium borate synthesis process for the increasing the possibility of borongypsum usage as a raw material. After the synthesis process, products are subjected to XRD and FT-IR to identify and characterize their crystal structure, respectively.

The Cadmium Adsorption Study by Using Seyitomer Fly Ash, Diatomite and Molasses in Wastewater

Fly ash is an important waste, produced in thermal power plants which causes very important environmental pollutions. For this reason the usage and evaluation the fly ash in various areas are very important. Nearly, 15 million tons/year of fly ash is produced in Turkey. In this study, usage of fly ash with diatomite and molasses for heavy metal (Cd) adsorption from wastewater is investigated. The samples of Seyitomer region fly ash were analyzed by X-ray fluorescence (XRF) and Scanning Electron Microscope (SEM) then diatomite (0 and 1% in terms of fly ash, w/w) and molasses (0-0.75 mL) were pelletized under 30 MPa of pressure for the usage of cadmium (Cd) adsorption in wastewater. After the adsorption process, samples of Seyitomer were analyzed using Optical Emission Spectroscopy (ICP-OES). As a result, it is seen that the usage of Seyitomer fly ash is proper for cadmium (Cd) adsorption and an optimum adsorption yield with 52% is found at a compound with Seyitomer fly ash (10 g), diatomite (0.5 g) and molasses (0.75 mL) at 2.5 h of reaction time, pH:4, 20ºC of reaction temperature and 300 rpm of stirring rate.

Survey to Assess the Feasibility of Executing the Web-Based Collaboration Process Using WBCS

The importance of the formal specification in the software life cycle is barely concealing to anyone. Formal specifications use mathematical notation to describe the properties of information system precisely, without unduly constraining the way in how these properties are achieved. Having a correct and quality software specification is not easy task. This study concerns with how a group of rectifiers can communicate with each other and work to prepare and produce a correct formal software specification. WBCS has been implemented based mainly in the proposed supported cooperative work model and a survey conducted on the existing Webbased collaborative writing tools. This paper aims to assess the feasibility of executing the web-based collaboration process using WBCS. The purpose of conducting this test is to test the system as a whole for functionality and fitness for use based on the evaluation test plan.

Power Quality Evaluation of Electrical Distribution Networks

Researches and concerns in power quality gained significant momentum in the field of power electronics systems over the last two decades globally. This sudden increase in the number of concerns over power quality problems is a result of the huge increase in the use of non-linear loads. In this paper, power quality evaluation of some distribution networks at Misurata - Libya has been done using a power quality and energy analyzer (Fluke 437 Series II). The results of this evaluation are used to minimize the problems of power quality. The analysis shows the main power quality problems that exist and the level of awareness of power quality issues with the aim of generating a start point which can be used as guidelines for researchers and end users in the field of power systems.

Review of Downscaling Methods in Climate Change and Their Role in Hydrological Studies

Recent perceived climate variability raises concerns with unprecedented hydrological phenomena and extremes. Distribution and circulation of the waters of the Earth become increasingly difficult to determine because of additional uncertainty related to anthropogenic emissions. The world wide observed changes in the large-scale hydrological cycle have been related to an increase in the observed temperature over several decades. Although the effect of change in climate on hydrology provides a general picture of possible hydrological global change, new tools and frameworks for modelling hydrological series with nonstationary characteristics at finer scales, are required for assessing climate change impacts. Of the downscaling techniques, dynamic downscaling is usually based on the use of Regional Climate Models (RCMs), which generate finer resolution output based on atmospheric physics over a region using General Circulation Model (GCM) fields as boundary conditions. However, RCMs are not expected to capture the observed spatial precipitation extremes at a fine cell scale or at a basin scale. Statistical downscaling derives a statistical or empirical relationship between the variables simulated by the GCMs, called predictors, and station-scale hydrologic variables, called predictands. The main focus of the paper is on the need for using statistical downscaling techniques for projection of local hydrometeorological variables under climate change scenarios. The projections can be then served as a means of input source to various hydrologic models to obtain streamflow, evapotranspiration, soil moisture and other hydrological variables of interest.

People Participation as Social Capital Form for Realizing Sustainable Ecotourism

As an entity of the tourism system, local communities were considered have better understanding of their region as well as influenced positively or negatively by the tourism activities in the region. This paper aimed to study role of community involvement in the development of ecotourism at Kintamani Bali from two perspectives of view, i.e. participation in the process of initiatives and participation in the utilizing the economic benefits of tourism. Thorough participation as an antecedent of social capital form, the sustainability of ecotourism at Kintamani could be expected.

Two Stage Fuzzy Methodology to Evaluate the Credit Risks of Investment Projects

The work proposes a decision support methodology for the credit risk minimization in selection of investment projects. The methodology provides two stages of projects’ evaluation. Preliminary selection of projects with minor credit risks is made using the Expertons Method. The second stage makes ranking of chosen projects using the Possibilistic Discrimination Analysis Method. The latter is a new modification of a well-known Method of Fuzzy Discrimination Analysis.

Generalized Vortex Lattice Method for Predicting Characteristics of Wings with Flap and Aileron Deflection

A generalized vortex lattice method for complex lifting surfaces with flap and aileron deflection is formulated. The method is not restricted by the linearized theory assumption and accounts for all standard geometric lifting surface parameters: camber, taper, sweep, washout, dihedral, in addition to flap and aileron deflection. Thickness is not accounted for since the physical lifting body is replaced by a lattice of panels located on the mean camber surface. This panel lattice setup and the treatment of different wake geometries is what distinguish the present work form the overwhelming majority of previous solutions based on the vortex lattice method. A MATLAB code implementing the proposed formulation is developed and validated by comparing our results to existing experimental and numerical ones and good agreement is demonstrated. It is then used to study the accuracy of the widely used classical vortex-lattice method. It is shown that the classical approach gives good agreement in the clean configuration but is off by as much as 30% when a flap or aileron deflection of 30° is imposed. This discrepancy is mainly due the linearized theory assumption associated with the conventional method. A comparison of the effect of four different wake geometries on the values of aerodynamic coefficients was also carried out and it is found that the choice of the wake shape had very little effect on the results.

Economic Analysis of Endogenous Growth Model with ICT Capital

This paper clarifies the role of ICT capital in economic growth. Albeit ICT remarkably contributes to economic growth, there are few studies on ICT capital in ICT sector from theoretical point of view. In this paper, production function of ICT which is used as input of intermediate good in final good and ICT sectors is incorporated into our model. In this setting, we analyze the role of ICT on balance growth path and show the possibility of general equilibrium solutions for this model. Through the simulation of the equilibrium solutions, we find that when ICT impacts on economy and economic growth increases, it is necessary that increases of efficiency at ICT sector and of accumulation of non-ICT and ICT capitals occur simultaneously.

Computational Methods in Official Statistics with an Example on Calculating and Predicting Diabetes Mellitus [DM] Prevalence in Different Age Groups within Australia in Future Years, in Light of the Aging Population

An analysis of the Australian Diabetes Screening Study estimated undiagnosed diabetes mellitus [DM] prevalence in a high risk general practice based cohort. DM prevalence varied from 9.4% to 18.1% depending upon the diagnostic criteria utilised with age being a highly significant risk factor. Utilising the gold standard oral glucose tolerance test, the prevalence of DM was 22-23% in those aged >= 70 years and

Cubic Trigonometric B-spline Approach to Numerical Solution of Wave Equation

The generalized wave equation models various problems in sciences and engineering. In this paper, a new three-time level implicit approach based on cubic trigonometric B-spline for the approximate solution of wave equation is developed. The usual finite difference approach is used to discretize the time derivative while cubic trigonometric B-spline is applied as an interpolating function in the space dimension. Von Neumann stability analysis is used to analyze the proposed method. Two problems are discussed to exhibit the feasibility and capability of the method. The absolute errors and maximum error are computed to assess the performance of the proposed method. The results were found to be in good agreement with known solutions and with existing schemes in literature.

A Web-Based Self-Learning Grammar for Spoken Language Understanding

One of the major goals of Spoken Dialog Systems (SDS) is to understand what the user utters. In the SDS domain, the Spoken Language Understanding (SLU) Module classifies user utterances by means of a pre-definite conceptual knowledge. The SLU module is able to recognize only the meaning previously included in its knowledge base. Due the vastity of that knowledge, the information storing is a very expensive process. Updating and managing the knowledge base are time-consuming and error-prone processes because of the rapidly growing number of entities like proper nouns and domain-specific nouns. This paper proposes a solution to the problem of Name Entity Recognition (NER) applied to a SDS domain. The proposed solution attempts to automatically recognize the meaning associated with an utterance by using the PANKOW (Pattern based Annotation through Knowledge On the Web) method at runtime. The method being proposed extracts information from the Web to increase the SLU knowledge module and reduces the development effort. In particular, the Google Search Engine is used to extract information from the Facebook social network.