Multi-Label Hierarchical Classification for Protein Function Prediction

Hierarchical classification is a problem with applications in many areas as protein function prediction where the dates are hierarchically structured. Therefore, it is necessary the development of algorithms able to induce hierarchical classification models. This paper presents experimenters using the algorithm for hierarchical classification called Multi-label Hierarchical Classification using a Competitive Neural Network (MHC-CNN). It was tested in ten datasets the Gene Ontology (GO) Cellular Component Domain. The results are compared with the Clus-HMC and Clus-HSC using the hF-Measure.

Dyadic Adjustment as a Mediator of the Relationship between Attachment, Attributional Style, and Violence in Male Batterers

This study examines the mediating effects of male dyadic adjustment on the relationships between attachment and attributional styles, and both psychological and physical husband violence. Based on data from 68 married violent men recruited through community organizations that work with violent men, regression analyses showed that husbands- dyadic adjustment mediates the associations between avoidant attachment and attributional style, and psychological aggression, but not physical violence. Scientific and clinical implications are discussed

Designing a Multilingual Auction Website for Selling Agricultural Products

The study aimed to identify the logical structure of data and particularities of developing and testing a website designed for selling farm products through online auctions. The research is based on a short literature review in the field and exploratory trials of some successful models from other industries, in order to identify the advantages of using such tool, as well as the optimal structure and functionality of an auction portal. In the last part, the study focuses on the results of testing the website by the potential beneficiaries. Conclusions of the study underlines that the particularities of some agricultural products could raise difficulties in the process of selling them through online auctions, but the use of such system it is perceived to bring significant improvements in the supply chain. The results of scientific investigations require a more detailed study regarding the importance of using quality standards for agricultural products sold via online auction, the impact that implementation of an online payment system could have on trade with agricultural products and problems which could arise in using the website in different countries.

Comparing Data Analysis, Communication and Information Technologies Expertise Levels in Undergraduate Psychology Students

Aims for this study: first, to compare the expertise level in data analysis, communication and information technologies in undergraduate psychology students. Second, to verify the factor structure of E-ETICA (Escala de Experticia en Tecnologias de la Informacion, la Comunicacion y el Análisis or Data Analysis, Communication and Information'Expertise Scale) which had shown an excellent internal consistency (α= 0.92) as well as a simple factor structure. Three factors, Complex, Basic Information and Communications Technologies and E-Searching and Download Abilities, explains 63% of variance. In the present study, 260 students (119 juniors and 141 seniors) were asked to respond to ETICA (16 items Likert scale of five points 1: null domain to 5: total domain). The results show that both junior and senior students report having very similar expertise level; however, E-ETICA presents a different factor structure for juniors and four factors explained also 63% of variance: Information E-Searching, Download and Process; Data analysis; Organization; and Communication technologies.

Evaluation of Chromium Contamination in the Sediments of Jen-Gen River Mouth, Taiwan

This study was conducted using the data collected at the mouth of Jen-Gen River to investigate and analyze chromium (Cr) contained in the sediments, and to evaluate the accumulation of Cr and the degree of its potential risk. The results show that samples collected at all monitoring stations near the mouth of Jen-Gen River contain 92–567 mg/kg of Cr with average of 366±166 mg/kg. The spatial distribution of Cr reveals that the Cr concentration is relatively high in the river mouth region, and gradually diminishes toward the harbor region. This indicates that upstream industrial and municipal wastewater discharges along the river bank are major sources of pollution. The accumulation factor and potential ecological risk index indicate that the sedimentation at Jen-Gen River mouth has the most serious degree of Cr accumulation and the highest ecological potential risk.

Theory of Planned Behaviour and Young Romanians’ Self-Reported Speed

Speeding represents one of the main concerns for road safety and it still is a subject for research. The need to address this problem and to understand why drivers over speed increases especially in Romania, where in 2011, speed was the main cause of car accidents. This article addresses this problem by using the theory of planned behaviour. A questionnaire was administered to a sample of young Romanian drivers (18 to 25 years) and several path analyses were made in order to verify if the model proposed by the theory of planned behaviour fits the data. One interesting result is that perceived behavioural control does not predict the intention to speed or self-reported driving speed, but subjective norms do. This implies that peers and social environment have a greater impact on young Romanian drivers than we thought.

Design of Multiple Clouds Based Global Performance Evaluation Service Broker System

According to dramatic growth of internet services, an easy and prompt service deployment has been important for internet service providers to successfully maintain time-to-market. Before global service deployment, they have to pay the big cost for service evaluation to make a decision of the proper system location, system scale, service delay and so on. But, intra-Lab evaluation tends to have big gaps in the measured data compared with the realistic situation, because it is very difficult to accurately expect the local service environment, network congestion, service delay, network bandwidth and other factors. Therefore, to resolve or ease the upper problems, we propose multiple cloud based GPES Broker system and use case that helps internet service providers to alleviate the above problems in beta release phase and to make a prompt decision for their service launching. By supporting more realistic and reliable evaluation information, the proposed GPES Broker system saves the service release cost and enables internet service provider to make a prompt decision about their service launching to various remote regions.

Experimental Determination of Reactions of Wind-Resistant Support of Circular Stacks in Various Configurations

Higher capacities of power plants together with increased awareness on environmental considerations have led to taller height of stacks. It is seen that strong wind can result in falling of stacks. So, aerodynamic consideration of stacks is very important in order to save the falling of stacks. One stack is not enough in industries and power sectors and two or three stacks are required for proper operation of the unit. It is very important to arrange the stacks in proper way to resist their downfall. The present experimental study concentrates on the mutual effect of three nearby stacks on each other at three different arrangements, viz. linear, side-by-side and triangular. The experiments find out the directions of resultant forces acting on the stacks in different configurations so that proper arrangement of supports can be made with respect to the wind directionality obtained from local meteorological data. One can also easily ascertain which stack is more vulnerable to wind in comparison to the others for a particular configuration. Thus, this study is important in studying the effect of wind force on three stacks in different arrangements and is very helpful in placing the supports in proper places in order to avoid failing of stack-like structures due to wind.

Critical Assessment of Scoring Schemes for Protein-Protein Docking Predictions

Protein-protein interactions (PPI) play a crucial role in many biological processes such as cell signalling, transcription, translation, replication, signal transduction, and drug targeting, etc. Structural information about protein-protein interaction is essential for understanding the molecular mechanisms of these processes. Structures of protein-protein complexes are still difficult to obtain by biophysical methods such as NMR and X-ray crystallography, and therefore protein-protein docking computation is considered an important approach for understanding protein-protein interactions. However, reliable prediction of the protein-protein complexes is still under way. In the past decades, several grid-based docking algorithms based on the Katchalski-Katzir scoring scheme were developed, e.g., FTDock, ZDOCK, HADDOCK, RosettaDock, HEX, etc. However, the success rate of protein-protein docking prediction is still far from ideal. In this work, we first propose a more practical measure for evaluating the success of protein-protein docking predictions,the rate of first success (RFS), which is similar to the concept of mean first passage time (MFPT). Accordingly, we have assessed the ZDOCK bound and unbound benchmarks 2.0 and 3.0. We also createda new benchmark set for protein-protein docking predictions, in which the complexes have experimentally determined binding affinity data. We performed free energy calculation based on the solution of non-linear Poisson-Boltzmann equation (nlPBE) to improve the binding mode prediction. We used the well-studied thebarnase-barstarsystem to validate the parameters for free energy calculations. Besides,thenlPBE-based free energy calculations were conducted for the badly predicted cases by ZDOCK and ZRANK. We found that direct molecular mechanics energetics cannot be used to discriminate the native binding pose from the decoys.Our results indicate that nlPBE-based calculations appeared to be one of the promising approaches for improving the success rate of binding pose predictions.

A New Approach for Classifying Large Number of Mixed Variables

The issue of classifying objects into one of predefined groups when the measured variables are mixed with different types of variables has been part of interest among statisticians in many years. Some methods for dealing with such situation have been introduced that include parametric, semi-parametric and nonparametric approaches. This paper attempts to discuss on a problem in classifying a data when the number of measured mixed variables is larger than the size of the sample. A propose idea that integrates a dimensionality reduction technique via principal component analysis and a discriminant function based on the location model is discussed. The study aims in offering practitioners another potential tool in a classification problem that is possible to be considered when the observed variables are mixed and too large.

Factors Influencing Knowledge Management Process Model: A Case Study of Manufacturing Industry in Thailand

The objectives of this research were to explore factors influencing knowledge management process in the manufacturing industry and develop a model to support knowledge management processes. The studied factors were technology infrastructure, human resource, knowledge sharing, and the culture of the organization. The knowledge management processes included discovery, capture, sharing, and application. Data were collected through questionnaires and analyzed using multiple linear regression and multiple correlation. The results found that technology infrastructure, human resource, knowledge sharing, and culture of the organization influenced the discovery and capture processes. However, knowledge sharing had no influence in sharing and application processes. A model to support knowledge management processes was developed, which indicated that sharing knowledge needed further improvement in the organization.

Numerical Analysis of Rapid Gas Decompression in Pure Nitrogen using 1D and 3D Transient Mathematical Models of Gas Flow in Pipes

The paper presents a numerical investigation on the rapid gas decompression in pure nitrogen which is made by using the one-dimensional (1D) and three-dimensional (3D) mathematical models of transient compressible non-isothermal fluid flow in pipes. A 1D transient mathematical model of compressible thermal multicomponent fluid mixture flow in pipes is presented. The set of the mass, momentum and enthalpy conservation equations for gas phase is solved in the model. Thermo-physical properties of multicomponent gas mixture are calculated by solving the Equation of State (EOS) model. The Soave-Redlich-Kwong (SRK-EOS) model is chosen. This model is successfully validated on the experimental data [1] and shows a good agreement with measurements. A 3D transient mathematical model of compressible thermal single-component gas flow in pipes, which is built by using the CFD Fluent code (ANSYS), is presented in the paper. The set of unsteady Reynolds-averaged conservation equations for gas phase is solved. Thermo-physical properties of single-component gas are calculated by solving the Real Gas Equation of State (EOS) model. The simplest case of gas decompression in pure nitrogen is simulated using both 1D and 3D models. The ability of both models to simulate the process of rapid decompression with a high order of agreement with each other is tested. Both, 1D and 3D numerical results show a good agreement between each other. The numerical investigation shows that 3D CFD model is very helpful in order to validate 1D simulation results if the experimental data is absent or limited.

Comparison between Associative Classification and Decision Tree for HCV Treatment Response Prediction

Combined therapy using Interferon and Ribavirin is the standard treatment in patients with chronic hepatitis C. However, the number of responders to this treatment is low, whereas its cost and side effects are high. Therefore, there is a clear need to predict patient’s response to the treatment based on clinical information to protect the patients from the bad drawbacks, Intolerable side effects and waste of money. Different machine learning techniques have been developed to fulfill this purpose. From these techniques are Associative Classification (AC) and Decision Tree (DT). The aim of this research is to compare the performance of these two techniques in the prediction of virological response to the standard treatment of HCV from clinical information. 200 patients treated with Interferon and Ribavirin; were analyzed using AC and DT. 150 cases had been used to train the classifiers and 50 cases had been used to test the classifiers. The experiment results showed that the two techniques had given acceptable results however the best accuracy for the AC reached 92% whereas for DT reached 80%.

A Rough Sets Approach for Relevant Internet/Web Online Searching

The internet is constantly expanding. Identifying web links of interest from web browsers requires users to visit each of the links listed, individually until a satisfactory link is found, therefore those users need to evaluate a considerable amount of links before finding their link of interest; this can be tedious and even unproductive. By incorporating web assistance, web users could be benefited from reduced time searching on relevant websites. In this paper, a rough set approach is presented, which facilitates classification of unlimited available e-vocabulary, to assist web users in reducing search times looking for relevant web sites. This approach includes two methods for identifying relevance data on web links based on the priority and percentage of relevance. As a result of these methods, a list of web sites is generated in priority sequence with an emphasis of the search criteria.

A Generator from Cascade Markov Model for Packet Loss and Subsequent Bit Error Description

In this paper we present a novel error model for packet loss and subsequent error description. The proposed model simulates the error performance of wireless communication link. The model is designed as two independent Markov chains, where the first one is used for packet generation and the second one generates correctly and incorrectly transmitted bits for received packets from the first chain. The statistical analyses of real communication on the wireless link are used for determination of model-s parameters. Using the obtained parameters and the implementation of the generator, we collected generated traffic. The obtained results generated by proposed model are compared with the real data collection.

Examining the Pearlite Growth Interface in a Fe-C-Mn Alloy

A method of collecting composition data and examining structural features of pearlite lamellae and the parent austenite at the growth interface in a 13wt. % manganese steel has been demonstrated with the use of Scanning Transmission Electron Microscopy (STEM). The combination of composition data and the structural features observed at the growth interface show that available theories of pearlite growth cannot explain all the observations.

Using Data Mining Techniques for Estimating Minimum, Maximum and Average Daily Temperature Values

Estimates of temperature values at a specific time of day, from daytime and daily profiles, are needed for a number of environmental, ecological, agricultural and technical applications, ranging from natural hazards assessments, crop growth forecasting to design of solar energy systems. The scope of this research is to investigate the efficiency of data mining techniques in estimating minimum, maximum and mean temperature values. For this reason, a number of experiments have been conducted with well-known regression algorithms using temperature data from the city of Patras in Greece. The performance of these algorithms has been evaluated using standard statistical indicators, such as Correlation Coefficient, Root Mean Squared Error, etc.

Server Virtualization Using User Behavior Model Focus on Provisioning Concept

Server provisioning is one of the most attractive topics in virtualization systems. Virtualization is a method of running multiple independent virtual operating systems on a single physical computer. It is a way of maximizing physical resources to maximize the investment in hardware. Additionally, it can help to consolidate servers, improve hardware utilization and reduce the consumption of power and physical space in the data center. However, management of heterogeneous workloads, especially for resource utilization of the server, or so called provisioning becomes a challenge. In this paper, a new concept for managing workloads based on user behavior is presented. The experimental results show that user behaviors are different in each type of service workload and time. Understanding user behaviors may improve the efficiency of management in provisioning concept. This preliminary study may be an approach to improve management of data centers running heterogeneous workloads for provisioning in virtualization system.

Effects of Network Dynamics on Routing Efficiency in P2P Networks

P2P Networks are highly dynamic structures since their nodes – peer users keep joining and leaving continuously. In the paper, we study the effects of network change rates on query routing efficiency. First we describe some background and an abstract system model. The chosen routing technique makes use of cached metadata from previous answer messages and also employs a mechanism for broken path detection and metadata maintenance. Several metrics are used to show that the protocol behaves quite well even with high rate of node departures, but above a certain threshold it literally breaks down and exhibits considerable efficiency degradation.

Improved Feature Processing for Iris Biometric Authentication System

Iris-based biometric authentication is gaining importance in recent times. Iris biometric processing however, is a complex process and computationally very expensive. In the overall processing of iris biometric in an iris-based biometric authentication system, feature processing is an important task. In feature processing, we extract iris features, which are ultimately used in matching. Since there is a large number of iris features and computational time increases as the number of features increases, it is therefore a challenge to develop an iris processing system with as few as possible number of features and at the same time without compromising the correctness. In this paper, we address this issue and present an approach to feature extraction and feature matching process. We apply Daubechies D4 wavelet with 4 levels to extract features from iris images. These features are encoded with 2 bits by quantizing into 4 quantization levels. With our proposed approach it is possible to represent an iris template with only 304 bits, whereas existing approaches require as many as 1024 bits. In addition, we assign different weights to different iris region to compare two iris templates which significantly increases the accuracy. Further, we match the iris template based on a weighted similarity measure. Experimental results on several iris databases substantiate the efficacy of our approach.