Advancing the Theory of Planned Behavior within Dietary and Physical Domains among Type 2 Diabetics: A Mixed Methods Approach

Many studies have applied the Theory of Planned Behavior (TPB) in predicting health behaviors among unique populations. However, a new paradigm is emerging where focus is now directed to modification and expansion of the TPB model rather than utilization of the traditional theory. This review proposes new models modified from the Theory of Planned Behavior and suggest an appropriate study design that can be used to test the models within physical activity and dietary practice domains among Type 2 diabetics in Kenya. The review was conducted by means of literature search in the field of nutrition behavior, health psychology and mixed methods using predetermined key words. The results identify pre-intention and post intention gaps within the TPB model that need to be filled. Additional psychosocial factors are proposed to be included in the TPB model to generate new models and the efficacy of these models tested using mixed methods design.

Using Combination of Optimized Recurrent Neural Network with Design of Experiments and Regression for Control Chart Forecasting

recurrent neural network (RNN) is an efficient tool for modeling production control process as well as modeling services. In this paper one RNN was combined with regression model and were employed in order to be checked whether the obtained data by the model in comparison with actual data, are valid for variable process control chart. Therefore, one maintenance process in workshop of Esfahan Oil Refining Co. (EORC) was taken for illustration of models. First, the regression was made for predicting the response time of process based upon determined factors, and then the error between actual and predicted response time as output and also the same factors as input were used in RNN. Finally, according to predicted data from combined model, it is scrutinized for test values in statistical process control whether forecasting efficiency is acceptable. Meanwhile, in training process of RNN, design of experiments was set so as to optimize the RNN.

Why Traditional Technology Acceptance Models Won't Work for Future Information Technologies?

This paper illustrates why existing technology acceptance models are only of limited use for predicting and explaining the adoption of future information and communication technologies. It starts with a general overview over technology adoption processes, and presents several theories for the acceptance as well as adoption of traditional information technologies. This is followed by an overview over the recent developments in the area of information and communication technologies. Based on the arguments elaborated in these sections, it is shown why the factors used to predict adoption in existing systems, will not be sufficient for explaining the adoption of future information and communication technologies.

Predicting Individual Investors- Intention to Invest: An Experimental Analysis of Attitude as a Mediator

The survival of publicly listed companies largely depends on their stocks being liquidly traded. This goal can be achieved when new investors are attracted to invest on companies- stocks. Among different groups of investors, individual investors are generally less able to objectively evaluate companies- risks and returns, and tend to be emotionally biased in their investing decisions. Therefore their decisions may be formed as a result of perceived risks and returns, and influenced by companies- images. This study finds that perceived risk, perceived returns and trust directly affect individual investors- trading decisions while attitude towards brand partially mediates the relationships. This finding suggests that, in courting individual investors, companies still need to perform financially while building a good image can result in their stocks being accepted quicker than the stocks of good performing companies with hidden images.

Comparison of Different Neural Network Approaches for the Prediction of Kidney Dysfunction

This paper presents the prediction of kidney dysfunction using different neural network (NN) approaches. Self organization Maps (SOM), Probabilistic Neural Network (PNN) and Multi Layer Perceptron Neural Network (MLPNN) trained with Back Propagation Algorithm (BPA) are used in this study. Six hundred and sixty three sets of analytical laboratory tests have been collected from one of the private clinical laboratories in Baghdad. For each subject, Serum urea and Serum creatinin levels have been analyzed and tested by using clinical laboratory measurements. The collected urea and cretinine levels are then used as inputs to the three NN models in which the training process is done by different neural approaches. SOM which is a class of unsupervised network whereas PNN and BPNN are considered as class of supervised networks. These networks are used as a classifier to predict whether kidney is normal or it will have a dysfunction. The accuracy of prediction, sensitivity and specificity were found for each type of the proposed networks .We conclude that PNN gives faster and more accurate prediction of kidney dysfunction and it works as promising tool for predicting of routine kidney dysfunction from the clinical laboratory data.

Predicting the Minimum Free Energy RNA Secondary Structures using Harmony Search Algorithm

The physical methods for RNA secondary structure prediction are time consuming and expensive, thus methods for computational prediction will be a proper alternative. Various algorithms have been used for RNA structure prediction including dynamic programming and metaheuristic algorithms. Musician's behaviorinspired harmony search is a recently developed metaheuristic algorithm which has been successful in a wide variety of complex optimization problems. This paper proposes a harmony search algorithm (HSRNAFold) to find RNA secondary structure with minimum free energy and similar to the native structure. HSRNAFold is compared with dynamic programming benchmark mfold and metaheuristic algorithms (RnaPredict, SetPSO and HelixPSO). The results showed that HSRNAFold is comparable to mfold and better than metaheuristics in finding the minimum free energies and the number of correct base pairs.

No-Reference Image Quality Assessment using Blur and Noise

Assessment for image quality traditionally needs its original image as a reference. The conventional method for assessment like Mean Square Error (MSE) or Peak Signal to Noise Ratio (PSNR) is invalid when there is no reference. In this paper, we present a new No-Reference (NR) assessment of image quality using blur and noise. The recent camera applications provide high quality images by help of digital Image Signal Processor (ISP). Since the images taken by the high performance of digital camera have few blocking and ringing artifacts, we only focus on the blur and noise for predicting the objective image quality. The experimental results show that the proposed assessment method gives high correlation with subjective Difference Mean Opinion Score (DMOS). Furthermore, the proposed method provides very low computational load in spatial domain and similar extraction of characteristics to human perceptional assessment.

A Study on the Relation of Corporate Governance and Pricing for Initial Public Offerings

The purpose of this study is to investigate the relationship between corporate governance and pricing for initial public offerings (IPOs). Empirical result finds that the prediction of pricing of IPOs with corporate governance added can have a rather higher degree of predicting accuracy than that of non governance added during the training and testing samples. Therefore, it can be observed that corporate governance mechanism can affect the pricing of IPOs

A Statistical Approach for Predicting and Optimizing Depth of Cut in AWJ Machining for 6063-T6 Al Alloy

In this paper, a set of experimental data has been used to assess the influence of abrasive water jet (AWJ) process parameters in cutting 6063-T6 aluminum alloy. The process variables considered here include nozzle diameter, jet traverse rate, jet pressure and abrasive flow rate. The effects of these input parameters are studied on depth of cut (h); one of most important characteristics of AWJ. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. In the next stage, the proposed model is embedded into a Simulated Annealing (SA) algorithm to optimize the AWJ process parameters. The objective is to determine a suitable set of process parameters that can produce a desired depth of cut, considering the ranges of the process parameters. Computational results prove the effectiveness of the proposed model and optimization procedure.

Knowledge Discovery Techniques for Talent Forecasting in Human Resource Application

Human Resource (HR) applications can be used to provide fair and consistent decisions, and to improve the effectiveness of decision making processes. Besides that, among the challenge for HR professionals is to manage organization talents, especially to ensure the right person for the right job at the right time. For that reason, in this article, we attempt to describe the potential to implement one of the talent management tasks i.e. identifying existing talent by predicting their performance as one of HR application for talent management. This study suggests the potential HR system architecture for talent forecasting by using past experience knowledge known as Knowledge Discovery in Database (KDD) or Data Mining. This article consists of three main parts; the first part deals with the overview of HR applications, the prediction techniques and application, the general view of Data mining and the basic concept of talent management in HRM. The second part is to understand the use of Data Mining technique in order to solve one of the talent management tasks, and the third part is to propose the potential HR system architecture for talent forecasting.

Energy Map Construction using Adaptive Alpha Grey Prediction Model in WSNs

Wireless Sensor Networks can be used to monitor the physical phenomenon in such areas where human approach is nearly impossible. Hence the limited power supply is the major constraint of the WSNs due to the use of non-rechargeable batteries in sensor nodes. A lot of researches are going on to reduce the energy consumption of sensor nodes. Energy map can be used with clustering, data dissemination and routing techniques to reduce the power consumption of WSNs. Energy map can also be used to know which part of the network is going to fail in near future. In this paper, Energy map is constructed using the prediction based approach. Adaptive alpha GM(1,1) model is used as the prediction model. GM(1,1) is being used worldwide in many applications for predicting future values of time series using some past values due to its high computational efficiency and accuracy.

Estimating Shortest Circuit Path Length Complexity

When binary decision diagrams are formed from uniformly distributed Monte Carlo data for a large number of variables, the complexity of the decision diagrams exhibits a predictable relationship to the number of variables and minterms. In the present work, a neural network model has been used to analyze the pattern of shortest path length for larger number of Monte Carlo data points. The neural model shows a strong descriptive power for the ISCAS benchmark data with an RMS error of 0.102 for the shortest path length complexity. Therefore, the model can be considered as a method of predicting path length complexities; this is expected to lead to minimum time complexity of very large-scale integrated circuitries and related computer-aided design tools that use binary decision diagrams.

Modelling Indoor Air Carbon Dioxide (CO2)Concentration using Neural Network

The use of neural networks is popular in various building applications such as prediction of heating load, ventilation rate and indoor temperature. Significant is, that only few papers deal with indoor carbon dioxide (CO2) prediction which is a very good indicator of indoor air quality (IAQ). In this study, a data-driven modelling method based on multilayer perceptron network for indoor air carbon dioxide in an apartment building is developed. Temperature and humidity measurements are used as input variables to the network. Motivation for this study derives from the following issues. First, measuring carbon dioxide is expensive and sensors power consumptions is high and secondly, this leads to short operating times of battery-powered sensors. The results show that predicting CO2 concentration based on relative humidity and temperature measurements, is difficult. Therefore, more additional information is needed.

A Hybrid Radial-Based Neuro-GA Multiobjective Design of Laminated Composite Plates under Moisture and Thermal Actions

In this paper, the optimum weight and cost of a laminated composite plate is seeked, while it undergoes the heaviest load prior to a complete failure. Various failure criteria are defined for such structures in the literature. In this work, the Tsai-Hill theory is used as the failure criterion. The theory of analysis was based on the Classical Lamination Theory (CLT). A newly type of Genetic Algorithm (GA) as an optimization technique with a direct use of real variables was employed. Yet, since the optimization via GAs is a long process, and the major time is consumed through the analysis, Radial Basis Function Neural Networks (RBFNN) was employed in predicting the output from the analysis. Thus, the process of optimization will be carried out through a hybrid neuro-GA environment, and the procedure will be carried out until a predicted optimum solution is achieved.

Integrating Decision Tree and Spatial Cluster Analysis for Landslide Susceptibility Zonation

Landslide susceptibility map delineates the potential zones for landslide occurrence. Previous works have applied multivariate methods and neural networks for mapping landslide susceptibility. This study proposed a new approach to integrate decision tree model and spatial cluster statistic for assessing landslide susceptibility spatially. A total of 2057 landslide cells were digitized for developing the landslide decision tree model. The relationships of landslides and instability factors were explicitly represented by using tree graphs in the model. The local Getis-Ord statistics were used to cluster cells with high landslide probability. The analytic result from the local Getis-Ord statistics was classed to create a map of landslide susceptibility zones. The map was validated using new landslide data with 482 cells. Results of validation show an accuracy rate of 86.1% in predicting new landslide occurrence. This indicates that the proposed approach is useful for improving landslide susceptibility mapping.

Optimal Design of Airfoil with High Aspect Ratio in Unmanned Aerial Vehicles

Shape optimization of the airfoil with high aspect ratio of long endurance unmanned aerial vehicle (UAV) is performed by the multi-objective optimization technology coupled with computational fluid dynamics (CFD). For predicting the aerodynamic characteristics around the airfoil the high-fidelity Navier-Stokes solver is employed and SMOGA (Simple Multi-Objective Genetic Algorithm), which is developed by authors, is used for solving the multi-objective optimization problem. To obtain the optimal solutions of the design variable (i.e., sectional airfoil profile, wing taper ratio and sweep) for high performance of UAVs, both the lift and lift-to-drag ratio are maximized whereas the pitching moment should be minimized, simultaneously. It is found that the lift force and lift-to-drag ratio are linearly dependent and a unique and dominant solution are existed. However, a trade-off phenomenon is observed between the lift-to-drag ratio and pitching moment. As the result of optimization, sixty-five (65) non-dominated Pareto individuals at the cutting edge of design spaces that is decided by airfoil shapes can be obtained.

Selecting Negative Examples for Protein-Protein Interaction

Proteomics is one of the largest areas of research for bioinformatics and medical science. An ambitious goal of proteomics is to elucidate the structure, interactions and functions of all proteins within cells and organisms. Predicting Protein-Protein Interaction (PPI) is one of the crucial and decisive problems in current research. Genomic data offer a great opportunity and at the same time a lot of challenges for the identification of these interactions. Many methods have already been proposed in this regard. In case of in-silico identification, most of the methods require both positive and negative examples of protein interaction and the perfection of these examples are very much crucial for the final prediction accuracy. Positive examples are relatively easy to obtain from well known databases. But the generation of negative examples is not a trivial task. Current PPI identification methods generate negative examples based on some assumptions, which are likely to affect their prediction accuracy. Hence, if more reliable negative examples are used, the PPI prediction methods may achieve even more accuracy. Focusing on this issue, a graph based negative example generation method is proposed, which is simple and more accurate than the existing approaches. An interaction graph of the protein sequences is created. The basic assumption is that the longer the shortest path between two protein-sequences in the interaction graph, the less is the possibility of their interaction. A well established PPI detection algorithm is employed with our negative examples and in most cases it increases the accuracy more than 10% in comparison with the negative pair selection method in that paper.

Effects of Input Speed on the Dynamic Response of Planar Multi-body Systems with Differently Located Frictionless Revolute Clearance Joints

This paper numerically investigates the effects of input speed on the overall dynamic characteristics of a multi-body system with differently located revolute clearance joints without friction. A typical planar slider-crank mechanism is used as a demonstration case in which the effects of the input speed on the dynamic performance of the mechanism with a revolute clearance joint between the crank and connecting rod, and between the connecting rod and slider are separately investigated with comprehensive observations numerically presented. It is observed that, changing the driving speed of a multibody system makes the behavior of the system to change from either periodic to chaotic, or chaotic to periodic depending on which joint has clearance. The location of the clearance revolute joint and the operating speed of a multi-body system play a crucial role in predicting accurately the dynamic responses of the system. Therefore the dynamic behavior of one clearance revolute joint cannot be used as a general case for a mechanical system.

Investigation of Dimethyl Ether Solubility in Liquid Hexadecane by UNIFAC Method

It is shown that a modified UNIFAC model can be applied to predict solubility of hydrocarbon gases and vapors in hydrocarbon solvents. Very good agreement with experimental data has been achieved. In this work we try to find best way for predicting dimethyl ether solubility in liquid paraffin by using group contribution theory.

Reliability of Chute-Feeders in Automatic Machines of High Production Capacity

Modern highly automated production systems faces problems of reliability. Machine function reliability results in changes of productivity rate and efficiency use of expensive industrial facilities. Predicting of reliability has become an important research and involves complex mathematical methods and calculation. The reliability of high productivity technological automatic machines that consists of complex mechanical, electrical and electronic components is important. The failure of these units results in major economic losses of production systems. The reliability of transport and feeding systems for automatic technological machines is also important, because failure of transport leads to stops of technological machines. This paper presents reliability engineering on the feeding system and its components for transporting a complex shape parts to automatic machines. It also discusses about the calculation of the reliability parameters of the feeding unit by applying the probability theory. Equations produced for calculating the limits of the geometrical sizes of feeders and the probability of sticking the transported parts into the chute represents the reliability of feeders as a function of its geometrical parameters.