Factorial Structure and Psychometric Validation of Ecotourism Experiential Value Construct: Insights from Taman Negara National Park, Malaysia

The purpose of this research is to disentangle and validate the underlying factorial-structure of Ecotourism Experiential Value (EEV) measurement scale and subsequently investigate its psychometric properties. The analysis was based on a sample of 225 eco-tourists, collected at the vicinity of Taman Negara National Park (TNNP) via interviewer-administered questionnaire. Exploratory factor analysis (EFA) was performed to determine the factorial structure of EEV. Subsequently, to confirm and validate the factorial structure and assess the psychometric properties of EEV, confirmatory factor analysis (CFA) was executed. In addition, to establish the nomological validity of EEV a structural model was developed to examine the effect of EEV on Total Eco-tourist Experience Quality (TEEQ). It is unveiled that EEV is a secondorder six-factorial structure construct and it scale has adequately met the psychometric criteria, thus could permit interpretation of results confidently. The findings have important implications for future research directions and management of ecotourism destination.

Coordinated Q–V Controller for Multi-machine Steam Power Plant: Design and Validation

This paper discusses coordinated reactive power - voltage (Q-V) control in a multi machine steam power plant. The drawbacks of manual Q-V control are briefly listed, and the design requirements for coordinated Q-V controller are specified. Theoretical background and mathematical model of the new controller are presented next followed by validation of developed Matlab/Simulink model through comparison with recorded responses in real steam power plant and description of practical realisation of the controller. Finally, the performance of commissioned controller is illustrated on several examples of coordinated Q-V control in real steam power plant and compared with manual control.

The Autoregresive Analysis for Wind Turbine Signal Postprocessing

Today modern simulations solutions in the wind turbine industry have achieved a high degree of complexity and detail in result. Limitations exist when it is time to validate model results against measurements. Regarding Model validation it is of special interest to identify mode frequencies and to differentiate them from the different excitations. A wind turbine is a complex device and measurements regarding any part of the assembly show a lot of noise. Input excitations are difficult or even impossible to measure due to the stochastic nature of the environment. Traditional techniques for frequency analysis or features extraction are widely used to analyze wind turbine sensor signals, but have several limitations specially attending to non stationary signals (Events). A new technique based on autoregresive analysis techniques is introduced here for a specific application, a comparison and examples related to different events in the wind turbine operations are presented.

Numerical Evaluation of the Contribution of Inertial and Aerodynamic Forces on VAWT Blade Loading

A two-dimensional numerical simulation of the contribution of both inertial and aerodynamic forces on the blade loads of a Vertical-Axis Wind Turbine (VAWT) is presented. After describing the computational model and the relative validation procedure, a complete campaign of simulations - based on full RANS unsteady calculations - is proposed for a three-bladed rotor architecture characterized by a NACA 0021 airfoil. For each analyzed angular velocity, the combined effect of pressure and viscous forces acting on every rotor blade are compared to the corresponding centrifugal forces, due to the revolution of the turbine, thus achieving a preliminary estimation of the correlation between overall rotor efficiency and structural blade loads.

Asymptotic Analysis of Instant Messaging Service with Relay Nodes

In this paper, we provide complete end-to-end delay analyses including the relay nodes for instant messages. Message Session Relay Protocol (MSRP) is used to provide congestion control for large messages in the Instant Messaging (IM) service. Large messages are broken into several chunks. These chunks may traverse through a maximum number of two relay nodes before reaching destination according to the IETF specification of the MSRP relay extensions. We discuss the current solutions of sending large instant messages and introduce a proposal to reduce message flows in the IM service. We consider virtual traffic parameter i.e., the relay nodes are stateless non-blocking for scalability purpose. This type of relay node is also assumed to have input rate at constant bit rate. We provide a new scheduling policy that schedules chunks according to their previous node?s delivery time stamp tags. Validation and analysis is shown for such scheduling policy. The performance analysis with the model introduced in this paper is simple and straight forward, which lead to reduced message flows in the IM service.

Validation on 3D Surface Roughness Algorithm for Measuring Roughness of Psoriasis Lesion

Psoriasis is a widespread skin disease affecting up to 2% population with plaque psoriasis accounting to about 80%. It can be identified as a red lesion and for the higher severity the lesion is usually covered with rough scale. Psoriasis Area Severity Index (PASI) scoring is the gold standard method for measuring psoriasis severity. Scaliness is one of PASI parameter that needs to be quantified in PASI scoring. Surface roughness of lesion can be used as a scaliness feature, since existing scale on lesion surface makes the lesion rougher. The dermatologist usually assesses the severity through their tactile sense, therefore direct contact between doctor and patient is required. The problem is the doctor may not assess the lesion objectively. In this paper, a digital image analysis technique is developed to objectively determine the scaliness of the psoriasis lesion and provide the PASI scaliness score. Psoriasis lesion is modelled by a rough surface. The rough surface is created by superimposing a smooth average (curve) surface with a triangular waveform. For roughness determination, a polynomial surface fitting is used to estimate average surface followed by a subtraction between rough and average surface to give elevation surface (surface deviations). Roughness index is calculated by using average roughness equation to the height map matrix. The roughness algorithm has been tested to 444 lesion models. From roughness validation result, only 6 models can not be accepted (percentage error is greater than 10%). These errors occur due the scanned image quality. Roughness algorithm is validated for roughness measurement on abrasive papers at flat surface. The Pearson-s correlation coefficient of grade value (G) of abrasive paper and Ra is -0.9488, its shows there is a strong relation between G and Ra. The algorithm needs to be improved by surface filtering, especially to overcome a problem with noisy data.

An Adaptive Dynamic Fracture for 3D Fatigue Crack Growth Using X-FEM

In recent years, a new numerical method has been developed, the extended finite element method (X-FEM). The objective of this work is to exploit the (X-FEM) for the treatment of the fracture mechanics problems on 3D geometries, where we showed the ability of this method to simulate the fatigue crack growth into two cases: edge and central crack. In the results we compared the six first natural frequencies of mode shapes uncracking with the cracking initiation in the structure, and showed the stress intensity factor (SIF) evolution function as crack size propagation into structure, the analytical validation of (SIF) is presented. For to evidence the aspects of this method, all result is compared between FEA and X-FEM.

Development and Initial Validation of the Social Competency Inventory for Tertiary Level Faculty Members

This study aimed to develop and initially validate an instrument that measures social competency among tertiary level faculty members. A review of extant literature on social competence was done. The review of extant literature led to the writing of the items in the initial instrument which was evaluated by 11 Subject Matter Experts (SMEs). The SMEs were either educators or psychologists. The results of the evaluations done by the SMEs served as bases for the creation of the pre-try-out instrument used in the first trial-run. Insights from the first trial-run participants led to the development of the main try-out instrument used in the final test administration. One Hundred Forty-one participants from five private Higher Education Institutions (HEIs) in the National Capital Region (NCR) and five private HEIs in Central Luzon in the Philippines participated in the final test administration. The reliability of the instrument was evaluated using Cronbach-s Coefficient Alpha formula and had a Cronbach-s Alpha of 0.92. On the other hand, Factor Analysis was used to evaluate the validity of the instrument and six factors were identified. The development of the final instrument was based on the results of the evaluation of the instrument-s reliability and validity. For purposes of recognition, the instrument was named “Social Competency Inventory for Tertiary Level Faculty Members (SCI-TLFM)."

2D Validation of a High-order Adaptive Cartesian-grid finite-volume Characteristic- flux Model with Embedded Boundaries

A Finite Volume method based on Characteristic Fluxes for compressible fluids is developed. An explicit cell-centered resolution is adopted, where second and third order accuracy is provided by using two different MUSCL schemes with Minmod, Sweby or Superbee limiters for the hyperbolic part. Few different times integrator is used and be describe in this paper. Resolution is performed on a generic unstructured Cartesian grid, where solid boundaries are handled by a Cut-Cell method. Interfaces are explicitely advected in a non-diffusive way, ensuring local mass conservation. An improved cell cutting has been developed to handle boundaries of arbitrary geometrical complexity. Instead of using a polygon clipping algorithm, we use the Voxel traversal algorithm coupled with a local floodfill scanline to intersect 2D or 3D boundary surface meshes with the fixed Cartesian grid. Small cells stability problem near the boundaries is solved using a fully conservative merging method. Inflow and outflow conditions are also implemented in the model. The solver is validated on 2D academic test cases, such as the flow past a cylinder. The latter test cases are performed both in the frame of the body and in a fixed frame where the body is moving across the mesh. Adaptive Cartesian grid is provided by Paramesh without complex geometries for the moment.

An Evaluation of Algorithms for Single-Echo Biosonar Target Classification

A recent neurospiking coding scheme for feature extraction from biosonar echoes of various plants is examined with avariety of stochastic classifiers. Feature vectors derived are employedin well-known stochastic classifiers, including nearest-neighborhood,single Gaussian and a Gaussian mixture with EM optimization.Classifiers' performances are evaluated by using cross-validation and bootstrapping techniques. It is shown that the various classifers perform equivalently and that the modified preprocessing configuration yields considerably improved results.

Ensembling Adaptively Constructed Polynomial Regression Models

The approach of subset selection in polynomial regression model building assumes that the chosen fixed full set of predefined basis functions contains a subset that is sufficient to describe the target relation sufficiently well. However, in most cases the necessary set of basis functions is not known and needs to be guessed – a potentially non-trivial (and long) trial and error process. In our research we consider a potentially more efficient approach – Adaptive Basis Function Construction (ABFC). It lets the model building method itself construct the basis functions necessary for creating a model of arbitrary complexity with adequate predictive performance. However, there are two issues that to some extent plague the methods of both the subset selection and the ABFC, especially when working with relatively small data samples: the selection bias and the selection instability. We try to correct these issues by model post-evaluation using Cross-Validation and model ensembling. To evaluate the proposed method, we empirically compare it to ABFC methods without ensembling, to a widely used method of subset selection, as well as to some other well-known regression modeling methods, using publicly available data sets.

Proposal of a Means for Reducing the Torque Variation on a Vertical-Axis Water Turbine by Increasing the Blade Number

This paper presents a means for reducing the torque variation during the revolution of a vertical-axis water turbine (VAWaterT) by increasing the blade number. For this purpose, twodimensional CFD analyses have been performed on a straight-bladed Darrieus-type rotor. After describing the computational model and the relative validation procedure, a complete campaign of simulations, based on full RANS unsteady calculations, is proposed for a three, four and five-bladed rotor architectures, characterized by a NACA 0025 airfoil. For each proposed rotor configuration, flow field characteristics are investigated at several values of tip speed ratio, allowing a quantification of the influence of blade number on flow geometric features and dynamic quantities, such as rotor torque and power. Finally, torque and power curves are compared for the three analyzed architectures, achieving a quantification of the effect of blade number on overall rotor performance.

Conceptual Design of Aeroelastic Demonstrator for Whirl Flutter Simulation

This paper deals with the conceptual design of the new aeroelastic demonstrator for the whirl flutter simulation. The paper gives a theoretical background of the whirl flutter phenomenon and describes the events of the whirl flutter occurrence in the aerospace practice. The second part is focused on the experimental research of the whirl flutter on aeroelastic similar models. Finally the concept of the new aeroelastic demonstrator is described. The demonstrator represents the wing and engine of the twin turboprop commuter aircraft including a driven propeller. It allows the changes of the main structural parameters influencing the whirl flutter stability characteristics. It is intended for the experimental investigation of the whirl flutter in the wind tunnel. The results will be utilized for validation of analytical methods and software tools.

Temporally Coherent 3D Animation Reconstruction from RGB-D Video Data

We present a new method to reconstruct a temporally coherent 3D animation from single or multi-view RGB-D video data using unbiased feature point sampling. Given RGB-D video data, in form of a 3D point cloud sequence, our method first extracts feature points using both color and depth information. In the subsequent steps, these feature points are used to match two 3D point clouds in consecutive frames independent of their resolution. Our new motion vectors based dynamic alignement method then fully reconstruct a spatio-temporally coherent 3D animation. We perform extensive quantitative validation using novel error functions to analyze the results. We show that despite the limiting factors of temporal and spatial noise associated to RGB-D data, it is possible to extract temporal coherence to faithfully reconstruct a temporally coherent 3D animation from RGB-D video data.

Numerical Simulation of Investment Casting of Gold Jewelry: Experiments and Validations

This paper proposes the numerical simulation of the investment casting of gold jewelry. It aims to study the behavior of fluid flow during mould filling and solidification and to optimize the process parameters, which lead to predict and control casting defects such as gas porosity and shrinkage porosity. A finite difference method, computer simulation software FLOW-3D was used to simulate the jewelry casting process. The simplified model was designed for both numerical simulation and real casting production. A set of sensor acquisitions were allocated on the different positions of the wax tree of the model to detect filling times, while a set of thermocouples were allocated to detect the temperature during casting and cooling. Those detected data were applied to validate the results of the numerical simulation to the results of the real casting. The resulting comparisons signify that the numerical simulation can be used as an effective tool in investment-casting-process optimization and casting-defect prediction.

Post-Cracking Behaviour of High Strength Fiber Concrete Prediction and Validation

Fracture process in mechanically loaded steel fiber reinforced high-strength (SFRHSC) concrete is characterized by fibers bridging the crack providing resistance to its opening. Structural SFRHSC fracture model was created; material fracture process was modeled, based on single fiber pull-out laws, which were determined experimentally (for straight fibers, fibers with end hooks (Dramix), and corrugated fibers (Tabix)) as well as obtained numerically ( using FEM simulations). For this purpose experimental program was realized and pull-out force versus pull-out fiber length was obtained (for fibers embedded into concrete at different depth and under different angle). Model predictions were validated by 15x15x60cm prisms 4 point bending tests. Fracture surfaces analysis was realized for broken prisms with the goal to improve elaborated model assumptions. Optimal SFRHSC structures were recognized.

A Multiclass BCMP Queueing Modeling and Simulation-Based Road Traffic Flow Analysis

Urban road network traffic has become one of the most studied research topics in the last decades. This is mainly due to the enlargement of the cities and the growing number of motor vehicles traveling in this road network. One of the most sensitive problems is to verify if the network is congestion-free. Another related problem is the automatic reconfiguration of the network without building new roads to alleviate congestions. These problems require an accurate model of the traffic to determine the steady state of the system. An alternative is to simulate the traffic to see if there are congestions and when and where they occur. One key issue is to find an adequate model for road intersections. Once the model established, either a large scale model is built or the intersection is represented by its performance measures and simulation for analysis. In both cases, it is important to seek the queueing model to represent the road intersection. In this paper, we propose to model the road intersection as a BCMP queueing network and we compare this analytical model against a simulation model for validation.

Computer Modeling of Drug Distribution after Intravitreal Administration

Intravitreal injection (IVI) is the most common treatment for eye posterior segment diseases such as endopthalmitis, retinitis, age-related macular degeneration, diabetic retinopathy, uveitis, and retinal detachment. Most of the drugs used to treat vitreoretinal diseases, have a narrow concentration range in which they are effective, and may be toxic at higher concentrations. Therefore, it is critical to know the drug distribution within the eye following intravitreal injection. Having knowledge of drug distribution, ophthalmologists can decide on drug injection frequency while minimizing damage to tissues. The goal of this study was to develop a computer model to predict intraocular concentrations and pharmacokinetics of intravitreally injected drugs. A finite volume model was created to predict distribution of two drugs with different physiochemical properties in the rabbit eye. The model parameters were obtained from literature review. To validate this numeric model, the in vivo data of spatial concentration profile from the lens to the retina were compared with the numeric data. The difference was less than 5% between the numerical and experimental data. This validation provides strong support for the numerical methodology and associated assumptions of the current study.

Real-time Haptic Modeling and Simulation for Prosthetic Insertion

In this work a surgical simulator is produced which enables a training otologist to conduct a virtual, real-time prosthetic insertion. The simulator provides the Ear, Nose and Throat surgeon with real-time visual and haptic responses during virtual cochlear implantation into a 3D model of the human Scala Tympani (ST). The parametric model is derived from measured data as published in the literature and accounts for human morphological variance, such as differences in cochlear shape, enabling patient-specific pre- operative assessment. Haptic modeling techniques use real physical data and insertion force measurements, to develop a force model which mimics the physical behavior of an implant as it collides with the ST walls during an insertion. Output force profiles are acquired from the insertion studies conducted in the work, to validate the haptic model. The simulator provides the user with real-time, quantitative insertion force information and associated electrode position as user inserts the virtual implant into the ST model. The information provided by this study may also be of use to implant manufacturers for design enhancements as well as for training specialists in optimal force administration, using the simulator. The paper reports on the methods for anatomical modeling and haptic algorithm development, with focus on simulator design, development, optimization and validation. The techniques may be transferrable to other medical applications that involve prosthetic device insertions where user vision is obstructed.

Rheological Modeling for Production of High Quality Polymeric

The fundamental defect inherent to the thermoforming technology is wall-thickness variation of the products due to inadequate thermal processing during production of polymer. A nonlinear viscoelastic rheological model is implemented for developing the process model. This model describes deformation process of a sheet in thermoforming process. Because of relaxation pause after plug-assist stage and also implementation of two stage thermoforming process have minor wall-thickness variation and consequently better mechanical properties of polymeric articles. For model validation, a comparative analysis of the theoretical and experimental data is presented.