Applying Gibbs Sampler for Multivariate Hierarchical Linear Model

Among various HLM techniques, the Multivariate Hierarchical Linear Model (MHLM) is desirable to use, particularly when multivariate criterion variables are collected and the covariance structure has information valuable for data analysis. In order to reflect prior information or to obtain stable results when the sample size and the number of groups are not sufficiently large, the Bayes method has often been employed in hierarchical data analysis. In these cases, although the Markov Chain Monte Carlo (MCMC) method is a rather powerful tool for parameter estimation, Procedures regarding MCMC have not been formulated for MHLM. For this reason, this research presents concrete procedures for parameter estimation through the use of the Gibbs samplers. Lastly, several future topics for the use of MCMC approach for HLM is discussed.

Grid Independence Study of Flow Past a Square Cylinder Using the Multi-Relaxation-Time Lattice Boltzmann Method

Numerical calculations of flow around a square cylinder are presented using the multi-relaxation-time lattice Boltzmann method at Reynolds number 150. The effects of upstream locations, downstream locations and blockage are investigated systematically. A detail analysis are given in terms of time-trace analysis of drag and lift coefficients, power spectra analysis of lift coefficient, vorticity contours visualizations and phase diagrams. A number of physical quantities mean drag coefficient, drag coefficient, Strouhal number and root-mean-square values of drag and lift coefficients are calculated and compared with the well resolved experimental data and numerical results available in open literature. The results had shown that the upstream, downstream and height of the computational domain are at least 7.5, 37.5 and 12 diameters of the cylinder, respectively.

Classification of the Bachet Elliptic Curves y2 = x3 + a3 in Fp, where p ≡ 1 (mod 6) is Prime

In this work, we first give in what fields Fp, the cubic root of unity lies in F*p, in Qp and in K*p where Qp and K*p denote the sets of quadratic and non-zero cubic residues modulo p. Then we use these to obtain some results on the classification of the Bachet elliptic curves y2 ≡ x3 +a3 modulo p, for p ≡ 1 (mod 6) is prime.

Mathematical Modeling for the Processes of Strain Hardening in Heterophase Materials with Nanoparticles

An investigation of the process of deformation hardening and evolution of deformation defect medium in dispersion-hardened materials with face centered cubic matrices and nanoparticles was done. Mathematical model including balance equation for the deformation defects was used.

On the Robust Stability of Homogeneous Perturbed Large-Scale Bilinear Systems with Time Delays and Constrained Inputs

The stability test problem for homogeneous large-scale perturbed bilinear time-delay systems subjected to constrained inputs is considered in this paper. Both nonlinear uncertainties and interval systems are discussed. By utilizing the Lyapunove equation approach associated with linear algebraic techniques, several delay-independent criteria are presented to guarantee the robust stability of the overall systems. The main feature of the presented results is that although the Lyapunov stability theorem is used, they do not involve any Lyapunov equation which may be unsolvable. Furthermore, it is seen the proposed schemes can be applied to solve the stability analysis problem of large-scale time-delay systems.

Comparison of Response Surface Designs in a Spherical Region

The objective of the research is to study and compare response surface designs: Central composite designs (CCD), Box- Behnken designs (BBD), Small composite designs (SCD), Hybrid designs, and Uniform shell designs (USD) over sets of reduced models when the design is in a spherical region for 3 and 4 design variables. The two optimality criteria ( D and G ) are considered which larger values imply a better design. The comparison of design optimality criteria of the response surface designs across the full second order model and sets of reduced models for 3 and 4 factors based on the two criteria are presented.

Preparation of Vanadium Powder by Hydrogenation and Dehydrogenation

Low oxygen content vanadium powder was prepared by hydrogenation dehydrogenization (HDH). The effect of purification treatment on hydrogen absorption kinetics of dendritic vanadium was tested, and the effects of milling technique on powder yield and grain size were studied. The crystal phase, oxygen and nitrgen content, and grain size of prepared powder were characterized and analyzed by X-ray diffraction (XRD), oxygen and nitrogen analyzer and grain size analyzer. The results show that the alkaline cleaning can improve the hydrogen absorption of vanadium. The yield of vanadium hydride powder can reach as high as 90% by 4h ball-milling, The resultant product also have an oxygen content less than 600μg/g, and the grain size is smaller than 37μm. Meanwhile, the XRD results show that the phase of hydride vanadium powder is mainly VH0.81. After a hydrogen desorption treatment in vacuum at 700Ôäâ, the phase of the powder converts into V and a little of V2H.

Estimating of the Renewal Function with Heavy-tailed Claims

We develop a new estimator of the renewal function for heavy-tailed claims amounts. Our approach is based on the peak over threshold method for estimating the tail of the distribution with a generalized Pareto distribution. The asymptotic normality of an appropriately centered and normalized estimator is established, and its performance illustrated in a simulation study.

Stepsize Control of the Finite Difference Method for Solving Ordinary Differential Equations

An important task in solving second order linear ordinary differential equations by the finite difference is to choose a suitable stepsize h. In this paper, by using the stochastic arithmetic, the CESTAC method and the CADNA library we present a procedure to estimate the optimal stepsize hopt, the stepsize which minimizes the global error consisting of truncation and round-off error.

A Markov Chain Model for Load-Balancing Based and Service Based RAT Selection Algorithms in Heterogeneous Networks

Next Generation Wireless Network (NGWN) is expected to be a heterogeneous network which integrates all different Radio Access Technologies (RATs) through a common platform. A major challenge is how to allocate users to the most suitable RAT for them. An optimized solution can lead to maximize the efficient use of radio resources, achieve better performance for service providers and provide Quality of Service (QoS) with low costs to users. Currently, Radio Resource Management (RRM) is implemented efficiently for the RAT that it was developed. However, it is not suitable for a heterogeneous network. Common RRM (CRRM) was proposed to manage radio resource utilization in the heterogeneous network. This paper presents a user level Markov model for a three co-located RAT networks. The load-balancing based and service based CRRM algorithms have been studied using the presented Markov model. A comparison for the performance of load-balancing based and service based CRRM algorithms is studied in terms of traffic distribution, new call blocking probability, vertical handover (VHO) call dropping probability and throughput.

Computational Algorithm for Obtaining Abelian Subalgebras in Lie Algebras

The set of all abelian subalgebras is computationally obtained for any given finite-dimensional Lie algebra, starting from the nonzero brackets in its law. More concretely, an algorithm is described and implemented to compute a basis for each nontrivial abelian subalgebra with the help of the symbolic computation package MAPLE. Finally, it is also shown a brief computational study for this implementation, considering both the computing time and the used memory.

Theory of Nanowire Radial p-n-Junction

We have developed an analytic model for the radial pn-junction in a nanowire (NW) core-shell structure utilizing as a new building block in different semiconductor devices. The potential distribution through the p-n-junction is calculated and the analytical expressions are derived to compute the depletion region widths. We show that the widths of space charge layers, surrounding the core, are the functions of core radius, which is the manifestation of so called classical size effect. The relationship between the depletion layer width and the built-in potential in the asymptotes of infinitely large core radius transforms to square-root dependence specific for conventional planar p-n-junctions. The explicit equation is derived to compute the capacitance of radial p-n-junction. The current-voltage behavior is also carefully determined taking into account the “short base" effects.

Unsteady Boundary Layer Flow over a Stretching Sheet in a Micropolar Fluid

Unsteady boundary layer flow of an incompressible micropolar fluid over a stretching sheet when the sheet is stretched in its own plane is studied in this paper. The stretching velocity is assumed to vary linearly with the distance along the sheet. Two equal and opposite forces are impulsively applied along the x-axis so that the sheet is stretched, keeping the origin fixed in a micropolar fluid. The transformed unsteady boundary layer equations are solved numerically using the Keller-box method for the whole transient from the initial state to final steady-state flow. Numerical results are obtained for the velocity and microrotation distributions as well as the skin friction coefficient for various values of the material parameter K. It is found that there is a smooth transition from the small-time solution to the large-time solution.

The Effect of Different Compression Schemes on Speech Signals

This paper studies the effect of different compression constraints and schemes presented in a new and flexible paradigm to achieve high compression ratios and acceptable signal to noise ratios of Arabic speech signals. Compression parameters are computed for variable frame sizes of a level 5 to 7 Discrete Wavelet Transform (DWT) representation of the signals for different analyzing mother wavelet functions. Results are obtained and compared for Global threshold and level dependent threshold techniques. The results obtained also include comparisons with Signal to Noise Ratios, Peak Signal to Noise Ratios and Normalized Root Mean Square Error.

Secure Resource Selection in Computational Grid Based on Quantitative Execution Trust

Grid computing provides a virtual framework for controlled sharing of resources across institutional boundaries. Recently, trust has been recognised as an important factor for selection of optimal resources in a grid. We introduce a new method that provides a quantitative trust value, based on the past interactions and present environment characteristics. This quantitative trust value is used to select a suitable resource for a job and eliminates run time failures arising from incompatible user-resource pairs. The proposed work will act as a tool to calculate the trust values of the various components of the grid and there by improves the success rate of the jobs submitted to the resource on the grid. The access to a resource not only depend on the identity and behaviour of the resource but also upon its context of transaction, time of transaction, connectivity bandwidth, availability of the resource and load on the resource. The quality of the recommender is also evaluated based on the accuracy of the feedback provided about a resource. The jobs are submitted for execution to the selected resource after finding the overall trust value of the resource. The overall trust value is computed with respect to the subjective and objective parameters.

On the Efficient Implementation of a Serial and Parallel Decomposition Algorithm for Fast Support Vector Machine Training Including a Multi-Parameter Kernel

This work deals with aspects of support vector machine learning for large-scale data mining tasks. Based on a decomposition algorithm for support vector machine training that can be run in serial as well as shared memory parallel mode we introduce a transformation of the training data that allows for the usage of an expensive generalized kernel without additional costs. We present experiments for the Gaussian kernel, but usage of other kernel functions is possible, too. In order to further speed up the decomposition algorithm we analyze the critical problem of working set selection for large training data sets. In addition, we analyze the influence of the working set sizes onto the scalability of the parallel decomposition scheme. Our tests and conclusions led to several modifications of the algorithm and the improvement of overall support vector machine learning performance. Our method allows for using extensive parameter search methods to optimize classification accuracy.

Exponentially Weighted Simultaneous Estimation of Several Quantiles

In this paper we propose new method for simultaneous generating multiple quantiles corresponding to given probability levels from data streams and massive data sets. This method provides a basis for development of single-pass low-storage quantile estimation algorithms, which differ in complexity, storage requirement and accuracy. We demonstrate that such algorithms may perform well even for heavy-tailed data.

Recursive Path-finding in a Dynamic Maze with Modified Tremaux's Algorithm

Number Link is a Japanese logic puzzle where pairs of same numbers are connected using lines. Number Link can be regarded as a dynamic multiple travelers, multiple entries and exits maze, where the walls and passages are dynamically changing as the travelers move. In this paper, we apply the Tremaux’s algorithm to solve Number Link puzzles of size 8x8, 10x10 and 15x20. The algorithm works well and produces a solution for puzzles of size 8x8 and 10x10. However, solving a puzzle of size 15x20 requires high computer processing power and is time consuming.

Increase of Organization in Complex Systems

Measures of complexity and entropy have not converged to a single quantitative description of levels of organization of complex systems. The need for such a measure is increasingly necessary in all disciplines studying complex systems. To address this problem, starting from the most fundamental principle in Physics, here a new measure for quantity of organization and rate of self-organization in complex systems based on the principle of least (stationary) action is applied to a model system - the central processing unit (CPU) of computers. The quantity of organization for several generations of CPUs shows a double exponential rate of change of organization with time. The exact functional dependence has a fine, S-shaped structure, revealing some of the mechanisms of self-organization. The principle of least action helps to explain the mechanism of increase of organization through quantity accumulation and constraint and curvature minimization with an attractor, the least average sum of actions of all elements and for all motions. This approach can help describe, quantify, measure, manage, design and predict future behavior of complex systems to achieve the highest rates of self organization to improve their quality. It can be applied to other complex systems from Physics, Chemistry, Biology, Ecology, Economics, Cities, network theory and others where complex systems are present.