3D Rendering of American Sign Language Finger-Spelling: A Comparative Study of Two Animation Techniques

In this paper we report a study aimed at determining the most effective animation technique for representing ASL (American Sign Language) finger-spelling. Specifically, in the study we compare two commonly used 3D computer animation methods (keyframe animation and motion capture) in order to ascertain which technique produces the most 'accurate', 'readable', and 'close to actual signing' (i.e. realistic) rendering of ASL finger-spelling. To accomplish this goal we have developed 20 animated clips of fingerspelled words and we have designed an experiment consisting of a web survey with rating questions. 71 subjects ages 19-45 participated in the study. Results showed that recognition of the words was correlated with the method used to animate the signs. In particular, keyframe technique produced the most accurate representation of the signs (i.e., participants were more likely to identify the words correctly in keyframed sequences rather than in motion captured ones). Further, findings showed that the animation method had an effect on the reported scores for readability and closeness to actual signing; the estimated marginal mean readability and closeness was greater for keyframed signs than for motion captured signs. To our knowledge, this is the first study aimed at measuring and comparing accuracy, readability and realism of ASL animations produced with different techniques.

A Markov Chain Model for Load-Balancing Based and Service Based RAT Selection Algorithms in Heterogeneous Networks

Next Generation Wireless Network (NGWN) is expected to be a heterogeneous network which integrates all different Radio Access Technologies (RATs) through a common platform. A major challenge is how to allocate users to the most suitable RAT for them. An optimized solution can lead to maximize the efficient use of radio resources, achieve better performance for service providers and provide Quality of Service (QoS) with low costs to users. Currently, Radio Resource Management (RRM) is implemented efficiently for the RAT that it was developed. However, it is not suitable for a heterogeneous network. Common RRM (CRRM) was proposed to manage radio resource utilization in the heterogeneous network. This paper presents a user level Markov model for a three co-located RAT networks. The load-balancing based and service based CRRM algorithms have been studied using the presented Markov model. A comparison for the performance of load-balancing based and service based CRRM algorithms is studied in terms of traffic distribution, new call blocking probability, vertical handover (VHO) call dropping probability and throughput.

Mechanisms Involved In Organic Solvent Resistance in Gram-Negative Bacteria

The high world interest given to the researches concerning the study of moderately halophilic solvent-tolerant bacteria isolated from marine polluted environments is due to their high biotechnological potential, and also to the perspective of their application in different remediation technologies. Using enrichment procedures, I isolated two moderately halophilic Gram-negative bacterial strains from seawater sample, which are tolerant to organic solvents. Cell tolerance, adhesion and cells viability of Aeromonas salmonicida IBBCt2 and Pseudomonas aeruginosa IBBCt3 in the presence of organic solvents depends not only on its physicochemical properties and its concentration, but also on the specific response of the cells, and the cellular response is not the same for these bacterial strains. n-hexane, n-heptane, propylbenzene, with log POW between 3.69 and 4.39, were less toxic for Aeromonas salmonicida IBBCt2 and Pseudomonas aeruginosa IBBCt3, compared with toluene, styrene, xylene isomers and ethylbenzene, with log POW between 2.64 and 3.17. The results indicated that Aeromonas salmonicida IBBCt2 is more susceptible to organic solvents than Pseudomonas aeruginosa IBBCt3. The mechanisms underlying solvent tolerance (e.g., the existance of the efflux pumps) in Aeromonas salmonicida IBBCt2 and Pseudomonas aeruginosa IBBCt3 it was also studied.

Suitability of Requirements Abstraction Model (RAM) Requirements for High-Level System Testing

The Requirements Abstraction Model (RAM) helps in managing abstraction in requirements by organizing them at four levels (product, feature, function and component). The RAM is adaptable and can be tailored to meet the needs of the various organizations. Because software requirements are an important source of information for developing high-level tests, organizations willing to adopt the RAM model need to know the suitability of the RAM requirements for developing high-level tests. To investigate this suitability, test cases from twenty randomly selected requirements were developed, analyzed and graded. Requirements were selected from the requirements document of a Course Management System, a web based software system that supports teachers and students in performing course related tasks. This paper describes the results of the requirements document analysis. The results show that requirements at lower levels in the RAM are suitable for developing executable tests whereas it is hard to develop from requirements at higher levels.

Lattice Boltzmann Simulation of Binary Mixture Diffusion Using Modern Graphics Processors

A highly optimized implementation of binary mixture diffusion with no initial bulk velocity on graphics processors is presented. The lattice Boltzmann model is employed for simulating the binary diffusion of oxygen and nitrogen into each other with different initial concentration distributions. Simulations have been performed using the latest proposed lattice Boltzmann model that satisfies both the indifferentiability principle and the H-theorem for multi-component gas mixtures. Contemporary numerical optimization techniques such as memory alignment and increasing the multiprocessor occupancy are exploited along with some novel optimization strategies to enhance the computational performance on graphics processors using the C for CUDA programming language. Speedup of more than two orders of magnitude over single-core processors is achieved on a variety of Graphical Processing Unit (GPU) devices ranging from conventional graphics cards to advanced, high-end GPUs, while the numerical results are in excellent agreement with the available analytical and numerical data in the literature.

Technological Innovation Capabilities and Firm Performance

Technological innovation capability (TIC) is defined as a comprehensive set of characteristics of a firm that facilities and supports its technological innovation strategies. An audit to evaluate the TICs of a firm may trigger improvement in its future practices. Such an audit can be used by the firm for self assessment or third-party independent assessment to identify problems of its capability status. This paper attempts to develop such an auditing framework that can help to determine the subtle links between innovation capabilities and business performance; and to enable the auditor to determine whether good practice is in place. The seven TICs in this study include learning, R&D, resources allocation, manufacturing, marketing, organization and strategic planning capabilities. Empirical data was acquired through a survey study of 200 manufacturing firms in the Hong Kong/Pearl River Delta (HK/PRD) region. Structural equation modelling was employed to examine the relationships among TICs and various performance indicators: sales performance, innovation performance, product performance, and sales growth. The results revealed that different TICs have different impacts on different performance measures. Organization capability was found to have the most influential impact. Hong Kong manufacturers are now facing the challenge of high-mix-low-volume customer orders. In order to cope with this change, good capability in organizing different activities among various departments is critical to the success of a company.

Assessment of Reliability and Quality Measures in Power Systems

The paper presents new results of a recent industry supported research and development study in which an efficient framework for evaluating practical and meaningful power system reliability and quality indices was applied. The system-wide integrated performance indices are capable of addressing and revealing areas of deficiencies and bottlenecks as well as redundancies in the composite generation-transmission-demand structure of large-scale power grids. The technique utilizes a linear programming formulation, which simulates practical operating actions and offers a general and comprehensive framework to assess the harmony and compatibility of generation, transmission and demand in a power system. Practical applications to a reduced system model as well as a portion of the Saudi power grid are also presented in the paper for demonstration purposes.

Study of Forging Process in 7075 Aluminum Alloy Professional Bicycle Pedal using Taguchi Method

The current of professional bicycle pedal-s manufacturing model mostly used casting, forging, die-casting processing methods, so the paper used 7075 aluminum alloy which is to produce the bicycle parts most commonly, and employs the rigid-plastic finite element (FE) DEFORMTM 3D software to simulate and to analyze the professional bicycle pedal design. First we use Solid works 2010 3D graphics software to design the professional bicycle pedal of the mold and appearance, then import finite element (FE) DEFORMTM 3D software for analysis. The paper used rigid-plastic model analytical methods, and assuming mode to be rigid body. A series of simulation analyses in which the variables depend on different temperature of forging billet, friction factors, forging speed, mold temperature are reveal to effective stress, effective strain, damage and die radial load distribution for forging bicycle pedal. The analysis results hope to provide professional bicycle pedal forming mold references to identified whether suit with the finite element results for high-strength design suitability of aluminum alloy.

Capacitor Placement in Radial Distribution System for Loss Reduction Using Artificial Bee Colony Algorithm

This paper presents a new method which applies an artificial bee colony algorithm (ABC) for capacitor placement in distribution systems with an objective of improving the voltage profile and reduction of power loss. The ABC algorithm is a new population based meta heuristic approach inspired by intelligent foraging behavior of honeybee swarm. The advantage of ABC algorithm is that it does not require external parameters such as cross over rate and mutation rate as in case of genetic algorithm and differential evolution and it is hard to determine these parameters in prior. The other advantage is that the global search ability in the algorithm is implemented by introducing neighborhood source production mechanism which is a similar to mutation process. To demonstrate the validity of the proposed algorithm, computer simulations are carried out on 69-bus system and compared the results with the other approach available in the literature. The proposed method has outperformed the other methods in terms of the quality of solution and computational efficiency.

Multiple Job Shop-Scheduling using Hybrid Heuristic Algorithm

In this paper, multi-processors job shop scheduling problems are solved by a heuristic algorithm based on the hybrid of priority dispatching rules according to an ant colony optimization algorithm. The objective function is to minimize the makespan, i.e. total completion time, in which a simultanous presence of various kinds of ferons is allowed. By using the suitable hybrid of priority dispatching rules, the process of finding the best solution will be improved. Ant colony optimization algorithm, not only promote the ability of this proposed algorithm, but also decreases the total working time because of decreasing in setup times and modifying the working production line. Thus, the similar work has the same production lines. Other advantage of this algorithm is that the similar machines (not the same) can be considered. So, these machines are able to process a job with different processing and setup times. According to this capability and from this algorithm evaluation point of view, a number of test problems are solved and the associated results are analyzed. The results show a significant decrease in throughput time. It also shows that, this algorithm is able to recognize the bottleneck machine and to schedule jobs in an efficient way.

Combined Feature Based Hyperspectral Image Classification Technique Using Support Vector Machines

A spatial classification technique incorporating a State of Art Feature Extraction algorithm is proposed in this paper for classifying a heterogeneous classes present in hyper spectral images. The classification accuracy can be improved if and only if both the feature extraction and classifier selection are proper. As the classes in the hyper spectral images are assumed to have different textures, textural classification is entertained. Run Length feature extraction is entailed along with the Principal Components and Independent Components. A Hyperspectral Image of Indiana Site taken by AVIRIS is inducted for the experiment. Among the original 220 bands, a subset of 120 bands is selected. Gray Level Run Length Matrix (GLRLM) is calculated for the selected forty bands. From GLRLMs the Run Length features for individual pixels are calculated. The Principle Components are calculated for other forty bands. Independent Components are calculated for next forty bands. As Principal & Independent Components have the ability to represent the textural content of pixels, they are treated as features. The summation of Run Length features, Principal Components, and Independent Components forms the Combined Features which are used for classification. SVM with Binary Hierarchical Tree is used to classify the hyper spectral image. Results are validated with ground truth and accuracies are calculated.

Reliability of Chute-Feeders in Automatic Machines of High Production Capacity

Modern highly automated production systems faces problems of reliability. Machine function reliability results in changes of productivity rate and efficiency use of expensive industrial facilities. Predicting of reliability has become an important research and involves complex mathematical methods and calculation. The reliability of high productivity technological automatic machines that consists of complex mechanical, electrical and electronic components is important. The failure of these units results in major economic losses of production systems. The reliability of transport and feeding systems for automatic technological machines is also important, because failure of transport leads to stops of technological machines. This paper presents reliability engineering on the feeding system and its components for transporting a complex shape parts to automatic machines. It also discusses about the calculation of the reliability parameters of the feeding unit by applying the probability theory. Equations produced for calculating the limits of the geometrical sizes of feeders and the probability of sticking the transported parts into the chute represents the reliability of feeders as a function of its geometrical parameters.

The Impact of the Type of Diversification of Listed Construction Enterprises in China on Corporation Performance

The construction industry is the pillar industry in China, accounting for about 6% of the gross domestic product. Along with changes in the external environment of the construction industry in China, the construction firm faces fierce competition. The paper aims to investigate the relationship between diversified types of construction firm and its performance in China. Based on generalist and specialist strategy in organizational ecology, we think a generalist organization can be applied to an enterprise with diversified developments, while specialist groups are extended to professional enterprises .This study takes advantage of annual financial data of listed construction firm to empirically verify the relationship between diversification and corporation performance establishing a regression equation to econometric analysis. We find that: 1) Specialization can significantly improve the level of profitability of listed construction firms, and there is a significant positive relationship with corporate performance; 2) The level of operating performance of listed construction enterprises which engage in unrelated diversification is higher than those with related diversification; 3) The relationship between state-owned construction firms and corporate performance is negative. The more the year of foundation is, the higher performance will be; however, the more the year of being listed, the lower performance will be.

Risk Level Evaluation for Power System Facilities in Smart Grid

Reliability Centered Maintenance(RCM) is one of most widely used methods in the modern power system to schedule a maintenance cycle and determine the priority of inspection. In order to apply the RCM method to the Smart Grid, a precedence study for the new structure of rearranged system should be performed due to introduction of additional installation such as renewable and sustainable energy resources, energy storage devices and advanced metering infrastructure. This paper proposes a new method to evaluate the priority of maintenance and inspection of the power system facilities in the Smart Grid using the Risk Priority Number. In order to calculate that risk index, it is required that the reliability block diagram should be analyzed for the Smart Grid system. Finally, the feasible technical method is discussed to estimate the risk potential as part of the RCM procedure.

IKEv1 and IKEv2: A Quantitative Analyses

Key management is a vital component in any modern security protocol. Due to scalability and practical implementation considerations automatic key management seems a natural choice in significantly large virtual private networks (VPNs). In this context IETF Internet Key Exchange (IKE) is the most promising protocol under permanent review. We have made a humble effort to pinpoint IKEv2 net gain over IKEv1 due to recent modifications in its original structure, along with a brief overview of salient improvements between the two versions. We have used US National Institute of Technology NIIST VPN simulator to get some comparisons of important performance metrics.

Pleurotus sajor-caju (PSC) Improves Nutrient Contents and Maintains Sensory Properties of Carbohydrate-based Products

The grey oyster mushroom, Pleurotus sajor-caju (PSC), is a common edible mushroom and is now grown commercially around the world for food. This fungus has been broadly used as food or food ingredients in various food products for a long time. To enhance the nutritional quality and sensory attributes of bakery-based products, PSC powder is used in the present study to partially replace wheat flour in baked product formulations. The nutrient content and sensory properties of rice-porridge and unleavened bread (paratha) incorporated with various levels of PSC powder were studied. These food items were formulated with either 0%, 2%, 4% or 6% of PSC powder. Results show PSC powder recorded β-glucan at 3.57g/100g. In sensory evaluation, consumers gave higher score to both rice-porridge and paratha bread containing 2-4% PSC compared to those that are not added with PSC powder. The paratha containing 4% PSC powder can be formulated with the intention in improving overall acceptability of paratha bread. Meanwhile, for rice-porridge, consumers prefer the formulated product added with 4% PSC powder. In conclusion, the addition of PSC powder to partially wheat flour can be recommended for the purpose of enhancing nutritional composition and maintaining the acceptability of carbohydrate-based products.

Secure Resource Selection in Computational Grid Based on Quantitative Execution Trust

Grid computing provides a virtual framework for controlled sharing of resources across institutional boundaries. Recently, trust has been recognised as an important factor for selection of optimal resources in a grid. We introduce a new method that provides a quantitative trust value, based on the past interactions and present environment characteristics. This quantitative trust value is used to select a suitable resource for a job and eliminates run time failures arising from incompatible user-resource pairs. The proposed work will act as a tool to calculate the trust values of the various components of the grid and there by improves the success rate of the jobs submitted to the resource on the grid. The access to a resource not only depend on the identity and behaviour of the resource but also upon its context of transaction, time of transaction, connectivity bandwidth, availability of the resource and load on the resource. The quality of the recommender is also evaluated based on the accuracy of the feedback provided about a resource. The jobs are submitted for execution to the selected resource after finding the overall trust value of the resource. The overall trust value is computed with respect to the subjective and objective parameters.

RDFGraph: New Data Modeling Tool for Semantic Web

The emerging Semantic Web has been attracted many researchers and developers. New applications have been developed on top of Semantic Web and many supporting tools introduced to improve its software development process. Metadata modeling is one of development process where supporting tools exists. The existing tools are lack of readability and easiness for a domain knowledge expert to graphically models a problem in semantic model. In this paper, a metadata modeling tool called RDFGraph is proposed. This tool is meant to solve those problems. RDFGraph is also designed to work with modern database management systems that support RDF and to improve the performance of the query execution process. The testing result shows that the rules used in RDFGraph follows the W3C standard and the graphical model produced in this tool is properly translated and correct.

Behavior of Optical Fiber Aged in CTAC Solutions

The evolution of silica optical fiber strength aged in cetyltrimethylammonium chloride solution (CTAC) has been investigated. If the solution containing surfactants presents appreciable changes in physical and chemical properties at the critical micelle concentration (CMC), a non negligible mechanical behavior fiber change is observed for silica fiber aged in cationic surfactants as CTAC which can lead to optical fiber reliability questioning. The purpose of this work is to study the mechanical behavior of silica coated and naked optical fibers in contact with CTAC solution at different concentrations. Result analysis proves that the immersion in CTAC drastically decreases the fiber strength and specially near the CMC point. Beyond CMC point, a small increase of fiber strength is analyzed and commented.

Emerging Wireless Standards - WiFi, ZigBee and WiMAX

The world of wireless telecommunications is rapidly evolving. Technologies under research and development promise to deliver more services to more users in less time. This paper presents the emerging technologies helping wireless systems grow from where we are today into our visions of the future. This paper will cover the applications and characteristics of emerging wireless technologies: Wireless Local Area Networks (WiFi-802.11n), Wireless Personal Area Networks (ZigBee) and Wireless Metropolitan Area Networks (WiMAX). The purpose of this paper is to explain the impending 802.11n standard and how it will enable WLANs to support emerging media-rich applications. The paper will also detail how 802.11n compares with existing WLAN standards and offer strategies for users considering higher-bandwidth alternatives. The emerging IEEE 802.15.4 (ZigBee) standard aims to provide low data rate wireless communications with high-precision ranging and localization, by employing UWB technologies for a low-power and low cost solution. WiMAX (Worldwide Interoperability for Microwave Access) is a standard for wireless data transmission covering a range similar to cellular phone towers. With high performance in both distance and throughput, WiMAX technology could be a boon to current Internet providers seeking to become the leader of next generation wireless Internet access. This paper also explores how these emerging technologies differ from one another.