Clustering of Variables Based On a Probabilistic Approach Defined on the Hypersphere

We consider n individuals described by p standardized variables, represented by points of the surface of the unit hypersphere Sn-1. For a previous choice of n individuals we suppose that the set of observables variables comes from a mixture of bipolar Watson distribution defined on the hypersphere. EM and Dynamic Clusters algorithms are used for identification of such mixture. We obtain estimates of parameters for each Watson component and then a partition of the set of variables into homogeneous groups of variables. Additionally we will present a factor analysis model where unobservable factors are just the maximum likelihood estimators of Watson directional parameters, exactly the first principal component of data matrix associated to each group previously identified. Such alternative model it will yield us to directly interpretable solutions (simple structure), avoiding factors rotations.

Anthropometric Correlates of Balance Performance in Non-Institutionalized Elderly

Purpose: The fear of falling is a major concern among the elderly. Sixty-five percent of individuals older than 60 years of age experience loss of balance often on a daily basis. Therefore, balance assessment in the elderly deserves special attention due to its importance in functional mobility and safety. This study aimed at assessing balance performance and comparing some anthropometric parameters among a Nigerian non-institutionalized elderly population. Methods: Sixty one elderly subjects (31 males and 30 females) participated in this study. Their ages ranged between 62 and 84 years. Ability to maintain balance was assessed using Functional Reach Test (FRT) and Sharpened Romberg Test (SRT). Anthropometric data including age, weight, height, arm length, leg length, bi-acromial breadth, foot length and trunk length were also collected. Analysis was done using Pearson’s Product Moment Correlation Coefficient and Independent T-test, while level of significance was set as p

Didactic Material Resources in the Teaching of National History and Geography: Selected Results of a Qualitative Survey

The paper is the first output of a larger research project conducted at the Faculty of Education of the University of Hradec Králové, which deals with an improved understanding of teachers' work in the subject of National History and Geography. Partial findings focusing on the use of didactic material resources in teaching are presented in this phase. With the regard to promotion of independent activity of students within learner based education, material equipment of schools with didactic aids is becoming increasingly important. This paper is based on qualitative research, where the possibilities and mainly the reasons for use of material didactic resources in teaching were investigated through semi-structured interviews. Attention was focused on ways of working with different teaching aids and their implementation in the educational process. It turns out that teachers accept current constructivist and humanistic approaches to education associated with the requirement to prepare students for life in an information society, and accordingly they adjust their teaching.

Quality Management in Public e-Administration

Since the late 1970s, quality management has become an important tool for achieving a high quality of public e-administration services in many countries. Very important part of quality management in e-administration is measurement of quality indicators related to this sector. Therefore, this paper gives a description of e-administration, including statistics about it and other examples from many countries worldwide, as well as the explanation of quality management in public e-administration. The paper also gives a list and description of quality indicators relevant to e-administration, as part of quality management within the e-administration. Through a literature review and best practices, the paper aims to analyze quality indicators measurement and other parts of good quality management when it comes to the public e-administration and consequently to show the usefulness of quality management in public e-administration in order to provide services of high quality.

Impact of Liquidity Crunch on Interbank Network

Most empirical studies have analyzed how liquidity risks faced by individual institutions turn into systemic risk. Recent banking crisis has highlighted the importance of grasping and controlling the systemic risk, and the acceptance by Central Banks to ease their monetary policies for saving default or illiquid banks. This last point shows that banks would pay less attention to liquidity risk which, in turn, can become a new important channel of loss. The financial regulation focuses on the most important and “systemic” banks in the global network. However, to quantify the expected loss associated with liquidity risk, it is worth to analyze sensitivity to this channel for the various elements of the global bank network. A small bank is not considered as potentially systemic; however the interaction of small banks all together can become a systemic element. This paper analyzes the impact of medium and small banks interaction on a set of banks which is considered as the core of the network. The proposed method uses the structure of agent-based model in a two-class environment. In first class, the data from actual balance sheets of 22 large and systemic banks (such as BNP Paribas or Barclays) are collected. In second one, to model a network as closely as possible to actual interbank market, 578 fictitious banks smaller than the ones belonging to first class have been split into two groups of small and medium ones. All banks are active on the European interbank network and have deposit and market activity. A simulation of 12 three month periods representing a midterm time interval three years is projected. In each period, there is a set of behavioral descriptions: repayment of matured loans, liquidation of deposits, income from securities, collection of new deposits, new demands of credit, and securities sale. The last two actions are part of refunding process developed in this paper. To strengthen reliability of proposed model, random parameters dynamics are managed with stochastic equations as rates the variations of which are generated by Vasicek model. The Central Bank is considered as the lender of last resort which allows banks to borrow at REPO rate and some ejection conditions of banks from the system are introduced. Liquidity crunch due to exogenous crisis is simulated in the first class and the loss impact on other bank classes is analyzed though aggregate values representing the aggregate of loans and/or the aggregate of borrowing between classes. It is mainly shown that the three groups of European interbank network do not have the same response, and that intermediate banks are the most sensitive to liquidity risk.

Isolation and Classification of Red Blood Cells in Anemic Microscopic Images

Red blood cells (RBCs) are among the most commonly and intensively studied type of blood cells in cell biology. Anemia is a lack of RBCs is characterized by its level compared to the normal hemoglobin level. In this study, a system based image processing methodology was developed to localize and extract RBCs from microscopic images. Also, the machine learning approach is adopted to classify the localized anemic RBCs images. Several textural and geometrical features are calculated for each extracted RBCs. The training set of features was analyzed using principal component analysis (PCA). With the proposed method, RBCs were isolated in 4.3secondsfrom an image containing 18 to 27 cells. The reasons behind using PCA are its low computation complexity and suitability to find the most discriminating features which can lead to accurate classification decisions. Our classifier algorithm yielded accuracy rates of 100%, 99.99%, and 96.50% for K-nearest neighbor (K-NN) algorithm, support vector machine (SVM), and neural network RBFNN, respectively. Classification was evaluated in highly sensitivity, specificity, and kappa statistical parameters. In conclusion, the classification results were obtained within short time period, and the results became better when PCA was used.

Critical Analysis of the Hong Kong International Convention on Ship Recycling

In May 2009, the International Maritime Organization (IMO) adopted the Hong Kong International Convention for the Safe and Environmentally Sound Recycling of Ships to address the growing concerns about the environmental, occupational health and safety risks related to ship recycling. The aim of the Hong Kong Convention is to provide a legally binding instrument which ensures that the process of ship recycling does not pose risks to human health, safety and to the environment. In this paper, critical analysis of the Hong Kong Convention has been carried out in order to study the effectiveness of the Convention to meet its objectives. The Convention has been studied in detail including its background, main features, major stakeholders, strengths and weaknesses. The Convention, though having several deficiencies, is a major breakthrough in not only recognizing but also dealing with the ill-practices associated with ship recycling.

Probabilistic Bhattacharya Based Active Contour Model in Structure Tensor Space

Object identification and segmentation application requires extraction of object in foreground from the background. In this paper the Bhattacharya distance based probabilistic approach is utilized with an active contour model (ACM) to segment an object from the background. In the proposed approach, the Bhattacharya histogram is calculated on non-linear structure tensor space. Based on the histogram, new formulation of active contour model is proposed to segment images. The results are tested on both color and gray images from the Berkeley image database. The experimental results show that the proposed model is applicable to both color and gray images as well as both texture images and natural images. Again in comparing to the Bhattacharya based ACM in ICA space, the proposed model is able to segment multiple object too.

Beneficiation of Pyrolitic Carbon Black

This research investigated treatment of crude carbon black produced from pyrolysis of waste tyres in order to evaluate its quality and possible industrial applications. A representative sample of crude carbon black was dry screened to determine the initial particle size distribution. This was followed by pulverizing the crude carbon black and leaching in hot concentrated sulphuric acid for the removal of heavy metals and other contaminants. Analysis of the refined carbon black showed a significant improvement of the product quality compared to crude carbon black. It was discovered that refined carbon black can be further classified into multiple high value products for various industrial applications such as filler, paint pigment, activated carbon and fuel briquettes.

A Study of Priority Evaluation and Resource Allocation for Revitalization of Cultural Heritages in the Urban Development

Proper maintenance and preservation of significant cultural heritages or historic buildings is necessary. It can not only enhance environmental benefits and a sense of community, but also preserve a city's history and people’s memory. It allows the next generation to be able to get a glimpse of our past, and achieve the goal of sustainable preserved cultural assets. However, the management of maintenance work has not been appropriate for many designated heritages or historic buildings so far. The planning and implementation of the reuse has yet to have a breakthrough specification. It leads the heritages to a mere formality of being “reserved”, instead of the real meaning of “conservation”. For the restoration and preservation of cultural heritages study issues, it is very important due to the consideration of historical significance, symbolism, and economic benefits effects. However, the decision makers such as the officials from public sector they often encounter which heritage should be prioritized to be restored first under the available limited budgets. Only very few techniques are available today to determine the appropriately restoration priorities for the diverse historical heritages, perhaps because of a lack of systematized decision-making aids been proposed before. In the past, the discussions of management and maintenance towards cultural assets were limited to the selection of reuse alternatives instead of the allocation of resources. In view of this, this research will adopt some integrated research methods to solve the existing problems that decision-makers might encounter when allocating resources in the management and maintenance of heritages and historic buildings. The purpose of this study is to develop a sustainable decision making model for local governments to resolve these problems. We propose an alternative decision support model to prioritize restoration needs within the limited budgets. The model is constructed based on fuzzy Delphi, fuzzy analysis network process (FANP) and goal programming (GP) methods. In order to avoid misallocate resources; this research proposes a precise procedure that can take multi-stakeholders views, limited costs and resources into consideration. Also, the combination of many factors and goals has been taken into account to find the highest priority and feasible solution results. To illustrate the approach we propose in this research, seven cultural heritages in Taipei city as one example has been used as an empirical study, and the results are in depth analyzed to explain the application of our proposed approach.

On the Computation of a Common n-finger Robotic Grasp for a Set of Objects

Industrial robotic arms utilize multiple end-effectors, each for a specific part and for a specific task. We propose a novel algorithm which will define a single end-effector’s configuration able to grasp a given set of objects with different geometries. The algorithm will have great benefit in production lines allowing a single robot to grasp various parts. Hence, reducing the number of endeffectors needed. Moreover, the algorithm will reduce end-effector design and manufacturing time and final product cost. The algorithm searches for a common grasp over the set of objects. The search algorithm maps all possible grasps for each object which satisfy a quality criterion and takes into account possible external wrenches (forces and torques) applied to the object. The mapped grasps are- represented by high-dimensional feature vectors which describes the shape of the gripper. We generate a database of all possible grasps for each object in the feature space. Then we use a search and classification algorithm for intersecting all possible grasps over all parts and finding a single common grasp suitable for all objects. We present simulations of planar and spatial objects to validate the feasibility of the approach.

Improved Robust Stability Criteria of a Class of Neutral Lur’e Systems with Interval Time-Varying Delays

This paper addresses the robust stability problem of a class of delayed neutral Lur’e systems. Combined with the property of convex function and double integral Jensen inequality, a new tripe integral Lyapunov functional is constructed to derive some new stability criteria. Compared with some related results, the new criteria established in this paper are less conservative. Finally, two numerical examples are presented to illustrate the validity of the main results.

The Implementation of the Multi-Agent Classification System (MACS) in Compliance with FIPA Specifications

The paper discusses the implementation of the MultiAgent classification System (MACS) and utilizing it to provide an automated and accurate classification of end users developing applications in the spreadsheet domain. However, different technologies have been brought together to build MACS. The strength of the system is the integration of the agent technology with the FIPA specifications together with other technologies, which are the .NET widows service based agents, the Windows Communication Foundation (WCF) services, the Service Oriented Architecture (SOA), and Oracle Data Mining (ODM). The Microsoft's .NET widows service based agents were utilized to develop the monitoring agents of MACS, the .NET WCF services together with SOA approach allowed the distribution and communication between agents over the WWW. The Monitoring Agents (MAs) were configured to execute automatically to monitor excel spreadsheets development activities by content. Data gathered by the Monitoring Agents from various resources over a period of time was collected and filtered by a Database Updater Agent (DUA) residing in the .NET client application of the system. This agent then transfers and stores the data in Oracle server database via Oracle stored procedures for further processing that leads to the classification of the end user developers.

Using Multi-Linguistic Techniques for Thailand Herb and Traditional Medicine Registration Systems

Thailand has evolved many unique culture and knowledge, and the leading is the Thai traditional medicine (TTM). Recently, a number of researchers have tried to save this indigenous knowledge. However, the system to do so has still been scant. To preserve this ancient knowledge, we therefore invented and integrated multi-linguistic techniques to create the system of the collected all of recipes. This application extracted the medical recipes from antique scriptures then normalized antiquarian words, primitive grammar and antiquated measurement of them to the modern ones. Then, we applied ingredient-duplication-calculation, proportion-similarity-calculation and score-ranking to examine duplicate recipes. We collected the questionnaires from registrants and people to investigate the users’ satisfaction. The satisfactory results were found. This application assists not only registrants to validating the copyright violation in TTM registration process but also people to cure their illness that aids both Thai people and all mankind to fight for intractable diseases.

A Car Parking Monitoring System Using Wireless Sensor Networks

This paper presents a car parking monitoring system using wireless sensor networks. Multiple sensor nodes and a sink node, a gateway, and a server constitute a wireless network for monitoring a parking lot. Each of the sensor nodes is equipped with a 3-axis AMR sensor and deployed in the center of a parking space. Each sensor node reads its sensor values periodically and transmits the data to the sink node if the current and immediate past sensor values show a difference exceeding a threshold value. The sensor nodes and sink node use the 448 MHz band for wireless communication. Since RF transmission only occurs when sensor values show abrupt changes, the number of RF transmission operations is reduced and battery power can be conserved. The data from the sensor nodes reach the server via the sink node and gateway. The server determines which parking spaces are taken by cars based upon the received sensor data and reference values. The reference values are average sensor values measured by each sensor node when the corresponding parking spot is not occupied by a vehicle. Because the decision making is done by the server, the computational burden of the sensor node is relieved, which helps reduce the duty cycle of the sensor node.

Effect of Processing Methods on Texture Evolution in AZ31 Mg Alloy Sheet

Textures of AZ31 Mg alloy sheets were evaluated by using neutron diffraction method in this study. The AZ31 sheets were fabricated either by conventional casting and subsequent hot rolling or strip casting. The effect of warm rolling was investigated using the AZ31 Mg alloy sheet produced by conventional casting. Warm rolling of 30% thickness reduction per pass was possible without any side-crack at temperatures as low as 200oC under the roll speed of 30 m/min. The initial microstructure of conventionally cast specimen was found to be partially recrystallized structures. Grain refinement was found to occur actively during the warm rolling. The (0002),(10-10) (10-11),and (10-12) complete pole figures were measured using the HANARO FCD (Neutron Four Circle Diffractometer) and ODF were calculated. The major texture of all specimens can be expressed by ND//(0001) fiber texture. Texture of hot rolled specimen showed the strongest fiber component, while that of strip cast sheet seemed to be similar to random distribution.

Technique for Voltage Control in Distribution System

This paper presents the techniques for voltage control in distribution system. It is integrated in the distribution management system. Voltage is an important parameter for the control of electrical power systems. The distribution network operators have the responsibility to regulate the voltage supplied to consumer within statutory limits. Traditionally, the On-Load Tap Changer (OLTC) transformer equipped with automatic voltage control (AVC) relays is the most popular and effective voltage control device. A static synchronous compensator (STATCOM) may be equipped with several controllers to perform multiple control functions. Static Var Compensation (SVC) is regulation slopes and available margins for var dispatch. The voltage control in distribution networks is established as a centralized analytical function in this paper. 

Using Strategic CSR to Achieve the Hybrid Middle Ground in Social Entrepreneurship: The Case of Telenor Hungary

To be considered a socially entrepreneurial organization today requires achieving what can be termed a “hybrid middle ground” equilibrium, comprising of economic as well as social sustainability. This middle ground requires some blend of both business and social commitments. In this paper, we use the case of Hungary's second ranked mobile operator, Telenor Hungary to illustrate an example of a company that is moving to the hybrid middle ground by transitioning from a for-profit company to a socially responsible business using the concept of strategic CSR. In this line of thinking, the organization explicitly supports programs and initiatives that have a direct link to the core business and bring operational and/or financial advantages for the company, while creating a positive social and/or environmental impact. The important lessons learned from the company transition are also discussed. 

Application of Particle Swarm Optimization for Economic Load Dispatch and Loss Reduction

This paper proposes a particle swarm optimization (PSO) technique to solve the economic load dispatch (ELD) problems. For the ELD problem in this work, the objective function is to minimize the total fuel cost of all generator units for a given daily load pattern while the main constraints are power balance and generation output of each units. Case study in the test system of 40-generation units with 6 load patterns is presented to demonstrate the performance of PSO in solving the ELD problem. It can be seen that the optimal solution given by PSO provides the minimum total cost of generation while satisfying all the constraints and benefiting greatly from saving in power loss reduction.

Reliability Approximation through the Discretization of Random Variables using Reversed Hazard Rate Function

Sometime it is difficult to determine the exact reliability for complex systems in analytical procedures. Approximate solution of this problem can be provided through discretization of random variables. In this paper we describe the usefulness of discretization of a random variable using the reversed hazard rate function of its continuous version. Discretization of the exponential distribution has been demonstrated. Applications of this approach have also been cited. Numerical calculations indicate that the proposed approach gives very good approximation of reliability of complex systems under stress-strength set-up. The performance of the proposed approach is better than the existing discrete concentration method of discretization. This approach is conceptually simple, handles analytic intractability and reduces computational time. The approach can be applied in manufacturing industries for producing high-reliable items.