The Socio-Technical Indicator Model: Socially-Sensitive CMC Technology, with an Implementation of Representative Moderation

Computer-mediated communication technologies which provide for virtual communities have typically evolved in a cross-dichotomous manner, such that technical constructs of the technology have evolved independently from the social environment of the community. The present paper analyses some limitations of current implementations of computer-mediated communication technology that are implied by such a dichotomy, and discusses their inhibiting effects on possible developments of virtual communities. A Socio-Technical Indicator Model is introduced that utilizes integrated feedback to describe, simulate and operationalise increasing representativeness within a variety of structurally and parametrically diverse systems. In illustration, applications of the model are briefly described for financial markets and for eco-systems. A detailed application is then provided to resolve the aforementioned technical limitations of moderation on the evolution of virtual communities. The application parameterises virtual communities to function as self-transforming social-technical systems which are sensitive to emergent and shifting community values as products of on-going communications within the collective.

Applications of Conic Optimization and Quadratic Programming in the Investigation of Index Arbitrage in the Thai Derivatives and Equity Markets

This research seeks to investigate the frequency and profitability of index arbitrage opportunities involving the SET50 futures, SET50 component stocks, and the ThaiDEX SET50 ETF (ticker symbol: TDEX). In particular, the frequency and profit of arbitrage are measured in the following three arbitrage tests: (1) SET50 futures vs. ThaiDEX SET50 ETF, (2) SET50 futures vs. SET50 component stocks, and (3) ThaiDEX SET50 ETF vs. SET50 component stocks are investigated. For tests (2) and (3), the problems involve conic optimization and quadratic programming as subproblems. This research is first to apply conic optimization and quadratic programming techniques in the context of index arbitrage and is first to investigate such index arbitrage in the Thai equity and derivatives markets. Thus, the contribution of this study is twofold. First, its results would help understand the contribution of the derivatives securities to the efficiency of the Thai markets. Second, the methodology employed in this study can be applied to other geographical markets, with minor adjustments.

The Finite Difference Scheme for the Suspended String Equation with the Nonlinear Damping Term

A numerical solution of the initial boundary value problem of the suspended string vibrating equation with the particular nonlinear damping term based on the finite difference scheme is presented in this paper. The investigation of how the second and third power terms of the nonlinear term affect the vibration characteristic. We compare the vibration amplitude as a result of the third power nonlinear damping with the second power obtained from previous report provided that the same initial shape and initial velocities are assumed. The comparison results show that the vibration amplitude is inversely proportional to the coefficient of the damping term for the third power nonlinear damping case, while the vibration amplitude is proportional to the coefficient of the damping term in the second power nonlinear damping case.

Optimal Model Order Selection for Transient Error Autoregressive Moving Average (TERA) MRI Reconstruction Method

An alternative approach to the use of Discrete Fourier Transform (DFT) for Magnetic Resonance Imaging (MRI) reconstruction is the use of parametric modeling technique. This method is suitable for problems in which the image can be modeled by explicit known source functions with a few adjustable parameters. Despite the success reported in the use of modeling technique as an alternative MRI reconstruction technique, two important problems constitutes challenges to the applicability of this method, these are estimation of Model order and model coefficient determination. In this paper, five of the suggested method of evaluating the model order have been evaluated, these are: The Final Prediction Error (FPE), Akaike Information Criterion (AIC), Residual Variance (RV), Minimum Description Length (MDL) and Hannan and Quinn (HNQ) criterion. These criteria were evaluated on MRI data sets based on the method of Transient Error Reconstruction Algorithm (TERA). The result for each criterion is compared to result obtained by the use of a fixed order technique and three measures of similarity were evaluated. Result obtained shows that the use of MDL gives the highest measure of similarity to that use by a fixed order technique.

A New Approach to Signal Processing for DC-Electromagnetic Flowmeters

Electromagnetic flowmeters with DC excitation are used for a wide range of fluid measurement tasks, but are rarely found in dosing applications with short measurement cycles due to the achievable accuracy. This paper will identify a number of factors that influence the accuracy of this sensor type when used for short-term measurements. Based on these results a new signal-processing algorithm will be described that overcomes the identified problems to some extend. This new method allows principally a higher accuracy of electromagnetic flowmeters with DC excitation than traditional methods.

A Study on Integrated Performance of Tap-Changing Transformer and SVC in Association with Power System Voltage Stability

Electricity market activities and a growing demand for electricity have led to heavily stressed power systems. This requires operation of the networks closer to their stability limits. Power system operation is affected by stability related problems, leading to unpredictable system behavior. Voltage stability refers to the ability of a power system to sustain appropriate voltage levels through large and small disturbances. Steady-state voltage stability is concerned with limits on the existence of steady-state operating points for the network. FACTS devices can be utilized to increase the transmission capacity, the stability margin and dynamic behavior or serve to ensure improved power quality. Their main capabilities are reactive power compensation, voltage control and power flow control. Among the FACTS controllers, Static Var Compensator (SVC) provides fast acting dynamic reactive compensation for voltage support during contingency events. In this paper, voltage stability assessment with appropriate representations of tap-changer transformers and SVC is investigated. Integrating both of these devices is the main topic of this paper. Effect of the presence of tap-changing transformers on static VAR compensator controller parameters and ratings necessary to stabilize load voltages at certain values are highlighted. The interrelation between transformer off nominal tap ratios and the SVC controller gains and droop slopes and the SVC rating are found. P-V curves are constructed to calculate loadability margins.

Deep Web Content Mining

The rapid expansion of the web is causing the constant growth of information, leading to several problems such as increased difficulty of extracting potentially useful knowledge. Web content mining confronts this problem gathering explicit information from different web sites for its access and knowledge discovery. Query interfaces of web databases share common building blocks. After extracting information with parsing approach, we use a new data mining algorithm to match a large number of schemas in databases at a time. Using this algorithm increases the speed of information matching. In addition, instead of simple 1:1 matching, they do complex (m:n) matching between query interfaces. In this paper we present a novel correlation mining algorithm that matches correlated attributes with smaller cost. This algorithm uses Jaccard measure to distinguish positive and negative correlated attributes. After that, system matches the user query with different query interfaces in special domain and finally chooses the nearest query interface with user query to answer to it.

Proteins Length and their Phenotypic Potential

Mendelian Disease Genes represent a collection of single points of failure for the various systems they constitute. Such genes have been shown, on average, to encode longer proteins than 'non-disease' proteins. Existing models suggest that this results from the increased likeli-hood of longer genes undergoing mutations. Here, we show that in saturated mutagenesis experiments performed on model organisms, where the likelihood of each gene mutating is one, a similar relationship between length and the probability of a gene being lethal was observed. We thus suggest an extended model demonstrating that the likelihood of a mutated gene to produce a severe phenotype is length-dependent. Using the occurrence of conserved domains, we bring evidence that this dependency results from a correlation between protein length and the number of functions it performs. We propose that protein length thus serves as a proxy for protein cardinality in different networks required for the organism's survival and well-being. We use this example to argue that the collection of Mendelian Disease Genes can, and should, be used to study the rules governing systems vulnerability in living organisms.

Representation of Coloured Petri Net in Abductive Logic Programming (CPN-LP) and Its Application in Modeling an Intelligent Agent

Coloured Petri net (CPN) has been widely adopted in various areas in Computer Science, including protocol specification, performance evaluation, distributed systems and coordination in multi-agent systems. It provides a graphical representation of a system and has a strong mathematical foundation for proving various properties. This paper proposes a novel representation of a coloured Petri net using an extension of logic programming called abductive logic programming (ALP), which is purely based on classical logic. Under such a representation, an implementation of a CPN could be directly obtained, in which every inference step could be treated as a kind of equivalence preserved transformation. We would describe how to implement a CPN under such a representation using common meta-programming techniques in Prolog. We call our framework CPN-LP and illustrate its applications in modeling an intelligent agent.

Concurrent Testing of ADC for Embedded System

Compaction testing methods allow at-speed detecting of errors while possessing low cost of implementation. Owing to this distinctive feature, compaction methods have been widely used for built-in testing, as well as external testing. In the latter case, the bandwidth requirements to the automated test equipment employed are relaxed which reduces the overall cost of testing. Concurrent compaction testing methods use operational signals to detect misbehavior of the device under test and do not require input test stimuli. These methods have been employed for digital systems only. In the present work, we extend the use of compaction methods for concurrent testing of analog-to-digital converters. We estimate tolerance bounds for the result of compaction and evaluate the aliasing rate.

Study about the Development of Small Towns in the Metropolitan Fringe in Developed Area of China–A Case Study of Sandun Town in Hangzhou

Due to the Rapid Urbanization in China, the influence of metropolises to surrounding areas grows by a tremendous speed in developed region. One of the most obvious influences is the expansion of the urban built-up areas which covers the land belongs to small towns. Around metropolitan fringe, the boundary between city and village becomes more and more obscure. So being the most sensitive area, the small towns on the fringe of metropolises have the special meaning on the research into the small towns- development. This paper chooses Sandun Town in Hangzhou of Zhejiang Province as an example, emphatically focus on aspects such as the central area proliferation, the industrial shift, the position effect, the subway effect and the commercial development, reviews a few problems of small towns in the future and the important problems in their planning by the analysis of the characteristics of the present conditions and the developing motive mechanism, so that guides small towns to develop properly by liking with these small towns and center metropolises.

Recognition and Reconstruction of Partially Occluded Objects

A new automatic system for the recognition and re¬construction of resealed and/or rotated partially occluded objects is presented. The objects to be recognized are described by 2D views and each view is occluded by several half-planes. The whole object views and their visible parts (linear cuts) are then stored in a database. To establish if a region R of an input image represents an object possibly occluded, the system generates a set of linear cuts of R and compare them with the elements in the database. Each linear cut of R is associated to the most similar database linear cut. R is recognized as an instance of the object 0 if the majority of the linear cuts of R are associated to a linear cut of views of 0. In the case of recognition, the system reconstructs the occluded part of R and determines the scale factor and the orientation in the image plane of the recognized object view. The system has been tested on two different datasets of objects, showing good performance both in terms of recognition and reconstruction accuracy.

Weight Functions for Signal Reconstruction Based On Level Crossings

Although the level crossing concept has been the subject of intensive investigation over the last few years, certain problems of great interest remain unsolved. One of these concern is distribution of threshold levels. This paper presents a new threshold level allocation schemes for level crossing based on nonuniform sampling. Intuitively, it is more reasonable if the information rich regions of the signal are sampled finer and those with sparse information are sampled coarser. To achieve this objective, we propose non-linear quantization functions which dynamically assign the number of quantization levels depending on the importance of the given amplitude range. Two new approaches to determine the importance of the given amplitude segment are presented. The proposed methods are based on exponential and logarithmic functions. Various aspects of proposed techniques are discussed and experimentally validated. Its efficacy is investigated by comparison with uniform sampling.

Quality of Life of Poor Residential Neighborhoods in Oshogbo, Nigeria

As a result of the high cost of housing, the increasing population is forced to live in substandard housing and unhealthy conditions giving rise to poor residential neighborhoods. The paper examines the causes and characteristics of poor residential neighborhood. The paper finds the problems that have influence poor neighborhoods to; poverty, growth of informal sector and housing shortage. The paper asserts that poor residential neighborhoods have adverse effects on the people. The secondary data was obtained from books, journals and seminar papers while primary data relating to building and environmental quality from structured questionnaire administered on sample of 500 household heads, from sampling frame of 5000 housing units. The study reveals that majority of the respondents are poor and employed in informal sector. The paper suggests urban renewal and slum upgrading programs as methods in dealing with the situation and an improvement in the socio-economic circumstances of the inhabitants.

Neuro-Hybrid Models for Automotive System Identification

In automotive systems almost all steps concerning the calibration of several control systems, e.g., low idle governor or boost pressure governor, are made with the vehicle because the timeto- production and cost requirements on the projects do not allow for the vehicle analysis necessary to build reliable models. Here is presented a procedure using parametric and NN (neural network) models that enables the generation of vehicle system models based on normal ECU engine control unit) vehicle measurements. These models are locally valid and permit pre and follow-up calibrations so that, only the final calibrations have to be done with the vehicle.

A Intelligent Inference Model about Complex Systems- Stability: Inspiration from Nature

A logic model for analyzing complex systems- stability is very useful to many areas of sciences. In the real world, we are enlightened from some natural phenomena such as “biosphere", “food chain", “ecological balance" etc. By research and practice, and taking advantage of the orthogonality and symmetry defined by the theory of multilateral matrices, we put forward a logic analysis model of stability of complex systems with three relations, and prove it by means of mathematics. This logic model is usually successful in analyzing stability of a complex system. The structure of the logic model is not only clear and simple, but also can be easily used to research and solve many stability problems of complex systems. As an application, some examples are given.

Unconditionally Secure Quantum Payment System

A potentially serious problem with current payment systems is that their underlying hard problems from number theory may be solved by either a quantum computer or unanticipated future advances in algorithms and hardware. A new quantum payment system is proposed in this paper. The suggested system makes use of fundamental principles of quantum mechanics to ensure the unconditional security without prior arrangements between customers and vendors. More specifically, the new system uses Greenberger-Home-Zeilinger (GHZ) states and Quantum Key Distribution to authenticate the vendors and guarantee the transaction integrity.

Multi-agent On-line Monitor for the Safety of Critical Systems

Operational safety of critical systems, such as nuclear power plants, industrial chemical processes and means of transportation, is a major concern for system engineers and operators. A means to assure that is on-line safety monitors that deliver three safety tasks; fault detection and diagnosis, alarm annunciation and fault controlling. While current monitors deliver these tasks, benefits and limitations in their approaches have at the same time been highlighted. Drawing from those benefits, this paper develops a distributed monitor based on semi-independent agents, i.e. a multiagent system, and monitoring knowledge derived from a safety assessment model of the monitored system. Agents are deployed hierarchically and provided with knowledge portions and collaboration protocols to reason and integrate over the operational conditions of the components of the monitored system. The monitor aims to address limitations arising from the large-scale, complicated behaviour and distributed nature of monitored systems and deliver the aforementioned three monitoring tasks effectively.

Lessons from Applying XP Methodology to Business Requirements Engineering in Developing Countries Context

Most standard software development methodologies are often not applied to software projects in many developing countries of the world. The approach generally practice is close to what eXtreme Programming (XP) is likely promoting, just keep coding and testing as the requirement evolves. XP is an agile software process development methodology that has inherent capability for improving efficiency of Business Software Development (BSD). XP can facilitate Business-to-Development (B2D) relationship due to its customer-oriented advocate. From practitioner point of view, we applied XP to BSD and result shows that customer involvement has positive impact on productivity, but can as well frustrate the success of the project. In an effort to promote software engineering practice in developing countries of Africa, we present the experiment performed, lessons learned, problems encountered and solution adopted in applying XP methodology to BSD.

Flexible Manufacturing System

Flexible manufacturing system is a system that is able to respond to changed conditions. In general, this flexibility is divided into two key categories and several subcategories. The first category is the so called machine flexibility which enables to make various products by the given machinery. The second category is routing flexibility enabling to execute the same operation by various machines. Flexible manufacturing systems usually consist of three main parts: CNC machine tools, transport system and control system. A higher level of flexible manufacturing systems is represented by the so called intelligent manufacturing systems.