Abstract: This paper focuses on analyzing medical diagnostic data using classification rules in data mining and context reduction in formal concept analysis. It helps in finding redundancies among the various medical examination tests used in diagnosis of a disease. Classification rules have been derived from positive and negative association rules using the Concept lattice structure of the Formal Concept Analysis. Context reduction technique given in Formal Concept Analysis along with classification rules has been used to find redundancies among the various medical examination tests. Also it finds out whether expensive medical tests can be replaced by some cheaper tests.
Abstract: In this work, we address theoretically the influence of red and white Gaussian noise for electronic energies and eigenstates of cylindrically shaped quantum dots. The stochastic effect can be imagined as resulting from crystal-growth statistical fluctuations in the quantum-dot material composition. In particular we obtain analytical expressions for the eigenvalue shifts and electronic envelope functions in the k . p formalism due to stochastic variations in the confining band-edge potential. It is shown that white noise in the band-edge potential leaves electronic properties almost unaffected while red noise may lead to changes in state energies and envelopefunction amplitudes of several percentages. In the latter case, the ensemble-averaged envelope function decays as a function of distance. It is also shown that, in a stochastic system, constant ensembleaveraged envelope functions are the only bounded solutions for the infinite quantum-wire problem and the energy spectrum is completely discrete. In other words, the infinite stochastic quantum wire behaves, ensemble-averaged, as an atom.
Abstract: One of the common problems encountered in software
engineering is addressing and responding to the changing nature of
requirements. While several approaches have been devised to address
this issue, ranging from instilling resistance to changing requirements
in order to mitigate impact to project schedules, to developing an
agile mindset towards requirements, the approach discussed in this
paper is one of conceptualizing the delta in requirement and
modeling it, in order to plan a response to it. To provide some
context here, change is first formally identified and categorized as
either formal change or informal change. While agile methodology
facilitates informal change, the approach discussed in this paper
seeks to develop the idea of facilitating formal change. To collect,
document meta-requirements that represent the phenomena of change
would be a pro-active measure towards building a realistic cognition
of the requirements entity that can further be harnessed in the
software engineering process.
Abstract: Recent advances in both the testing and verification of software based on formal specifications of the system to be built have reached a point where the ideas can be applied in a powerful way in the design of agent-based systems. The software engineering research has highlighted a number of important issues: the importance of the type of modeling technique used; the careful design of the model to enable powerful testing techniques to be used; the automated verification of the behavioural properties of the system; the need to provide a mechanism for translating the formal models into executable software in a simple and transparent way. This paper introduces the use of the X-machine formalism as a tool for modeling biology inspired agents proposing the use of the techniques built around X-machine models for the construction of effective, and reliable agent-based software systems.
Abstract: This paper focuses on issues of engagement by staff in professional development related to the delivery of e-learning. The paper reports on findings drawn from a New Zealand research project which is producing a sector-wide framework for professional development in tertiary e-learning. The research findings indicate that staff engaged in e-learning in tertiary institutions is not making the most effective use of the professional development opportunities available to them; rather they seem to gain their knowledge and support from a variety of informal means. This is despite an emphasis on the provision of professional development opportunities by both Government Policies and Institutions themselves. The conclusion drawn from the findings is that institutional approaches to professional development for e-learning do not yet fully reflect the demands and constraints that working in a digital context impose.
Abstract: We study the problem of reconstructing a three dimensional binary matrices whose interiors are only accessible through few projections. Such question is prominently motivated by the demand in material science for developing tool for reconstruction of crystalline structures from their images obtained by high-resolution transmission electron microscopy. Various approaches have been suggested to reconstruct 3D-object (crystalline structure) by reconstructing slice of the 3D-object. To handle the ill-posedness of the problem, a priori information such as convexity, connectivity and periodicity are used to limit the number of possible solutions. Formally, 3Dobject (crystalline structure) having a priory information is modeled by a class of 3D-binary matrices satisfying a priori information. We consider 3D-binary matrices with periodicity constraints, and we propose a polynomial time algorithm to reconstruct 3D-binary matrices with periodicity constraints from two orthogonal projections.
Abstract: Recently the usefulness of Concept Abduction, a novel non-monotonic inference service for Description Logics (DLs), has been argued in the context of ontology-based applications such as semantic matchmaking and resource retrieval. Based on tableau calculus, a method has been proposed to realize this reasoning task in ALN, a description logic that supports simple cardinality restrictions as well as other basic constructors. However, in many ontology-based systems, the representation of ontology would require expressive formalisms for capturing domain-specific constraints, this language is not sufficient. In order to increase the applicability of the abductive reasoning method in such contexts, we would like to present in the scope of this paper an extension of the tableaux-based algorithm for dealing with concepts represented inALCQ, the description logic that extends ALN with full concept negation and quantified number restrictions.
Abstract: Business Process Modeling (BPM) is the first and
most important step in business process management lifecycle. Graph
based formalism and rule based formalism are the two most
predominant formalisms on which process modeling languages are
developed. BPM technology continues to face challenges in coping
with dynamic business environments where requirements and goals
are constantly changing at the execution time. Graph based
formalisms incur problems to react to dynamic changes in Business
Process (BP) at the runtime instances. In this research, an adaptive
and flexible framework based on the integration between Object
Oriented diagramming technique and Petri Net modeling language is
proposed in order to support change management techniques for
BPM and increase the representation capability for Object Oriented
modeling for the dynamic changes in the runtime instances. The
proposed framework is applied in a higher education environment to
achieve flexible, updatable and dynamic BP.
Abstract: As computer network technology becomes
increasingly complex, it becomes necessary to place greater
requirements on the validity of developing standards and the
resulting technology. Communication networks are based on large
amounts of protocols. The validity of these protocols have to be
proved either individually or in an integral fashion. One strategy for
achieving this is to apply the growing field of formal methods.
Formal methods research defines systems in high order logic so that
automated reasoning can be applied for verification. In this research
we represent and implement a formerly announced multicast protocol
in Prolog language so that certain properties of the protocol can be
verified. It is shown that by using this approach some minor faults in
the protocol were found and repaired. Describing the protocol as
facts and rules also have other benefits i.e. leads to a process-able
knowledge. This knowledge can be transferred as ontology between
systems in KQML format. Since the Prolog language can increase its
knowledge base every time, this method can also be used to learn an
intelligent network.
Abstract: e-mail has become an important means of electronic
communication but the viability of its usage is marred by Unsolicited
Bulk e-mail (UBE) messages. UBE consists of many types
like pornographic, virus infected and 'cry-for-help' messages as well
as fake and fraudulent offers for jobs, winnings and medicines. UBE
poses technical and socio-economic challenges to usage of e-mails.
To meet this challenge and combat this menace, we need to
understand UBE. Towards this end, the current paper presents a
content-based textual analysis of more than 2700 body enhancement
medicinal UBE. Technically, this is an application of Text Parsing
and Tokenization for an un-structured textual document and we
approach it using Bag Of Words (BOW) and Vector Space Document
Model techniques. We have attempted to identify the most
frequently occurring lexis in the UBE documents that advertise
various products for body enhancement. The analysis of such top
100 lexis is also presented. We exhibit the relationship between
occurrence of a word from the identified lexis-set in the given UBE
and the probability that the given UBE will be the one advertising for
fake medicinal product. To the best of our knowledge and survey of
related literature, this is the first formal attempt for identification of
most frequently occurring lexis in such UBE by its textual analysis.
Finally, this is a sincere attempt to bring about alertness against and
mitigate the threat of such luring but fake UBE.
Abstract: Organizations face challenges supporting knowledge
workers due to their particular requirements for an environment
supportive of their self-guided learning activities which are important
to increase their productivity and to develop creative solutions to
non-routine problems. Face-to-face knowledge sharing remains
crucial in spite of a large number of knowledge management
instruments that aim at supporting a more impersonal transfer of
knowledge. This paper first describes the main criteria for a
conceptual and technical solution targeted at flexible management of
office space that aims at assigning those knowledge workers to the
same room that are most likely to thrive when being brought together
thus enhancing their knowledge work productivity. The paper
reflects on lessons learned from the implementation and operation of
such a solution in a project-focused organization and derives several
implications for future extensions that target to foster problem
solving, informal learning and personal development.
Abstract: Web services are pieces of software that can be invoked via a standardized protocol. They can be combined via formalized taskflow languages. The Open Knowledge system is a fully distributed system using P2P technology, that allows users to publish the setaskflows, and programmers to register their web services or publish implementations of them, for the roles described in these workflows.Besides this, the system offers the functionality to select a peer that could coordinate such an interaction model and inform web services when it is their 'turn'. In this paper we describe the architecture and implementation of the Open Knowledge Kernel which provides the core functionality of the Open Knowledge system.
Abstract: This paper study the behavior of the solution at the crack edges for an elliptical crack with developing cusps, Ω in the plane elasticity subjected to shear loading. The problem of finding the resulting shear stress can be formulated as a hypersingular integral equation over Ω and it is then transformed into a similar equation over a circular region, D, using conformal mapping. An appropriate collocation points are chosen on the region D to reduce the hypersingular integral equation into a system of linear equations with (2N+1)(N+1) unknown coefficients, which will later be used in the determination of shear stress intensity factors and maximum shear stress intensity. Numerical solution for the considered problem are compared with the existing asymptotic solution, and displayed graphically. Our results give a very good agreement to the existing asymptotic solutions.
Abstract: Droplet size distributions in the cold spray of a fuel
are important in observed combustion behavior. Specification of
droplet size and velocity distributions in the immediate downstream
of injectors is also essential as boundary conditions for advanced
computational fluid dynamics (CFD) and two-phase spray transport
calculations. This paper describes the development of a new model to
be incorporated into maximum entropy principle (MEP) formalism
for prediction of droplet size distribution in droplet formation region.
The MEP approach can predict the most likely droplet size and
velocity distributions under a set of constraints expressing the
available information related to the distribution.
In this article, by considering the mechanisms of turbulence
generation inside the nozzle and wave growth on jet surface, it is
attempted to provide a logical framework coupling the flow inside the
nozzle to the resulting atomization process. The purpose of this paper
is to describe the formulation of this new model and to incorporate it
into the maximum entropy principle (MEP) by coupling sub-models
together using source terms of momentum and energy. Comparison
between the model prediction and experimental data for a gas turbine
swirling nozzle and an annular spray indicate good agreement
between model and experiment.
Abstract: To determine the length of engagement threads of a bolt installed in a tapped part in order to avoid the threads stripping remains a very current problem in the design of the thread assemblies. It does not exist a calculation method formalized for the cases where the bolt is screwed directly in a ductile material. In this article, we study the behavior of the threads stripping of a loaded assembly by using a modelling by finite elements and a rupture criterion by damage. This modelling enables us to study the different parameters likely to influence the behavior of this bolted connection. We study in particular, the influence of couple of materials constituting the connection, of the bolt-s diameter and the geometrical characteristics of the tapped part, like the external diameter and the length of engagement threads. We established an experiments design to know the most significant parameters. That enables us to propose a simple expression making possible to calculate the resistance of the threads whatever the metallic materials of the bolt and the tapped part. We carried out stripping tests in order to validate our model. The estimated results are very close to those obtained by the tests.
Abstract: A separation-kernel-based operating system (OS) has been designed for use in secure embedded systems by applying formal methods to the design of the separation-kernel part. The separation kernel is a small OS kernel that provides an abstract distributed environment on a single CPU. The design of the separation kernel was verified using two formal methods, the B method and the Spin model checker. A newly designed semi-formal method, the extended state transition method, was also applied. An OS comprising the separation-kernel part and additional OS services on top of the separation kernel was prototyped on the Intel IA-32 architecture. Developing and testing of a prototype embedded application, a point-of-sale application, on the prototype OS demonstrated that the proposed architecture and the use of formal methods to design its kernel part are effective for achieving a secure embedded system having a high-assurance separation kernel.
Abstract: Some quality control tools use non metric subjective information coming from experts, who qualify the intensity of relations existing inside processes, but without quantifying them. In this paper we have developed a quality control analytic tool, measuring the impact or strength of the relationship between process operations and product characteristics. The tool includes two models: a qualitative model, allowing relationships description and analysis; and a formal quantitative model, by means of which relationship quantification is achieved. In the first one, concepts from the Graphs Theory were applied to identify those process elements which can be sources of variation, that is, those quality characteristics or operations that have some sort of prelacy over the others and that should become control items. Also the most dependent elements can be identified, that is those elements receiving the effects of elements identified as variation sources. If controls are focused in those dependent elements, efficiency of control is compromised by the fact that we are controlling effects, not causes. The second model applied adapts the multivariate statistical technique of Covariance Structural Analysis. This approach allowed us to quantify the relationships. The computer package LISREL was used to obtain statistics and to validate the model.
Abstract: An automatic speech recognition system for the
formal Arabic language is needed. The Quran is the most formal
spoken book in Arabic, it is spoken all over the world. In this
research, an automatic speech recognizer for Quranic based speakerindependent
was developed and tested. The system was developed
based on the tri-phone Hidden Markov Model and Maximum
Likelihood Linear Regression (MLLR). The MLLR computes a set
of transformations which reduces the mismatch between an initial
model set and the adaptation data. It uses the regression class tree, as
well as, estimates a set of linear transformations for the mean and
variance parameters of a Gaussian mixture HMM system. The 30th
Chapter of the Quran, with five of the most famous readers of the
Quran, was used for the training and testing of the data. The chapter
includes about 2000 distinct words. The advantages of using the
Quranic verses as the database in this developed recognizer are the
uniqueness of the words and the high level of orderliness between
verses. The level of accuracy from the tested data ranged 68 to 85%.
Abstract: Recently, a model multi-agent e-commerce system based on mobile buyer agents and transfer of strategy modules was proposed. In this paper a different approach to code mobility is introduced, where agent mobility is replaced by local agent creation supplemented by similar code mobility as in the original proposal. UML diagrams of agents involved in the new approach to mobility and the augmented system activity diagram are presented and discussed.
Abstract: The UML modeling of complex distributed systems often is a great challenge due to the large amount of parallel real-time operating components. In this paper the problems of verification of such systems are discussed. ECPN, an Extended Colored Petri Net is defined to formally describe state transitions of components and interactions among components. The relationship between sequence diagrams and Free Choice Petri Nets is investigated. Free Choice Petri Net theory helps verifying the liveness of sequence diagrams. By converting sequence diagrams to ECPNs and then comparing behaviors of sequence diagram ECPNs and statecharts, the consistency among models is analyzed. Finally, a verification process for an example model is demonstrated.