Abstract: Automated discovery of Rule is, due to its applicability, one of the most fundamental and important method in KDD. It has been an active research area in the recent past. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form: Decision If < condition> Generality Specificity . HPRs systems are capable of handling taxonomical structures inherent in the knowledge about the real world. This paper focuses on the issue of mining Quantified rules with crisp hierarchical structure using Genetic Programming (GP) approach to knowledge discovery. The post-processing scheme presented in this work uses Quantified production rules as initial individuals of GP and discovers hierarchical structure. In proposed approach rules are quantified by using Dempster Shafer theory. Suitable genetic operators are proposed for the suggested encoding. Based on the Subsumption Matrix(SM), an appropriate fitness function is suggested. Finally, Quantified Hierarchical Production Rules (HPRs) are generated from the discovered hierarchy, using Dempster Shafer theory. Experimental results are presented to demonstrate the performance of the proposed algorithm.
Abstract: In Virtual organization, Knowledge Discovery (KD)
service contains distributed data resources and computing grid nodes.
Computational grid is integrated with data grid to form Knowledge
Grid, which implements Apriori algorithm for mining association
rule on grid network. This paper describes development of parallel
and distributed version of Apriori algorithm on Globus Toolkit using
Message Passing Interface extended with Grid Services (MPICHG2).
The creation of Knowledge Grid on top of data and
computational grid is to support decision making in real time
applications. In this paper, the case study describes design and
implementation of local and global mining of frequent item sets. The
experiments were conducted on different configurations of grid
network and computation time was recorded for each operation. We
analyzed our result with various grid configurations and it shows
speedup of computation time is almost superlinear.
Abstract: A data warehouse (DW) is a system which has value and role for decision-making by querying. Queries to DW are critical regarding to their complexity and length. They often access millions of tuples, and involve joins between relations and aggregations. Materialized views are able to provide the better performance for DW queries. However, these views have maintenance cost, so materialization of all views is not possible. An important challenge of DW environment is materialized view selection because we have to realize the trade-off between performance and view maintenance. Therefore, in this paper, we introduce a new approach aimed to solve this challenge based on Two-Phase Optimization (2PO), which is a combination of Simulated Annealing (SA) and Iterative Improvement (II), with the use of Multiple View Processing Plan (MVPP). Our experiments show that 2PO outperform the original algorithms in terms of query processing cost and view maintenance cost.
Abstract: Young patients suffering from Cerebral Palsy are
facing difficult choices concerning heavy surgeries. Diagnosis settled
by surgeons can be complex and on the other hand decision for
patient about getting or not such a surgery involves important
reflection effort. Proposed software combining prediction for
surgeries and post surgery kinematic values, and from 3D model
representing the patient is an innovative tool helpful for both patients
and medicine professionals. Beginning with analysis and
classification of kinematics values from Data Base extracted from
gait analysis in 3 separated clusters, it is possible to determine close
similarity between patients. Prediction surgery best adapted to
improve a patient gait is then determined by operating a suitable
preconditioned neural network. Finally, patient 3D modeling based
on kinematic values analysis, is animated thanks to post surgery
kinematic vectors characterizing the closest patient selected from
patients clustering.
Abstract: Segmentation of a color image composed of different
kinds of regions can be a hard problem, namely to compute for an
exact texture fields. The decision of the optimum number of
segmentation areas in an image when it contains similar and/or un
stationary texture fields. A novel neighborhood-based segmentation
approach is proposed. A genetic algorithm is used in the proposed
segment-pass optimization process. In this pass, an energy function,
which is defined based on Markov Random Fields, is minimized. In
this paper we use an adaptive threshold estimation method for image
thresholding in the wavelet domain based on the generalized
Gaussian distribution (GGD) modeling of sub band coefficients. This
method called Normal Shrink is computationally more efficient and
adaptive because the parameters required for estimating the threshold
depend on sub band data energy that used in the pre-stage of
segmentation. A quad tree is employed to implement the multi
resolution framework, which enables the use of different strategies at
different resolution levels, and hence, the computation can be
accelerated. The experimental results using the proposed
segmentation approach are very encouraging.
Abstract: increased competition and increased costs of
designing made it important for the firms to identify the right
products and the right methods for manufacturing the products. Firms
should focus on customers and identify customer demands directly to
design the right products. Several management methods and
techniques that are currently available improve one or more functions
or processes in an industry and do not take the complete product life
cycle into consideration. On the other hand target costing is a method
/ philosophy that takes financial, manufacturing and customer aspects
into consideration during designing phase and helps firms in making
product design decisions to increase the profit / value of the
company. It uses various techniques to identify customer demands, to
decrease costs of manufacturing and finally to achieve strategic goals.
Target Costing forms an integral part of total product design /
redesign based on strategic plans.
Abstract: The efficient use of available licensed spectrum is
becoming more and more critical with increasing demand and usage
of the radio spectrum. This paper shows how the use of spectrum as
well as dynamic spectrum management can be effectively managed
and spectrum allocation schemes in the wireless communication
systems be implemented and used, in future. This paper would be an
attempt towards better utilization of the spectrum. This research will
focus on the decision-making process mainly, with an
assumption that the radio environment has already been sensed and
the QoS requirements for the application have been specified either
by the sensed radio environment or by the secondary user itself. We
identify and study the characteristic parameters of Cognitive Radio
and use Genetic Algorithm for spectrum allocation. Performance
evaluation is done using MATLAB toolboxes.
Abstract: Using maximal consistent blocks of tolerance relation
on the universe in incomplete decision table, the concepts of join block
and meet block are introduced and studied. Including tolerance class,
other blocks such as tolerant kernel and compatible kernel of an object
are also discussed at the same time. Upper and lower approximations
based on those blocks are also defined. Default definite decision rules
acquired from incomplete decision table are proposed in the paper. An
incremental algorithm to update default definite decision rules is
suggested for effective mining tasks from incomplete decision table
into which data is appended. Through an example, we demonstrate
how default definite decision rules based on maximal consistent
blocks, join blocks and meet blocks are acquired and how optimization
is done in support of discernibility matrix and discernibility function
in the incomplete decision table.
Abstract: Due to the excess of a vehicle operation through its life, some elements may face failure and deteriorate with time. This leads us to carry out maintenance, repair, tune up or full overhaul. After a certain period, the vehicle elements deteriorations increase with time which causes a very high increase of doing the maintenance operations and their costs. However, the logic decision at this point is to replace the current vehicle by a new one with minimum failure and maximum income. The importance of studying vehicle replacement problems come from the increase of stopping days due to many deteriorations in the vehicle parts. These deteriorations increase year after year causing an increase of operating costs and decrease the vehicle income. Vehicle replacement aims to determine the optimum time to keep, maintain, overhaul, renew and replace vehicles. This leads to an improvement in vehicle income, total operating costs, maintenance cost, fuel and oil costs, ton-kilometers, vehicle and engine performance, vehicle noise, vibration, and pollution. The aim of this paper is to find the optimum replacement policies of Kuwait Passenger Transport Company (KPTCP) fleet of busses. The objective of these policies is to maximize the busses pure profits. The dynamic programming (D.P.) technique is used to generate the busses optimal replacement policies
Abstract: In this paper a new approach to prioritize urban planning projects in an efficient and reliable way is presented. It is based on environmental pressure indices and multicriteria decision methods. The paper introduces a rigorous method with acceptable complexity of rank ordering urban development proposals according to their environmental pressure. The technique combines the use of Environmental Pressure Indicators, the aggregation of indicators in an Environmental Pressure Index by means of the Analytic Network Process method and interpreting the information obtained from the experts during the decision-making process. The ANP method allows the aggregation of the experts- judgments on each of the indicators into one Environmental Pressure Index. In addition, ANP is based on utility ratio functions which are the most appropriate for the analysis of uncertain data, like experts- estimations. Finally, unlike the other multicriteria techniques, ANP allows the decision problem to be modelled using the relationships among dependent criteria. The method has been applied to the proposal for urban development of La Carlota airport in Caracas (Venezuela). The Venezuelan Government would like to see a recreational project develop on the abandoned area and mean a significant improvement for the capital. There are currently three options on their table which are currently under evaluation. They include a Health Club, a Residential area and a Theme Park. The participating experts coincided in the appreciation that the method proposed in this paper is useful and an improvement from traditional techniques such as environmental impact studies, lifecycle analysis, etc. They find the results obtained coherent, the process seems sufficiently rigorous and precise, and the use of resources is significantly less than in other methods.
Abstract: Automatic reusability appraisal could be helpful in
evaluating the quality of developed or developing reusable software
components and in identification of reusable components from
existing legacy systems; that can save cost of developing the software
from scratch. But the issue of how to identify reusable components
from existing systems has remained relatively unexplored. In this
paper, we have mentioned two-tier approach by studying the
structural attributes as well as usability or relevancy of the
component to a particular domain. Latent semantic analysis is used
for the feature vector representation of various software domains. It
exploits the fact that FeatureVector codes can be seen as documents
containing terms -the idenifiers present in the components- and so
text modeling methods that capture co-occurrence information in
low-dimensional spaces can be used. Further, we devised Neuro-
Fuzzy hybrid Inference System, which takes structural metric values
as input and calculates the reusability of the software component.
Decision tree algorithm is used to decide initial set of fuzzy rules for
the Neuro-fuzzy system. The results obtained are convincing enough
to propose the system for economical identification and retrieval of
reusable software components.
Abstract: Several recent studies have shown that the
transparency of financial reporting have a significant influence on investor-s decisions. Thus, regulation authorities and professional
organizations (IFAC) have emphasized the role of XBRL (eXtensible Business Reporting Language) and interactive data as a means of
promoting transparency and monitoring corporate reporting. In this
context, this paper has as objective the analysis of interactive reporting through XBRL and its use as a support in the process of
taking decisions in corporate governance, namely the potential of interactive reports in XBRL to increase the transparency and
monitoring process of corporate governance.
Abstract: Data mining techniques have been used in medical
research for many years and have been known to be effective. In order
to solve such problems as long-waiting time, congestion, and delayed
patient care, faced by emergency departments, this study concentrates
on building a hybrid methodology, combining data mining techniques
such as association rules and classification trees. The methodology is
applied to real-world emergency data collected from a hospital and is
evaluated by comparing with other techniques. The methodology is
expected to help physicians to make a faster and more accurate
classification of chest pain diseases.
Abstract: Recent changes in food retailing structure have led to the development of large supercenters in suburban areas of the United States. These changes have led some authors to suggest that there are food deserts in some urban areas, where food is difficult to access, especially for disadvantaged consumers. This study tests the food desert hypothesis by comparing the distance from food retailers to food secure and food insecure households in one urban, Midwest neighborhood. This study utilizes GIS to compare household survey respondent locations against the location of various types of area food retailers. Results of this study indicate no apparent difference between food secure and insecure households in the reported importance of distance on the decision to shop at various retailers. However, there were differences in the spatial relationship between households and retailers. Food insecure households tended to be located slightly farther from large food retailers and slightly closer to convenience stores. Furthermore, food insecure households reported traveling slightly farther to their primary food retailer. The differences between the two groups was, however, relatively small.
Abstract: Manufacturing companies are facing a broad variety
of challenges caused by a dynamic production environment. To
succeed in such an environment, it is crucial to minimize the loss of
time required to trigger the adaptation process of a company-s
production structures. This paper presents an approach for the
continuous monitoring of production structures by neurologic
principles. It enhances classical monitoring concepts, which are
principally focused on reactive strategies, and enables companies to
act proactively. Thereby, strategic aspects regarding the
harmonization of certain life cycles are integrated into the decision
making process for triggering the reconfiguration process of the
production structure.
Abstract: Binary Decision Diagrams (BDDs) are useful data
structures for symbolic Boolean manipulations. BDDs are used in
many tasks in VLSI/CAD, such as equivalence checking, property
checking, logic synthesis, and false paths. In this paper we describe a
new approach for the realization of a BDD package. To perform
manipulations of Boolean functions, the proposed approach does not
depend on the recursive synthesis operation of the IF-Then-Else
(ITE). Instead of using the ITE operation, the basic synthesis
algorithm is done using Boolean NOR operation.
Abstract: Khatunabad area is situated geologically in Urmieh-
Dokhtar magmatic belt in NW of Iran. In this research, studied area
has been investigated in order to recognize the potential copper and
molybdenum-bearing target areas. The survey layers include the
lithologic units, alteration, geochemical result, tectonics and copper
and molybdenum occurrence. As an accurate decision can have a
considerable effect on exploration plans, so in this efforts have been
made to make use of new combination method. For this purpose, the
analytical hierarchy process was used and revealed highly potential
copper and molybdenum mineralization areas. Based on achieved
results, geological perspective in north of studied area is appropriate
for advance stage, especially for subsurface exploration in future.
Abstract: Smoke discharging is a main reason of air pollution
problem from industrial plants. The obstacle of a building has an
affect with the air pollutant discharge. In this research, a mathematical
model of the smoke dispersion from two sources and one source with
a structural obstacle is considered. The governing equation of the
model is an isothermal mass transfer model in a viscous fluid. The
finite element method is used to approximate the solutions of the
model. The triangular linear elements have been used for discretising
the domain, and time integration has been carried out by semi-implicit
finite difference method. The simulations of smoke dispersion in
cases of one chimney and two chimneys are presented. The maximum
calculated smoke concentration of both cases are compared. It is then
used to make the decision for smoke discharging and air pollutant
control problems on industrial area.
Abstract: This paper discusses site selection process for
biological soil conservation planning. It was supported by a valuefocused
approach and spatial multi-criteria evaluation techniques. A
first set of spatial criteria was used to design a number of potential
sites. Next, a new set of spatial and non-spatial criteria was
employed, including the natural factors and the financial costs,
together with the degree of suitability for the Bonkuh watershed to
biological soil conservation planning and to recommend the most
acceptable program. The whole process was facilitated by a new
software tool that supports spatial multiple criteria evaluation, or
SMCE in GIS software (ILWIS). The application of this tool,
combined with a continual feedback by the public attentions, has
provided an effective methodology to solve complex decisional
problem in biological soil conservation planning.
Abstract: The use of a Geographic Information System (GIS) in
roadway lighting to show the state of street-lighting and nighttime
accident is demonstrated. Geographical maps were generated
showing colored streets based on how much of the street's length is
illuminated. The night to daytime accidents ratio at intersections
were found along with the state of lighting at those intersections.
The result is a method to show the state of street-lighting at roads and
intersections and a quick guide for decision makers to implement
strategies for better street-lighting to reduce night time traffic
accidents in a particular district.