Abstract: In this paper, we investigated the characteristic of a
clinical dataseton the feature selection and classification
measurements which deal with missing values problem.And also
posed the appropriated techniques to achieve the aim of the activity;
in this research aims to find features that have high effect to mortality
and mortality time frame. We quantify the complexity of a clinical
dataset. According to the complexity of the dataset, we proposed the
data mining processto cope their complexity; missing values, high
dimensionality, and the prediction problem by using the methods of
missing value replacement, feature selection, and classification.The
experimental results will extend to develop the prediction model for
cardiology.
Abstract: Software Reusability is primary attribute of software
quality. There are metrics for identifying the quality of reusable
components but the function that makes use of these metrics to find
reusability of software components is still not clear. These metrics if
identified in the design phase or even in the coding phase can help us
to reduce the rework by improving quality of reuse of the component
and hence improve the productivity due to probabilistic increase in
the reuse level. In this paper, we have devised the framework of
metrics that uses McCabe-s Cyclometric Complexity Measure for
Complexity measurement, Regularity Metric, Halstead Software
Science Indicator for Volume indication, Reuse Frequency metric
and Coupling Metric values of the software component as input
attributes and calculated reusability of the software component. Here,
comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA
approaches is performed to evaluate the reusability of software
components and Fuzzy-GA results outperform the other used
approaches. The developed reusability model has produced high
precision results as expected by the human experts.
Abstract: With a rapid growth in 3D graphics technology over the last few years, people are desired to see more flexible reacting motions of a biped in animations. In particular, it is impossible to anticipate all reacting motions of a biped while facing a perturbation. In this paper, we propose a three-level tracking method for animating a 3D humanoid character. First, we take the laws of physics into account to attach physical attributes, such as mass, gravity, friction, collision, contact, and torque, to bones and joints of a character. The next step is to employ PD controller to follow a reference motion as closely as possible. Once the character cannot tolerate a strong perturbation to prevent itself from falling down, we are capable of tracking a desirable falling-down action to avoid any falling condition inaccuracy. From the experimental results, we demonstrate the effectiveness and flexibility of the proposed method in comparison with conventional data-driven approaches.
Abstract: This study aims to assess the potential of solar energy technology for improving access to water and hence the livelihood strategies of rural communities in Baja California Sur, Mexico. It focuses on livestock ranches and photovoltaic water-pumptechnology as well as other water extraction methods. The methodology used are the Sustainable Livelihoods and the Appropriate Technology approaches. A household survey was applied in June of 2006 to 32 ranches in the municipality, of which 22 used PV pumps; and semi-structured interviews were conducted. Findings indicate that solar pumps have in fact helped people improve their quality of life by allowing them to pursue a different livelihood strategy and that improved access to water -not necessarily as more water but as less effort to extract and collect it- does not automatically imply overexploitation of the resource; consumption is based on basic needs as well as on storage and pumping capacity. Justification for such systems lies in the avoidance of logistical problems associated to fossil fuels, PV pumps proved to be the most beneficial when substituting gasoline or diesel equipment but of dubious advantage if intended to replace wind or gravity systems. Solar water pumping technology-s main obstacle to dissemination are high investment and repairs costs and it is therefore not suitable for all cases even when insolation rates and water availability are adequate. In cases where affordability is not an obstacle it has become an important asset that contributes –by means of reduced expenses, less effort and saved time- to the improvement of livestock, the main livelihood provider for these ranches.
Abstract: Web-based systems have become increasingly
important due to the fact that the Internet and the World Wide Web
have become ubiquitous, surpassing all other technological
developments in our history. The Internet and especially companies
websites has rapidly evolved in their scope and extent of use, from
being a little more than fixed advertising material, i.e. a "web
presences", which had no particular influence for the company's
business, to being one of the most essential parts of the company's
core business.
Traditional software engineering approaches with process models
such as, for example, CMM and Waterfall models, do not work very
well since web system development differs from traditional
development. The development differs in several ways, for example,
there is a large gap between traditional software engineering designs
and concepts and the low-level implementation model, many of the
web based system development activities are business oriented (for
example web application are sales-oriented, web application and
intranets are content-oriented) and not engineering-oriented.
This paper aims to introduce Increment Iterative extreme
Programming (IIXP) methodology for developing web based
systems. In difference to the other existence methodologies, this
methodology is combination of different traditional and modern
software engineering and web engineering principles.
Abstract: This paper proposes the use of metrics in design space exploration that highlight where in the structure of the model and at what point in the behaviour, prevention is needed against transient faults. Previous approaches to tackle transient faults focused on recovery after detection. Almost no research has been directed towards preventive measures. But in real-time systems, hard deadlines are performance requirements that absolutely must be met and a missed deadline constitutes an erroneous action and a possible system failure. This paper proposes the use of metrics to assess the system design to flag where transient faults may have significant impact. These tools then allow the design to be changed to minimize that impact, and they also flag where particular design techniques – such as coding of communications or memories – need to be applied in later stages of design.
Abstract: This paper presents an optimal and unsupervised satellite image segmentation approach based on Pearson system and k-Means Clustering Algorithm Initialization. Such method could be considered as original by the fact that it utilised K-Means clustering algorithm for an optimal initialisation of image class number on one hand and it exploited Pearson system for an optimal statistical distributions- affectation of each considered class on the other hand. Satellite image exploitation requires the use of different approaches, especially those founded on the unsupervised statistical segmentation principle. Such approaches necessitate definition of several parameters like image class number, class variables- estimation and generalised mixture distributions. Use of statistical images- attributes assured convincing and promoting results under the condition of having an optimal initialisation step with appropriated statistical distributions- affectation. Pearson system associated with a k-means clustering algorithm and Stochastic Expectation-Maximization 'SEM' algorithm could be adapted to such problem. For each image-s class, Pearson system attributes one distribution type according to different parameters and especially the Skewness 'β1' and the kurtosis 'β2'. The different adapted algorithms, K-Means clustering algorithm, SEM algorithm and Pearson system algorithm, are then applied to satellite image segmentation problem. Efficiency of those combined algorithms was firstly validated with the Mean Quadratic Error 'MQE' evaluation, and secondly with visual inspection along several comparisons of these unsupervised images- segmentation.
Abstract: This paper summarizes and compares approaches to
solving the knapsack problem and its known application in capital
budgeting. The first approach uses deterministic methods and can be
applied to small-size tasks with a single constraint. We can also
apply commercial software systems such as the GAMS modelling
system. However, because of NP-completeness of the problem, more
complex problem instances must be solved by means of heuristic
techniques to achieve an approximation of the exact solution in a
reasonable amount of time. We show the problem representation and
parameter settings for a genetic algorithm framework.
Abstract: This paper regards the phenomena of intensive suburbanization and urbanization in Olomouc city and in Olomouc region in general for the period of 1986–2009. A Remote Sensing approach that involves tracking of changes in Land Cover units is proposed to quantify the urbanization state and trends in temporal and spatial aspects. It actually consisted of two approaches, Experiment 1 and Experiment 2 which implied two different image classification solutions in order to provide Land Cover maps for each 1986–2009 time split available in the Landsat image set. Experiment 1 dealt with the unsupervised classification, while Experiment 2 involved semi- supervised classification, using a combination of object-based and pixel-based classifiers. The resulting Land Cover maps were subsequently quantified for the proportion of urban area unit and its trend through time, and also for the urban area unit stability, yielding the relation of spatial and temporal development of the urban area unit. Some outcomes seem promising but there is indisputably room for improvements of source data and also processing and filtering.
Abstract: Higher education has an important role to play in
advocating environmentalism. Given this responsibility, the goal of
higher education should therefore be to develop graduates with the
knowledge, skills and values related to environmentalism. However,
research indicates that there is a lack of consciousness amongst
graduates on the need to be more environmentally aware, especially
when it comes to applying the appropriate knowledge and skills
related to environmentalism. Although institutions of higher learning
do include environmental parameters within their undergraduate and
postgraduate academic programme structures, the environmental
boundaries are usually confined to specific engineering majors within
an engineering programme. This makes environmental knowledge,
skills and values exclusive to certain quarters of the higher education
system. The incorporation of environmental literacy within higher
education institutions as a whole is of utmost pertinence if a nation-s
human capital is to be nurtured to become change agents for the
preservation of environment. This paper discusses approaches that
can be adapted by institutions of higher learning to include
environmental literacy within the graduate-s higher learning
experience.
Abstract: Motion capture devices have been utilized in
producing several contents, such as movies and video games. However,
since motion capture devices are expensive and inconvenient to use,
motions segmented from captured data was recycled and synthesized
to utilize it in another contents, but the motions were generally
segmented by contents producers in manual. Therefore, automatic
motion segmentation is recently getting a lot of attentions. Previous
approaches are divided into on-line and off-line, where on-line
approaches segment motions based on similarities between
neighboring frames and off-line approaches segment motions by
capturing the global characteristics in feature space. In this paper, we
propose a graph-based high-level motion segmentation method. Since
high-level motions consist of several repeated frames within temporal
distances, we consider all similarities among all frames within the
temporal distance. This is achieved by constructing a graph, where
each vertex represents a frame and the edges between the frames are
weighted by their similarity. Then, normalized cuts algorithm is used
to partition the constructed graph into several sub-graphs by globally
finding minimum cuts. In the experiments, the results using the
proposed method showed better performance than PCA-based method
in on-line and GMM-based method in off-line, as the proposed method
globally segment motions from the graph constructed based
similarities between neighboring frames as well as similarities among
all frames within temporal distances.
Abstract: In this paper, we present a matrix game-theoretic cross-layer optimization formulation to maximize the network lifetime in wireless ad hoc networks with network coding. To this end, we introduce a cross-layer formulation of general NUM (network utility maximization) that accommodates routing, scheduling, and stream control from different layers in the coded networks. Specifically, for the scheduling problem and then the objective function involved, we develop a matrix game with the strategy sets of the players corresponding to hyperlinks and transmission modes, and design the payoffs specific to the lifetime. In particular, with the inherit merit that matrix game can be solved with linear programming, our cross-layer programming formulation can benefit from both game-based and NUM-based approaches at the same time by cooperating the programming model for the matrix game with that for the other layers in a consistent framework. Finally, our numerical example demonstrates its performance results on a well-known wireless butterfly network to verify the cross-layer optimization scheme.
Abstract: FAQ system can make user find answer to the problem that puzzles them. But now the research on Chinese FAQ system is still on the theoretical stage. This paper presents an approach to semantic inference for FAQ mining. To enhance the efficiency, a small pool of the candidate question-answering pairs retrieved from the system for the follow-up work according to the concept of the agriculture domain extracted from user input .Input queries or questions are converted into four parts, the question word segment (QWS), the verb segment (VS), the concept of agricultural areas segment (CS), the auxiliary segment (AS). A semantic matching method is presented to estimate the similarity between the semantic segments of the query and the questions in the pool of the candidate. A thesaurus constructed from the HowNet, a Chinese knowledge base, is adopted for word similarity measure in the matcher. The questions are classified into eleven intension categories using predefined question stemming keywords. For FAQ mining, given a query, the question part and answer part in an FAQ question-answer pair is matched with the input query, respectively. Finally, the probabilities estimated from these two parts are integrated and used to choose the most likely answer for the input query. These approaches are experimented on an agriculture FAQ system. Experimental results indicate that the proposed approach outperformed the FAQ-Finder system in agriculture FAQ retrieval.
Abstract: The management of the health-care wastes is one of
the most important problems in Istanbul, a city with more than 12
million inhabitants, as it is in most of the developing countries.
Negligence in appropriate treatment and final disposal of the healthcare
wastes can lead to adverse impacts to public health and to the
environment. This paper employs a fuzzy multi-criteria group
decision making approach, which is based on the principles of fusion
of fuzzy information, 2-tuple linguistic representation model, and
technique for order preference by similarity to ideal solution
(TOPSIS), to evaluate health-care waste (HCW) treatment
alternatives for Istanbul. The evaluation criteria are determined
employing nominal group technique (NGT), which is a method of
systematically developing a consensus of group opinion. The
employed method is apt to manage information assessed using multigranularity
linguistic information in a decision making problem with
multiple information sources. The decision making framework
employs ordered weighted averaging (OWA) operator that
encompasses several operators as the aggregation operator since it
can implement different aggregation rules by changing the order
weights. The aggregation process is based on the unification of
information by means of fuzzy sets on a basic linguistic term set
(BLTS). Then, the unified information is transformed into linguistic
2-tuples in a way to rectify the problem of loss information of other
fuzzy linguistic approaches.
Abstract: Economic models are complex dynamic systems with a lot of uncertainties and fuzzy data. Conventional modeling approaches using well known methods and techniques cannot provide realistic and satisfactory answers to today-s challenging economic problems. Qualitative modeling using fuzzy logic and intelligent system theories can be used to model macroeconomic models. Fuzzy Cognitive maps (FCM) is a new method been used to model the dynamic behavior of complex systems. For the first time FCMs and the Mamdani Model of Intelligent control is used to model macroeconomic models. This new model is referred as the Mamdani Rule-Based Fuzzy Cognitive Map (MBFCM) and provides the academic and research community with a new promising integrated advanced computational model. A new economic model is developed for a qualitative approach to Macroeconomic modeling. Fuzzy Controllers for such models are designed. Simulation results for an economic scenario are provided and extensively discussed
Abstract: Natural language processing systems pose a unique
challenge for software architectural design as system complexity has
increased continually and systems cannot be easily constructed from
loosely coupled modules. Lexical, syntactic, semantic, and pragmatic
aspects of linguistic information are tightly coupled in a manner that
requires separation of concerns in a special way in design,
implementation and maintenance. An aspect oriented software
architecture is proposed in this paper after critically reviewing
relevant architectural issues. For the purpose of this paper, the
syntactic aspect is characterized by an augmented context-free
grammar. The semantic aspect is composed of multiple perspectives
including denotational, operational, axiomatic and case frame
approaches. Case frame semantics matured in India from deep
thematic analysis. It is argued that lexical, syntactic, semantic and
pragmatic aspects work together in a mutually dependent way and
their synergy is best represented in the aspect oriented approach. The
software architecture is presented with an augmented Unified
Modeling Language.
Abstract: In this paper we apply one of approaches in category of heuristic methods as Genetic Algorithms for obtaining approximate solution of optimal control problems. The firs we convert optimal control problem to a quasi Assignment Problem by defining some usual characters as defined in Genetic algorithm applications. Then we obtain approximate optimal control function as an piecewise constant function. Finally the numerical examples are given.
Abstract: Problem-based learning (PBL) is one of the student
centered approaches and has been considered by a number of higher
educational institutions in many parts of the world as a method of
delivery. This paper presents a creative thinking approach for
implementing Problem-based Learning in Mechanics of Structure
within a Malaysian Polytechnics environment. In the learning
process, students learn how to analyze the problem given among the
students and sharing classroom knowledge into practice. Further,
through this course-s emphasis on problem-based learning, students
acquire creative thinking skills and professional skills as they tackle
complex, interdisciplinary and real-situation problems. Once the
creative ideas are generated, there are useful additional techniques
for tender ideas that will grow into a productive concept or solution.
The combination of creative skills and technical abilities will enable
the students to be ready to “hit-the-ground-running" and produce in
industry when they graduate.
Abstract: Deoxyribonucleic Acid or DNA computing has
emerged as an interdisciplinary field that draws together chemistry,
molecular biology, computer science and mathematics. Thus, in this
paper, the possibility of DNA-based computing to solve an absolute
1-center problem by molecular manipulations is presented. This is
truly the first attempt to solve such a problem by DNA-based
computing approach. Since, part of the procedures involve with
shortest path computation, research works on DNA computing for
shortest path Traveling Salesman Problem, in short, TSP are reviewed.
These approaches are studied and only the appropriate one is adapted
in designing the computation procedures. This DNA-based
computation is designed in such a way that every path is encoded by
oligonucleotides and the path-s length is directly proportional to the
length of oligonucleotides. Using these properties, gel electrophoresis
is performed in order to separate the respective DNA molecules
according to their length. One expectation arise from this paper is that
it is possible to verify the instance absolute 1-center problem using
DNA computing by laboratory experiments.
Abstract: Recognition of Indian languages scripts is challenging problems. In Optical Character Recognition [OCR], a character or symbol to be recognized can be machine printed or handwritten characters/numerals. There are several approaches that deal with problem of recognition of numerals/character depending on the type of feature extracted and different way of extracting them. This paper proposes a recognition scheme for handwritten Hindi (devnagiri) numerals; most admired one in Indian subcontinent. Our work focused on a technique in feature extraction i.e. global based approach using end-points information, which is extracted from images of isolated numerals. These feature vectors are fed to neuro-memetic model [18] that has been trained to recognize a Hindi numeral. The archetype of system has been tested on varieties of image of numerals. . In proposed scheme data sets are fed to neuro-memetic algorithm, which identifies the rule with highest fitness value of nearly 100 % & template associates with this rule is nothing but identified numerals. Experimentation result shows that recognition rate is 92-97 % compared to other models.