Abstract: This paper provides an in-depth study of Wireless
Sensor Network (WSN) application to monitor and control the
swiftlet habitat. A set of system design is designed and developed
that includes the hardware design of the nodes, Graphical User
Interface (GUI) software, sensor network, and interconnectivity for
remote data access and management. System architecture is proposed
to address the requirements for habitat monitoring. Such applicationdriven
design provides and identify important areas of further work
in data sampling, communications and networking. For this
monitoring system, a sensor node (MTS400), IRIS and Micaz radio
transceivers, and a USB interfaced gateway base station of Crossbow
(Xbow) Technology WSN are employed. The GUI of this monitoring
system is written using a Laboratory Virtual Instrumentation
Engineering Workbench (LabVIEW) along with Xbow Technology
drivers provided by National Instrument. As a result, this monitoring
system is capable of collecting data and presents it in both tables and
waveform charts for further analysis. This system is also able to send
notification message by email provided Internet connectivity is
available whenever changes on habitat at remote sites (swiftlet farms)
occur. Other functions that have been implemented in this system
are the database system for record and management purposes; remote
access through the internet using LogMeIn software. Finally, this
research draws a conclusion that a WSN for monitoring swiftlet
habitat can be effectively used to monitor and manage swiftlet
farming industry in Sarawak.
Abstract: In the recent past Learning Classifier Systems have
been successfully used for data mining. Learning Classifier System
(LCS) is basically a machine learning technique which combines
evolutionary computing, reinforcement learning, supervised or
unsupervised learning and heuristics to produce adaptive systems. A
LCS learns by interacting with an environment from which it
receives feedback in the form of numerical reward. Learning is
achieved by trying to maximize the amount of reward received. All
LCSs models more or less, comprise four main components; a finite
population of condition–action rules, called classifiers; the
performance component, which governs the interaction with the
environment; the credit assignment component, which distributes the
reward received from the environment to the classifiers accountable
for the rewards obtained; the discovery component, which is
responsible for discovering better rules and improving existing ones
through a genetic algorithm. The concatenate of the production rules
in the LCS form the genotype, and therefore the GA should operate
on a population of classifier systems. This approach is known as the
'Pittsburgh' Classifier Systems. Other LCS that perform their GA at
the rule level within a population are known as 'Mitchigan' Classifier
Systems. The most predominant representation of the discovered
knowledge is the standard production rules (PRs) in the form of IF P
THEN D. The PRs, however, are unable to handle exceptions and do
not exhibit variable precision. The Censored Production Rules
(CPRs), an extension of PRs, were proposed by Michalski and
Winston that exhibit variable precision and supports an efficient
mechanism for handling exceptions. A CPR is an augmented
production rule of the form: IF P THEN D UNLESS C, where
Censor C is an exception to the rule. Such rules are employed in
situations, in which conditional statement IF P THEN D holds
frequently and the assertion C holds rarely. By using a rule of this
type we are free to ignore the exception conditions, when the
resources needed to establish its presence are tight or there is simply
no information available as to whether it holds or not. Thus, the IF P
THEN D part of CPR expresses important information, while the
UNLESS C part acts only as a switch and changes the polarity of D
to ~D. In this paper Pittsburgh style LCSs approach is used for
automated discovery of CPRs. An appropriate encoding scheme is
suggested to represent a chromosome consisting of fixed size set of
CPRs. Suitable genetic operators are designed for the set of CPRs
and individual CPRs and also appropriate fitness function is proposed
that incorporates basic constraints on CPR. Experimental results are
presented to demonstrate the performance of the proposed learning
classifier system.
Abstract: Link reliability and transmitted power are two important design constraints in wireless network design. Error control coding (ECC) is a classic approach used to increase link reliability and to lower the required transmitted power. It provides coding gain, resulting in transmitter energy savings at the cost of added decoder power consumption. But the choice of ECC is very critical in the case of wireless sensor network (WSN). Since the WSNs are energy constraint in nature, both the BER and power consumption has to be taken into count. This paper develops a step by step approach in finding suitable error control codes for WSNs. Several simulations are taken considering different error control codes and the result shows that the RS(31,21) fits both in BER and power consumption criteria.
Abstract: A Comparison and evaluation of the different
condition monitoring (CM) techniques was applied experimentally
on RC e.g. Dynamic cylinder pressure and crankshaft Instantaneous
Angular Speed (IAS), for the detection and diagnosis of valve faults
in a two - stage reciprocating compressor for a programme of
condition monitoring which can successfully detect and diagnose a
fault in machine. Leakage in the valve plate was introduced
experimentally into a two-stage reciprocating compressor. The effect
of the faults on compressor performance was monitored and the
differences with the normal, healthy performance noted as a fault
signature been used for the detection and diagnosis of faults.
The paper concludes with what is considered to be a unique
approach to condition monitoring. First, each of the two most useful
techniques is used to produce a Truth Table which details the
circumstances in which each method can be used to detect and
diagnose a fault. The two Truth Tables are then combined into a
single Decision Table to provide a unique and reliable method of
detection and diagnosis of each of the individual faults introduced
into the compressor. This gives accurate diagnosis of compressor
faults.
Abstract: We aimed to investigate how can target and optimize
pulmonary delivery distribution by changing physicochemical
characteristics of instilled liquid.Therefore, we created a new liquids
group:
a. eligible for desired distribution within lung because of
assorted physicochemical characteristics
b. capable of being augmented with a broad range of
chemicals inertly
c. no interference on respiratory function
d. compatible with airway surface liquid
We developed forty types of new liquid,were composed of
Carboxymethylcellulose sodium,Glycerin and different types of
Polysorbates.Viscosity was measured using a Programmable
Rheometer and surface tension by KRUSS Tensiometer.We
subsequently examined the liquids and delivery protocols by simple
and branched glass capillary tube models of airways.Eventually,we
explored pulmonary distribution of liquids being augmented with
technetium-99m in mechanically ventilated rabbits.We used a single
head large field of view gamma camera.Kinematic viscosity between
0.265Stokes and 0.289Stokes,density between 1g/cm3 and 1.5g/cm3
and surface tension between 25dyn/cm and 35dyn/cm were the most
acceptable.
Abstract: The temporal nature of negative selection is an under exploited area. In a negative selection system, newly generated antibodies go through a maturing phase, and the survivors of the phase then wait to be activated by the incoming antigens after certain number of matches. These without having enough matches will age and die, while these with enough matches (i.e., being activated) will become active detectors. A currently active detector may also age and die if it cannot find any match in a pre-defined (lengthy) period of time. Therefore, what matters in a negative selection system is the dynamics of the involved parties in the current time window, not the whole time duration, which may be up to eternity. This property has the potential to define the uniqueness of negative selection in comparison with the other approaches. On the other hand, a negative selection system is only trained with “normal" data samples. It has to learn and discover unknown “abnormal" data patterns on the fly by itself. Consequently, it is more appreciate to utilize negation selection as a system for pattern discovery and recognition rather than just pattern recognition. In this paper, we study the potential of using negative selection in discovering unknown temporal patterns.
Abstract: Let R be a ring and n a fixed positive integer, we
investigate the properties of n-strongly Gorenstein projective, injective
and flat modules. Using the homological theory , we prove that
the tensor product of an n-strongly Gorenstein projective (flat) right
R -module and projective (flat) left R-module is also n-strongly
Gorenstein projective (flat). Let R be a coherent ring ,we prove that
the character module of an n -strongly Gorenstein flat left R -module
is an n-strongly Gorenstein injective right R -module . At last, let
R be a commutative ring and S a multiplicatively closed set of R ,
we establish the relation between n -strongly Gorenstein projective
(injective , flat ) R -modules and n-strongly Gorenstein projective
(injective , flat ) S−1R-modules. All conclusions in this paper is
helpful for the research of Gorenstein dimensions in future.
Abstract: Mining sequential patterns from large customer transaction databases has been recognized as a key research topic in database systems. However, the previous works more focused on mining sequential patterns at a single concept level. In this study, we introduced concept hierarchies into this problem and present several algorithms for discovering multiple-level sequential patterns based on the hierarchies. An experiment was conducted to assess the performance of the proposed algorithms. The performances of the algorithms were measured by the relative time spent on completing the mining tasks on two different datasets. The experimental results showed that the performance depends on the characteristics of the datasets and the pre-defined threshold of minimal support for each level of the concept hierarchy. Based on the experimental results, some suggestions were also given for how to select appropriate algorithm for a certain datasets.
Abstract: Sliding mode control with a fuzzy boundary layer is presented to hydraulic position control problem in this paper. A nonlinear hydraulic servomechanism which has an asymmetric cylinder is modeled and simulated first, then the proposed control scheme is applied to this model versus the conventional sliding mode control. Simulation results proved that the chattering free position control is achieved by tuning the fuzzy scaling factors properly.
Abstract: In recent years, scanning probe atomic force
microscopy SPM AFM has gained acceptance over a wide spectrum
of research and science applications. Most fields focuses on physical,
chemical, biological while less attention is devoted to manufacturing
and machining aspects. The purpose of the current study is to assess
the possible implementation of the SPM AFM features and its
NanoScope software in general machining applications with special
attention to the tribological aspects of cutting tool. The surface
morphology of coated and uncoated as-received carbide inserts is
examined, analyzed, and characterized through the determination of
the appropriate scanning setting, the suitable data type imaging
techniques and the most representative data analysis parameters
using the MultiMode SPM AFM in contact mode. The NanoScope
operating software is used to capture realtime three data types
images: “Height", “Deflection" and “Friction". Three scan sizes are
independently performed: 2, 6, and 12 μm with a 2.5 μm vertical
range (Z). Offline mode analysis includes the determination of three
functional topographical parameters: surface “Roughness", power
spectral density “PSD" and “Section". The 12 μm scan size in
association with “Height" imaging is found efficient to capture every
tiny features and tribological aspects of the examined surface. Also,
“Friction" analysis is found to produce a comprehensive explanation
about the lateral characteristics of the scanned surface. Configuration
of many surface defects and drawbacks has been precisely detected
and analyzed.
Abstract: Trust management and Reputation models are
becoming integral part of Internet based applications such as CSCW,
E-commerce and Grid Computing. Also the trust dimension is a
significant social structure and key to social relations within a
collaborative community. Collaborative Decision Making (CDM) is
a difficult task in the context of distributed environment (information
across different geographical locations) and multidisciplinary
decisions are involved such as Virtual Organization (VO). To aid
team decision making in VO, Decision Support System and social
network analysis approaches are integrated. In such situations social
learning helps an organization in terms of relationship, team
formation, partner selection etc. In this paper we focus on trust
learning. Trust learning is an important activity in terms of
information exchange, negotiation, collaboration and trust
assessment for cooperation among virtual team members. In this
paper we have proposed a reinforcement learning which enhances the
trust decision making capability of interacting agents during
collaboration in problem solving activity. Trust computational model
with learning that we present is adapted for best alternate selection of
new project in the organization. We verify our model in a multi-agent
simulation where the agents in the community learn to identify
trustworthy members, inconsistent behavior and conflicting behavior
of agents.
Abstract: A simple mobile engine-driven pneumatic paddy
collector made of locally available materials using local
manufacturing technology was designed, fabricated, and tested for
collecting and bagging of paddy dried on concrete pavement. The
pneumatic paddy collector had the following major components:
radial flat bladed type centrifugal fan, power transmission system,
bagging area, frame and the conveyance system. Results showed
significant differences on the collecting capacity, noise level, and fuel
consumption when rotational speed of the air mover shaft was varied.
Other parameters such as collecting efficiency, air velocity,
augmented cracked grain percentage, and germination rate were not
significantly affected by varying rotational speed of the air mover
shaft. The pneumatic paddy collector had a collecting efficiency of
99.33 % with a collecting capacity of 2685.00 kg/h at maximum
rotational speed of centrifugal fan shaft of about 4200 rpm. The
machine entailed an investment cost of P 62,829.25. The break-even
weight of paddy was 510,606.75 kg/yr at a collecting cost of 0.11
P/kg of paddy. Utilizing the machine for 400 hours per year
generated an income of P 23,887.73. The projected time needed to
recover cost of the machine based on 2685 kg/h collecting capacity
was 2.63 year.
Abstract: A traffic light gives security from traffic congestion,reducing the traffic jam, and organizing the traffic flow. Furthermore,increasing congestion level in public road networks is a growingproblem in many countries. Using Intelligent Transportation Systemsto provide emergency vehicles a green light at intersections canreduce driver confusion, reduce conflicts, and improve emergencyresponse times. Nowadays, the technology of wireless sensornetworks can solve many problems and can offer a good managementof the crossroad. In this paper, we develop a new approach based onthe technique of clustering and the graphical possibilistic fusionmodeling. So, the proposed model is elaborated in three phases. Thefirst one consists to decompose the environment into clusters,following by the fusion intra and inter clusters processes. Finally, wewill show some experimental results by simulation that proves theefficiency of our proposed approach.KeywordsTraffic light, Wireless sensor network, Controller,Possibilistic network/Bayesain network.
Abstract: IETF defines mobility support in IPv6, i.e. MIPv6, to
allow nodes to remain reachable while moving around in the IPv6
internet. When a node moves and visits a foreign network, it is still
reachable through the indirect packet forwarding from its home
network. This triangular routing feature provides node mobility but
increases the communication latency between nodes. This deficiency
can be overcome by using a Binding Update (BU) scheme, which let
nodes keep up-to-date IP addresses and communicate with each other
through direct IP routing. To further protect the security of BU, a
Return Routability (RR) procedure was developed. However, it has
been found that RR procedure is vulnerable to many attacks. In this
paper, we will propose a lightweight RR procedure based on
geometric computing. In consideration of the inherent limitation of
computing resources in mobile node, the proposed scheme is
developed to minimize the cost of computations and to eliminate the
overhead of state maintenance during binding updates. Compared with
other CGA-based BU schemes, our scheme is more efficient and
doesn-t need nonce tables in nodes.
Abstract: There are several approaches for handling multiclass classification. Aside from one-against-one (OAO) and one-against-all (OAA), hierarchical classification technique is also commonly used. A binary classification tree is a hierarchical classification structure that breaks down a k-class problem into binary sub-problems, each solved by a binary classifier. In each node, a set of classes is divided into two subsets. A good class partition should be able to group similar classes together. Many algorithms measure similarity in term of distance between class centroids. Classes are grouped together by a clustering algorithm when distances between their centroids are small. In this paper, we present a binary classification tree with tuned observation-based clustering (BCT-TOB) that finds a class partition by performing clustering on observations instead of class centroids. A merging step is introduced to merge any insignificant class split. The experiment shows that performance of BCT-TOB is comparable to other algorithms.
Abstract: For stable bipedal gait generation on the level floor,
efficient restoring of mechanical energy lost by heel collision at
the ground is necessary. Parametric excitation principle is one of
the solutions. We dealt with the robot-s total center of mass as
an inverted pendulum to consider the total dynamics of the robot.
Parametrically excited walking requires the use of continuous target
trajectory that is close to discontinuous optimal trajectory. In this
paper, we proposed the new target trajectory based on a position
in the walking direction. We surveyed relations between walking
performance and the parameters that form the target trajectory via
numerical simulations. As a result, it was found that our target
trajectory has the similar characteristics of a parametrically excited
inverted pendulum.
Abstract: Information sharing and exchange, rather than
information processing, is what characterizes information
technology in the 21st century. Ontologies, as shared common
understanding, gain increasing attention, as they appear as the
most promising solution to enable information sharing both at
a semantic level and in a machine-processable way. Domain
Ontology-based modeling has been exploited to provide
shareability and information exchange among diversified,
heterogeneous applications of enterprises.
Contextual ontologies are “an explicit specification of
contextual conceptualization". That is: ontology is
characterized by concepts that have multiple representations
and they may exist in several contexts. Hence, contextual
ontologies are a set of concepts and relationships, which are
seen from different perspectives. Contextualization is to allow
for ontologies to be partitioned according to their contexts.
The need for contextual ontologies in enterprise modeling
has become crucial due to the nature of today's competitive
market. Information resources in enterprise is distributed and
diversified and is in need to be shared and communicated
locally through the intranet and globally though the internet.
This paper discusses the roles that ontologies play in an
enterprise modeling, and how ontologies assist in building a
conceptual model in order to provide communicative and
interoperable information systems. The issue of enterprise
modeling based on contextual domain ontology is also
investigated, and a framework is proposed for an enterprise
model that consists of various applications.
Abstract: This study examines the possibility to apply the theory of multidimensional accounting (momentum accounting) in a Brazilian Navy-s Services Provider Military Organization (Organização Militar Prestadora de Serviços - OMPS). In general, the core of the said theory is the fact that Accounting does not recognize the inertia of transactions occurring in an entity, and that occur repeatedly in some cases, regardless of the implementation of new actions by its managers. The study evaluates the possibility of greater use of information recorded in the financial statements of the unit of analysis, within the strategic decisions of the organization. As a research strategy, we adopted the case study. The results infer that it is possible to use the theory in the context of a multidimensional OMPS, promoting useful information for decision-making and thereby contributing to the strengthening of the necessary alignment of its administration with the current desires of the Brazilian society.
Abstract: In this paper, the modified optimal sliding mode control with a proposed method to design a sliding surface is presented. Because of the inability of the previous approach of the sliding mode method to design a bounded and suitable input, the new variation is proposed in the sliding manifold to obviate problems in a structural system. Although the sliding mode control is a powerful method to reject disturbances and noises, the chattering problem is not good for actuators. To decrease the chattering phenomena, the optimal control is added to the sliding mode control. Not only the proposed method can decline the intense variations in the inputs of the system but also it can produce the efficient responses respect to the sliding mode control and optimal control that are shown by performing some numerical simulations.
Abstract: High quality requirements analysis is one of the most
crucial activities to ensure the success of a software project, so that
requirements verification for software system becomes more and more
important in Requirements Engineering (RE) and it is one of the most
helpful strategies for improving the quality of software system.
Related works show that requirement elicitation and analysis can be
facilitated by ontological approaches and semantic web technologies.
In this paper, we proposed a hybrid method which aims to verify
requirements with structural and formal semantics to detect
interactions. The proposed method is twofold: one is for modeling
requirements with the semantic web language OWL, to construct a
semantic context; the other is a set of interaction detection rules which
are derived from scenario-based analysis and represented with
semantic web rule language (SWRL). SWRL based rules are working
with rule engines like Jess to reason in semantic context for
requirements thus to detect interactions. The benefits of the proposed
method lie in three aspects: the method (i) provides systematic steps
for modeling requirements with an ontological approach, (ii) offers
synergy of requirements elicitation and domain engineering for
knowledge sharing, and (3)the proposed rules can systematically assist
in requirements interaction detection.