Abstract: This paper presents a system for tracking the movement of laparoscopic instruments which is based on an orthogonal system of webcams and video image processing. The movements are captured with two webcams placed orthogonally inside of the physical trainer. On the image, the instruments were detected by using color markers placed on the distal tip of each instrument. The 3D position of the tip of the instrument within the work space was obtained by linear triangulation method. Preliminary results showed linearity and repeatability in the motion tracking with a resolution of 0.616 mm in each axis; the accuracy of the system showed a 3D instrument positioning error of 1.009 ± 0.101 mm. This tool is a portable and low-cost alternative to traditional tracking devices and a trustable method for the objective evaluation of the surgeon’s surgical skills.
Abstract: One of object oriented software developing problem
is the difficulty of searching the appropriate and suitable objects for
starting the system. In this work, ontologies appear in the part of
supporting the object discovering in the initial of object oriented
software developing. There are many researches try to demonstrate
that there is a great potential between object model and ontologies.
Constructing ontology from object model is called ontology
engineering can be done; On the other hand, this research is aiming to
support the idea of building object model from ontology is also
promising and practical. Ontology classes are available online in any
specific areas, which can be searched by semantic search engine.
There are also many helping tools to do so; one of them which are
used in this research is Protégé ontology editor and Visual Paradigm.
To put them together give a great outcome. This research will be
shown how it works efficiently with the real case study by using
ontology classes in travel/tourism domain area. It needs to combine
classes, properties, and relationships from more than two ontologies
in order to generate the object model. In this paper presents a simple
methodology framework which explains the process of discovering
objects. The results show that this framework has great value while
there is possible for expansion. Reusing of existing ontologies offers
a much cheaper alternative than building new ones from scratch.
More ontologies are becoming available on the web, and online
ontologies libraries for storing and indexing ontologies are increasing
in number and demand. Semantic and Ontologies search engines have
also started to appear, to facilitate search and retrieval of online
ontologies.
Abstract: Nowadays, many manufacturing companies try to
reinforce their competitiveness or find a breakthrough by considering
collaboration. In Korea, more than 900 manufacturing companies are
using web-based collaboration systems developed by the
government-led project, referred to as i-Manufacturing. The system
supports some similar functions of Product Data Management (PDM)
as well as Project Management System (PMS). A web-based
collaboration system provides many useful functions for collaborative
works. This system, however, does not support new linking services
between buyers and suppliers. Therefore, in order to find new
collaborative partners, this paper proposes a framework which creates
new connections between buyers and suppliers facilitating their
collaboration, referred to as Excellent Manufacturer Scouting System
(EMSS). EMSS plays a role as a bridge between overseas buyers and
suppliers. As a part of study on EMSS, we also propose an evaluation
method of manufacturability of potential partners with six main factors.
Based on the results of evaluation, buyers may get a good guideline to
choose their new partners before getting into negotiation processes
with them.
Abstract: Cooperative learning has been defined as learners working together as a team to solve a problem to complete a task or to accomplish a common goal, which emphasizes the importance of interactions among members to promote the whole learning performance. With the popularity of society networks, cooperative learning is no longer limited to traditional classroom teaching activities. Since society networks facilitate to organize online learners, to establish common shared visions, and to advance learning interaction, the online community and online learning community have triggered the establishment of web-based societies. Numerous research literatures have indicated that the collaborative learning community is a critical issue to enhance learning performance. Hence, this paper proposes a learning community recommendation approach to facilitate that a learner joins the appropriate learning communities, which is based on k-nearest neighbor (kNN) classification. To demonstrate the viability of the proposed approach, the proposed approach is implemented for 117 students to recommend learning communities. The experimental results indicate that the proposed approach can effectively recommend appropriate learning communities for learners.
Abstract: A number of mass spectrometry applications are already available as web-based and windows-based systems to calculate isotope pattern and to display the mass spectrum based on the specific molecular formula besides providing necessary information. These applications were evaluated and compared with our new alternative application called Theoretical Isotope Generator (TIG) in terms of its functionality and features provided to prove this new application is working better and performing well. TIG provides extra features than others, complete with several functionality such as drawing, normalizing and zooming the generated graph that convey with the molecular information in a number of formats by providing the details of the calculation and molecules. Thus, any chemist, students, lecturers and researchers from anywhere could use TIG to gain related information on molecules and their relative intensity.
Abstract: This paper studies, maps and explains the interactions between downloaders and uploaders pertaining to the Internet film piracy. This study also covers several motivational factors that influence users to upload or download movies, and thus to engage in film piracy over the Internet. The essay also proposes a model that describes user behavior including their relationships and influences. Moreover, proposed theoretical interactions and motivational factors are applied to the real world scenario, using examples of a data storage webpage server Ulozto.net and webpage Piratebay.org gathering information about downloadable BitTorrents. Moreover, the theory is further supported by description of behavior of real Internet uploaders.
Abstract: Knowledge sharing enables the information or
knowledge to be transmitted from one source to another. This paper
demonstrates the needs of having the online book catalogue which
can be used to facilitate disseminating information on textbook used
in the university. This project is aimed to give access to the students
and lecturers to the list of books in the bookstore and at the same
time to allow book reviewing without having to visit the bookstore
physically. Research is carried out according to the boundaries which
accounts to current process of new book purchasing, current system
used by the bookstore and current process the lecturers go through
for reviewing textbooks. The questionnaire is used to gather the
requirements and it is distributed to 100 students and 40 lecturers.
This project has enabled the improvement of a manual process to be
carried out automatically, through a web based platform. It is shown
based on the user acceptance survey carried out that target groups
found that this web service is feasible to be implemented in
Universiti Teknologi PETRONAS (UTP), and they have shown
positive signs of interest in utilizing it in the future.
Abstract: We present an implementation of an Online Exhibition System (OES) web service(s) that reflects our experiences with using web service development packages and software process models. The system provides major functionality that exists in similar packages. While developing such a complex web service, we gained insightful experience (i) in the traditional software development processes: waterfall model and evolutionary development and their fitness to web services development, (ii) in the fitness and effectiveness of a major web services development kit.
Abstract: Because of its global reach, reduction of time
restraints, and ability to reduce costs and increase sales, use of the
Internet, the World Wide Web (WWW), and related technologies
can be a competitive tool in the arsenal of small and medium-sized
enterprises (SMEs). Countries the world over are interested in the
successful adoption of the Internet by SMEs. Because a vast
majority of jobs come from that sector, greater financial success of
SMEs translates into greater job growth and, subsequently, higher
tax revenue to the government. This research investigated the level
of Internet usage for business solutions by small and medium
enterprises in Jordan. Through the survey of a random sample of
100 firms with less than 500 employees and from data obtained
from this survey that formed the basis for our study, we found that
a majority of respondents use the Internet in business activities ,
the adoption of the Internet as a business tool is limited to a
brochure where Web site which primarily provides one way. As
such, there wasn't interactive information about the company and
its products and services.
Abstract: The Tropical Data Hub (TDH) is a virtual research environment that provides researchers with an e-research infrastructure to congregate significant tropical data sets for data reuse, integration, searching, and correlation. However, researchers often require data and metadata synthesis across disciplines for crossdomain analyses and knowledge discovery. A triplestore offers a semantic layer to achieve a more intelligent method of search to support the synthesis requirements by automating latent linkages in the data and metadata. Presently, the benchmarks to aid the decision of which triplestore is best suited for use in an application environment like the TDH are limited to performance. This paper describes a new evaluation tool developed to analyze both features and performance. The tool comprises a weighted decision matrix to evaluate the interoperability, functionality, performance, and support availability of a range of integrated and native triplestores to rank them according to requirements of the TDH.
Abstract: A Finite Volume method based on Characteristic Fluxes for compressible fluids is developed. An explicit cell-centered resolution is adopted, where second and third order accuracy is provided by using two different MUSCL schemes with Minmod, Sweby or Superbee limiters for the hyperbolic part. Few different times integrator is used and be describe in this paper. Resolution is performed on a generic unstructured Cartesian grid, where solid boundaries are handled by a Cut-Cell method. Interfaces are explicitely advected in a non-diffusive way, ensuring local mass conservation. An improved cell cutting has been developed to handle boundaries of arbitrary geometrical complexity. Instead of using a polygon clipping algorithm, we use the Voxel traversal algorithm coupled with a local floodfill scanline to intersect 2D or 3D boundary surface meshes with the fixed Cartesian grid. Small cells stability problem near the boundaries is solved using a fully conservative merging method. Inflow and outflow conditions are also implemented in the model. The solver is validated on 2D academic test cases, such as the flow past a cylinder. The latter test cases are performed both in the frame of the body and in a fixed frame where the body is moving across the mesh. Adaptive Cartesian grid is provided by Paramesh without complex geometries for the moment.
Abstract: Data mining and knowledge engineering have become a tough task due to the availability of large amount of data in the web nowadays. Validity and reliability of data also become a main debate in knowledge acquisition. Besides, acquiring knowledge from different languages has become another concern. There are many language translators and corpora developed but the function of these translators and corpora are usually limited to certain languages and domains. Furthermore, search results from engines with traditional 'keyword' approach are no longer satisfying. More intelligent knowledge engineering agents are needed. To address to these problems, a system known as Multilingual Word Semantic Network is proposed. This system adapted semantic network to organize words according to concepts and relations. The system also uses open source as the development philosophy to enable the native language speakers and experts to contribute their knowledge to the system. The contributed words are then defined and linked using lexical and semantic relations. Thus, related words and derivatives can be identified and linked. From the outcome of the system implementation, it contributes to the development of semantic web and knowledge engineering.
Abstract: This paper has as its main aim to analyse how
corporate web pages can become an essential tool in order to detect
strategic trends by firms or sectors, and even a primary source for
benchmarking. This technique has made it possible to identify the key
issues in the strategic management of the most excellent large Spanish
firms and also to describe trends in their long-range planning, a way of
working that can be generalised to any country or firm group. More
precisely, two objectives were sought. The first one consisted in showing
the way in which corporate websites make it possible to obtain direct
information about the strategic variables which can define firms. This
tool is dynamic (since web pages are constantly updated) as well as
direct and reliable, since the information comes from the firm itself, not
from comments of third parties (such as journalists, academicians,
consultants...). When this information is analysed for a group of firms,
one can observe their characteristics in terms of both managerial tasks
and business management. As for the second objective, the methodology
proposed served to describe the corporate profile of the large Spanish
enterprises included in the Ibex35 (the Ibex35 or Iberia Index is the
reference index in the Spanish Stock Exchange and gathers periodically
the 35 most outstanding Spanish firms). An attempt is therefore made to
define the long-range planning that would be characteristic of the largest
Spanish firms.
Abstract: In recent years with the rapid development of Internet and the Web, more and more web applications have been deployed in many fields and organizations such as finance, military, and government. Together with that, hackers have found more subtle ways to attack web applications. According to international statistics, SQL Injection is one of the most popular vulnerabilities of web applications. The consequences of this type of attacks are quite dangerous, such as sensitive information could be stolen or authentication systems might be by-passed. To mitigate the situation, several techniques have been adopted. In this research, a security solution is proposed using Artificial Neural Network to protect web applications against this type of attacks. The solution has been experimented on sample datasets and has given promising result. The solution has also been developed in a prototypic web application firewall called ANNbWAF.
Abstract: A Web-services based grid infrastructure is evolving to be readily available in the near future. In this approach, the Web services are inherited (encapsulated or functioned) into the same existing Grid services class. In practice there is not much difference between the existing Web and grid infrastructure. Grid services emerged as stateful web services. In this paper, we present the key components of web-services based grid and also how the resource discovery is performed on web-services based grid considering resource discovery, as a critical service, to be provided by any type of grid.
Abstract: A virtual collaborative classroom was created at East Carolina University, using videoconference technology via regular internet to bring students from 18 different countries, 2 at a time, to the ECU classroom in real time to learn about each other-s culture. Students from two countries are partnered one on one, they meet for 4-5 weeks, and submit a joint paper. Then the same process is repeated for two other countries. Lectures and student discussions are managed with pre-determined topics and questions. Classes are conducted in English and reading assignments are placed on the website. Administratively all partners are independent, students pay fees and get credits at their home institution. Familiarity with technology, knowledge in cultural understanding and attitude change were assessed, only attitude changes are reported in this paper. After taking this course, all students stated their comfort level in working with, and their desire to interact with, culturally different others grew stronger and their xenophobia and isolationist attitudes decreased.
Abstract: Currently searching through internet is very popular especially in a field of academic. A huge of educational information such as research papers are overload for user. So community-base web sites have been developed to help user search information more easily from process of customizing a web site to need each specifies user or set of user. In this paper propose to use association rule analyze the community group on research paper bookmarking. A set of design goals for community group frameworks is developed and discussed. Additionally Researcher analyzes the initial relation by using association rule discovery between the antecedent and the consequent of a rule in the groups of user for generate the idea to improve ranking search result and development recommender system.
Abstract: In recent years, many researches to mine the exploding Web world, especially User Generated Content (UGC) such as
weblogs, for knowledge about various phenomena and events in the physical world have been done actively, and also Web services
with the Web-mined knowledge have begun to be developed for
the public. However, there are few detailed investigations on how accurately Web-mined data reflect physical-world data. It must be
problematic to idolatrously utilize the Web-mined data in public Web services without ensuring their accuracy sufficiently. Therefore,
this paper introduces the simplest Web Sensor and spatiotemporallynormalized
Web Sensor to extract spatiotemporal data about a target
phenomenon from weblogs searched by keyword(s) representing the
target phenomenon, and tries to validate the potential and reliability of the Web-sensed spatiotemporal data by four kinds of granularity
analyses of coefficient correlation with temperature, rainfall, snowfall,
and earthquake statistics per day by region of Japan Meteorological
Agency as physical-world data: spatial granularity (region-s population
density), temporal granularity (time period, e.g., per day vs. per week), representation granularity (e.g., “rain" vs. “heavy rain"), and
media granularity (weblogs vs. microblogs such as Tweets).
Abstract: Currently, web usage make a huge data from a lot of
user attention. In general, proxy server is a system to support web
usage from user and can manage system by using hit rates. This
research tries to improve hit rates in proxy system by applying data
mining technique. The data set are collected from proxy servers in the
university and are investigated relationship based on several features.
The model is used to predict the future access websites. Association
rule technique is applied to get the relation among Date, Time, Main
Group web, Sub Group web, and Domain name for created model.
The results showed that this technique can predict web content for the
next day, moreover the future accesses of websites increased from
38.15% to 85.57 %.
This model can predict web page access which tends to increase
the efficient of proxy servers as a result. In additional, the
performance of internet access will be improved and help to reduce
traffic in networks.
Abstract: EGOTHOR is a search engine that indexes the Web
and allows us to search the Web documents. Its hit list contains URL
and title of the hits, and also some snippet which tries to shortly
show a match. The snippet can be almost always assembled by an
algorithm that has a full knowledge of the original document (mostly
HTML page). It implies that the search engine is required to store
the full text of the documents as a part of the index.
Such a requirement leads us to pick up an appropriate compression
algorithm which would reduce the space demand. One of the solutions
could be to use common compression methods, for instance gzip or
bzip2, but it might be preferable if we develop a new method which
would take advantage of the document structure, or rather, the textual
character of the documents.
There already exist a special compression text algorithms and
methods for a compression of XML documents. The aim of this
paper is an integration of the two approaches to achieve an optimal
level of the compression ratio