Abstract: Public sector tendering has traditionally been conducted using manual paper-based processes which are known to be inefficient, less transparent and more prone to manipulations and errors. The advent of the Internet and the World Wide Web has led to the development of numerous e-Tendering systems that addressed some of the problems associated with the manual paper-based tendering system. However, most of these systems rarely support the evaluation of tenders and where they do it is mostly based on the single decision maker which is not suitable in public sector tendering, where for the sake of objectivity, transparency, and fairness, it is required that the evaluation is conducted through a tender evaluation committee. Currently, in Nigeria, the public tendering process in general and the evaluation of tenders, in particular, are largely conducted using manual paper-based processes. Automating these manual-based processes to digital-based processes can help in enhancing the proficiency of public sector tendering in Nigeria. This paper is part of a larger study to develop an electronic tendering system that supports the whole tendering lifecycle based on Nigerian procurement law. Specifically, this paper presents the design and implementation of part of the system that supports group evaluation of tenders based on a technique called fuzzy multi-attributes group decision making. The system was developed using Object-Oriented methodologies and Unified Modelling Language and hypothetically applied in the evaluation of technical and financial proposals submitted by bidders. The system was validated by professionals with extensive experiences in public sector procurement. The results of the validation showed that the system called NPS-eTender has an average rating of 74% with respect to correct and accurate modelling of the existing manual tendering domain and an average rating of 67.6% with respect to its potential to enhance the proficiency of public sector tendering in Nigeria. Thus, based on the results of the validation, the automation of the evaluation process to support tender evaluation committee is achievable and can lead to a more proficient public sector tendering system.
Abstract: In the course of recent decades, medical imaging has
been dominated by the use of costly film media for review and
archival of medical investigation, however due to developments in
networks technologies and common acceptance of a standard digital
imaging and communication in medicine (DICOM) another approach
in light of World Wide Web was produced. Web technologies
successfully used in telemedicine applications, the combination of
web technologies together with DICOM used to design a web-based
and open source DICOM viewer. The Web server allowance to
inquiry and recovery of images and the images viewed/manipulated
inside a Web browser without need for any preinstalling software.
The dynamic site page for medical images visualization and
processing created by using JavaScript and HTML5 advancements.
The XAMPP ‘apache server’ is used to create a local web server for
testing and deployment of the dynamic site. The web-based viewer
connected to multiples devices through local area network (LAN) to
distribute the images inside healthcare facilities. The system offers a
few focal points over ordinary picture archiving and communication
systems (PACS): easy to introduce, maintain and independently
platforms that allow images to display and manipulated efficiently,
the system also user-friendly and easy to integrate with an existing
system that have already been making use of web technologies. The
wavelet-based image compression technique on which 2-D discrete
wavelet transform used to decompose the image then wavelet
coefficients are transmitted by entropy encoding after threshold to
decrease transmission time, stockpiling cost and capacity. The
performance of compression was estimated by using images quality
metrics such as mean square error ‘MSE’, peak signal to noise ratio
‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when
‘coif3’ wavelet filter is used.
Abstract: With a continuously increasing speed of information exchange on the World Wide Web, retailers in the E-Commerce sector are faced with immense possibilities regarding different online purchase processes like dynamic price settings. By use of Dynamic Pricing, retailers are able to set short time price changes in order to optimize producer surplus. The empirical research illustrates the basics of Dynamic Pricing and identifies six influencing factors of Dynamic Pricing. The results of a structural equation modeling approach show five main drivers increasing the potential of dynamic price settings in the E-Commerce. Influencing factors are the knowledge of customers’ individual willingness to pay, rising sales, the possibility of customization, the data volume and the information about competitors’ pricing strategy.
Abstract: Finding relevant information on the World Wide Web is becoming highly challenging day by day. Web usage mining is used for the extraction of relevant and useful knowledge, such as user behaviour patterns, from web access log records. Web access log records all the requests for individual files that the users have requested from the website. Web usage mining is important for Customer Relationship Management (CRM), as it can ensure customer satisfaction as far as the interaction between the customer and the organization is concerned. Web usage mining is helpful in improving website structure or design as per the user’s requirement by analyzing the access log file of a website through a log analyzer tool. The focus of this paper is to enhance the accessibility and usability of a guitar selling web site by analyzing their access log through Deep Log Analyzer tool. The results show that the maximum number of users is from the United States and that they use Opera 9.8 web browser and the Windows XP operating system.
Abstract: By the development of World Wide Web, the usage rate of Internet has rapidly grown globally; and provided a basis for the emergence of electronic business. As well as other sectors, the banking sector has adopted the use of internet with the developments in information and communication technologies. Due to the public disclosure and transparency principle of Corporate Governance, the importance of information disclosure of banks on their web sites has increased significantly. For the purpose of this study, a Bank Disclosure Attribute Index (BDAI) in Turkey has been constructed through classifying the information disclosure on banks’ web sites into general, financial, investors and corporate governance attributes. All 47 banks in Turkish Banking System have been evaluated according to the index with the aim of providing a comparison between banks. By Chi Square Test, Pearson Correlation, T-Test, and ANOVA statistical tools, it has been concluded that the majority of banks in Turkey have shared information on their web sites adequately with respect to their total index score. Although there is a positive correlation between various types of information on banks’ web sites, there is no uniformity among them. Also, no significant difference between various types of information disclosure and bank types has been observed. Compared with the total index score averages of the five largest banks in Turkey, there are some banks that need to improve the content of their web sites.
Abstract: The explosion of the World Wide Web and the
electronic trend of university teaching have transformed the learning
style to become more learner-centered, which has popularized the
digital delivery of mediated lectures as an alternative or an adjunct to
traditional lectures. Despite its potential and popularity, virtual
lectures have not been adopted yet in Jordanian universities. This
research aimed to fill this gap by studying the factors that influence
students’ willingness to accept virtual lectures in one Jordanian
University. A quantitative approach was followed, by obtaining 216
survey responses and statistically applying the UTAUT model with
some modifications. Results revealed that performance expectancy,
effort expectancy, social influences, and self-efficacy could
significantly influence students’ attitudes towards virtual lectures.
Additionally, Facilitating conditions and attitudes towards virtual
lectures were found with significant influence on students’ intention
to take virtual lectures. Research implications and future work were
specified afterwards.
Abstract: The world wide web network is a network with a
complex topology, the main properties of which are the distribution
of degrees in power law, A low clustering coefficient and a weak
average distance. Modeling the web as a graph allows locating the
information in little time and consequently offering a help in the
construction of the research engine. Here, we present a model based
on the already existing probabilistic graphs with all the aforesaid
characteristics. This work will consist in studying the web in order to
know its structuring thus it will enable us to modelize it more easily
and propose a possible algorithm for its exploration.
Abstract: The continuous growth in the size of the World Wide Web has resulted in intricate Web sites, demanding enhanced user skills and more sophisticated tools to help the Web user to find the desired information. In order to make Web more user friendly, it is necessary to provide personalized services and recommendations to the Web user. For discovering interesting and frequent navigation patterns from Web server logs many Web usage mining techniques have been applied. The recommendation accuracy of usage based techniques can be improved by integrating Web site content and site structure in the personalization process.
Herein, we propose semantically enriched Web Usage Mining method for Personalization (SWUMP), an extension to solely usage based technique. This approach is a combination of the fields of Web Usage Mining and Semantic Web. In the proposed method, we envisage enriching the undirected graph derived from usage data with rich semantic information extracted from the Web pages and the Web site structure. The experimental results show that the SWUMP generates accurate recommendations and is able to achieve 10-20% better accuracy than the solely usage based model. The SWUMP addresses the new item problem inherent to solely usage based techniques.
Abstract: Web-based systems have become increasingly
important due to the fact that the Internet and the World Wide Web
have become ubiquitous, surpassing all other technological
developments in our history. The Internet and especially companies
websites has rapidly evolved in their scope and extent of use, from
being a little more than fixed advertising material, i.e. a "web
presences", which had no particular influence for the company's
business, to being one of the most essential parts of the company's
core business.
Traditional software engineering approaches with process models
such as, for example, CMM and Waterfall models, do not work very
well since web system development differs from traditional
development. The development differs in several ways, for example,
there is a large gap between traditional software engineering designs
and concepts and the low-level implementation model, many of the
web based system development activities are business oriented (for
example web application are sales-oriented, web application and
intranets are content-oriented) and not engineering-oriented.
This paper aims to introduce Increment Iterative extreme
Programming (IIXP) methodology for developing web based
systems. In difference to the other existence methodologies, this
methodology is combination of different traditional and modern
software engineering and web engineering principles.
Abstract: Success is a European project that will implement several clean transport offers in three European cities and evaluate the environmental impacts. The goal of these measures is to improve urban mobility or the displacement of residents inside cities. For e.g. park and ride, electric vehicles, hybrid bus and bike sharing etc. A list of 28 criteria and 60 measures has been established for evaluation of these transport projects. The evaluation criteria can be grouped into: Transport, environment, social, economic and fuel consumption. This article proposes a decision support system based that encapsulates a hybrid approach based on fuzzy logic, multicriteria analysis and belief theory for the evaluation of impacts of urban mobility solutions. A web-based tool called DeSSIA (Decision Support System for Impacts Assessment) has been developed that treats complex data. The tool has several functionalities starting from data integration (import of data), evaluation of projects and finishes by graphical display of results. The tool development is based on the concept of MVC (Model, View, and Controller). The MVC is a conception model adapted to the creation of software's which impose separation between data, their treatment and presentation. Effort is laid on the ergonomic aspects of the application. It has codes compatible with the latest norms (XHTML, CSS) and has been validated by W3C (World Wide Web Consortium). The main ergonomic aspect focuses on the usability of the application, ease of learning and adoption. By the usage of technologies such as AJAX (XML and Java Script asynchrones), the application is more rapid and convivial. The positive points of our approach are that it treats heterogeneous data (qualitative, quantitative) from various information sources (human experts, survey, sensors, model etc.).
Abstract: Due to the tremendous amount of information provided
by the World Wide Web (WWW) developing methods for mining
the structure of web-based documents is of considerable interest. In
this paper we present a similarity measure for graphs representing
web-based hypertext structures. Our similarity measure is mainly
based on a novel representation of a graph as linear integer strings,
whose components represent structural properties of the graph. The
similarity of two graphs is then defined as the optimal alignment of
the underlying property strings. In this paper we apply the well known
technique of sequence alignments for solving a novel and challenging
problem: Measuring the structural similarity of generalized trees.
In other words: We first transform our graphs considered as high
dimensional objects in linear structures. Then we derive similarity
values from the alignments of the property strings in order to
measure the structural similarity of generalized trees. Hence, we
transform a graph similarity problem to a string similarity problem for
developing a efficient graph similarity measure. We demonstrate that
our similarity measure captures important structural information by
applying it to two different test sets consisting of graphs representing
web-based document structures.
Abstract: Internet is nowadays included to all National Curriculums of the elementary school. A comparative study of their
goals leads to the conclusion that a complete curriculum should aim to student-s acquisition of the abilities to navigate and search for
information and additionally to emphasize on the evaluation of the information provided by the World Wide Web. In a constructivistic knowledge framework the design of a course has to take under
consideration the conceptual representations of students. The following paper presents the conceptual representation of students of eleven years old, attending the Sixth Grade of Greek Elementary School about World Wide Web and their use in the design and
implementation of an innovative course.
Abstract: This paper discusses the approach of real-time
controlling of the energy management system using the data
acquisition tool of LabVIEW. The main idea of this inspiration was
to interface the Station (PC) with the system and publish the data on
internet using LabVIEW. In this venture, controlling and switching of
3 phase AC loads are effectively and efficiently done. The phases are
also sensed through devices. In case of any failure the attached
generator starts functioning automatically. The computer sends
command to the system and system respond to the request. The
modern feature is to access and control the system world-wide using
world wide web (internet). This controlling can be done at any time
from anywhere to effectively use the energy especially in developing
countries where energy management is a big problem. In this system
totally integrated devices are used to operate via remote location.
Abstract: With the tremendous growth of World Wide Web
(WWW) data, there is an emerging need for effective information
retrieval at the document level. Several query languages such as
XML-QL, XPath, XQL, Quilt and XQuery are proposed in recent
years to provide faster way of querying XML data, but they still lack of
generality and efficiency. Our approach towards evolving a framework
for querying semistructured documents is based on formal query
algebra. Two elements are introduced in the proposed framework:
first, a generic and flexible data model for logical representation of
semistructured data and second, a set of operators for the manipulation
of objects defined in the data model. In additional to accommodating
several peculiarities of semistructured data, our model offers novel
features such as bidirectional paths for navigational querying and
partitions for data transformation that are not available in other
proposals.
Abstract: This paper describes a paradigmatic approach to develop architecture of secure systems by describing the requirements from four different points of view: that of the owner, the administrator, the user, and the network. Deriving requirements and developing architecture implies the joint elicitation and describing the problem and the structure of the solution. The view points proposed in this paper are those we consider as requirements towards their contributions as major parties in the design, implementation, usage and maintenance of secure systems. The dramatic growth of the technology of Internet and the applications deployed in World Wide Web have lead to the situation where the security has become a very important concern in the development of secure systems. Many security approaches are currently being used in organizations. In spite of the widespread use of many different security solutions, the security remains a problem. It is argued that the approach that is described in this paper for the development of secure architecture is practical by all means. The models representing these multiple points of view are termed the requirements model (views of owner and administrator) and the operations model (views of user and network). In this paper, this multiple view paradigm is explained by first describing the specific requirements and or characteristics of secure systems (particularly in the domain of networks) and the secure architecture / system development methodology.
Abstract: Nowadays, computer worms, viruses and Trojan horse
become popular, and they are collectively called malware. Those
malware just spoiled computers by deleting or rewriting important
files a decade ago. However, recent malware seems to be born to earn
money. Some of malware work for collecting personal information so
that malicious people can find secret information such as password for
online banking, evidence for a scandal or contact address which relates
with the target. Moreover, relation between money and malware
becomes more complex. Many kinds of malware bear bots to get
springboards. Meanwhile, for ordinary internet users,
countermeasures against malware come up against a blank wall.
Pattern matching becomes too much waste of computer resources,
since matching tools have to deal with a lot of patterns derived from
subspecies. Virus making tools can automatically bear subspecies of
malware. Moreover, metamorphic and polymorphic malware are no
longer special. Recently there appears malware checking sites that
check contents in place of users' PC. However, there appears a new
type of malicious sites that avoids check by malware checking sites. In
this paper, existing protocols and methods related with the web are
reconsidered in terms of protection from current attacks, and new
protocol and method are indicated for the purpose of security of the
web.
Abstract: This paper proposes an auto-classification algorithm
of Web pages using Data mining techniques. We consider the
problem of discovering association rules between terms in a set of
Web pages belonging to a category in a search engine database, and
present an auto-classification algorithm for solving this problem that
are fundamentally based on Apriori algorithm. The proposed
technique has two phases. The first phase is a training phase where
human experts determines the categories of different Web pages, and
the supervised Data mining algorithm will combine these categories
with appropriate weighted index terms according to the highest
supported rules among the most frequent words. The second phase is
the categorization phase where a web crawler will crawl through the
World Wide Web to build a database categorized according to the
result of the data mining approach. This database contains URLs and
their categories.
Abstract: As the web continues to grow exponentially, the idea
of crawling the entire web on a regular basis becomes less and less
feasible, so the need to include information on specific domain,
domain-specific search engines was proposed. As more information
becomes available on the World Wide Web, it becomes more difficult
to provide effective search tools for information access. Today,
people access web information through two main kinds of search
interfaces: Browsers (clicking and following hyperlinks) and Query
Engines (queries in the form of a set of keywords showing the topic
of interest) [2]. Better support is needed for expressing one's
information need and returning high quality search results by web
search tools. There appears to be a need for systems that do reasoning
under uncertainty and are flexible enough to recover from the
contradictions, inconsistencies, and irregularities that such reasoning
involves. In a multi-view problem, the features of the domain can be
partitioned into disjoint subsets (views) that are sufficient to learn the
target concept. Semi-supervised, multi-view algorithms, which
reduce the amount of labeled data required for learning, rely on the
assumptions that the views are compatible and uncorrelated. This
paper describes the use of semi-structured machine learning approach
with Active learning for the “Domain Specific Search Engines". A
domain-specific search engine is “An information access system that
allows access to all the information on the web that is relevant to a
particular domain. The proposed work shows that with the help of
this approach relevant data can be extracted with the minimum
queries fired by the user. It requires small number of labeled data and
pool of unlabelled data on which the learning algorithm is applied to
extract the required data.
Abstract: The purpose of semantic web research is to transform
the Web from a linked document repository into a distributed knowledge base and application platform, thus allowing the vast range of available information and services to be more efficiently
exploited. As a first step in this transformation, languages such as
OWL have been developed. Although fully realizing the Semantic Web still seems some way off, OWL has already been very
successful and has rapidly become a defacto standard for ontology
development in fields as diverse as geography, geology, astronomy,
agriculture, defence and the life sciences. The aim of this paper is to classify key concepts of Semantic Web as well as introducing a new
practical approach which uses these concepts to outperform Word Wide Web.
Abstract: Recognizing the increasing importance of using the
Internet to conduct business, this paper looks at some related matters
associated with small businesses making a decision of whether or not
to have a Website and go online. Small businesses in Saudi Arabia
struggle to have this decision. For organizations, to fully go online,
conduct business and provide online information services, they need
to connect their database to the Web. Some issues related to doing
that might be beyond the capabilities of most small businesses in
Saudi Arabia, such as Website management, technical issues and
security concerns. Here we focus on a small business firm in Saudi
Arabia (Case Study), discussing the issues related to going online
decision and the firm's options of what to do and how to do it. The
paper suggested some valuable solutions of connecting databases to
the Web. It also discusses some of the important issues related to
online information services and e-commerce, mainly Web hosting
options and security issues.