Abstract: The world wide web coupled with the ever-increasing
sophistication of online technologies and software applications puts
greater emphasis on the need of even more sophisticated and
consistent quality requirements modeling than traditional software
applications. Web sites and Web applications (WebApps) are
becoming more information driven and content-oriented raising the
concern about their information quality (InQ). The consistent and
consolidated modeling of InQ requirements for WebApps at different
stages of the life cycle still poses a challenge. This paper proposes an
approach to specify InQ requirements for WebApps by reusing and
extending the ISO 25012:2008(E) data quality model. We also
discuss learnability aspect of information quality for the WebApps.
The proposed ISO 25012 based InQ framework is a step towards a
standardized approach to evaluate WebApps InQ.
Abstract: Current OCR technology does not allow to
accurately recognizing small text images, such as those found
in web images. Our goal is to investigate new approaches to
recognize very low resolution text images containing antialiased
character shapes.
This paper presents a preliminary study on the variability of
such characters and the feasibility to discriminate them by
using geometrical features. In a first stage we analyze the
distribution of these features. In a second stage we present a
study on the discriminative power for recognizing isolated
characters, using various rendering methods and font
properties. Finally we present interesting results of our
evaluation tests leading to our conclusion and future focus.
Abstract: Although the World Wide Web is considered the
largest source of information there exists nowadays, due to its
inherent dynamic characteristics, the task of finding useful and
qualified information can become a very frustrating experience. This
study presents a research on the information mining systems in the
Web; and proposes an implementation of these systems by means of
components that can be built using the technology of Web services.
This implies that they can encompass features offered by a services
oriented architecture (SOA) and specific components may be used by
other tools, independent of platforms or programming languages.
Hence, the main objective of this work is to provide an architecture
to Web mining systems, divided into stages, where each step is a
component that will incorporate the characteristics of SOA. The
separation of these steps was designed based upon the existing
literature. Interesting results were obtained and are shown here.
Abstract: Online news websites are one of the main and wide areas of Mass Media. Since the nineties several Jordanian newspapers were introduced to the World Wide Web to reach various and large numbers of audiances. Examples of these newspapers that have online version are Al-Rai, Ad-Dustor and AlGhad. Other pure online news websites include Ammon and Rum. The main aim of this study is to evaluate online newspaper websites using two assessment measures; usability and web content. This aim is achieved by using a questionnaire based evaluation which is based on the definition of usability and web content in the ISO document as the standard number 9241-part 11. The results are obtained based on 204 audiences- responses. The results of the research showed that the usability factor is relatively good for all Jordanian online newspapers whereas the web content factor is moderate.
Abstract: With the explosive growth of information sources available on the World Wide Web, it has become increasingly difficult to identify the relevant pieces of information, since web pages are often cluttered with irrelevant content like advertisements, navigation-panels, copyright notices etc., surrounding the main content of the web page. Hence, tools for the mining of data regions, data records and data items need to be developed in order to provide value-added services. Currently available automatic techniques to mine data regions from web pages are still unsatisfactory because of their poor performance and tag-dependence. In this paper a novel method to extract data items from the web pages automatically is proposed. It comprises of two steps: (1) Identification and Extraction of the data regions based on visual clues information. (2) Identification of data records and extraction of data items from a data region. For step1, a novel and more effective method is proposed based on visual clues, which finds the data regions formed by all types of tags using visual clues. For step2 a more effective method namely, Extraction of Data Items from web Pages (EDIP), is adopted to mine data items. The EDIP technique is a list-based approach in which the list is a linear data structure. The proposed technique is able to mine the non-contiguous data records and can correctly identify data regions, irrespective of the type of tag in which it is bound. Our experimental results show that the proposed technique performs better than the existing techniques.
Abstract: The development of Web has affected different aspects of our lives, such as communication, sharing knowledge, searching for jobs, social activities, etc. The web portal as a gateway in the World Wide Web is a starting point for people who are connecting to the Internet. The web portal as the type of knowledge management system provides a rich space to share and search information as well as communication services like free email or content provision for the users. This research aims to discover the university needs to the web portal as a necessary tool for students in the universities to help them in getting the required information. A survey was conducted to gather students' requirements which can be incorporated in to portal to be developed.
Abstract: E-tailing websites are often perceived to be static, impersonal and distant. However, with the movement of the World Wide Web to Web 2.0 in recent years, these online websites have been found to display personalities akin to 'humanistic' qualities and project impressions much like its retailing counterpart i.e. salespeople. This paper examines the personality of e-tailing websites and their impact on consumers- initial trust towards the sites. A total of 239 Internet users participated in this field experiment study which utilized 6 online book retailers- websites that the participants had not previously visited before. Analysis revealed that out of four website personalities (sincerity, competence, excitement and sophistication) only sincerity and competence are able to exert an influence in building consumers- trust upon their first visit to the website. The implications of the findings are further elaborated in this paper.
Abstract: With the proliferation of World Wide Web,
development of web-based technologies and the growth in web
content, the structure of a website becomes more complex and web
navigation becomes a critical issue to both web designers and users.
In this paper we define the content and web pages as two important
and influential factors in website navigation and paraphrase the
enhancement in the website navigation as making some useful
changes in the link structure of the website based on the
aforementioned factors. Then we suggest a new method for
proposing the changes using fuzzy approach to optimize the website
architecture. Applying the proposed method to a real case of Iranian
Civil Aviation Organization (CAO) website, we discuss the results of
the novel approach at the final section.
Abstract: This paper offers suggestions for educators at all levels about how to better prepare our students for the future, by building on the past. The discussion begins with a summary of changes in the World Wide Web, especially as the term Web 3.0 is being heard. The bulk of the discussion is retrospective and concerned with an overview of traditional teaching and research approaches as they evolved during the 20th century beginning with those grounded in the Cartesian reality of IA Richards- (1929) Practical Criticism. The paper concludes with a proposal of five strategies which incorporate timeless elements from the past as well as cutting-edge elements from today, in order to better prepare our students for the future.
Abstract: Phishing, or stealing of sensitive information on the
web, has dealt a major blow to Internet Security in recent times. Most
of the existing anti-phishing solutions fail to handle the fuzziness
involved in phish detection, thus leading to a large number of false
positives. This fuzziness is attributed to the use of highly flexible and
at the same time, highly ambiguous HTML language. We introduce a
new perspective against phishing, that tries to systematically prove,
whether a given page is phished or not, using the corresponding
original page as the basis of the comparison. It analyzes the layout of
the pages under consideration to determine the percentage distortion
between them, indicative of any form of malicious alteration. The
system design represents an intelligent system, employing dynamic
assessment which accurately identifies brand new phishing attacks
and will prove effective in reducing the number of false positives.
This framework could potentially be used as a knowledge base, in
educating the internet users against phishing.
Abstract: The explosive growth of World Wide Web has posed
a challenging problem in extracting relevant data. Traditional web
crawlers focus only on the surface web while the deep web keeps
expanding behind the scene. Deep web pages are created
dynamically as a result of queries posed to specific web databases.
The structure of the deep web pages makes it impossible for
traditional web crawlers to access deep web contents. This paper,
Deep iCrawl, gives a novel and vision-based approach for extracting
data from the deep web. Deep iCrawl splits the process into two
phases. The first phase includes Query analysis and Query translation
and the second covers vision-based extraction of data from the
dynamically created deep web pages. There are several established
approaches for the extraction of deep web pages but the proposed
method aims at overcoming the inherent limitations of the former.
This paper also aims at comparing the data items and presenting them
in the required order.
Abstract: This paper aims to address the new trend of social
commerce as electronic commerce leverages Web 2.0 technologies
and online social media. The infusions of new technologies on the
World Wide Web connect users in their homes and workplaces,
thus transforming social formations and business transactions. An
in-depth study of the growth and success of a social commerce site,
Facebook was conducted. The investigation is finalized with a triad
relational model which reflects socioeconomic life in the Internet
today. The following three concepts work jointly to form a global
community that has already started to take the place of traditional
commerce and socialization: Web 2.0 technology, E-commerce,
and online social media. A discussion of the research findings
indicates that social commerce networks are sustainable because of
the various incentives given to users as they collaborate with others
regardless of their identity and location. The focus of this article is
to increase understanding on quickly developing Web 2.0 based
social media and their subsequent effects on the emerging social
commerce.
Abstract: Because of its global reach, reduction of time
restraints, and ability to reduce costs and increase sales, use of the
Internet, the World Wide Web (WWW), and related technologies
can be a competitive tool in the arsenal of small and medium-sized
enterprises (SMEs). Countries the world over are interested in the
successful adoption of the Internet by SMEs. Because a vast
majority of jobs come from that sector, greater financial success of
SMEs translates into greater job growth and, subsequently, higher
tax revenue to the government. This research investigated the level
of Internet usage for business solutions by small and medium
enterprises in Jordan. Through the survey of a random sample of
100 firms with less than 500 employees and from data obtained
from this survey that formed the basis for our study, we found that
a majority of respondents use the Internet in business activities ,
the adoption of the Internet as a business tool is limited to a
brochure where Web site which primarily provides one way. As
such, there wasn't interactive information about the company and
its products and services.
Abstract: In order to make surfing the internet faster, and to save redundant processing load with each request for the same web page, many caching techniques have been developed to reduce latency of retrieving data on World Wide Web. In this paper we will give a quick overview of existing web caching techniques used for dynamic web pages then we will introduce a design and implementation model that take advantage of “URL Rewriting" feature in some popular web servers, e.g. Apache, to provide an effective approach of caching dynamic web pages.
Abstract: The ever increasing use of World Wide Web in the
existing network, results in poor performance. Several techniques
have been developed for reducing web traffic by compressing the size
of the file, saving the web pages at the client side, changing the burst
nature of traffic into constant rate etc. No single method was
adequate enough to access the document instantly through the
Internet. In this paper, adaptive hybrid algorithms are developed for
reducing web traffic. Intelligent agents are used for monitoring the
web traffic. Depending upon the bandwidth usage, user-s preferences,
server and browser capabilities, intelligent agents use the best
techniques to achieve maximum traffic reduction. Web caching,
compression, filtering, optimization of HTML tags, and traffic
dispersion are incorporated into this adaptive selection. Using this
new hybrid technique, latency is reduced to 20 – 60 % and cache hit
ratio is increased 40 – 82 %.
Abstract: Internet security attack could endanger the privacy of
World Wide Web users and the integrity of their data. The attack can
be carried out on today's most secure systems- browsers, including
Netscape Navigator and Microsoft Internet Explorer. There are too
many types, methods and mechanisms of attack where new attack
techniques and exploits are constantly being developed and
discovered. In this paper, various types of internet security attack
mechanisms are explored and it is pointed out that when different
types of attacks are combined together, network security can suffer
disastrous consequences.