Corporate Governance Practices and Analysts Forecast Accuracy Evidence for Romania

In the last few years, several steps were taken in order to improve the quality of corporate governance for Romanian listed companies. Higher standards of corporate governance is documented in the literature to lead to a better information environment, and, consequently, to increase analysts forecast accuracy. Accordingly, the purpose of this paper is to investigate the extent to which corporate governance policies affect analysts forecasts for companies listed on Bucharest Stock Exchange. The results showed that there is indeed a negative correlation between a corporate governance index – used as a proxy for the quality of corporate governance practices - and analysts forecast errors.

Enabling Integration across Heterogeneous Care Networks

The paper shows how the CASMAS modeling language, and its associated pervasive computing architecture, can be used to facilitate continuity of care by providing members of patientcentered communities of care with a support to cooperation and knowledge sharing through the usage of electronic documents and digital devices. We consider a scenario of clearly fragmented care to show how proper mechanisms can be defined to facilitate a better integration of practices and information across heterogeneous care networks. The scenario is declined in terms of architectural components and cooperation-oriented mechanisms that make the support reactive to the evolution of the context where these communities operate.

Skew Detection Technique for Binary Document Images based on Hough Transform

Document image processing has become an increasingly important technology in the automation of office documentation tasks. During document scanning, skew is inevitably introduced into the incoming document image. Since the algorithm for layout analysis and character recognition are generally very sensitive to the page skew. Hence, skew detection and correction in document images are the critical steps before layout analysis. In this paper, a novel skew detection method is presented for binary document images. The method considered the some selected characters of the text which may be subjected to thinning and Hough transform to estimate skew angle accurately. Several experiments have been conducted on various types of documents such as documents containing English Documents, Journals, Text-Book, Different Languages and Document with different fonts, Documents with different resolutions, to reveal the robustness of the proposed method. The experimental results revealed that the proposed method is accurate compared to the results of well-known existing methods.

Digital Hypertexts vs. Traditional Books: An Inquiry into Non-Linearity

The current study begins with an awareness that today-s media environment is characterized by technological development and a new way of reading caused by the introduction of the Internet. The researcher conducted a meta analysis framed within Technological Determinism to investigate the process of hypertext reading, its differences from linear reading and the effects such differences can have on people-s ways of mentally structuring their world. The relationship between literacy and the comprehension achieved by reading hypertexts is also investigated. The results show hypertexts are not always user friendly. People experience hyperlinks as interruptions that distract their attention generating comprehension and disorientation. On one hand hypertextual jumping reading generates interruptions that finally make people lose their concentration. On the other hand hypertexts fascinate people who would rather read a document in such a format even though the outcome is often frustrating and affects their ability to elaborate and retain information.

The Transfer of Energy Technologies in a Developing Country Context Towards Improved Practice from Past Successes and Failures

Technology transfer of renewable energy technologies is very often unsuccessful in the developing world. Aside from challenges that have social, economic, financial, institutional and environmental dimensions, technology transfer has generally been misunderstood, and largely seen as mere delivery of high tech equipment from developed to developing countries or within the developing world from R&D institutions to society. Technology transfer entails much more, including, but not limited to: entire systems and their component parts, know-how, goods and services, equipment, and organisational and managerial procedures. Means to facilitate the successful transfer of energy technologies, including the sharing of lessons are subsequently extremely important for developing countries as they grapple with increasing energy needs to sustain adequate economic growth and development. Improving the success of technology transfer is an ongoing process as more projects are implemented, new problems are encountered and new lessons are learnt. Renewable energy is also critical to improve the quality of lives of the majority of people in developing countries. In rural areas energy is primarily traditional biomass. The consumption activities typically occur in an inefficient manner, thus working against the notion of sustainable development. This paper explores the implementation of technology transfer in the developing world (sub-Saharan Africa). The focus is necessarily on RETs since most rural energy initiatives are RETs-based. Additionally, it aims to highlight some lessons drawn from the cited RE projects and identifies notable differences where energy technology transfer was judged to be successful. This is done through a literature review based on a selection of documented case studies which are judged against the definition provided for technology transfer. This paper also puts forth research recommendations that might contribute to improved technology transfer in the developing world. Key findings of this paper include: Technology transfer cannot be complete without satisfying pre-conditions such as: affordability, maintenance (and associated plans), knowledge and skills transfer, appropriate know how, ownership and commitment, ability to adapt technology, sound business principles such as financial viability and sustainability, project management, relevance and many others. It is also shown that lessons are learnt in both successful and unsuccessful projects.

Methodology of Realization for Supervisor and Simulator Dedicated to a Semiconductor Research and Production Factory

In the micro and nano-technology industry, the «clean-rooms» dedicated to manufacturing chip, are equipped with the most sophisticated equipment-tools. There use a large number of resources in according to strict specifications for an optimum working and result. The distribution of «utilities» to the production is assured by teams who use a supervision tool. The studies show the interest to control the various parameters of production or/and distribution, in real time, through a reliable and effective supervision tool. This document looks at a large part of the functions that the supervisor must assure, with complementary functionalities to help the diagnosis and simulation that prove very useful in our case where the supervised installations are complexed and in constant evolution.

Suitability of Requirements Abstraction Model (RAM) Requirements for High-Level System Testing

The Requirements Abstraction Model (RAM) helps in managing abstraction in requirements by organizing them at four levels (product, feature, function and component). The RAM is adaptable and can be tailored to meet the needs of the various organizations. Because software requirements are an important source of information for developing high-level tests, organizations willing to adopt the RAM model need to know the suitability of the RAM requirements for developing high-level tests. To investigate this suitability, test cases from twenty randomly selected requirements were developed, analyzed and graded. Requirements were selected from the requirements document of a Course Management System, a web based software system that supports teachers and students in performing course related tasks. This paper describes the results of the requirements document analysis. The results show that requirements at lower levels in the RAM are suitable for developing executable tests whereas it is hard to develop from requirements at higher levels.

Virtual or Virtually U: Educational Institutions in Second Life

Educational institutions are increasingly exploring the affordances of 3D virtual worlds for instruction and research, but few studies have been done to document current practices and uses of this emerging technology. This observational survey examines the virtual presences of 170 accredited educational institutions found in one such 3D virtual world called Second Life®, created by San- Francisco based Linden Lab®. The study focuses on what educational institutions look like in this virtual environment, the types of spaces educational institutions are creating or simulating, and what types of activities are being conducted.

Malaysia under the Purview of the United Nations and Agenda 21

Developing a nation geared by the principle of sustainable development has been one of the piers in moulding a greater nation for Malaysia since its independence. This is seen by the act of joining the United Nations in 1957, just a month after gaining their independence. The United Nations is an international organization that aims to unite the nations worldwide based on justice, human dignity and human well-being. Malaysia has established a local body called the United Nations Malaysia which collaborates with the government to accomplish the aim of supporting sustainable development in Malaysia. Agenda 21 is an international document produced from the Earth Summit providing guidelines of implementing sustainable development globally, nationally and locally. Initiatives of applying Agenda 21 in Malaysia have been taken by the government and non-profit organizations to expose issues regarding sustainable development and providing environmental education to the community to increase awareness towards environmental protection.

Himmapan Creatures: The Tactile Texture Designed for the Blind

The main purpose of this research aimed to create tactile texture designed media for the blind used for extra learning outside classrooms in order to enhance imagination of the blind about Himmapan creatures, furthermore, the main objective of the research focused on improving the visual disabled perception to be equal to normal people. The target group of the research is blinded students studying in The Bangkok school for the blind between grade 4-6 in the second semester of 2011 who are able to read the braille language. The research methodology consisted of the field study and the documentary study related to the blind, tactile texture designed media and Himmapan creatures. 10 pictures of tactile texture designed media were created in the designing process which began after the analysis had conducted based the primary and secondary data. The works had presented to experts in the visual disabled field who evaluated the works. After approval, the works used as prototype to teach the blind. KeywordsBlind, Himmapan Creatures, Tactile Texture.

Language and Retrieval Accuracy

One of the major challenges in the Information Retrieval field is handling the massive amount of information available to Internet users. Existing ranking techniques and strategies that govern the retrieval process fall short of expected accuracy. Often relevant documents are buried deep in the list of documents returned by the search engine. In order to improve retrieval accuracy we examine the issue of language effect on the retrieval process. Then, we propose a solution for a more biased, user-centric relevance for retrieved data. The results demonstrate that using indices based on variations of the same language enhances the accuracy of search engines for individual users.

Automatic Reusability Appraisal of Software Components using Neuro-fuzzy Approach

Automatic reusability appraisal could be helpful in evaluating the quality of developed or developing reusable software components and in identification of reusable components from existing legacy systems; that can save cost of developing the software from scratch. But the issue of how to identify reusable components from existing systems has remained relatively unexplored. In this paper, we have mentioned two-tier approach by studying the structural attributes as well as usability or relevancy of the component to a particular domain. Latent semantic analysis is used for the feature vector representation of various software domains. It exploits the fact that FeatureVector codes can be seen as documents containing terms -the idenifiers present in the components- and so text modeling methods that capture co-occurrence information in low-dimensional spaces can be used. Further, we devised Neuro- Fuzzy hybrid Inference System, which takes structural metric values as input and calculates the reusability of the software component. Decision tree algorithm is used to decide initial set of fuzzy rules for the Neuro-fuzzy system. The results obtained are convincing enough to propose the system for economical identification and retrieval of reusable software components.

Advanced Information Extraction with n-gram based LSI

Number of documents being created increases at an increasing pace while most of them being in already known topics and little of them introducing new concepts. This fact has started a new era in information retrieval discipline where the requirements have their own specialties. That is digging into topics and concepts and finding out subtopics or relations between topics. Up to now IR researches were interested in retrieving documents about a general topic or clustering documents under generic subjects. However these conventional approaches can-t go deep into content of documents which makes it difficult for people to reach to right documents they were searching. So we need new ways of mining document sets where the critic point is to know much about the contents of the documents. As a solution we are proposing to enhance LSI, one of the proven IR techniques by supporting its vector space with n-gram forms of words. Positive results we have obtained are shown in two different application area of IR domain; querying a document database, clustering documents in the document database.

Determining the Criteria and their Importance Level of Calibration Supplier Selection

Quality control is the crucial step for ISO 9001 Quality System Management Standard for companies. While measuring the quality level of both raw material and semi product/product, the calibration of the measuring device is an essential requirement. Calibration suppliers are in the service sector and therefore the calibration supplier selection is becoming a worthy topic for improving service quality. This study presents the results of a questionnaire about the selection criteria of a calibration supplier. The questionnaire was applied to 103 companies and the results are discussed in this paper. The analysis was made with MINITAB 14.0 statistical programs. “Competence of documentations" and “technical capability" are defined as the prerequisites because of the ISO/IEC17025:2005 standard. Also “warranties and complaint policy", “communication", “service features", “quality" and “performance history" are defined as very important criteria for calibration supplier selection.

Data Annotation Models and Annotation Query Language

This paper presents data annotation models at five levels of granularity (database, relation, column, tuple, and cell) of relational data to address the problem of unsuitability of most relational databases to express annotations. These models do not require any structural and schematic changes to the underlying database. These models are also flexible, extensible, customizable, database-neutral, and platform-independent. This paper also presents an SQL-like query language, named Annotation Query Language (AnQL), to query annotation documents. AnQL is simple to understand and exploits the already-existent wide knowledge and skill set of SQL.

Software Architecture Recovery

The advent of modern technology shadows its impetus repercussions on successful Legacy systems making them obsolete with time. These systems have evolved the large organizations in major problems in terms of new business requirements, response time, financial depreciation and maintenance. Major difficulty is due to constant system evolution and incomplete, inconsistent and obsolete documents which a legacy system tends to have. The myriad dimensions of these systems can only be explored by incorporating reverse engineering, in this context, is the best method to extract useful artifacts and by exploring these artifacts for reengineering existing legacy systems to meet new requirements of organizations. A case study is conducted on six different type of software systems having source code in different programming languages using the architectural recovery framework.

Using the Semantic Web in Ubiquitous and Mobile Computing: the Morfeo Experience

With the advent of emerging personal computing paradigms such as ubiquitous and mobile computing, Web contents are becoming accessible from a wide range of mobile devices. Since these devices do not have the same rendering capabilities, Web contents need to be adapted for transparent access from a variety of client agents. Such content adaptation results in better rendering and faster delivery to the client device. Nevertheless, Web content adaptation sets new challenges for semantic markup. This paper presents an advanced components platform, called MorfeoSMC, enabling the development of mobility applications and services according to a channel model based on Services Oriented Architecture (SOA) principles. It then goes on to describe the potential for integration with the Semantic Web through a novel framework of external semantic annotation of mobile Web contents. The role of semantic annotation in this framework is to describe the contents of individual documents themselves, assuring the preservation of the semantics during the process of adapting content rendering, as well as to exploit these semantic annotations in a novel user profile-aware content adaptation process. Semantic Web content adaptation is a way of adding value to and facilitates repurposing of Web contents (enhanced browsing, Web Services location and access, etc).

Towards Clustering of Web-based Document Structures

Methods for organizing web data into groups in order to analyze web-based hypertext data and facilitate data availability are very important in terms of the number of documents available online. Thereby, the task of clustering web-based document structures has many applications, e.g., improving information retrieval on the web, better understanding of user navigation behavior, improving web users requests servicing, and increasing web information accessibility. In this paper we investigate a new approach for clustering web-based hypertexts on the basis of their graph structures. The hypertexts will be represented as so called generalized trees which are more general than usual directed rooted trees, e.g., DOM-Trees. As a important preprocessing step we measure the structural similarity between the generalized trees on the basis of a similarity measure d. Then, we apply agglomerative clustering to the obtained similarity matrix in order to create clusters of hypertext graph patterns representing navigation structures. In the present paper we will run our approach on a data set of hypertext structures and obtain good results in Web Structure Mining. Furthermore we outline the application of our approach in Web Usage Mining as future work.

A Study on the User Experience Design of Mobile Twitter Application

The number of people using SNS with their mobile devices is soaring. This research focuses on the Twitter service that has the most third-party applications and delved into the fact that there were not sufficient studies on the UX design aspects of Twitter applications. Among social network services which have emerged as a major social topic lately, this research try to analyze the UX design of the Twitter application which is also called micro-blogging service. Therefore this research sets its goal to draw components of the UX design aspect of the Tweeter application on which there are not enough analysis yet. Moreover, this research suggests improvement of mobile application which will assure better users- experience. In order to analyze the UX design aspect of the mobile twitter application, with relevant document and user research, evaluating factors of the UX Design which would affect users- experience were organized. The subjects for cases were selected among six paid and free social networking applications that had been consistently ranked from 1st to 100th in the Korean application store during May, 2012 after closely monitoring the rank. From May 15th to May 11th in 2012, in accordance with the evaluating standard, surveys were conducted in a form of interviews with 20 subjects who have used the Twitter application to find out problems and solutions for the UX design of the mobile Twitter application.

ORank: An Ontology Based System for Ranking Documents

Increasing growth of information volume in the internet causes an increasing need to develop new (semi)automatic methods for retrieval of documents and ranking them according to their relevance to the user query. In this paper, after a brief review on ranking models, a new ontology based approach for ranking HTML documents is proposed and evaluated in various circumstances. Our approach is a combination of conceptual, statistical and linguistic methods. This combination reserves the precision of ranking without loosing the speed. Our approach exploits natural language processing techniques for extracting phrases and stemming words. Then an ontology based conceptual method will be used to annotate documents and expand the query. To expand a query the spread activation algorithm is improved so that the expansion can be done in various aspects. The annotated documents and the expanded query will be processed to compute the relevance degree exploiting statistical methods. The outstanding features of our approach are (1) combining conceptual, statistical and linguistic features of documents, (2) expanding the query with its related concepts before comparing to documents, (3) extracting and using both words and phrases to compute relevance degree, (4) improving the spread activation algorithm to do the expansion based on weighted combination of different conceptual relationships and (5) allowing variable document vector dimensions. A ranking system called ORank is developed to implement and test the proposed model. The test results will be included at the end of the paper.