Achieving Business and IT Alignment from Organisational Learning Perspectives

Business and IT alignment has continued as a top concern for business and IT executives for almost three decades. Many researchers have conducted empirical studies on the relationship between business-IT alignment and performance. Yet, these approaches, lacking a social perspective, have had little impact on sustaining performance and competitive advantage. In addition to the limited alignment literature that explores organisational learning that is represented in shared understanding, communication, cognitive maps and experiences. Hence, this paper proposes an integrated process that enables social and intellectual dimensions through the concept of organisational learning. In particular, the feedback and feedforward process which provide a value creation across dynamic multilevel of learning. This mechanism enables on-going effectiveness through development of individuals, groups and organisations, which improves the quality of business and IT strategies and drives to performance.

Process-Oriented Learning Requirements for Employees and for Organizations

Using activity theory, organisational theory and didactics as theoretical foundations, a comprehensive model of the organisational dimensions relevant for learning and knowledge transfer will be developed. In a second step, a Learning Assessment Guideline will be elaborated. This guideline will be designed to permit a targeted analysis of organisations to identify the status quo in those areas crucial to the implementation of learning and knowledge transfer. In addition, this self-analysis tool will enable learning managers to select adequate didactic models for e- and blended learning. As part of the European Integrated Project "Process-oriented Learning and Information Exchange" (PROLIX), this model of organisational prerequisites for learning and knowledge transfer will be empirically tested in four profit and non-profit organisations in Great Britain, Germany and France (to be finalized in autumn 2006). The findings concern not only the capability of the model of organisational dimensions, but also the predominant perceptions of and obstacles to learning in organisations.

The Influencing Factors and the Approach to Enhance the Standard of E-Commerce for Small and Medium Enterprises in Bangkok

The objectives of this research paper were to study the influencing factors that contributed to the success of electronic commerce (e-commerce) and to study the approach to enhance the standard of e-commerce for small and medium enterprises (SME). The research paper focused the study on only sole proprietorship SMEs in Bangkok, Thailand. The factors contributed to the success of SME included business management, learning in the organization, business collaboration, and the quality of website. A quantitative and qualitative mixed research methodology was used. In terms of quantitative method, a questionnaire was used to collect data from 251 sole proprietorships. The System Equation Model (SEM) was utilized as the tool for data analysis. In terms of qualitative method, an in-depth interview, a dialogue with experts in the field of ecommerce for SMEs, and content analysis were used. By using the adjusted causal relationship structure model, it was revealed that the factors affecting the success of e-commerce for SMEs were found to be congruent with the empirical data. The hypothesis testing indicated that business management influenced the learning in the organization, the learning in the organization influenced business collaboration and the quality of the website, and these factors, in turn, influenced the success of SMEs. Moreover, the approach to enhance the standard of SMEs revealed that the majority of respondents wanted to enhance the standard of SMEs to a high level in the category of safety of e-commerce system, basic structure of e-commerce, development of staff potentials, assistance of budget and tax reduction, and law improvement regarding the e-commerce respectively.

Low Resolution Single Neural Network Based Face Recognition

This research paper deals with the implementation of face recognition using neural network (recognition classifier) on low-resolution images. The proposed system contains two parts, preprocessing and face classification. The preprocessing part converts original images into blurry image using average filter and equalizes the histogram of those image (lighting normalization). The bi-cubic interpolation function is applied onto equalized image to get resized image. The resized image is actually low-resolution image providing faster processing for training and testing. The preprocessed image becomes the input to neural network classifier, which uses back-propagation algorithm to recognize the familiar faces. The crux of proposed algorithm is its beauty to use single neural network as classifier, which produces straightforward approach towards face recognition. The single neural network consists of three layers with Log sigmoid, Hyperbolic tangent sigmoid and Linear transfer function respectively. The training function, which is incorporated in our work, is Gradient descent with momentum (adaptive learning rate) back propagation. The proposed algorithm was trained on ORL (Olivetti Research Laboratory) database with 5 training images. The empirical results provide the accuracy of 94.50%, 93.00% and 90.25% for 20, 30 and 40 subjects respectively, with time delay of 0.0934 sec per image.

Estimating Cost of R&D Activities for Feasibility Study of Public R&D Investment

Since the feasibility study of R&D programs have been initiated for efficient public R&D investments, year 2008, feasibility studies have improved in terms of precision. Although experience related to these studies of R&D programs have increased to a certain point, still methodological improvement is required. The feasibility studies of R&D programs are consisted of various viewpoints, such as technology, policy, and economics. This research is to provide improvement methods to the economic perspective; especially the cost estimation process of R&D activities. First of all, the fundamental concept of cost estimation is reviewed. After the review, a statistical and econometric analysis method is applied as empirical analysis. Conclusively, limitations and further research directions are provided.

High Performance in Parallel Data Integration: An Empirical Evaluation of the Ratio Between Processing Time and Number of Physical Nodes

Many studies have shown that parallelization decreases efficiency [1], [2]. There are many reasons for these decrements. This paper investigates those which appear in the context of parallel data integration. Integration processes generally cannot be allocated to packages of identical size (i. e. tasks of identical complexity). The reason for this is unknown heterogeneous input data which result in variable task lengths. Process delay is defined by the slowest processing node. It leads to a detrimental effect on the total processing time. With a real world example, this study will show that while process delay does initially increase with the introduction of more nodes it ultimately decreases again after a certain point. The example will make use of the cloud computing platform Hadoop and be run inside Amazon-s EC2 compute cloud. A stochastic model will be set up which can explain this effect.

Impact of Computer-Mediated Communication on Virtual Teams- Performance: An Empirical Study

In a complex project environment, project teams face multi-dimensional communication problems that can ultimately lead to project breakdown. Team Performance varies in Face-to-Face (FTF) environment versus groups working remotely in a computermediated communication (CMC) environment. A brief review of the Input_Process_Output model suggested by James E. Driskell, Paul H. Radtke and Eduardo Salas in “Virtual Teams: Effects of Technological Mediation on Team Performance (2003)", has been done to develop the basis of this research. This model theoretically analyzes the effects of technological mediation on team processes, such as, cohesiveness, status and authority relations, counternormative behavior and communication. An empirical study described in this paper has been undertaken to test the “cohesiveness" of diverse project teams in a multi-national organization. This study uses both quantitative and qualitative techniques for data gathering and analysis. These techniques include interviews, questionnaires for data collection and graphical data representation for analyzing the collected data. Computer-mediated technology may impact team performance because of difference in cohesiveness among teams and this difference may be moderated by factors, such as, the type of communication environment, the type of task and the temporal context of the team. Based on the reviewed model, sets of hypotheses are devised and tested. This research, reports on a study that compared team cohesiveness among virtual teams using CMC and non-CMC communication mediums. The findings suggest that CMC can help virtual teams increase team cohesiveness among their members, making CMC an effective medium for increasing productivity and team performance.

Ranking - Convex Risk Minimization

The problem of ranking (rank regression) has become popular in the machine learning community. This theory relates to problems, in which one has to predict (guess) the order between objects on the basis of vectors describing their observed features. In many ranking algorithms a convex loss function is used instead of the 0-1 loss. It makes these procedures computationally efficient. Hence, convex risk minimizers and their statistical properties are investigated in this paper. Fast rates of convergence are obtained under conditions, that look similarly to the ones from the classification theory. Methods used in this paper come from the theory of U-processes as well as empirical processes.

Examining Herzberg-s Two Factor Theory in a Large Chinese Chemical Fiber Company

The validity of Herzberg-s Two-Factor Theory of Motivation was tested empirically by surveying 2372 chemical fiber employees in 2012. In the valid sample of 1875 respondents, the degree of overall job satisfaction was more than moderate. The most highly valued components of job satisfaction were: “corporate image," “collaborative working atmosphere," and “supervisor-s expertise"; whereas the lowest mean score was 34.65 for “job rotation and promotion." The top three job retention options rated by the participants were “good image of the enterprise," “good compensation," and “workplace is close to my residence." The overall evaluation of the level of thriving facilitation workplace reached almost to “mostly agree." For those participants who chose at least one motivator as their job retention options had significantly greater job satisfaction than those who chose only hygiene factors as their retention options. Therefore, Herzberg-s Two-Factor Theory of Motivation was proven valid in this study.

Experimental Investigation of the Maximum Axial Force in the Folding Process of Aluminum Square Columns

In this paper, a semi empirical formula is presented based on the experimental results to predict the first pick (maximum force) value in the instantaneous folding force- axial distance diagram of a square column. To achieve this purpose, the maximum value of the folding force was assumed to be a function of the average folding force. Using the experimental results, the maximum value of the force necessary to initiate the first fold in a square column was obtained with respect to the geometrical quantities and material properties. Finally, the results obtained from the semi empirical relation in this paper, were compared to the experimental results which showed a good correlation.

Multivariate School Travel Demand Regression Based on Trip Attraction

Since primary school trips usually start from home, attention by many scholars have been focused on the home end for data gathering. Thereafter category analysis has often been relied upon when predicting school travel demands. In this paper, school end was relied on for data gathering and multivariate regression for future travel demand prediction. 9859 pupils were surveyed by way of questionnaires at 21 primary schools. The town was divided into 5 zones. The study was carried out in Skudai Town, Malaysia. Based on the hypothesis that the number of primary school trip ends are expected to be the same because school trips are fixed, the choice of trip end would have inconsequential effect on the outcome. The study compared empirical data for home and school trip end productions and attractions. Variance from both data results was insignificant, although some claims from home based family survey were found to be grossly exaggerated. Data from the school trip ends was relied on for travel demand prediction because of its completeness. Accessibility, trip attraction and trip production were then related to school trip rates under daylight and dry weather conditions. The paper concluded that, accessibility is an important parameter when predicting demand for future school trip rates.

The Empirical Survey on the Effect of Using Media in Explosive Forming of Tubular Shells

The special and unique advantages of explosive forming, has developed its use in different industries. Considering the important influence of improving the current explosive forming techniques on increasing the efficiency and control over the explosive forming procedure, the effects of air and water as the energy-conveying medium, and also their differences will be illustrated in this paper. Hence, a large number of explosive forming tests have been conducted on two sizes of thin walled cylindrical shells by using air and water as the working medium. Comparative diagrams of the maximum radial deflection of work-pieces of the same size, as a function of the scaled distance, show that for the points with the same values of scaled distance, the maximum radial deformation caused by the under water explosive loading is 4 to 5 times more than the deflection of the shells under explosive forming, while using air. Results of this experimental research have also been compared with other studies which show that using water as the energy conveying media increases the efficiency up to 4.8 times. The effect of the media on failure modes of the shells, and the necking mechanism of the walls of the specimens, while being explosively loaded, are also discussed in this issue. Measuring the tested specimens shows that, the increase in the internal volume has been accompanied by necking of the walls, which finally results in the radial rupture of the structure.

The Determinants and Outcomes of Pathological Internet use (PIU) among Urban Millennial Teens: A Theoretical Framework

The rapid adoption of Internet has turned the Millennial Teens- life like a lightning speed. Empirical evidence has illustrated that Pathological Internet Use (PIU) among them ensure long-term success to the market players in the children industry. However, it creates concerns among their care takers as it generates mental disorder among some of them. The purpose of this paper is to examine the determinants of PIU and identify its outcomes among urban Millennial Teens. It aims to develop a theoretical framework based on a modified Media System Dependency (MSD) Theory that integrates important systems and components that determine and resulted from PIU.

Combining Bagging and Additive Regression

Bagging and boosting are among the most popular re-sampling ensemble methods that generate and combine a diversity of regression models using the same learning algorithm as base-learner. Boosting algorithms are considered stronger than bagging on noise-free data. However, there are strong empirical indications that bagging is much more robust than boosting in noisy settings. For this reason, in this work we built an ensemble using an averaging methodology of bagging and boosting ensembles with 10 sub-learners in each one. We performed a comparison with simple bagging and boosting ensembles with 25 sub-learners on standard benchmark datasets and the proposed ensemble gave better accuracy.

EMD-Based Signal Noise Reduction

This paper introduces a new signal denoising based on the Empirical mode decomposition (EMD) framework. The method is a fully data driven approach. Noisy signal is decomposed adaptively into oscillatory components called Intrinsic mode functions (IMFs) by means of a process called sifting. The EMD denoising involves filtering or thresholding each IMF and reconstructs the estimated signal using the processed IMFs. The EMD can be combined with a filtering approach or with nonlinear transformation. In this work the Savitzky-Golay filter and shoftthresholding are investigated. For thresholding, IMF samples are shrinked or scaled below a threshold value. The standard deviation of the noise is estimated for every IMF. The threshold is derived for the Gaussian white noise. The method is tested on simulated and real data and compared with averaging, median and wavelet approaches.

Stock Market Integration Measurement: Investigation of Malaysia and Singapore Stock Markets

This paper tests the level of market integration between Malaysia and Singapore stock markets with the world market. Kalman Filter (KF) methodology is used on the International Capital Asset Pricing Model (ICAPM) and the pricing errors estimated within the framework of ICAPM are used as a measure of market integration or segmentation. The advantage of the KF technique is that it allows for time-varying coefficients in estimating ICAPM and hence able to capture the varying degree of market integration. Empirical results show clear evidence of varying degree of market integration for both case of Malaysia and Singapore. Furthermore, the results show that the changes in the level of market integration are found to coincide with certain economic events that have taken placed. The findings certainly provide evidence on the practicability of the KF technique to estimate stock markets integration. In the comparison between Malaysia and Singapore stock market, the result shows that the trends of the market integration indices for Malaysia and Singapore look similar through time but the magnitude is notably different with the Malaysia stock market showing greater degree of market integration. Finally, significant evidence of varying degree of market integration shows the inappropriate use of OLS in estimating the level of market integration.

Development of Coronal Field and Solar Wind Components for MHD Interplanetary Simulations

The connection between solar activity and adverse phenomena in the Earth’s environment that can affect space and ground based technologies has spurred interest in Space Weather (SW) research. A great effort has been put on the development of suitable models that can provide advanced forecast of SW events. With the progress in computational technology, it is becoming possible to develop operational large scale physics based models which can incorporate the most important physical processes and domains of the Sun-Earth system. In order to enhance our SW prediction capabilities we are developing advanced numerical tools. With operational requirements in mind, our goal is to develop a modular simulation framework of propagation of the disturbances from the Sun through interplanetary space to the Earth. Here, we report and discuss on the development of coronal field and solar wind components for a large scale MHD code. The model for these components is based on a potential field source surface model and an empirical Wang-Sheeley-Arge solar wind relation. 

The Impact of Stakeholder Communication Strategies on Consumers- Acceptance and Financial Performance: In the Case of Fertilizer Industry in Malaysia

There has been a growing emphasis in communication management from simple coordination of promotional tools to a complex strategic process. This study will examine the current marketing communications and engagement strategies used in addressing the key stakeholders. In the case of fertilizer industry in Malaysia, there has been little empirical research on stakeholder communication when major challenges facing the modern corporation is the need to communicate its identity, its values and products in order to distinguish itself from competitors. The study will employ both quantitative and qualitative methods and the use of Structural Equation Modeling (SEM) to establish a causal relationship amongst the key factors of stakeholder communication strategies and increment in consumers- choice/acceptance and impact on financial performance. One of the major contributions is a conceptual framework for communication strategies and engagement in increasing consumers- acceptance level and the firm-s financial performance.

Knowledge Sharing Behavior in E-Communities: from the Perspective of Transaction Cost Theory

This study aims to examine the factors affecting knowledge sharing behavior in knowledge-based electronic communities (e-communities) because quantity and quality of knowledge shared among the members play a critical role in the community-s sustainability. Past research has suggested three perspectives that may affect the quantity and quality of knowledge shared: economics, social psychology, and social ecology. In this study, we strongly believe that an economic perspective may be suitable to validate factors influencing newly registered members- knowledge contribution at the beginning of relationship development. Accordingly, this study proposes a model to validate the factors influencing members- knowledge sharing based on Transaction Cost Theory. By doing so, we may empirically test our hypotheses in various types of e-communities to determine the generalizability of our research models.

Measuring the Comprehensibility of a UML-B Model and a B Model

Software maintenance, which involves making enhancements, modifications and corrections to existing software systems, consumes more than half of developer time. Specification comprehensibility plays an important role in software maintenance as it permits the understanding of the system properties more easily and quickly. The use of formal notation such as B increases a specification-s precision and consistency. However, the notation is regarded as being difficult to comprehend. Semi-formal notation such as the Unified Modelling Language (UML) is perceived as more accessible but it lacks formality. Perhaps by combining both notations could produce a specification that is not only accurate and consistent but also accessible to users. This paper presents an experiment conducted on a model that integrates the use of both UML and B notations, namely UML-B, versus a B model alone. The objective of the experiment was to evaluate the comprehensibility of a UML-B model compared to a traditional B model. The measurement used in the experiment focused on the efficiency in performing the comprehension tasks. The experiment employed a cross-over design and was conducted on forty-one subjects, including undergraduate and masters students. The results show that the notation used in the UML-B model is more comprehensible than the B model.