Meaning Chasing Kiddies: Children-s Perception of Metaphors Used in Printed Advertisements

Today-s children, who are born into a more colorful, more creative, more abstract and more accessible communication environment than their ancestors as a result of dizzying advances in technology, have an interesting capacity to perceive and make sense of the world. Millennium children, who live in an environment where all kinds of efforts by marketing communication are more intensive than ever are, from their early childhood on, subject to all kinds of persuasive messages. As regards advertising communication, it outperforms all the other marketing communication efforts in creating little consumer individuals and, as a result of processing of codes and signs, plays a significant part in building a world of seeing, thinking and understanding for children. Children who are raised with metaphorical expressions such as tales and riddles also meet that fast and effective meaning communication in advertisements. Children-s perception of metaphors, which help grasp the “product and its promise" both verbally and visually and facilitate association between them is the subject of this study. Stimulating and activating imagination, metaphors have unique advantages in promoting the product and its promise especially in regard to print advertisements, which have certain limitations. This study deals comparatively with both literal and metaphoric versions of print advertisements belonging to various product groups and attempts to discover to what extent advertisements are liked, recalled, perceived and are persuasive. The sample group of the study, which was conducted in two elementary schools situated in areas that had different socioeconomic features, consisted of children aged 12.

Information Filtering using Index Word Selection based on the Topics

We have proposed an information filtering system using index word selection from a document set based on the topics included in a set of documents. This method narrows down the particularly characteristic words in a document set and the topics are obtained by Sparse Non-negative Matrix Factorization. In information filtering, a document is often represented with the vector in which the elements correspond to the weight of the index words, and the dimension of the vector becomes larger as the number of documents is increased. Therefore, it is possible that useless words as index words for the information filtering are included. In order to address the problem, the dimension needs to be reduced. Our proposal reduces the dimension by selecting index words based on the topics included in a document set. We have applied the Sparse Non-negative Matrix Factorization to the document set to obtain these topics. The filtering is carried out based on a centroid of the learning document set. The centroid is regarded as the user-s interest. In addition, the centroid is represented with a document vector whose elements consist of the weight of the selected index words. Using the English test collection MEDLINE, thus, we confirm the effectiveness of our proposal. Hence, our proposed selection can confirm the improvement of the recommendation accuracy from the other previous methods when selecting the appropriate number of index words. In addition, we discussed the selected index words by our proposal and we found our proposal was able to select the index words covered some minor topics included in the document set.

Cumulative Learning based on Dynamic Clustering of Hierarchical Production Rules(HPRs)

An important structuring mechanism for knowledge bases is building clusters based on the content of their knowledge objects. The objects are clustered based on the principle of maximizing the intraclass similarity and minimizing the interclass similarity. Clustering can also facilitate taxonomy formation, that is, the organization of observations into a hierarchy of classes that group similar events together. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form Decision If < condition> Generality Specificity . HPRs systems are capable of handling taxonomical structures inherent in the knowledge about the real world. In this paper, a set of related HPRs is called a cluster and is represented by a HPR-tree. This paper discusses an algorithm based on cumulative learning scenario for dynamic structuring of clusters. The proposed scheme incrementally incorporates new knowledge into the set of clusters from the previous episodes and also maintains summary of clusters as Synopsis to be used in the future episodes. Examples are given to demonstrate the behaviour of the proposed scheme. The suggested incremental structuring of clusters would be useful in mining data streams.

Statistical Process Optimization Through Multi-Response Surface Methodology

In recent years, response surface methodology (RSM) has brought many attentions of many quality engineers in different industries. Most of the published literature on robust design methodology is basically concerned with optimization of a single response or quality characteristic which is often most critical to consumers. For most products, however, quality is multidimensional, so it is common to observe multiple responses in an experimental situation. Through this paper interested person will be familiarize with this methodology via surveying of the most cited technical papers. It is believed that the proposed procedure in this study can resolve a complex parameter design problem with more than two responses. It can be applied to those areas where there are large data sets and a number of responses are to be optimized simultaneously. In addition, the proposed procedure is relatively simple and can be implemented easily by using ready-made standard statistical packages.

Optimization of the Headspace Solid-Phase Microextraction Gas Chromatography for Volatile Compounds Determination in Phytophthora Cinnamomi Rands

Phytophthora cinnamomi (P. c) is a plant pathogenic oomycete that is capable of damaging plants in commercial production systems and natural ecosystems worldwide. The most common methods for the detection and diagnosis of P. c infection are expensive, elaborate and time consuming. This study was carried out to examine whether species specific and life cycle specific volatile organic compounds (VOCs) can be absorbed by solid-phase microextraction fibers and detected by gas chromatography that are produced by P. c and another oomycete Pythium dissotocum. A headspace solid-phase microextraction (HS-SPME) together with gas chromatography (GC) method was developed and optimized for the identification of the VOCs released by P. c. The optimized parameters included type of fiber, exposure time, desorption temperature and desorption time. Optimization was achieved with the analytes of P. c+V8A and V8A alone. To perform the HS-SPME, six types of fiber were assayed and compared: 7μm Polydimethylsiloxane (PDMS), 100μm Polydimethylsiloxane (PDMS), 50/30μm Divinylbenzene/CarboxenTM/Polydimethylsiloxane DVB/CAR/PDMS), 65μm Polydimethylsiloxane/Divinylbenzene (PDMS/DVB), 85μm Polyacrylate (PA) fibre and 85μm CarboxenTM/ Polydimethylsiloxane (Carboxen™/PDMS). In a comparison of the efficacy of the fibers, the bipolar fiber DVB/CAR/PDMS had a higher extraction efficiency than the other fibers. An exposure time of 16h with DVB/CAR/PDMS fiber in the sample headspace was enough to reach the maximum extraction efficiency. A desorption time of 3min in the GC injector with the desorption temperature of 250°C was enough for the fiber to desorb the compounds of interest. The chromatograms and morphology study confirmed that the VOCs from P. c+V8A had distinct differences from V8A alone, as did different life cycle stages of P. c and different taxa such as Pythium dissotocum. The study proved that P. c has species and life cycle specific VOCs, which in turn demonstrated the feasibility of this method as means of

Theory of Planned Behaviour and Young Romanians’ Self-Reported Speed

Speeding represents one of the main concerns for road safety and it still is a subject for research. The need to address this problem and to understand why drivers over speed increases especially in Romania, where in 2011, speed was the main cause of car accidents. This article addresses this problem by using the theory of planned behaviour. A questionnaire was administered to a sample of young Romanian drivers (18 to 25 years) and several path analyses were made in order to verify if the model proposed by the theory of planned behaviour fits the data. One interesting result is that perceived behavioural control does not predict the intention to speed or self-reported driving speed, but subjective norms do. This implies that peers and social environment have a greater impact on young Romanian drivers than we thought.

A Method to Improve Test Process in Federal Enterprise Architecture Framework Using ISTQB Framework

Enterprise Architecture (EA) is a framework for description, coordination and alignment of all activities across the organization in order to achieve strategic goals using ICT enablers. A number of EA-compatible frameworks have been developed. We, in this paper, mainly focus on Federal Enterprise Architecture Framework (FEAF) since its reference models are plentiful. Among these models we are interested here in its business reference model (BRM). The test process is one important subject of an EA project which is to somewhat overlooked. This lack of attention may cause drawbacks or even failure of an enterprise architecture project. To address this issue we intend to use International Software Testing Qualification Board (ISTQB) framework and standard test suites to present a method to improve EA testing process. The main challenge is how to communicate between the concepts of EA and ISTQB. In this paper, we propose a method for integrating these concepts.

A New Approach for Classifying Large Number of Mixed Variables

The issue of classifying objects into one of predefined groups when the measured variables are mixed with different types of variables has been part of interest among statisticians in many years. Some methods for dealing with such situation have been introduced that include parametric, semi-parametric and nonparametric approaches. This paper attempts to discuss on a problem in classifying a data when the number of measured mixed variables is larger than the size of the sample. A propose idea that integrates a dimensionality reduction technique via principal component analysis and a discriminant function based on the location model is discussed. The study aims in offering practitioners another potential tool in a classification problem that is possible to be considered when the observed variables are mixed and too large.

A Rough Sets Approach for Relevant Internet/Web Online Searching

The internet is constantly expanding. Identifying web links of interest from web browsers requires users to visit each of the links listed, individually until a satisfactory link is found, therefore those users need to evaluate a considerable amount of links before finding their link of interest; this can be tedious and even unproductive. By incorporating web assistance, web users could be benefited from reduced time searching on relevant websites. In this paper, a rough set approach is presented, which facilitates classification of unlimited available e-vocabulary, to assist web users in reducing search times looking for relevant web sites. This approach includes two methods for identifying relevance data on web links based on the priority and percentage of relevance. As a result of these methods, a list of web sites is generated in priority sequence with an emphasis of the search criteria.

Waste Management, Strategies and Situation in South Africa: An Overview

This paper highlights some interesting facts on South African-s waste situation and management strategies, in particular the Integrated Waste Management. South Africa supports a waste hierarchy by promoting cleaner production, waste minimisation, reuse, recycling and waste treatment with disposal and remediation as the last preferred options in waste management. The drivers for waste management techniques are identified as increased demand for waste service provision; increased demand for waste minimisation; recycling and recovery; land use, physical and environmental limitations; and socio-economic and demographic factors. The South African government recognizes the importance of scientific research as outlined on the white paper on Integrated Pollution and Waste Management (IP and WM) (DEAT, 2000).

The Investors- Reaction to Investment Rating Change Announcements

This study investigates the investors- behavioral reaction to the investment rating change announcements from the views of behavioral finance. The empirical results indicate that self-interest does affect the intention of securities firms to release investment ratings for individual stocks. In addition, behavioral pitfalls are also found in the response of retail investors to investment rating change announcements.

Evaluation of Market Limitations in the Case of Ecosystem Services

Biodiversity crisis is one of the many crises that started at the turn of the millennia. Concrete form of expression is still disputed, but there is a relatively high consensus regarding the high rate of degradation and the urgent need for action. The strategy of action outlines a strong economic component, together with the recognition of market mechanisms as the most effective policies to protect biodiversity. In this context, biodiversity and ecosystem services are natural assets that play a key role in economic strategies and technological development to promote development and prosperity. Developing and strengthening policies for transition to an economy based on efficient use of resources is the way forward. To emphasize the co-viability specific to the connection economyecosystem services, scientific approach aimed on one hand how to implement policies for nature conservation and on the other hand, the concepts underlying the economic expression of ecosystem services- value, in the context of current technology. Following the analysis of business opportunities associated with changes in ecosystem services was concluded that development of market mechanisms for nature conservation is a trend that is increasingly stronger individualized within recent years. Although there are still many controversial issues that have already given rise to an obvious bias, international organizations and national governments have initiated and implemented in cooperation or independently such mechanisms. Consequently, they created the conditions for convergence between private interests and social interests of nature conservation, so there are opportunities for ongoing business development which leads, among other things, the positive effects on biodiversity. Finally, points out that markets fail to quantify the value of most ecosystem services. Existing price signals reflect at best, only a proportion of the total amount corresponding provision of food, water or fuel.

Dynamic Shock Bank Liquidity Analysis

Simulations are developed in this paper with usual DSGE model equations. The model is based on simplified version of Smets-Wouters equations in use at European Central Bank which implies 10 macro-economic variables: consumption, investment, wages, inflation, capital stock, interest rates, production, capital accumulation, labour and credit rate, and allows take into consideration the banking system. Throughout the simulations, this model will be used to evaluate the impact of rate shocks recounting the actions of the European Central Bank during 2008.

Building a Personalized Multidimensional Intelligent Learning System

Currently, most of distance learning courses can only deliver standard material to students. Students receive course content passively which leads to the neglect of the goal of education – “to suit the teaching to the ability of students". Providing appropriate course content according to students- ability is the main goal of this paper. Except offering a series of conventional learning services, abundant information available, and instant message delivery, a complete online learning environment should be able to distinguish between students- ability and provide learning courses that best suit their ability. However, if a distance learning site contains well-designed course content and design but fails to provide adaptive courses, students will gradually loss their interests and confidence in learning and result in ineffective learning or discontinued learning. In this paper, an intelligent tutoring system is proposed and it consists of several modules working cooperatively in order to build an adaptive learning environment for distance education. The operation of the system is based on the result of Self-Organizing Map (SOM) to divide students into different groups according to their learning ability and learning interests and then provide them with suitable course content. Accordingly, the problem of information overload and internet traffic problem can be solved because the amount of traffic accessing the same content is reduced.

A Kernel Based Rejection Method for Supervised Classification

In this paper we are interested in classification problems with a performance constraint on error probability. In such problems if the constraint cannot be satisfied, then a rejection option is introduced. For binary labelled classification, a number of SVM based methods with rejection option have been proposed over the past few years. All of these methods use two thresholds on the SVM output. However, in previous works, we have shown on synthetic data that using thresholds on the output of the optimal SVM may lead to poor results for classification tasks with performance constraint. In this paper a new method for supervised classification with rejection option is proposed. It consists in two different classifiers jointly optimized to minimize the rejection probability subject to a given constraint on error rate. This method uses a new kernel based linear learning machine that we have recently presented. This learning machine is characterized by its simplicity and high training speed which makes the simultaneous optimization of the two classifiers computationally reasonable. The proposed classification method with rejection option is compared to a SVM based rejection method proposed in recent literature. Experiments show the superiority of the proposed method.

Designing Pictogram for Food Portion Size

The objective of this paper is to investigate a new approach based on the idea of pictograms for food portion size. This approach adopts the model of the United States Pharmacopeia- Drug Information (USP-DI). The representation of each food portion size composed of three parts: frame, the connotation of dietary portion sizes and layout. To investigate users- comprehension based on this approach, two experiments were conducted, included 122 Taiwanese people, 60 male and 62 female with ages between 16 and 64 (divided into age groups of 16-30, 31-45 and 46-64). In Experiment 1, the mean correcting rate of the understanding level of food items is 48.54% (S.D.= 95.08) and the mean response time 2.89sec (S.D.=2.14). The difference on the correct rates for different age groups is significant (P*=0.00

Improving Spatiotemporal Change Detection: A High Level Fusion Approach for Discovering Uncertain Knowledge from Satellite Image Database

This paper investigates the problem of tracking spa¬tiotemporal changes of a satellite image through the use of Knowledge Discovery in Database (KDD). The purpose of this study is to help a given user effectively discover interesting knowledge and then build prediction and decision models. Unfortunately, the KDD process for spatiotemporal data is always marked by several types of imperfections. In our paper, we take these imperfections into consideration in order to provide more accurate decisions. To achieve this objective, different KDD methods are used to discover knowledge in satellite image databases. Each method presents a different point of view of spatiotemporal evolution of a query model (which represents an extracted object from a satellite image). In order to combine these methods, we use the evidence fusion theory which considerably improves the spatiotemporal knowledge discovery process and increases our belief in the spatiotemporal model change. Experimental results of satellite images representing the region of Auckland in New Zealand depict the improvement in the overall change detection as compared to using classical methods.

The Effects of Roots Action of Tropical Green Roofs–Replication of German FLL in Singapore

Green Roofs offers numerous advantages, including lowering ambient temperature, which is of increasing interest due to global warming concerns. However, there are technical problems pertaining to waterproofing to be resolved. Currently, the only recognized green roof waterproofing test is the German standard FLL. This paper examines the potential of replicating the test in tropical climate and reducing the test duration by using pre-grown plants. A three year old sample and a new setup were used for this experimental study. The new setup was prepared with close reference to the FLL standards and was compared against the three year old sample. Results showed that the waterproofing membrane was damaged by plant roots in both setups. Joints integrity was also challenged.

On Simulation based WSN Multi-Parametric Performance Analysis

Optimum communication and performance in Wireless Sensor Networks, constitute multi-facet challenges due to the specific networking characteristics as well as the scarce resource availability. Furthermore, it is becoming increasingly apparent that isolated layer based approaches often do not meet the demands posed by WSNs applications due to omission of critical inter-layer interactions and dependencies. As a counterpart, cross-layer is receiving high interest aiming to exploit these interactions and increase network performance. However, in order to clearly identify existing dependencies, comprehensive performance studies are required evaluating the effect of different critical network parameters on system level performance and behavior.This paper-s main objective is to address the need for multi-parametric performance evaluations considering critical network parameters using a well known network simulator, offering useful and practical conclusions and guidelines. The results reveal strong dependencies among considered parameters which can be utilized by and drive future research efforts, towards designing and implementing highly efficient protocols and architectures.

Gravitino Dark Matter in (nearly) SLagy D3/D7 m-Split SUSY

In the context of large volume Big Divisor (nearly) SLagy D3/D7 μ-Split SUSY [1], after an explicit identification of first generation of SM leptons and quarks with fermionic superpartners of four Wilson line moduli, we discuss the identification of gravitino as a potential dark matter candidate by explicitly calculating the decay life times of gravitino (LSP) to be greater than age of universe and lifetimes of decays of the co-NLSPs (the first generation squark/slepton and a neutralino) to the LSP (the gravitino) to be very small to respect BBN constraints. Interested in non-thermal production mechanism of gravitino, we evaluate the relic abundance of gravitino LSP in terms of that of the co-NLSP-s by evaluating their (co-)annihilation cross sections and hence show that the former satisfies the requirement for a potential Dark Matter candidate. We also show that it is possible to obtain a 125 GeV light Higgs in our setup.