Analyses of Socio-Cognitive Identity Styles by Slovak Adolescents

The contribution deals with analysis of identity style at adolescents (N=463) at the age from 16 to 19 (the average age is 17,7 years). We used the Identity Style Inventory by Berzonsky, distinguishing three basic, measured identity styles: informational, normative, diffuse-avoidant identity style and also commitment. The informational identity style influencing on personal adaptability, coping strategies, quality of life and the normative identity style, it means the style in which an individual takes on models of authorities at self-defining were found to have the highest representation in the studied group of adolescents by higher scores at girls in comparison with boys. The normative identity style positively correlates with the informational identity style. The diffuse-avoidant identity style was found to be positively associated with maladaptive decisional strategies, neuroticism and depressive reactions. There is the style, in which the individual shifts aside defining his personality. In our research sample the lowest score represents it and negatively correlates with commitment, it means with coping strategies, thrust in oneself and the surrounding world. The age of adolescents did not significantly differentiate representation of identity style. We were finding the model, in which informational and normative identity style had positive relationship and the informational and diffuseavoidant style had negative relationship, which were determinated with commitment. In the same time the commitment is influenced with other outside factors.

FCA-based Conceptual Knowledge Discovery in Folksonomy

The tagging data of (users, tags and resources) constitutes a folksonomy that is the user-driven and bottom-up approach to organizing and classifying information on the Web. Tagging data stored in the folksonomy include a lot of very useful information and knowledge. However, appropriate approach for analyzing tagging data and discovering hidden knowledge from them still remains one of the main problems on the folksonomy mining researches. In this paper, we have proposed a folksonomy data mining approach based on FCA for discovering hidden knowledge easily from folksonomy. Also we have demonstrated how our proposed approach can be applied in the collaborative tagging system through our experiment. Our proposed approach can be applied to some interesting areas such as social network analysis, semantic web mining and so on.

Universal Metadata Definition

The need to have standards has always been a priority of all the disciplines in the world. Today, standards such as XML and USB are trying to create a universal interface for their respective areas. The information regarding every family in the discipline addressed, must have a lot in common, known as Metadata. A lot of work has been done in specific domains such as IEEE LOM and MPEG-7 but they do not appeal to the universality of creating Metadata for all entities, where we take an entity (object) as, not restricted to Software Terms. This paper tries to address this problem of universal Metadata Definition which may lead to increase in precision of search.

Semantic Modeling of Management Information: Enabling Automatic Reasoning on DMTF-CIM

CIM is the standard formalism for modeling management information developed by the Distributed Management Task Force (DMTF) in the context of its WBEM proposal, designed to provide a conceptual view of the managed environment. In this paper, we propose the inclusion of formal knowledge representation techniques, based on Description Logics (DLs) and the Web Ontology Language (OWL), in CIM-based conceptual modeling, and then we examine the benefits of such a decision. The proposal is specified as a CIM metamodel level mapping to a highly expressive subset of DLs capable of capturing all the semantics of the models. The paper shows how the proposed mapping can be used for automatic reasoning about the management information models, as a design aid, by means of new-generation CASE tools, thanks to the use of state-of-the-art automatic reasoning systems that support the proposed logic and use algorithms that are sound and complete with respect to the semantics. Such a CASE tool framework has been developed by the authors and its architecture is also introduced. The proposed formalization is not only useful at design time, but also at run time through the use of rational autonomous agents, in response to a need recently recognized by the DMTF.

Training on the Ceasing Intention of Betelnut Addiction

According to the governmental data, the cases of oral cancers doubled in the past 10 years. This had brought heavy burden to the patients- family, the society, and the country. The literature generally evidenced the betel nut contained particular chemicals that can cause oral cancers. Research in Taiwan had also proofed that 90 percent of oral cancer patients had experience of betel nut chewing. It is thus important to educate the betel-nut hobbyists to cease such a hazardous behavior. A program was then organized to establish several training classes across different areas specific to help ceasing this particular habit. Purpose of this research was to explore the attitude and intention toward ceasing betel-nut chewing before and after attending the training classes. 50 samples were taken from a ceasing class with average age at 45 years old with high school education (54%). 74% of the respondents were male in service or agricultural industries. Experiences in betel-nut chewing were 5-20 years with a dose of 1-20 pieces per day. The data had shown that 60% of the respondents had cigarette smoking habit, and 30% of the respondents were concurrently alcoholic dependent. Research results indicated that the attitude, intentions, and the knowledge on oral cancers were found significant different between before and after attendance. This provided evidence for the effectiveness of the training class. However, we do not perform follow-up after the class. Noteworthy is the test result also shown that participants who were drivers as occupation, or habitual smokers or alcoholic dependents would be less willing to quit the betel-nut chewing. The test results indicated as well that the educational levels and the type of occupation may have significant impacts on an individual-s decisions in taking betel-nut or substance abuse.

Towards Benchmarking English Residential Gas Consumption

The UK Government has emphasized the role of Local Authorities as a key player in its flagship residential energy efficiency strategies, by identifying and targeting areas for energy efficiency improvements. Residential energy consumption in England is characterized by significant geographical variation in energy demand, which makes centralized targeting of areas for energy efficiency intervention difficult. This paper draws on research which aims to understand how demographic, social, economic, urban form and climatic factors influence the geographical variations in English residential gas consumption. The paper reports the findings of a multiple regression model that shows how 64% of the geographical variation in residential gas consumption is accounted for by variations in these factors. Results from this study, after further refinement and validation, can be used by Local Authorities to identify areas within their boundaries that have higher than expected gas consumption, these may be prime targets for energy efficiency initiatives.

Evolutionary Techniques for Model Order Reduction of Large Scale Linear Systems

Recently, genetic algorithms (GA) and particle swarm optimization (PSO) technique have attracted considerable attention among various modern heuristic optimization techniques. The GA has been popular in academia and the industry mainly because of its intuitiveness, ease of implementation, and the ability to effectively solve highly non-linear, mixed integer optimization problems that are typical of complex engineering systems. PSO technique is a relatively recent heuristic search method whose mechanics are inspired by the swarming or collaborative behavior of biological populations. In this paper both PSO and GA optimization are employed for finding stable reduced order models of single-input- single-output large-scale linear systems. Both the techniques guarantee stability of reduced order model if the original high order model is stable. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example from literature and the results are compared with recently published conventional model reduction technique.

Research on the Predict Method of Random Vibration Cumulative Fatigue Damage Life Based on the Finite Element Analysis

Aiming at most of the aviation products are facing the problem of fatigue fracture in vibration environment, we makes use of the testing result of a bracket, analysis for the structure with ANSYS-Workbench, predict the life of the bracket by different ways, and compared with the testing result. With the research on analysis methods, make an organic combination of simulation analysis and testing, Not only ensure the accuracy of simulation analysis and life predict, but also make a dynamic supervision of product life process, promote the application of finite element simulation analysis in engineering practice.

Determining the Best Method of Stability Landslide by Using of DSS (Case Study: Landslide in Hasan Salaran, Kurdistan Province in Iran)

One of the processes of slope that occurs every year in Iran and some parts of world and cause a lot of criminal and financial harms is called landslide. They are plenty of method to stability landslide in soil and rock slides. The use of the best method with the least cost and in the shortest time is important for researchers. In this research, determining the best method of stability is investigated by using of Decision Support systems. DSS is made for this purpose and was used (for Hasan Salaran area in Kurdistan). Field study data from topography, slope, geology, geometry of landslide and the related features was used. The related data entered decision making managements programs (DSS) (ALES).Analysis of mass stability indicated the instability potential at present. Research results show that surface and sub surface drainage the best method of stabilizing. Analysis of stability shows that acceptable increase in security coefficient is a consequence of drainage.

Development of Performance Indicators in Operational Level for Pre-hospital EMS in Thailand

The objective of this research is to develop the performance indicators (PIs) in operational level for the Pre-hospital Emergency Medical Service (EMS) system employing in Thailand. This research started with ascertaining the current pre-hospital care system. The team analyzed the strategies of Narerthorn, a government unit under the ministry of public health, and the existing PIs of the pre-hospital care. Afterwards, the current National Strategic Plan of EMS development (2008-2012) of the Emergency Medical Institute of Thailand (EMIT) was considered using strategic analysis to developed Strategy Map (SM) and identified the Success Factors (SFs). The analysis results from strategy map and SFs were used to develop the Performance Indicators (PIs). To verify the set of PIs, the team has interviewed with the relevant practitioners for the possibilities to implement the PIs. To this paper, it was to ascertain that all the developed PIs support the objectives of the strategic plan. Nevertheless, the results showed that the operational level PIs suited only with the first dimension of National Strategic Plan (infrastructure and information technology development). Besides, the SF was the infrastructure development (to contribute the EMS system to people throughout with standard and efficiency both in normally and disaster conditions). Finally, twenty-nine indicators were developed from the analysis results of SM and SFs.

Six Sigma Solutions and its Benefit-Cost Ratio for Quality Improvement

This is an application research presenting the improvement of production quality using the six sigma solutions and the analyses of benefit-cost ratio. The case of interest is the production of tile-concrete. Such production has faced with the problem of high nonconforming products from an inappropriate surface coating and had low process capability based on the strength property of tile. Surface coating and tile strength are the most critical to quality of this product. The improvements followed five stages of six sigma solutions. After the improvement, the production yield was improved to 80% as target required and the defective products from coating process was remarkably reduced from 29.40% to 4.09%. The process capability based on the strength quality was increased from 0.87 to 1.08 as customer oriented. The improvement was able to save the materials loss for 3.24 millions baht or 0.11 million dollars. The benefits from the improvement were analyzed from (1) the reduction of the numbers of non conforming tile using its factory price for surface coating improvement and (2) the materials saved from the increment of process capability. The benefit-cost ratio of overall improvement was high as 7.03. It was non valuable investment in define, measure, analyses and the initial of improve stages after that it kept increasing. This was due to there were no benefits in define, measure, and analyze stages of six sigma since these three stages mainly determine the cause of problem and its effects rather than improve the process. The benefit-cost ratio starts existing in the improve stage and go on. Within each stage, the individual benefitcost ratio was much higher than the accumulative one as there was an accumulation of cost since the first stage of six sigma. The consideration of the benefit-cost ratio during the improvement project helps make decisions for cost saving of similar activities during the improvement and for new project. In conclusion, the determination of benefit-cost ratio behavior through out six sigma implementation period provides the useful data for managing quality improvement for the optimal effectiveness. This is the additional outcome from the regular proceeding of six sigma.

Design of High Torque Elbow Joint for Above Elbow Prosthesis

Above Elbow Prosthesis is one of the most commonly amputated or missing limbs. The research is done for modelling techniques of upper limb prosthesis and design of high torque, light weight and compact in size elbow actuator. The purposed actuator consists of a DC motor, planetary gear set and a harmonic drive. The calculations show that the actuator is good enough to be used in real life powered prosthetic upper limb or rehabilitation exoskeleton.

Concurrent Approach to Data Parallel Model using Java

Parallel programming models exist as an abstraction of hardware and memory architectures. There are several parallel programming models in commonly use; they are shared memory model, thread model, message passing model, data parallel model, hybrid model, Flynn-s models, embarrassingly parallel computations model, pipelined computations model. These models are not specific to a particular type of machine or memory architecture. This paper expresses the model program for concurrent approach to data parallel model through java programming.

An Overview of Islanding Detection Methods in Photovoltaic Systems

The issue of unintentional islanding in PV grid interconnection still remains as a challenge in grid-connected photovoltaic (PV) systems. This paper discusses the overview of popularly used anti-islanding detection methods, practically applied in PV grid-connected systems. Anti-islanding methods generally can be classified into four major groups, which include passive methods, active methods, hybrid methods and communication base methods. Active methods have been the preferred detection technique over the years due to very small non-detected zone (NDZ) in small scale distribution generation. Passive method is comparatively simpler than active method in terms of circuitry and operations. However, it suffers from large NDZ that significantly reduces its performance. Communication base methods inherit the advantages of active and passive methods with reduced drawbacks. Hybrid method which evolved from the combination of both active and passive methods has been proven to achieve accurate anti-islanding detection by many researchers. For each of the studied anti-islanding methods, the operation analysis is described while the advantages and disadvantages are compared and discussed. It is difficult to pinpoint a generic method for a specific application, because most of the methods discussed are governed by the nature of application and system dependent elements. This study concludes that the setup and operation cost is the vital factor for anti-islanding method selection in order to achieve minimal compromising between cost and system quality.

Combining Color and Layout Features for the Identification of Low-resolution Documents

This paper proposes a method, combining color and layout features, for identifying documents captured from lowresolution handheld devices. On one hand, the document image color density surface is estimated and represented with an equivalent ellipse and on the other hand, the document shallow layout structure is computed and hierarchically represented. The combined color and layout features are arranged in a symbolic file, which is unique for each document and is called the document-s visual signature. Our identification method first uses the color information in the signatures in order to focus the search space on documents having a similar color distribution, and finally selects the document having the most similar layout structure in the remaining search space. Finally, our experiment considers slide documents, which are often captured using handheld devices.

An Evolutionary Statistical Learning Theory

Statistical learning theory was developed by Vapnik. It is a learning theory based on Vapnik-Chervonenkis dimension. It also has been used in learning models as good analytical tools. In general, a learning theory has had several problems. Some of them are local optima and over-fitting problems. As well, statistical learning theory has same problems because the kernel type, kernel parameters, and regularization constant C are determined subjectively by the art of researchers. So, we propose an evolutionary statistical learning theory to settle the problems of original statistical learning theory. Combining evolutionary computing into statistical learning theory, our theory is constructed. We verify improved performances of an evolutionary statistical learning theory using data sets from KDD cup.

Artificial Intelligence Model to Predict Surface Roughness of Ti-15-3 Alloy in EDM Process

Conventionally the selection of parameters depends intensely on the operator-s experience or conservative technological data provided by the EDM equipment manufacturers that assign inconsistent machining performance. The parameter settings given by the manufacturers are only relevant with common steel grades. A single parameter change influences the process in a complex way. Hence, the present research proposes artificial neural network (ANN) models for the prediction of surface roughness on first commenced Ti-15-3 alloy in electrical discharge machining (EDM) process. The proposed models use peak current, pulse on time, pulse off time and servo voltage as input parameters. Multilayer perceptron (MLP) with three hidden layer feedforward networks are applied. An assessment is carried out with the models of distinct hidden layer. Training of the models is performed with data from an extensive series of experiments utilizing copper electrode as positive polarity. The predictions based on the above developed models have been verified with another set of experiments and are found to be in good agreement with the experimental results. Beside this they can be exercised as precious tools for the process planning for EDM.

From Mother Tongue Education to Multilingual Higher Education

Through the time, the higher education has changed the learning system since mother tongue to bilingual, and in this new century has been coming develop a multilingual education. All as part of globalization process of the countries and the education. Nevertheless, this change only has been effectively in countries of the first world, the rest have been lagging. Therefore, these countries require strengthen their higher education systems through models that give way to multilingual and bilingual education. In this way, shows a new model adapted from a systemic form to allow a higher bilingual and multilingual education in Latin America. This systematization aims to increase the skills and competencies student’s, decrease the time learning of a second tongue, add to multilingualism in the American Latin Universities, also, contribute to position the region´s countries in a better global status, and stimulate the development of new research in this area.

Judges System for Classifiers Specialization

In this paper we designed and implemented a new ensemble of classifiers based on a sequence of classifiers which were specialized in regions of the training dataset where errors of its trained homologous are concentrated. In order to separate this regions, and to determine the aptitude of each classifier to properly respond to a new case, it was used another set of classifiers built hierarchically. We explored a selection based variant to combine the base classifiers. We validated this model with different base classifiers using 37 training datasets. It was carried out a statistical comparison of these models with the well known Bagging and Boosting, obtaining significantly superior results with the hierarchical ensemble using Multilayer Perceptron as base classifier. Therefore, we demonstrated the efficacy of the proposed ensemble, as well as its applicability to general problems.

Managing Iterations in Product Design and Development

The inherent iterative nature of product design and development poses significant challenge to reduce the product design and development time (PD). In order to shorten the time to market, organizations have adopted concurrent development where multiple specialized tasks and design activities are carried out in parallel. Iterative nature of work coupled with the overlap of activities can result in unpredictable time to completion and significant rework. Many of the products have missed the time to market window due to unanticipated or rather unplanned iteration and rework. The iterative and often overlapped processes introduce greater amounts of ambiguity in design and development, where the traditional methods and tools of project management provide less value. In this context, identifying critical metrics to understand the iteration probability is an open research area where significant contribution can be made given that iteration has been the key driver of cost and schedule risk in PD projects. Two important questions that the proposed study attempts to address are: Can we predict and identify the number of iterations in a product development flow? Can we provide managerial insights for a better control over iteration? The proposal introduces the concept of decision points and using this concept intends to develop metrics that can provide managerial insights into iteration predictability. By characterizing the product development flow as a network of decision points, the proposed research intends to delve further into iteration probability and attempts to provide more clarity.