Evaluation of Eulerian and Lagrangian Method in Analysis of Concrete Gravity Dam Including Dam Water Foundation Interaction

Because of the reservoir effect, dynamic analysis of concrete dams is more involved than other common structures. This problem is mostly sourced by the differences between reservoir water, dam body and foundation material behaviors. To account for the reservoir effect in dynamic analysis of concrete gravity dams, two methods are generally employed. Eulerian method in reservoir modeling gives rise to a set of coupled equations, whereas in Lagrangian method, the same equations for dam and foundation structure are used. The Purpose of this paper is to evaluate and study possible advantages and disadvantages of both methods. Specifically, application of the above methods in the analysis of dam-foundationreservoir systems is leveraged to calculate the hydrodynamic pressure on dam faces. Within the frame work of dam- foundationreservoir systems, dam displacement under earthquake for various dimensions and characteristics are also studied. The results of both Lagrangian and Eulerian methods in effects of loading frequency, boundary condition and foundation elasticity modulus are quantitatively evaluated and compared. Our analyses show that each method has individual advantages and disadvantages. As such, in any particular case, one of the two methods may prove more suitable as presented in the results section of this study.

Simulation of Large Deformations of Rubbers by the RKPM Method

In this paper processes including large deformations of a rubber with hyperelastic material behavior are simulated by the RKPM method. Due to the loss of kronecker delta properties in the mesh less shape functions, the imposition of essential boundary conditions consumes significant CPU time in mesh free computations. In this work transformation method is used for imposition of essential boundary conditions. A RKPM material shape function is used in this analysis. The support of the material shape functions covers the same set of particles during material deformation and hence the transformation matrix is formed only once at the initial stages. A computer program in MATLAB is developed for simulations.

Applying Lean Principles, Tools and Techniques in Set Parts Supply Implementation

Lean, which was initially developed by Toyota, is widely implemented in other companies to improve competitiveness. This research is an attempt to identify the adoption of lean in the production system of Malaysian car manufacturer, Proton using case study approach. To gain the in-depth information regarding lean implementation, an activity on the assembly line called Set Parts Supply (SPS) was studied. The result indicates that by using lean principles, tools and techniques in the implementation of SPS enabled to achieve the goals on safety, quality, cost, delivery and morale. The implementation increased the size of the workspace, improved the quality of assembly and the delivery of parts supply, reduced the manpower, achieved cost savings on electricity and also increased the motivation of manpower in respect of attendance at work. A framework of SPS implementation is suggested as a contribution for lean practices in production system.

Military Families’ Attachment to the Royal Guards Community of Dusit District, Bangkok Metropolitan

The objective of this research is to study the people’s level of participation in activities of the community, their satisfaction towards the community, the attachment they have to the community, factors that influence the attachment, as well as the characteristics of the relationships of military families’ of the Royal Guards community of Dusit District. The method used was non-probability sampling by quota sampling according to people’s age. The determined age group was 18 years or older. One set of a sample group was done per family. The questionnaires were conducted by 287 people. Snowball sampling was also used by interviewing people of the community, starting from the Royal Guards Community’s leader, then by 20 of the community’s well-respected persons. The data was analyzed by using descriptive statistics, such as arithmetic mean and standard deviation, as well as by inferential statistics, such as Independent - Samples T test (T-test), One-Way ANOVA (F-test), Chi-Square. Descriptive analysis according to the structure of the interview content was also used. The results of the research is that the participation of the population in the Royal Guards Community in various activities is at a medium level, with the average participation level during Mother’s and Father’s Days. The people’s general level of satisfaction towards the premises of the Royal Guards Community is at the highest level. The people were most satisfied with the transportation within the community and in contacting with people from outside the premises. The access to the community is convenient and there are various entrances. The attachment of the people to the Royal Guards Community in general and by each category is at a high level. The feeling that the community is their home rated the highest average. Factors that influence the attachment of the people of the Royal Guards Community are age, status, profession, income, length of stay in the community, membership of social groups, having neighbors they feel close and familiar with, and as well as the benefits they receive from the community. In addition, it was found that people that participate in activities have a high level of positive relationship towards the attachment of the people to the Royal Guards Community. The satisfaction of the community has a very high level of positive relationship with the attachment of the people to the Royal Guards Community. The characteristics of the attachment of military families’ is that they live in big houses that everyone has to protect and care for, starting from the leader of the family as well as all members. Therefore, they all love the community they live in. The characteristics that show the participation of activities within the community and the high level of satisfaction towards the premises of the community will enable the people to be more attached to the community. The people feel that everyone is close neighbors within the community, as if they are one big family.

Semantic Modeling of Management Information: Enabling Automatic Reasoning on DMTF-CIM

CIM is the standard formalism for modeling management information developed by the Distributed Management Task Force (DMTF) in the context of its WBEM proposal, designed to provide a conceptual view of the managed environment. In this paper, we propose the inclusion of formal knowledge representation techniques, based on Description Logics (DLs) and the Web Ontology Language (OWL), in CIM-based conceptual modeling, and then we examine the benefits of such a decision. The proposal is specified as a CIM metamodel level mapping to a highly expressive subset of DLs capable of capturing all the semantics of the models. The paper shows how the proposed mapping can be used for automatic reasoning about the management information models, as a design aid, by means of new-generation CASE tools, thanks to the use of state-of-the-art automatic reasoning systems that support the proposed logic and use algorithms that are sound and complete with respect to the semantics. Such a CASE tool framework has been developed by the authors and its architecture is also introduced. The proposed formalization is not only useful at design time, but also at run time through the use of rational autonomous agents, in response to a need recently recognized by the DMTF.

Knowledge Based Concept Analysis Method using Concept Maps and UML: Security Notion Case

One of the most ancient humankind concerns is knowledge formalization i.e. what a concept is. Concept Analysis, a branch of analytical philosophy, relies on the purpose of decompose the elements, relations and meanings of a concept. This paper aims at presenting a method to make a concept analysis obtaining a knowledge representation suitable to be processed by a computer system using either object-oriented or ontology technologies. Security notion is, usually, known as a set of different concepts related to “some kind of protection". Our method concludes that a more general framework for the concept, despite it is dynamic, is possible and any particular definition (instantiation) depends on the elements used by its construction instead of the concept itself.

Development of Performance Indicators in Operational Level for Pre-hospital EMS in Thailand

The objective of this research is to develop the performance indicators (PIs) in operational level for the Pre-hospital Emergency Medical Service (EMS) system employing in Thailand. This research started with ascertaining the current pre-hospital care system. The team analyzed the strategies of Narerthorn, a government unit under the ministry of public health, and the existing PIs of the pre-hospital care. Afterwards, the current National Strategic Plan of EMS development (2008-2012) of the Emergency Medical Institute of Thailand (EMIT) was considered using strategic analysis to developed Strategy Map (SM) and identified the Success Factors (SFs). The analysis results from strategy map and SFs were used to develop the Performance Indicators (PIs). To verify the set of PIs, the team has interviewed with the relevant practitioners for the possibilities to implement the PIs. To this paper, it was to ascertain that all the developed PIs support the objectives of the strategic plan. Nevertheless, the results showed that the operational level PIs suited only with the first dimension of National Strategic Plan (infrastructure and information technology development). Besides, the SF was the infrastructure development (to contribute the EMS system to people throughout with standard and efficiency both in normally and disaster conditions). Finally, twenty-nine indicators were developed from the analysis results of SM and SFs.

Design of High Torque Elbow Joint for Above Elbow Prosthesis

Above Elbow Prosthesis is one of the most commonly amputated or missing limbs. The research is done for modelling techniques of upper limb prosthesis and design of high torque, light weight and compact in size elbow actuator. The purposed actuator consists of a DC motor, planetary gear set and a harmonic drive. The calculations show that the actuator is good enough to be used in real life powered prosthetic upper limb or rehabilitation exoskeleton.

Electrical Resistivity of Subsurface: Field and Laboratory Assessment

The objective of this paper is to study the electrical resistivity complexity between field and laboratory measurement, in order to improve the effectiveness of data interpretation for geophysical ground resistivity survey. The geological outcrop in Penang, Malaysia with an obvious layering contact was chosen as the study site. Two dimensional geoelectrical resistivity imaging were used in this study to maps the resistivity distribution of subsurface, whereas few subsurface sample were obtained for laboratory advance. In this study, resistivity of samples in original conditions is measured in laboratory by using time domain low-voltage technique, particularly for granite core sample and soil resistivity measuring set for soil sample. The experimentation results from both schemes are studied, analyzed, calibrated and verified, including basis and correlation, degree of tolerance and characteristics of substance. Consequently, the significant different between both schemes is explained comprehensively within this paper.

Comparative Analysis of the Public Funding for Greek Universities: An Ordinal DEA/MCDM Approach

This study performs a comparative analysis of the 21 Greek Universities in terms of their public funding, awarded for covering their operating expenditure. First it introduces a DEA/MCDM model that allocates the fund into four expenditure factors in the most favorable way for each university. Then, it presents a common, consensual assessment model to reallocate the amounts, remaining in the same level of total public budget. From the analysis it derives that a number of universities cannot justify the public funding in terms of their size and operational workload. For them, the sufficient reduction of their public funding amount is estimated as a future target. Due to the lack of precise data for a number of expenditure criteria, the analysis is based on a mixed crisp-ordinal data set.

Meta-Classification using SVM Classifiers for Text Documents

Text categorization is the problem of classifying text documents into a set of predefined classes. In this paper, we investigated three approaches to build a meta-classifier in order to increase the classification accuracy. The basic idea is to learn a metaclassifier to optimally select the best component classifier for each data point. The experimental results show that combining classifiers can significantly improve the accuracy of classification and that our meta-classification strategy gives better results than each individual classifier. For 7083 Reuters text documents we obtained a classification accuracies up to 92.04%.

An Overview of Islanding Detection Methods in Photovoltaic Systems

The issue of unintentional islanding in PV grid interconnection still remains as a challenge in grid-connected photovoltaic (PV) systems. This paper discusses the overview of popularly used anti-islanding detection methods, practically applied in PV grid-connected systems. Anti-islanding methods generally can be classified into four major groups, which include passive methods, active methods, hybrid methods and communication base methods. Active methods have been the preferred detection technique over the years due to very small non-detected zone (NDZ) in small scale distribution generation. Passive method is comparatively simpler than active method in terms of circuitry and operations. However, it suffers from large NDZ that significantly reduces its performance. Communication base methods inherit the advantages of active and passive methods with reduced drawbacks. Hybrid method which evolved from the combination of both active and passive methods has been proven to achieve accurate anti-islanding detection by many researchers. For each of the studied anti-islanding methods, the operation analysis is described while the advantages and disadvantages are compared and discussed. It is difficult to pinpoint a generic method for a specific application, because most of the methods discussed are governed by the nature of application and system dependent elements. This study concludes that the setup and operation cost is the vital factor for anti-islanding method selection in order to achieve minimal compromising between cost and system quality.

Pin type Clamping Attachment for Remote Setup of Machining Process

Sharing the manufacturing facility through remote operation and monitoring of a machining process is challenge for effective use the production facility. Several automation tools in term of hardware and software are necessary for successfully remote operation of a machine. This paper presents a prototype of workpiece holding attachment for remote operation of milling process by self configuration the workpiece setup. The prototype is designed with mechanism to reorient the work surface into machining spindle direction with high positioning accuracy. Variety of parts geometry is hold by attachment to perform single setup machining. Pin type with array pattern additionally clamps the workpiece surface from two opposite directions for increasing the machining rigidity. Optimum pins configuration for conforming the workpiece geometry with minimum deformation is determined through hybrid algorithms, Genetic Algorithms (GA) and Particle Swarm Optimization (PSO). Prototype with intelligent optimization technique enables to hold several variety of workpiece geometry which is suitable for machining low of repetitive production in remote operation.

An Evolutionary Statistical Learning Theory

Statistical learning theory was developed by Vapnik. It is a learning theory based on Vapnik-Chervonenkis dimension. It also has been used in learning models as good analytical tools. In general, a learning theory has had several problems. Some of them are local optima and over-fitting problems. As well, statistical learning theory has same problems because the kernel type, kernel parameters, and regularization constant C are determined subjectively by the art of researchers. So, we propose an evolutionary statistical learning theory to settle the problems of original statistical learning theory. Combining evolutionary computing into statistical learning theory, our theory is constructed. We verify improved performances of an evolutionary statistical learning theory using data sets from KDD cup.

Estimating Development Time of Software Projects Using a Neuro Fuzzy Approach

Software estimation accuracy is among the greatest challenges for software developers. This study aimed at building and evaluating a neuro-fuzzy model to estimate software projects development time. The forty-one modules developed from ten programs were used as dataset. Our proposed approach is compared with fuzzy logic and neural network model and Results show that the value of MMRE (Mean of Magnitude of Relative Error) applying neuro-fuzzy was substantially lower than MMRE applying fuzzy logic and neural network.

Artificial Intelligence Model to Predict Surface Roughness of Ti-15-3 Alloy in EDM Process

Conventionally the selection of parameters depends intensely on the operator-s experience or conservative technological data provided by the EDM equipment manufacturers that assign inconsistent machining performance. The parameter settings given by the manufacturers are only relevant with common steel grades. A single parameter change influences the process in a complex way. Hence, the present research proposes artificial neural network (ANN) models for the prediction of surface roughness on first commenced Ti-15-3 alloy in electrical discharge machining (EDM) process. The proposed models use peak current, pulse on time, pulse off time and servo voltage as input parameters. Multilayer perceptron (MLP) with three hidden layer feedforward networks are applied. An assessment is carried out with the models of distinct hidden layer. Training of the models is performed with data from an extensive series of experiments utilizing copper electrode as positive polarity. The predictions based on the above developed models have been verified with another set of experiments and are found to be in good agreement with the experimental results. Beside this they can be exercised as precious tools for the process planning for EDM.

Image Similarity: A Genetic Algorithm Based Approach

The paper proposes an approach using genetic algorithm for computing the region based image similarity. The image is denoted using a set of segmented regions reflecting color and texture properties of an image. An image is associated with a family of image features corresponding to the regions. The resemblance of two images is then defined as the overall similarity between two families of image features, and quantified by a similarity measure, which integrates properties of all the regions in the images. A genetic algorithm is applied to decide the most plausible matching. The performance of the proposed method is illustrated using examples from an image database of general-purpose images, and is shown to produce good results.

Prediction the Deformation in Upsetting Process by Neural Network and Finite Element

In this paper back-propagation artificial neural network (BPANN) is employed to predict the deformation of the upsetting process. To prepare a training set for BPANN, some finite element simulations were carried out. The input data for the artificial neural network are a set of parameters generated randomly (aspect ratio d/h, material properties, temperature and coefficient of friction). The output data are the coefficient of polynomial that fitted on barreling curves. Neural network was trained using barreling curves generated by finite element simulations of the upsetting and the corresponding material parameters. This technique was tested for three different specimens and can be successfully employed to predict the deformation of the upsetting process

Automatic Design Algorithm for the Tower Crane Foundations

Foundation of tower crane serves to ensure stability against vertical and horizontal forces. If foundation stress is not sufficient, tower crane may be subject to overturning, shearing or foundation settlement. Therefore, engineering review of stable support is a highly critical part of foundation design. However, there are not many professionals who can conduct engineering review of tower crane foundation and, if any, they have information only on a small number of cranes in which they have hands-on experience. It is also customary to rely on empirical knowledge and tower crane renter-s recommendations rather than designing foundation on the basis of engineering knowledge. Therefore, a foundation design automation system considering not only lifting conditions but also overturning risk, shearing and vertical force may facilitate production of foolproof foundation design for experts and enable even non-experts to utilize professional knowledge that only experts can access now. This study proposes Automatic Design Algorithm for the Tower Crane Foundations considering load and horizontal force.

Combining Fuzzy Logic and Neural Networks in Modeling Landfill Gas Production

Heterogeneity of solid waste characteristics as well as the complex processes taking place within the landfill ecosystem motivated the implementation of soft computing methodologies such as artificial neural networks (ANN), fuzzy logic (FL), and their combination. The present work uses a hybrid ANN-FL model that employs knowledge-based FL to describe the process qualitatively and implements the learning algorithm of ANN to optimize model parameters. The model was developed to simulate and predict the landfill gas production at a given time based on operational parameters. The experimental data used were compiled from lab-scale experiment that involved various operating scenarios. The developed model was validated and statistically analyzed using F-test, linear regression between actual and predicted data, and mean squared error measures. Overall, the simulated landfill gas production rates demonstrated reasonable agreement with actual data. The discussion focused on the effect of the size of training datasets and number of training epochs.