Material Failure Process Simulation by Improve Finite Elements with Embedded Discontinuities

This paper shows the advantages of the material failure process simulation by improve finite elements with embedded discontinuities, using a new definition of traction vector, dependent on the discontinuity length and the angle. Particularly, two families of this kind of elements are compared: kinematically optimal symmetric and statically and kinematically optimal non-symmetric. The constitutive model to describe the behavior of the material in the symmetric formulation is a traction-displacement jump relationship equipped with softening after reaching the failure surface. To show the validity of this symmetric formulation, representative numerical examples illustrating the performance of the proposed formulation are presented. It is shown that the non-symmetric family may over or underestimate the energy required to create a discontinuity, as this effect is related with the total length of the discontinuity, fact that is not noticed when the discontinuity path is a straight line.

Tuning of PV Array Layout Configurations for Maximum Power Delivery

In this paper, an approach for finding optimized layouts for connecting PV units delivering maximum array output power is suggested. The approach is based on considering the different varying parameters of PV units that might be extracted from a general two-diode model. These are mainly, solar irradiation, reverse saturation currents, ideality factors, series and shunt resistances in addition to operating temperature. The approach has been tested on 19 possible 2×3 configurations and allowed to determine the optimized configurations as well as examine the effects of the different units- parameters on the maximum output power. Thus, using this approach, standard arrays with n×m units can be configured for maximum generated power and allows designing PV based systems having reduced surfaces to fit specific required power, as it is the case for solar cars and other mobile systems.

The Study on the Conversed Remediation between Old and New Media in Case of Smart Phone and PC in South Korea

After Apple's first introduction its smart phone, iPhone in the end of 2009 in Korea, the number of Korean smarphone users had been rapidly increasing so that the half of Korean population became smart phone users as of February, 2012. Currently, smart phones are positioned as a major digital media with powerful influences in Korea. And, now, Koreans are leaning new information, enjoying games and communicating other people every time and everywhere. As smart phone devices' performances increased, the number of usable services became more while adequate GUI developments are required to implement various functions with smart phones. The strategy to provide similar experiences on smart phones through familiar features based on employment of existing media's functions mostly contributed to smart phones' popularization in connection with smart phone devices' iconic GUIs. The spread of Smart phone increased mobile web accesses. Therefore, the attempts to implement PC's web in the smart phone's web are continuously made. The mobile web GUI provides familiar experiences to users through designs adequately utilizing the smart phone's GUIs. As the number of users familiarized to smart phones and mobile web GUIs, opposite to reversed remediation from many parts of PCs, PCs are starting to adapt smart phone GUIs. This study defines this phenomenon as the reversed remediation, and reviews the reversed remediation cases of Smart phone GUI' characteristics of PCs. For this purpose, the established study issues are as under: · what is the reversed remediation? · what are the smart phone GUI's characteristics? · what kind of interrelationship exist s between the smart phone and PC's web site? It is meaningful in the forecast of the future GUI's change by understanding of characteristics in the paradigm changes of PC and smart phone's GUI designs. This also will be helpful to establish strategies for digital devices' development and design.

Optimal Controller Design for Linear Magnetic Levitation Rail System

In many applications, magnetic suspension systems are required to operate over large variations in air gap. As a result, the nonlinearities inherent in most types of suspensions have a significant impact on performance. Specifically, it may be difficult to design a linear controller which gives satisfactory performance, stability, and disturbance rejection over a wide range of operating points. in this paper an optimal controller based on discontinuous mathematical model of the system for an electromagnetic suspension system which is applied in magnetic trains has been designed . Simulations show that the new controller can adapt well to the variance of suspension mass and gap, and keep its dynamic performance, thus it is superior to the classic controller.

Intelligent Caching in on-demand Routing Protocol for Mobile Adhoc Networks

An on-demand routing protocol for wireless ad hoc networks is one that searches for and attempts to discover a route to some destination node only when a sending node originates a data packet addressed to that node. In order to avoid the need for such a route discovery to be performed before each data packet is sent, such routing protocols must cache routes previously discovered. This paper presents an analysis of the effect of intelligent caching in a non clustered network, using on-demand routing protocols in wireless ad hoc networks. The analysis carried out is based on the Dynamic Source Routing protocol (DSR), which operates entirely on-demand. DSR uses the cache in every node to save the paths that are learnt during route discovery procedure. In this implementation, caching these paths only at intermediate nodes and using the paths from these caches when required is tried. This technique helps in storing more number of routes that are learnt without erasing the entries in the cache, to store a new route that is learnt. The simulation results on DSR have shown that this technique drastically increases the available memory for caching the routes discovered without affecting the performance of the DSR routing protocol in any way, except for a small increase in end to end delay.

Tuning of Thermal FEA Using Krylov Parametric MOR for Subsea Application

A dead leg is a typical subsea production system component. CFD is required to model heat transfer within the dead leg. Unfortunately its solution is time demanding and thus not suitable for fast prediction or repeated simulations. Therefore there is a need to create a thermal FEA model, mimicking the heat flows and temperatures seen in CFD cool down simulations. This paper describes the conventional way of tuning and a new automated way using parametric model order reduction (PMOR) together with an optimization algorithm. The tuned FE analyses replicate the steady state CFD parameters within a maximum error in heat flow of 6 % and 3 % using manual and PMOR method respectively. During cool down, the relative error of the tuned FEA models with respect to temperature is below 5% comparing to the CFD. In addition, the PMOR method obtained the correct FEA setup five times faster than the manually tuned FEA.

Performance Evaluation of a Limited Round-Robin System

Performance of a limited Round-Robin (RR) rule is studied in order to clarify the characteristics of a realistic sharing model of a processor. Under the limited RR rule, the processor allocates to each request a fixed amount of time, called a quantum, in a fixed order. The sum of the requests being allocated these quanta is kept below a fixed value. Arriving requests that cannot be allocated quanta because of such a restriction are queued or rejected. Practical performance measures, such as the relationship between the mean sojourn time, the mean number of requests, or the loss probability and the quantum size are evaluated via simulation. In the evaluation, the requested service time of an arriving request is converted into a quantum number. One of these quanta is included in an RR cycle, which means a series of quanta allocated to each request in a fixed order. The service time of the arriving request can be evaluated using the number of RR cycles required to complete the service, the number of requests receiving service, and the quantum size. Then an increase or decrease in the number of quanta that are necessary before service is completed is reevaluated at the arrival or departure of other requests. Tracking these events and calculations enables us to analyze the performance of our limited RR rule. In particular, we obtain the most suitable quantum size, which minimizes the mean sojourn time, for the case in which the switching time for each quantum is considered.

Motion Planning and Control of Autonomous Robots in a Two-dimensional Plane

This paper proposes a solution to the motion planning and control problem of a point-mass robot which is required to move safely to a designated target in a priori known workspace cluttered with fixed elliptical obstacles of arbitrary position and sizes. A tailored and unique algorithm for target convergence and obstacle avoidance is proposed that will work for any number of fixed obstacles. The control laws proposed in this paper also ensures that the equilibrium point of the given system is asymptotically stable. Computer simulations with the proposed technique and applications to a planar (RP) manipulator will be presented.

Enabling Factors towards Safety Improvement for Industrialised Building System (IBS)

The utilisation of Industrial Building System (IBS) in construction industry will lead to a safe site condition since minimum numbers of workers are required to be on-site, timely material delivery, systematic component storage, reduction of construction material and waste. These matters are being promoted in the Construction Industry Master Plan (CIMP 2006-2015). However, the enabling factors of IBS that will foster a safer working environment are indefinite; on that basis a research has been conducted. The purpose of this paper is to discuss and identify the relevant factors towards safety improvement for IBS. A quantitative research by way of questionnaire surveys have been conducted to 314 construction companies. The target group was Grade 5 to Grade 7 contractors registered with Construction Industry Development Board (CIDB) which specialise in IBS. The findings disclosed seven factors linked to the safety improvement of IBS construction site in Malaysia. The factors were historical, economic, psychological, technical, procedural, organisational and the environmental factors. From the findings, a psychological factor ranked as the highest and most crucial factor contributing to safer IBS construction site. The psychological factor included the self-awareness and influences from workmates behaviour. Followed by organisational factors, where project management style will encourage the safety efforts. From the procedural factors, it was also found that training was one of the significant factors to improve safety culture of IBS construction site. Another important finding that formed as a part of the environmental factor was storage of IBS components, in which proper planning of the layout would able to contribute to a safer site condition. To conclude, in order to improve safety of IBS construction site, a welltrained and skilled workers are required for IBS projects, thus proper training is permissible and should be emphasised.

Star-Hexagon Transformer Supported UPQC

A new topology of unified power quality conditioner (UPQC) is proposed for different power quality (PQ) improvement in a three-phase four-wire (3P-4W) distribution system. For neutral current mitigation, a star-hexagon transformer is connected in shunt near the load along with three-leg voltage source inverters (VSIs) based UPQC. For the mitigation of source neutral current, the uses of passive elements are advantageous over the active compensation due to ruggedness and less complexity of control. In addition to this, by connecting a star-hexagon transformer for neutral current mitigation the over all rating of the UPQC is reduced. The performance of the proposed topology of 3P-4W UPQC is evaluated for power-factor correction, load balancing, neutral current mitigation and mitigation of voltage and currents harmonics. A simple control algorithm based on Unit Vector Template (UVT) technique is used as a control strategy of UPQC for mitigation of different PQ problems. In this control scheme, the current/voltage control is applied over the fundamental supply currents/voltages instead of fast changing APFs currents/voltages, thereby reducing the computational delay. Moreover, no extra control is required for neutral source current compensation; hence the numbers of current sensors are reduced. The performance of the proposed topology of UPQC is analyzed through simulations results using MATLAB software with its Simulink and Power System Block set toolboxes.

Evaluation of Risk Attributes Driven by Periodically Changing System Functionality

Modeling of the distributed systems allows us to represent the whole its functionality. The working system instance rarely fulfils the whole functionality represented by model; usually some parts of this functionality should be accessible periodically. The reporting system based on the Data Warehouse concept seams to be an intuitive example of the system that some of its functionality is required only from time to time. Analyzing an enterprise risk associated with the periodical change of the system functionality, we should consider not only the inaccessibility of the components (object) but also their functions (methods), and the impact of such a situation on the system functionality from the business point of view. In the paper we suggest that the risk attributes should be estimated from risk attributes specified at the requirements level (Use Case in the UML model) on the base of the information about the structure of the model (presented at other levels of the UML model). We argue that it is desirable to consider the influence of periodical changes in requirements on the enterprise risk estimation. Finally, the proposition of such a solution basing on the UML system model is presented.

Ec-A: A Task Allocation Algorithm for Energy Minimization in Multiprocessor Systems

With the necessity of increased processing capacity with less energy consumption; power aware multiprocessor system has gained more attention in the recent future. One of the additional challenges that is to be solved in a multi-processor system when compared to uni-processor system is job allocation. This paper presents a novel task dependent job allocation algorithm: Energy centric- Allocation (Ec-A) and Rate Monotonic (RM) scheduling to minimize energy consumption in a multiprocessor system. A simulation analysis is carried out to verify the performance increase with reduction in energy consumption and required number of processors in the system.

Computation of Probability Coefficients using Binary Decision Diagram and their Application in Test Vector Generation

This paper deals with efficient computation of probability coefficients which offers computational simplicity as compared to spectral coefficients. It eliminates the need of inner product evaluations in determination of signature of a combinational circuit realizing given Boolean function. The method for computation of probability coefficients using transform matrix, fast transform method and using BDD is given. Theoretical relations for achievable computational advantage in terms of required additions in computing all 2n probability coefficients of n variable function have been developed. It is shown that for n ≥ 5, only 50% additions are needed to compute all probability coefficients as compared to spectral coefficients. The fault detection techniques based on spectral signature can be used with probability signature also to offer computational advantage.

An Efficient Data Mining Approach on Compressed Transactions

In an era of knowledge explosion, the growth of data increases rapidly day by day. Since data storage is a limited resource, how to reduce the data space in the process becomes a challenge issue. Data compression provides a good solution which can lower the required space. Data mining has many useful applications in recent years because it can help users discover interesting knowledge in large databases. However, existing compression algorithms are not appropriate for data mining. In [1, 2], two different approaches were proposed to compress databases and then perform the data mining process. However, they all lack the ability to decompress the data to their original state and improve the data mining performance. In this research a new approach called Mining Merged Transactions with the Quantification Table (M2TQT) was proposed to solve these problems. M2TQT uses the relationship of transactions to merge related transactions and builds a quantification table to prune the candidate itemsets which are impossible to become frequent in order to improve the performance of mining association rules. The experiments show that M2TQT performs better than existing approaches.

Position Control of an AC Servo Motor Using VHDL and FPGA

In this paper, a new method of controlling position of AC Servomotor using Field Programmable Gate Array (FPGA). FPGA controller is used to generate direction and the number of pulses required to rotate for a given angle. Pulses are sent as a square wave, the number of pulses determines the angle of rotation and frequency of square wave determines the speed of rotation. The proposed control scheme has been realized using XILINX FPGA SPARTAN XC3S400 and tested using MUMA012PIS model Alternating Current (AC) servomotor. Experimental results show that the position of the AC Servo motor can be controlled effectively. KeywordsAlternating Current (AC), Field Programmable Gate Array (FPGA), Liquid Crystal Display (LCD).

Communication and Quality in Distributed Agile Development: An Empirical Case Study

Through inward perceptions, we intuitively expect distributed software development to increase the risks associated with achieving cost, schedule, and quality goals. To compound this problem, agile software development (ASD) insists one of the main ingredients of its success is cohesive communication attributed to collocation of the development team. The following study identified the degree of communication richness needed to achieve comparable software quality (reduce pre-release defects) between distributed and collocated teams. This paper explores the relevancy of communication richness in various development phases and its impact on quality. Through examination of a large distributed agile development project, this investigation seeks to understand the levels of communication required within each ASD phase to produce comparable quality results achieved by collocated teams. Obviously, a multitude of factors affects the outcome of software projects. However, within distributed agile software development teams, the mode of communication is one of the critical components required to achieve team cohesiveness and effectiveness. As such, this study constructs a distributed agile communication model (DAC-M) for potential application to similar distributed agile development efforts using the measurement of the suitable level of communication. The results of the study show that less rich communication methods, in the appropriate phase, might be satisfactory to achieve equivalent quality in distributed ASD efforts.

An Agent Oriented Architecture to Supply Dynamic Document Generation in ERP Systems

One of the most important aspects expected from an ERP system is to mange user\administrator manual documents dynamically. Since an ERP package is frequently changed during its implementation in customer sites, it is often needed to add new documents and/or apply required changes to existing documents in order to cover new or changed capabilities. The worse is that since these changes occur continuously, the corresponding documents should be updated dynamically; otherwise, implementing the ERP package in the organization encounters serious risks. In this paper, we propose a new architecture which is based on the agent oriented vision and supplies the dynamic document generation expected from ERP systems using several independent but cooperative agents. Beside the dynamic document generation which is the main issue of this paper, the presented architecture will address some aspects of intelligence and learning capabilities existing in ERP.

Response of King Abdulla Canal Water to the Upgrade of As Samra WWTP

The response of King Abdulla Canal (KAC) water to the upgrade of As Samra Wastewater Treatment Plant which discharges its effluent to the Zarqa River is investigated. Time series quality data that extends between October 2005 and December 2009 obtained by a state of the art telemetric monitoring system were analyzed for COD, EC, TP and TN at two monitoring stations located upstream and downstream of the confluence of the Zarqa River with KAC. The samples- means and the t-test showed that there has been significant improvement in the quality of the KAC water for COD, and TP. However, the improvement in the TN was found statistically insignificant, whereas the EC of the KAC was unaffected by the upgrade. Comparing the selected parameters with the standards and guidelines for using treated wastewater in irrigation showed that the KAC water has improved towards meeting the required standards and guidelines for treated wastewater reuse in irrigation.

Response Quality Evaluation in Heterogeneous Question Answering System: A Black-box Approach

The evaluation of the question answering system is a major research area that needs much attention. Before the rise of domain-oriented question answering systems based on natural language understanding and reasoning, evaluation is never a problem as information retrieval-based metrics are readily available for use. However, when question answering systems began to be more domains specific, evaluation becomes a real issue. This is especially true when understanding and reasoning is required to cater for a wider variety of questions and at the same time achieve higher quality responses The research in this paper discusses the inappropriateness of the existing measure for response quality evaluation and in a later part, the call for new standard measures and the related considerations are brought forward. As a short-term solution for evaluating response quality of heterogeneous systems, and to demonstrate the challenges in evaluating systems of different nature, this research presents a black-box approach using observation, classification scheme and a scoring mechanism to assess and rank three example systems (i.e. AnswerBus, START and NaLURI).

Comparison of Evolutionary Algorithms and their Hybrids Applied to MarioAI

Researchers have been applying artificial/ computational intelligence (AI/CI) methods to computer games. In this research field, further researchesare required to compare AI/CI methods with respect to each game application. In thispaper, we report our experimental result on the comparison of evolution strategy, genetic algorithm and their hybrids, applied to evolving controller agents for MarioAI. GA revealed its advantage in our experiment, whereas the expected ability of ES in exploiting (fine-tuning) solutions was not clearly observed. The blend crossover operator and the mutation operator of GA might contribute well to explore the vast search space.