Seismic Behaviour of Romanian Ortodox Churches, Modeling of Failure Modes by Rigid Blocks

Historic religious buildings located in seismic areas have developed different failure mechanisms. Simulation of failure modes is done with computer programs through a nonlinear dynamic analysis or simplified using the method of failure blocks. Currently there are simulation methodologies of failure modes based on the failure rigid blocks method only for Roman Catholic churches type. Due to differences of shape in plan, elevation and construction systems between Orthodox churches and Catholic churches, for the first time there were initiated researches in the development of this simulation methodology for Orthodox churches. In this article are presented the first results from the researches. The theoretical results were compared with real failure modes recorded at an Orthodox church from Banat region, severely damaged by earthquakes in 1991. Simulated seismic response, using a computer program based on finite element method was confirmed by cracks after earthquakes. The consolidation of the church was made according to these theoretical results, realizing a rigid floor connecting all the failure blocks.

An Environmental Impact Tool to Assess National Energy Scenarios

The Long-range Energy and Alternatives Planning (LEAP) energy planning system has been developed for South Africa, for the 2005 base year and a limited number of plausible future scenarios that may have significant implications (negative or positive) in terms of environmental impacts. The system quantifies the national energy demand for the domestic, commercial, transport, industry and agriculture sectors, the supply of electricity and liquid fuels, and the resulting emissions. The South African National Energy Research Institute (SANERI) identified the need to develop an environmental assessment tool, based on the LEAP energy planning system, to provide decision-makers and stakeholders with the necessary understanding of the environmental impacts associated with different energy scenarios. A comprehensive analysis of indicators that are used internationally and in South Africa was done and the available data was accessed to select a reasonable number of indicators that could be utilized in energy planning. A consultative process was followed to determine the needs of different stakeholders on the required indicators and also the most suitable form of reporting. This paper demonstrates the application of Energy Environmental Sustainability Indicators (EESIs) as part of the developed tool, which assists with the identification of the environmental consequences of energy generation and use scenarios and thereby promotes sustainability, since environmental considerations can then be integrated into the preparation and adoption of policies, plans, programs and projects. Recommendations are made to refine the tool further for South Africa.

Analysis of Program PRIME at Brazil

Policies that support entrepreneurship are keys to the generation of new business. In Brazil, seed capital, installation of technology parks, programs and zero interest financing, economic subsidy as Program First Innovative Company (PRIME) are examples of incentive policies. For the implementation of PRIME, in particular the Brazilian Innovation Agency (FINEP) decentralized operationalization so that business incubators could select innovative projects. This paper analyzes the program PRIME Business Incubator Center of the State of Sergipe (CISE) after calculating the mean and standard deviation of the grades obtained by companies in the factors of innovation, market potential, financial return economic, market strategy and staff and application of the Mann-Whitney test.

Virtual Learning Process Environment: Cohort Analytics for Learning and Learning Processes

Traditional higher-education classrooms allow lecturers to observe students- behaviours and responses to a particular pedagogy during learning in a way that can influence changes to the pedagogical approach. Within current e-learning systems it is difficult to perform continuous analysis of the cohort-s behavioural tendency, making real-time pedagogical decisions difficult. This paper presents a Virtual Learning Process Environment (VLPE) based on the Business Process Management (BPM) conceptual framework. Within the VLPE, course designers can model various education pedagogies in the form of learning process workflows using an intuitive flow diagram interface. These diagrams are used to visually track the learning progresses of a cohort of students. This helps assess the effectiveness of the chosen pedagogy, providing the information required to improve course design. A case scenario of a cohort of students is presented and quantitative statistical analysis of their learning process performance is gathered and displayed in realtime using dashboards.

Strategic Software Development: Productivity Comparisons of General Development Programs

Productivity has been one of the major concerns with the increasingly high cost of software development. Choosing the right development language with high productivity is one approach to reduce development costs. Working on the large database with 4106 projects ever developed, we found the factors significant to productivity. After the removal of the effects of other factors on productivity, we compare the productivity differences of the ten general development programs. The study supports the fact that fourth-generation languages are more productive than thirdgeneration languages.

The Empirical Survey on the Effect of Using Media in Explosive Forming of Tubular Shells

The special and unique advantages of explosive forming, has developed its use in different industries. Considering the important influence of improving the current explosive forming techniques on increasing the efficiency and control over the explosive forming procedure, the effects of air and water as the energy-conveying medium, and also their differences will be illustrated in this paper. Hence, a large number of explosive forming tests have been conducted on two sizes of thin walled cylindrical shells by using air and water as the working medium. Comparative diagrams of the maximum radial deflection of work-pieces of the same size, as a function of the scaled distance, show that for the points with the same values of scaled distance, the maximum radial deformation caused by the under water explosive loading is 4 to 5 times more than the deflection of the shells under explosive forming, while using air. Results of this experimental research have also been compared with other studies which show that using water as the energy conveying media increases the efficiency up to 4.8 times. The effect of the media on failure modes of the shells, and the necking mechanism of the walls of the specimens, while being explosively loaded, are also discussed in this issue. Measuring the tested specimens shows that, the increase in the internal volume has been accompanied by necking of the walls, which finally results in the radial rupture of the structure.

Implementation of a New Neural Network Function Block to Programmable Logic Controllers Library Function

Programmable logic controllers are the main controllers in the today's industries; they are used for several applications in industrial control systems and there are lots of examples exist from the PLC applications in industries especially in big companies and plants such as refineries, power plants, petrochemical companies, steel companies, and food and production companies. In the PLCs there are some functions in the function library in software that can be used in PLC programs as basic program elements. The aim of this project are introducing and implementing a new function block of a neural network to the function library of PLC. This block can be applied for some control applications or nonlinear functions calculations after it has been trained for these applications. The implemented neural network is a Perceptron neural network with three layers, three input nodes and one output node. The block can be used in manual or automatic mode. In this paper the structure of the implemented function block, the parameters and the training method of the network are presented by considering the especial method of PLC programming and its complexities. Finally the application of the new block is compared with a classic simulated block and the results are presented.

Target Detection with Improved Image Texture Feature Coding Method and Support Vector Machine

An image texture analysis and target recognition approach of using an improved image texture feature coding method (TFCM) and Support Vector Machine (SVM) for target detection is presented. With our proposed target detection framework, targets of interest can be detected accurately. Cascade-Sliding-Window technique was also developed for automated target localization. Application to mammogram showed that over 88% of normal mammograms and 80% of abnormal mammograms can be correctly identified. The approach was also successfully applied to Synthetic Aperture Radar (SAR) and Ground Penetrating Radar (GPR) images for target detection.

Effect of Valve Pressure Drop in Exergy Analysis of C2+ Recovery Plants Refrigeration Cycles

This paper provides an exergy analysis of the multistage refrigeration cycle used for C2+ recovery plant. The behavior of an industrial refrigeration cycle with refrigerant propane has been investigated by the exergy method. A computational model based on the exergy analysis is presented for the investigation of the effects of the valves on the exergy losses, the second law of efficiency, and the coefficient of performance (COP) of a vapor compression refrigeration cycle. The equations of exergy destruction and exergetic efficiency for the main cycle components such as evaporators, condensers, compressors, and expansion valves are developed. The relations for the total exergy destruction in the cycle and the cycle exergetic efficiency are obtained. An ethane recovery unit with its refrigeration cycle has been simulated to prepare the exergy analysis. Using a typical actual work input value; the exergetic efficiency of the refrigeration cycle is determined to be 39.90% indicating a great potential for improvements. The simulation results reveal that the exergetic efficiencies of the heat exchanger and expansion sections get the lowest rank among the other compartments of refrigeration cycle. Refrigeration calculations have been carried out through the analysis of T–S and P–H diagrams where coefficient of performance (COP) was obtained as 1.85. The novelty of this article includes the effect and sensitivity analysis of molar flow, pressure drops and temperature on the exergy efficiency and coefficient of performance of the cycle.

Automatic Translation of Ada-ECATNet Using Rewriting Logic

One major difficulty that faces developers of concurrent and distributed software is analysis for concurrency based faults like deadlocks. Petri nets are used extensively in the verification of correctness of concurrent programs. ECATNets are a category of algebraic Petri nets based on a sound combination of algebraic abstract types and high-level Petri nets. ECATNets have 'sound' and 'complete' semantics because of their integration in rewriting logic and its programming language Maude. Rewriting logic is considered as one of very powerful logics in terms of description, verification and programming of concurrent systems We proposed previously a method for translating Ada-95 tasking programs to ECATNets formalism (Ada-ECATNet) and we showed that ECATNets formalism provides a more compact translation for Ada programs compared to the other approaches based on simple Petri nets or Colored Petri nets. We showed also previously how the ECATNet formalism offers to Ada many validation and verification tools like simulation, Model Checking, accessibility analysis and static analysis. In this paper, we describe the implementation of our translation of the Ada programs into ECATNets.

Implementation of a “DIVA“ Concept withspecific Elisa Kits; When Subunit H5 Avian Influenza Vaccine is used

The main objective of this study was to demonstrate that differentiation of infected and vaccinated animals (DIVA) strategy using different ELISA tests is possible when a subunit vaccine (Haemagglutinin protein) is used to prevent Avian influenza. Special emphasis was placed on the differentiation in the serological response to different components of the AIV (Nucleoprotein, Neuraminidase, Haemagglutinin, Nucleocapsid) between chickens that were vaccinated with a whole virus kill vaccine and recombinant vaccine. Furthermore, the potential use of this DIVA strategy using ELISA assays to detect Neuraminidase 1 (N1) was analyzed as strategy in countries where the field virus is H5N1 and the vaccine used is formulated with H5N2. Detection of AIV-s antibodies to any component in serum was negative for all animals on the study days 0-13. At study day 14 the titers of antibodies against Nucleoprotein (NP) and Nucleocapsid (NC) rose in the experimental groups vaccinated with Volvac® AI KV and were negatives during all the trial in the experimental groups vaccinated with a subunit H5; significant statistically differences were observed between these groups (p < 0.05). The seroconversion either Haemagglutinin or Neuraminidase was evident after 21 days post-vaccination in the experimental groups vaccinated with the respective viral fraction. Regarding the main aim of this study and according with the results that were obtained, use a combination of different ELISA test as a DIVA strategy is feasible when the vaccination is carry out with a subunit H5 vaccine. Also is possible to use the ELISA kit to detect Neuraminidase (either N1 or N2) as a DIVA concept in countries where H5N1 is present and the vaccination programs are done with H5N2 vaccine.

Decision-Making Criteria of PPP Projects: Stakeholder Theoretic Perspective

Any decision-making is based on certain theory. Taking the public rental housing in Chongqing municipality as an example, this essay states that the stakeholder theory can provide innovative criteria and evaluation methods for Public Private Partnership (PPP) projects. It gives an analysis of how to choose decision-making criteria for different stakeholders in the PPP model and what measures to take to meet the criteria to form “symbiotic" decision-making mode through contracts and to boost the application of PPP model in large-scale public programs in China.

Binary Decision Diagrams: An Improved Variable Ordering using Graph Representation of Boolean Functions

This paper presents an improved variable ordering method to obtain the minimum number of nodes in Reduced Ordered Binary Decision Diagrams (ROBDD). The proposed method uses the graph topology to find the best variable ordering. Therefore the input Boolean function is converted to a unidirectional graph. Three levels of graph parameters are used to increase the probability of having a good variable ordering. The initial level uses the total number of nodes (NN) in all the paths, the total number of paths (NP) and the maximum number of nodes among all paths (MNNAP). The second and third levels use two extra parameters: The shortest path among two variables (SP) and the sum of shortest path from one variable to all the other variables (SSP). A permutation of the graph parameters is performed at each level for each variable order and the number of nodes is recorded. Experimental results are promising; the proposed method is found to be more effective in finding the variable ordering for the majority of benchmark circuits.

Extended Well-Founded Semantics in Bilattices

One of the most used assumptions in logic programming and deductive databases is the so-called Closed World Assumption (CWA), according to which the atoms that cannot be inferred from the programs are considered to be false (i.e. a pessimistic assumption). One of the most successful semantics of conventional logic programs based on the CWA is the well-founded semantics. However, the CWA is not applicable in all circumstances when information is handled. That is, the well-founded semantics, if conventionally defined, would behave inadequately in different cases. The solution we adopt in this paper is to extend the well-founded semantics in order for it to be based also on other assumptions. The basis of (default) negative information in the well-founded semantics is given by the so-called unfounded sets. We extend this concept by considering optimistic, pessimistic, skeptical and paraconsistent assumptions, used to complete missing information from a program. Our semantics, called extended well-founded semantics, expresses also imperfect information considered to be missing/incomplete, uncertain and/or inconsistent, by using bilattices as multivalued logics. We provide a method of computing the extended well-founded semantics and show that Kripke-Kleene semantics is captured by considering a skeptical assumption. We show also that the complexity of the computation of our semantics is polynomial time.

Conflict, Confusion, Choice: A Phenomenological Approach to Acts of Corruption

Public sector corruption has long-term and damaging effects that are deep and broad. Addressing corruption relies on understanding the drivers that precipitate acts of corruption and developing educational programs that target areas of vulnerability. This paper provides an innovative approach to explore the nature of corruption by drawing on the perceptions and ideas of a group of public servants who have been part of a corruption investigation. The paper examines these reflections through the ideas of Pierre Bourdieu and Alfred Schutz to point to some of the steps that can lead to corrupt activity. The paper demonstrates that phenomenological inquiry is useful in the exploration of corruption and, as a theoretical framework, it highlights that corruption emerges through a combination of conflict, doubt and uncertainty. The paper calls for anti-corruption education programs to be attentive to way in which these conditions can influence the steps into corruption.

Using Interpretive Structural Modeling to Determine the Relationships among Knowledge Management Criteria inside Malaysian Organizations

This paper is concerned with the establishment of relationships among knowledge management (KM) criteria that will ensure an essential foundation to evaluate KM outcomes. The major issue under investigation is to assess the popularity of criteria within organizations and to establish a structure of criteria for measuring KM results. An empirical survey was conducted among Malaysian organizations to investigate KM criteria for measuring success of KM initiatives. Therefore, knowledge workers as the respondents were targeted to establish a structure of criteria for evaluating KM outcomes. An established structure of criteria based on the Interpretive Structural Modeling (ISM) is used to map criteria relationships inside organizations. This structure is portrayed to identify that how these set of criteria are related. This network schema should be investigated and implemented to promote innovation and improve enterprise performance. To the researchers, this survey has significant insights into relationship between KM programs and business success.

Determining the Best Method of Stability Landslide by Using of DSS (Case Study: Landslide in Hasan Salaran, Kurdistan Province in Iran)

One of the processes of slope that occurs every year in Iran and some parts of world and cause a lot of criminal and financial harms is called landslide. They are plenty of method to stability landslide in soil and rock slides. The use of the best method with the least cost and in the shortest time is important for researchers. In this research, determining the best method of stability is investigated by using of Decision Support systems. DSS is made for this purpose and was used (for Hasan Salaran area in Kurdistan). Field study data from topography, slope, geology, geometry of landslide and the related features was used. The related data entered decision making managements programs (DSS) (ALES).Analysis of mass stability indicated the instability potential at present. Research results show that surface and sub surface drainage the best method of stabilizing. Analysis of stability shows that acceptable increase in security coefficient is a consequence of drainage.

Estimating Development Time of Software Projects Using a Neuro Fuzzy Approach

Software estimation accuracy is among the greatest challenges for software developers. This study aimed at building and evaluating a neuro-fuzzy model to estimate software projects development time. The forty-one modules developed from ten programs were used as dataset. Our proposed approach is compared with fuzzy logic and neural network model and Results show that the value of MMRE (Mean of Magnitude of Relative Error) applying neuro-fuzzy was substantially lower than MMRE applying fuzzy logic and neural network.

FHOJ: A New Java Benchmark Framework

There are some existing Java benchmarks, application benchmarks as well as micro benchmarks or mixture both of them,such as: Java Grande, Spec98, CaffeMark, HBech, etc. But none of them deal with behaviors of multi tasks operating systems. As a result, the achieved outputs are not satisfied for performance evaluation engineers. Behaviors of multi tasks operating systems are based on a schedule management which is employed in these systems. Different processes can have different priority to share the same resources. The time is measured by estimating from applications started to it is finished does not reflect the real time value which the system need for running those programs. New approach to this problem should be done. Having said that, in this paper we present a new Java benchmark, named FHOJ benchmark, which directly deals with multi tasks behaviors of a system. Our study shows that in some cases, results from FHOJ benchmark are far more reliable in comparison with some existing Java benchmarks.

64 bit Computer Architectures for Space Applications – A study

The more recent satellite projects/programs makes extensive usage of real – time embedded systems. 16 bit processors which meet the Mil-Std-1750 standard architecture have been used in on-board systems. Most of the Space Applications have been written in ADA. From a futuristic point of view, 32 bit/ 64 bit processors are needed in the area of spacecraft computing and therefore an effort is desirable in the study and survey of 64 bit architectures for space applications. This will also result in significant technology development in terms of VLSI and software tools for ADA (as the legacy code is in ADA). There are several basic requirements for a special processor for this purpose. They include Radiation Hardened (RadHard) devices, very low power dissipation, compatibility with existing operational systems, scalable architectures for higher computational needs, reliability, higher memory and I/O bandwidth, predictability, realtime operating system and manufacturability of such processors. Further on, these may include selection of FPGA devices, selection of EDA tool chains, design flow, partitioning of the design, pin count, performance evaluation, timing analysis etc. This project deals with a brief study of 32 and 64 bit processors readily available in the market and designing/ fabricating a 64 bit RISC processor named RISC MicroProcessor with added functionalities of an extended double precision floating point unit and a 32 bit signal processing unit acting as co-processors. In this paper, we emphasize the ease and importance of using Open Core (OpenSparc T1 Verilog RTL) and Open “Source" EDA tools such as Icarus to develop FPGA based prototypes quickly. Commercial tools such as Xilinx ISE for Synthesis are also used when appropriate.