Lease Agreement in the European Countries

This paper present lease agreement regulations in selected European countries. The lease agreement has a long history and now is one of the main ways to manage agricultural lands in Europe. The analysis of individual regulations, which has been done, indicates that this agreement is very important to build social relations in agriculture and society. This article provides an analysis of the legal regulations concerning the lease in France, Spain, Switzerland, Ukraine and Italy. Article is example of study of the legal regulations and can be used for legal changes in individual countries.

Effective Defect Prevention Approach in Software Process for Achieving Better Quality Levels

Defect prevention is the most vital but habitually neglected facet of software quality assurance in any project. If functional at all stages of software development, it can condense the time, overheads and wherewithal entailed to engineer a high quality product. The key challenge of an IT industry is to engineer a software product with minimum post deployment defects. This effort is an analysis based on data obtained for five selected projects from leading software companies of varying software production competence. The main aim of this paper is to provide information on various methods and practices supporting defect detection and prevention leading to thriving software generation. The defect prevention technique unearths 99% of defects. Inspection is found to be an essential technique in generating ideal software generation in factories through enhanced methodologies of abetted and unaided inspection schedules. On an average 13 % to 15% of inspection and 25% - 30% of testing out of whole project effort time is required for 99% - 99.75% of defect elimination. A comparison of the end results for the five selected projects between the companies is also brought about throwing light on the possibility of a particular company to position itself with an appropriate complementary ratio of inspection testing.

A Study of Calcination and Carbonation of Cockle Shell

Calcium oxide (CaO) as carbon dioxide (CO2) adsorbent at the elevated temperature has been very well-received thus far. The CaO can be synthesized from natural calcium carbonate (CaCO3) sources through the reversible calcination-carbonation process. In the study, cockle shell has been selected as CaO precursors. The objectives of the study are to investigate the performance of calcination and carbonation with respect to different temperature, heating rate, particle size and the duration time. Overall, better performance is shown at the calcination temperature of 850oC for 40 minutes, heating rate of 20oC/min, particle size of < 0.125mm and the carbonation temperature is at 650oC. The synthesized materials have been characterized by nitrogen physisorption and surface morphology analysis. The effectiveness of the synthesized cockle shell in capturing CO2 (0.72 kg CO2/kg adsorbent) which is comparable to the commercialized adsorbent (0.60 kg CO2/kg adsorbent) makes them as the most promising materials for CO2 capture.

Reliability Assessment of Bangladesh Power System Using Recursive Algorithm

An electric utility-s main concern is to plan, design, operate and maintain its power supply to provide an acceptable level of reliability to its users. This clearly requires that standards of reliability be specified and used in all three sectors of the power system, i.e., generation, transmission and distribution. That is why reliability of a power system is always a major concern to power system planners. This paper presents the reliability analysis of Bangladesh Power System (BPS). Reliability index, loss of load probability (LOLP) of BPS is evaluated using recursive algorithm and considering no de-rated states of generators. BPS has sixty one generators and a total installed capacity of 5275 MW. The maximum demand of BPS is about 5000 MW. The relevant data of the generators and hourly load profiles are collected from the National Load Dispatch Center (NLDC) of Bangladesh and reliability index 'LOLP' is assessed for the period of last ten years.

Evaluation of Graph-based Analysis for Forest Fire Detections

Spatial outliers in remotely sensed imageries represent observed quantities showing unusual values compared to their neighbor pixel values. There have been various methods to detect the spatial outliers based on spatial autocorrelations in statistics and data mining. These methods may be applied in detecting forest fire pixels in the MODIS imageries from NASA-s AQUA satellite. This is because the forest fire detection can be referred to as finding spatial outliers using spatial variation of brightness temperature. This point is what distinguishes our approach from the traditional fire detection methods. In this paper, we propose a graph-based forest fire detection algorithm which is based on spatial outlier detection methods, and test the proposed algorithm to evaluate its applicability. For this the ordinary scatter plot and Moran-s scatter plot were used. In order to evaluate the proposed algorithm, the results were compared with the MODIS fire product provided by the NASA MODIS Science Team, which showed the possibility of the proposed algorithm in detecting the fire pixels.

Prospective Mathematics Teachers' Views about Using Flash Animations in Mathematics Lessons

The purpose of the study is to determine secondary prospective mathematics teachers- views related to using flash animations in mathematics lessons and to reveal how the sample presentations towards different mathematical concepts altered their views. This is a case study involving three secondary prospective mathematics teachers from a state university in Turkey. The data gathered from two semi-structural interviews. Findings revealed that these animations help understand mathematics meaningfully, relate mathematics and real world, visualization, and comprehend the importance of mathematics. The analysis of the data indicated that the sample presentations enhanced participants- views about using flash animations in mathematics lessons.

Analysis Fraction Flow of Water versus Cumulative Oil Recoveries Using Buckley Leverett Method

To derive the fractional flow equation oil displacement will be assumed to take place under the so-called diffusive flow condition. The constraints are that fluid saturations at any point in the linear displacement path are uniformly distributed with respect to thickness; this allows the displacement to be described mathematically in one dimension. The simultaneous flow of oil and water can be modeled using thickness averaged relative permeability, along the centerline of the reservoir. The condition for fluid potential equilibrium is simply that of hydrostatic equilibrium for which the saturation distribution can be determined as a function of capillary pressure and therefore, height. That is the fluids are distributed in accordance with capillary-gravity equilibrium. This paper focused on the fraction flow of water versus cumulative oil recoveries using Buckley Leverett method. Several field cases have been developed to aid in analysis. Producing watercut (at surface conditions) will be compared with the cumulative oil recovery at breakthrough for the flowing fluid.

Thermal and Morphological Evaluation of Chemically Pretreated Sugarcane Bagasse

Enzymatic hydrolysis is one of the major steps involved in the conversion from sugarcane bagasse to yield ethanol. This process offers potential for yields and selectivity higher, lower energy costs and milder operating conditions than chemical processes. However, the presence of some factors such as lignin content, crystallinity degree of the cellulose, and particle sizes, limits the digestibility of the cellulose present in the lignocellulosic biomasses. Pretreatment aims to improve the access of the enzyme to the substrate. In this study sugarcane bagasse was submitted chemical pretreatment that consisted of two consecutive steps, the first with dilute sulfuric acid (1 % (v/v) H2SO4), and the second with alkaline solutions with different concentrations of NaOH (1, 2, 3 and 4 % (w/v)). Thermal Analysis (TG/ DTG and DTA) was used to evaluate hemicellulose, cellulose and lignin contents in the samples. Scanning Electron Microscopy (SEM) was used to evaluate the morphological structures of the in natura and chemically treated samples. Results showed that pretreatments were effective in chemical degradation of lignocellulosic materials of the samples, and also was possible to observe the morphological changes occurring in the biomasses after pretreatments.

Approach to Implementation of Power Management with Load Prioritizations in Modern Civil Aircraft

Any use of energy in industrial productive activities is combined with various environment impacts. Withintransportation, this fact was not only found among land transport, railways and maritime transport, but also in the air transport industry. An effective climate protection requires strategies and measures for reducing all greenhouses gas emissions, in particular carbon dioxide, and must take into account the economic, ecologic and social aspects. It seem simperative now to develop and manufacture environmentally friendly products and systems, to reduce consumption and use less resource, and to save energy and power. Today-sproducts could better serve these requirements taking into account the integration of a power management system into the electrical power system.This paper gives an overview of an approach ofpower management with load prioritization in modernaircraft. Load dimensioning and load management strategies on current civil aircraft will be presented and used as a basis for the proposed approach.

Encryption Efficiency Analysis and Security Evaluation of RC6 Block Cipher for Digital Images

This paper investigates the encryption efficiency of RC6 block cipher application to digital images, providing a new mathematical measure for encryption efficiency, which we will call the encryption quality instead of visual inspection, The encryption quality of RC6 block cipher is investigated among its several design parameters such as word size, number of rounds, and secret key length and the optimal choices for the best values of such design parameters are given. Also, the security analysis of RC6 block cipher for digital images is investigated from strict cryptographic viewpoint. The security estimations of RC6 block cipher for digital images against brute-force, statistical, and differential attacks are explored. Experiments are made to test the security of RC6 block cipher for digital images against all aforementioned types of attacks. Experiments and results verify and prove that RC6 block cipher is highly secure for real-time image encryption from cryptographic viewpoint. Thorough experimental tests are carried out with detailed analysis, demonstrating the high security of RC6 block cipher algorithm. So, RC6 block cipher can be considered to be a real-time secure symmetric encryption for digital images.

A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation

Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.

Novel Direct Flux and Torque Control of Optimally Designed 6 Phase Reluctance Machine with Special Current Waveform

In this paper the principle, basic torque theory and design optimisation of a six-phase reluctance dc machine are considered. A trapezoidal phase current waveform for the machine drive is proposed and evaluated to minimise ripple torque. Low cost normal laminated salient-pole rotors with and without slits and chamfered poles are investigated. The six-phase machine is optimised in multi-dimensions by linking the finite-element analysis method directly with an optimisation algorithm; the objective function is to maximise the torque per copper losses of the machine. The armature reaction effect is investigated in detail and found to be severe. The measured and calculated torque performances of a 35 kW optimum designed six-phase reluctance dc machine drive are presented.

Dynamic Threshold Adjustment Approach For Neural Networks

The use of neural networks for recognition application is generally constrained by their inherent parameters inflexibility after the training phase. This means no adaptation is accommodated for input variations that have any influence on the network parameters. Attempts were made in this work to design a neural network that includes an additional mechanism that adjusts the threshold values according to the input pattern variations. The new approach is based on splitting the whole network into two subnets; main traditional net and a supportive net. The first deals with the required output of trained patterns with predefined settings, while the second tolerates output generation dynamically with tuning capability for any newly applied input. This tuning comes in the form of an adjustment to the threshold values. Two levels of supportive net were studied; one implements an extended additional layer with adjustable neuronal threshold setting mechanism, while the second implements an auxiliary net with traditional architecture performs dynamic adjustment to the threshold value of the main net that is constructed in dual-layer architecture. Experiment results and analysis of the proposed designs have given quite satisfactory conducts. The supportive layer approach achieved over 90% recognition rate, while the multiple network technique shows more effective and acceptable level of recognition. However, this is achieved at the price of network complexity and computation time. Recognition generalization may be also improved by accommodating capabilities involving all the innate structures in conjugation with Intelligence abilities with the needs of further advanced learning phases.

Phase Error Accumulation Methodology for On-Chip Cell Characterization

This paper describes the design of new method of propagation delay measurement in micro and nanostructures during characterization of ASIC standard library cell. Providing more accuracy timing information about library cell to the design team we can improve a quality of timing analysis inside of ASIC design flow process. Also, this information could be very useful for semiconductor foundry team to make correction in technology process. By comparison of the propagation delay in the CMOS element and result of analog SPICE simulation. It was implemented as digital IP core for semiconductor manufacturing process. Specialized method helps to observe the propagation time delay in one element of the standard-cell library with up-to picoseconds accuracy and less. Thus, the special useful solutions for VLSI schematic to parameters extraction, basic cell layout verification, design simulation and verification are announced.

Design for Manufacturability and Concurrent Engineering for Product Development

In the 1980s, companies began to feel the effect of three major influences on their product development: newer and innovative technologies, increasing product complexity and larger organizations. And therefore companies were forced to look for new product development methods. This paper tries to focus on the two of new product development methods (DFM and CE). The aim of this paper is to see and analyze different product development methods specifically on Design for Manufacturability and Concurrent Engineering. Companies can achieve and be benefited by minimizing product life cycle, cost and meeting delivery schedule. This paper also presents simplified models that can be modified and used by different companies based on the companies- objective and requirements. Methodologies that are followed to do this research are case studies. Two companies were taken and analysed on the product development process. Historical data, interview were conducted on these companies in addition to that, Survey of literatures and previous research works on similar topics has been done during this research. This paper also tries to show the implementation cost benefit analysis and tries to calculate the implementation time. From this research, it has been found that the two companies did not achieve the delivery time to the customer. Some of most frequently coming products are analyzed and 50% to 80 % of their products are not delivered on time to the customers. The companies are following the traditional way of product development that is sequentially design and production method, which highly affect time to market. In the case study it is found that by implementing these new methods and by forming multi disciplinary team in designing and quality inspection; the company can reduce the workflow steps from 40 to 30.

Feature Based Unsupervised Intrusion Detection

The goal of a network-based intrusion detection system is to classify activities of network traffics into two major categories: normal and attack (intrusive) activities. Nowadays, data mining and machine learning plays an important role in many sciences; including intrusion detection system (IDS) using both supervised and unsupervised techniques. However, one of the essential steps of data mining is feature selection that helps in improving the efficiency, performance and prediction rate of proposed approach. This paper applies unsupervised K-means clustering algorithm with information gain (IG) for feature selection and reduction to build a network intrusion detection system. For our experimental analysis, we have used the new NSL-KDD dataset, which is a modified dataset for KDDCup 1999 intrusion detection benchmark dataset. With a split of 60.0% for the training set and the remainder for the testing set, a 2 class classifications have been implemented (Normal, Attack). Weka framework which is a java based open source software consists of a collection of machine learning algorithms for data mining tasks has been used in the testing process. The experimental results show that the proposed approach is very accurate with low false positive rate and high true positive rate and it takes less learning time in comparison with using the full features of the dataset with the same algorithm.

A Simple Adaptive Algorithm for Norm-Constrained Optimization

In this paper we propose a simple adaptive algorithm iteratively solving the unit-norm constrained optimization problem. Instead of conventional parameter norm based normalization, the proposed algorithm incorporates scalar normalization which is computationally much simpler. The analysis of stationary point is presented to show that the proposed algorithm indeed solves the constrained optimization problem. The simulation results illustrate that the proposed algorithm performs as good as conventional ones while being computationally simpler.

Analysis of the Long-term Effect of Office Lighting Environment on Human Reponses

This study aims to discuss the effect of illumination and the color temperature of the lighting source under the office lighting environment on human psychological and physiological responses. In this study, 21 healthy participants were selected, and the Ryodoraku measurement system was utilized to measure their skin resistance change.The findings indicated that the effect of the color temperature of the lighting source on human physiological responses is significant within 90 min after turning the lights on; while after 90 min the effect of illumination on human physiological responses is higher than that of the color temperature. Moreover, the cardiovascular, digestive and endocrine systems are prone to be affected by the indoor lighting environment. During the long-term exposure to high intensity of illumination and high color temperature (2000Lux -6500K), the effect on the psychological responses turned moderate after the human visual system adopted to the lighting environment. However, the effect of the Ryodoraku value on human physiological responses was more significant with the increase of perceptive time. The effect of long time exposure to a lighting environment on the physiological responses is greater than its effect on the psychological responses. This conclusion is different from the traditional public viewpoint that the effect on the psychological responses is greater.

Groundwater Quality and the Sources of Pollution in Baghan Watershed, Iran

The protection of groundwater resources is the great important many semiarid and arid environments. Baghan watershed is located in the north of Kangan in the Boshehr province in Iran. The groundwater resources have a vital role in supplying agricultural, drinking, domestic and industrial water demand in Baghan watershed. For our investigation into the water quality we collected 30 samples to chemical and physical analysis. The result showed the marl and evaporation deposits that contain anhydrite and gypsum is the main source of groundwater pollution, and one part of the groundwater was polluted by oil and gas industrial. Another part of the groundwater was contaminated by urban waste water. The electrical conductivity and captions and anions increased around of towns and gas refinery. Although the negative impact of untreated domestic wastewater is relatively low but the results showed strongly the negative impact of wastewater refinery is very considerable. This negative impact increased in downstream due to shallow aquifer. Additionally, the agents that adversely affect the quality of groundwater come from a variety of sources, including geology, domestic wastewater and the Jam refinery in Baghan watershed.

Negative Impact of Bacteria Legionella Pneumophila in Hot Water Distribution Systems on Human Health

Safe drinking water is one of the biggest issues facing the planet this century. The primary aim of this paper is to present our research focused on theoretical and experimental analysis of potable water and in-building water distribution systems from the point of view of microbiological risk on the basis of confrontation between the theoretical analysis and synthesis of gathered information in conditions of the Slovak Republic. The presence of the bacteria Legionella in water systems, especially in hot water distribution system, represents in terms of health protection of inhabitants the crucial problem which cannot be overlooked. Legionella pneumophila discovery, its classification and its influence on installations inside buildings are relatively new. There are a lot of guidelines and regulations developed in many individual countries for the design, operation and maintenance for tap water systems to avoid the growth of bacteria Legionella pneumophila, but in Slovakia we don-t have any. The goal of this paper is to show the necessity of prevention and regulations for installations inside buildings verified by simulation methods.