FPGA Implementation of a Vision-Based Blind Spot Warning System

Vision-based intelligent vehicle applications often require large amounts of memory to handle video streaming and image processing, which in turn increases complexity of hardware and software. This paper presents an FPGA implement of a vision-based blind spot warning system. Using video frames, the information of the blind spot area turns into one-dimensional information. Analysis of the estimated entropy of image allows the detection of an object in time. This idea has been implemented in the XtremeDSP video starter kit. The blind spot warning system uses only 13% of its logic resources and 95k bits block memory, and its frame rate is over 30 frames per sec (fps).

Analysis of Sequence Moves in Successful Chess Openings Using Data Mining with Association Rules

Chess is one of the indoor games, which improves the level of human confidence, concentration, planning skills and knowledge. The main objective of this paper is to help the chess players to improve their chess openings using data mining techniques. Budding Chess Players usually do practices by analyzing various existing openings. When they analyze and correlate thousands of openings it becomes tedious and complex for them. The work done in this paper is to analyze the best lines of Blackmar- Diemer Gambit(BDG) which opens with White D4... using data mining analysis. It is carried out on the collection of winning games by applying association rules. The first step of this analysis is assigning variables to each different sequence moves. In the second step, the sequence association rules were generated to calculate support and confidence factor which help us to find the best subsequence chess moves that may lead to winning position.

Effective Online Staff Training: Is This Possible?

The purpose of this paper is to consider the introduction of online courses to replace the current classroom-based staff training. The current training is practical, and must be completed before access to the financial computer system is authorized. The long term objective is to measure the efficacy, effectiveness and efficiency of the training, and to establish whether a transfer of knowledge back to the workplace has occurred. This paper begins with an overview explaining the importance of staff training in an evolving, competitive business environment and defines the problem facing this particular organization. A summary of the literature review is followed by a brief discussion of the research methodology and objective. The implementation of the alpha version of the online course is then described. This paper may be of interest to those seeking insights into, or new theory regarding, practical interventions of online learning in the real world.

Sustainable and Ecological Designs of the Built Environment

This paper reviews designs of the built environment from a sustainability perspective, emphasizing their importance in achieving ecological and sustainable economic objectives. The built environment has traditionally resulted in loss of biodiversity, extinction of some species, climate change, excessive water use, land degradation, space depletion, waste accumulation, energy consumption and environmental pollution. Materials used like plastics, metals, bricks, concrete, cement, natural aggregates, glass and plaster have wreaked havoc on the earth´s resources, since they have high levels of embodied energy hence not sustainable. Additional resources are consumed during use and disposal phases. Proposed designs for sustainability solutions include: ecological sanitation and eco-efficiency systems that ensure social, economic, environmental and technical sustainability. Renewable materials and energy systems, passive cooling and heating systems and material and energy reduction, reuse and recycling can improve the sector. These ideas are intended to inform the field of ecological design of the built environment.

Cost Optimization of Concentric Braced Steel Building Structures

Seismic design may require non-conventional concept, due to the fact that the stiffness and layout of the structure have a great effect on the overall structural behaviour, on the seismic load intensity as well as on the internal force distribution. To find an economical and optimal structural configuration the key issue is the optimal design of the lateral load resisting system. This paper focuses on the optimal design of regular, concentric braced frame (CBF) multi-storey steel building structures. The optimal configurations are determined by a numerical method using genetic algorithm approach, developed by the authors. Aim is to find structural configurations with minimum structural cost. The design constraints of objective function are assigned in accordance with Eurocode 3 and Eurocode 8 guidelines. In this paper the results are presented for various building geometries, different seismic intensities, and levels of energy dissipation.

Design Optimization of Aerocapture with Aerodynamic-Environment-Adaptive Variable Geometry Flexible Aeroshell

This paper proposes the concept of aerocapture with aerodynamic-environment-adaptive variable geometry flexible aeroshell that vehicle deploys. The flexible membrane is composed of thin-layer film or textile as its aeroshell in order to solve some problems obstructing realization of aerocapture technique. Multi-objective optimization study is conducted to investigate solutions and derive design guidelines. As a result, solutions which can avoid aerodynamic heating and enlarge the corridor width up to 10% are obtained successfully, so that the effectiveness of this concept can be demonstrated. The deformation-use optimum solution changes its drag coefficient from 1.6 to 1.1, along with the change in dynamic pressure. Moreover, optimization results show that deformation-use solution requires the membrane for which upper temperature limit and strain limit are more than 700 K and 120%, respectively, and elasticity (Young-s modulus) is of order of 106 Pa.

A Perceptually Optimized Wavelet Embedded Zero Tree Image Coder

In this paper, we propose a Perceptually Optimized Embedded ZeroTree Image Coder (POEZIC) that introduces a perceptual weighting to wavelet transform coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to the coding quality obtained using the SPIHT algorithm only. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEZIC quality assessment. Our POEZIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) luminance masking and Contrast masking, 2) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting, 3) the Wavelet Error Sensitivity WES used to reduce the perceptual quantization errors. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.

Evaluation of the Burden of Taxation Received by Households in Lithuania

Many foreign and Lithuanian scientists, analyzing the evaluation of the tax system in respect of the burden of taxation, agree that the latter, in principle, depends on how many individuals and what units of the residents constitute a household. Therefore, the aim of scientific research is to substantiate or to deny the significance of a household, but not a resident, as a statistical unit, during the evaluation of tax system, to be precise, determination of the value of the burden of taxation. A performed scientific research revealed that evaluation of the tax system in respect of a household, but not a resident, as a statistical unit, allows not only to evaluate the efficiency of the tax system more objectively, but also to forecast practicably existing poverty line, burden of taxation, and to capacitate the initiation of efficient decisions in social and tax fields creating the environment of existence.

Reasoning with Dynamic Domains and Computer Security

Representing objects in a dynamic domain is essential in commonsense reasoning under some circumstances. Classical logics and their nonmonotonic consequences, however, are usually not able to deal with reasoning with dynamic domains due to the fact that every constant in the logical language denotes some existing object in the static domain. In this paper, we explore a logical formalization which allows us to represent nonexisting objects in commonsense reasoning. A formal system named N-theory is proposed for this purpose and its possible application in computer security is briefly discussed.

Automated Separation of Organic Liquids through Their Boiling Points

This paper discuss the separation of the miscible liquids by means of fractional distillation. For complete separation of liquids, the process of heating, condensation, separation and storage is done automatically to achieve the objective. PIC micro-controller has been used to control each and every process of the work. The controller also controls the storage process by activating and deactivating the conveyors. The liquids are heated which on reaching their respective boiling points evaporate and enter the condensation chamber where they convert into liquid. The liquids are then directed to their respective tanks by means of stepper motor which moves in three directions, each movement into different tank. The tank on filling sends the signal to controller which then opens the solenoid valves. The tank is emptied into the beakers below the nozzle. As the beaker filled, the nozzle closes and the conveyors come into operation. The filled beaker is replaced by an empty beaker from behind. The work can be used in oil industries, chemical industries and paint industries.

Effective Traffic Lights Recognition Method for Real Time Driving Assistance Systemin the Daytime

This paper presents an effective traffic lights recognition method at the daytime. First, Potential Traffic Lights Detector (PTLD) use whole color source of YCbCr channel image and make each binary image of green and red traffic lights. After PTLD step, Shape Filter (SF) use to remove noise such as traffic sign, street tree, vehicle, and building. At this time, noise removal properties consist of information of blobs of binary image; length, area, area of boundary box, etc. Finally, after an intermediate association step witch goal is to define relevant candidates region from the previously detected traffic lights, Adaptive Multi-class Classifier (AMC) is executed. The classification method uses Haar-like feature and Adaboost algorithm. For simulation, we are implemented through Intel Core CPU with 2.80 GHz and 4 GB RAM and tested in the urban and rural roads. Through the test, we are compared with our method and standard object-recognition learning processes and proved that it reached up to 94 % of detection rate which is better than the results achieved with cascade classifiers. Computation time of our proposed method is 15 ms.

Design of Nonlinear Observer by Using Augmented Linear System based on Formal Linearization of Polynomial Type

The objective of this study is to propose an observer design for nonlinear systems by using an augmented linear system derived by application of a formal linearization method. A given nonlinear differential equation is linearized by the formal linearization method which is based on Taylor expansion considering up to the higher order terms, and a measurement equation is transformed into an augmented linear one. To this augmented dimensional linear system, a linear estimation theory is applied and a nonlinear observer is derived. As an application of this method, an estimation problem of transient state of electric power systems is studied, and its numerical experiments indicate that this observer design shows remarkable performances for nonlinear systems.

Brain MRI Segmentation and Lesions Detection by EM Algorithm

In Multiple Sclerosis, pathological changes in the brain results in deviations in signal intensity on Magnetic Resonance Images (MRI). Quantitative analysis of these changes and their correlation with clinical finding provides important information for diagnosis. This constitutes the objective of our work. A new approach is developed. After the enhancement of images contrast and the brain extraction by mathematical morphology algorithm, we proceed to the brain segmentation. Our approach is based on building statistical model from data itself, for normal brain MRI and including clustering tissue type. Then we detect signal abnormalities (MS lesions) as a rejection class containing voxels that are not explained by the built model. We validate the method on MR images of Multiple Sclerosis patients by comparing its results with those of human expert segmentation.

Assessing the Effect of the Shift of Rural Labor towards Non-Agricultural Sectors on Rice Cultivation in the African Environment: Evidence from Sierra Leone

The crop rice is the staple food of most Sierra Leone with no close substitute. However, its cultivation has been on its last legs over the years. The decline in the domestic rice cultivation has had vicious socio-economic implications such as hiking consumer prices, balance of payment dilemmas with debt burden. The objective of this study is thus, to assess the effect of the shift of rural labour towards non-agricultural sectors on rice cultivation. The tools utilized for analyzing the problem under consideration involved a thorough descriptive statistics and generalized linear model using OLS technique. Increased rural population was established positive and significant in affecting rice cultivation. Fertilizer utilization was insignificant in rice cultivation. For reducing the shift of rural labor force towards nonagricultural sectors, the government should make the agricultural sector very lucrative.

Evaluation of a New Method for Detection of Kidney Stone during Laparoscopy Using 3D Conceptual Modeling

Minimally invasive surgery (MIS) is now being widely used as a preferred choice for various types of operations. The need to detect various tactile properties, justifies the key role of tactile sensing that is currently missing in MIS. In this regard, Laparoscopy is one of the methods of minimally invasive surgery that can be used in kidney stone removal surgeries. At this moment, determination of the exact location of stone during laparoscopy is one of the limitations of this method that no scientific solution has been found for so far. Artificial tactile sensing is a new method for obtaining the characteristics of a hard object embedded in a soft tissue. Artificial palpation is an important application of artificial tactile sensing that can be used in different types of surgeries. In this study, a new method for determining the exact location of stone during laparoscopy is presented. In the present study, the effects of stone existence on the surface of kidney were investigated using conceptual 3D model of kidney containing a simulated stone. Having imitated palpation and modeled it conceptually, indications of stone existence that appear on the surface of kidney were determined. A number of different cases were created and solved by the software and using stress distribution contours and stress graphs, it is illustrated that the created stress patterns on the surface of kidney show not only the existence of stone inside, but also its exact location. So three-dimensional analysis leads to a novel method of predicting the exact location of stone and can be directly applied to the incorporation of tactile sensing in artificial palpation, helping surgeons in non-invasive procedures.

Vehicle Velocity Estimation for Traffic Surveillance System

This paper describes an algorithm to estimate realtime vehicle velocity using image processing technique from the known camera calibration parameters. The presented algorithm involves several main steps. First, the moving object is extracted by utilizing frame differencing technique. Second, the object tracking method is applied and the speed is estimated based on the displacement of the object-s centroid. Several assumptions are listed to simplify the transformation of 2D images from 3D real-world images. The results obtained from the experiment have been compared to the estimated ground truth. From this experiment, it exhibits that the proposed algorithm has achieved the velocity accuracy estimation of about ± 1.7 km/h.

Solving Part Type Selection and Loading Problem in Flexible Manufacturing System Using Real Coded Genetic Algorithms – Part II: Optimization

This paper presents modeling and optimization of two NP-hard problems in flexible manufacturing system (FMS), part type selection problem and loading problem. Due to the complexity and extent of the problems, the paper was split into two parts. The first part of the papers has discussed the modeling of the problems and showed how the real coded genetic algorithms (RCGA) can be applied to solve the problems. This second part discusses the effectiveness of the RCGA which uses an array of real numbers as chromosome representation. The novel proposed chromosome representation produces only feasible solutions which minimize a computational time needed by GA to push its population toward feasible search space or repair infeasible chromosomes. The proposed RCGA improves the FMS performance by considering two objectives, maximizing system throughput and maintaining the balance of the system (minimizing system unbalance). The resulted objective values are compared to the optimum values produced by branch-and-bound method. The experiments show that the proposed RCGA could reach near optimum solutions in a reasonable amount of time.

On Formalizing Predefined OCL Properties

The ability of UML to handle the modeling process of complex industrial software applications has increased its popularity to the extent of becoming the de-facto language in serving the design purpose. Although, its rich graphical notation naturally oriented towards the object-oriented concept, facilitates the understandability, it hardly successes to report all domainspecific aspects in a satisfactory way. OCL, as the standard language for expressing additional constraints on UML models, has great potential to help improve expressiveness. Unfortunately, it suffers from a weak formalism due to its poor semantic resulting in many obstacles towards the build of tools support and thus its application in the industry field. For this reason, many researches were established to formalize OCL expressions using a more rigorous approach. Our contribution join this work in a complementary way since it focuses specifically on OCL predefined properties which constitute an important part in the construction of OCL expressions. Using formal methods, we mainly succeed in expressing rigorously OCL predefined functions.

Mining Image Features in an Automatic Two-Dimensional Shape Recognition System

The number of features required to represent an image can be very huge. Using all available features to recognize objects can suffer from curse dimensionality. Feature selection and extraction is the pre-processing step of image mining. Main issues in analyzing images is the effective identification of features and another one is extracting them. The mining problem that has been focused is the grouping of features for different shapes. Experiments have been conducted by using shape outline as the features. Shape outline readings are put through normalization and dimensionality reduction process using an eigenvector based method to produce a new set of readings. After this pre-processing step data will be grouped through their shapes. Through statistical analysis, these readings together with peak measures a robust classification and recognition process is achieved. Tests showed that the suggested methods are able to automatically recognize objects through their shapes. Finally, experiments also demonstrate the system invariance to rotation, translation, scale, reflection and to a small degree of distortion.

Bayesian Network Model for Students- Laboratory Work Performance Assessment: An Empirical Investigation of the Optimal Construction Approach

There are three approaches to complete Bayesian Network (BN) model construction: total expert-centred, total datacentred, and semi data-centred. These three approaches constitute the basis of the empirical investigation undertaken and reported in this paper. The objective is to determine, amongst these three approaches, which is the optimal approach for the construction of a BN-based model for the performance assessment of students- laboratory work in a virtual electronic laboratory environment. BN models were constructed using all three approaches, with respect to the focus domain, and compared using a set of optimality criteria. In addition, the impact of the size and source of the training, on the performance of total data-centred and semi data-centred models was investigated. The results of the investigation provide additional insight for BN model constructors and contribute to literature providing supportive evidence for the conceptual feasibility and efficiency of structure and parameter learning from data. In addition, the results highlight other interesting themes.