Abstract: Motion detection is a basic operation in the selection of significant segments of the video signals. For an effective Human Computer Intelligent Interaction, the computer needs to recognize the motion and track the moving object. Here an efficient neural network system is proposed for motion detection from the static background. This method mainly consists of four parts like Frame Separation, Rough Motion Detection, Network Formation and Training, Object Tracking. This paper can be used to verify real time detections in such a way that it can be used in defense applications, bio-medical applications and robotics. This can also be used for obtaining detection information related to the size, location and direction of motion of moving objects for assessment purposes. The time taken for video tracking by this Neural Network is only few seconds.
Abstract: The aim of this research is to develop a fast and
reliable surveillance system based on a personal digital assistant
(PDA) device. This is to extend the capability of the device to detect
moving objects which is already available in personal computers.
Secondly, to compare the performance between Background
subtraction (BS) and Temporal Frame Differencing (TFD) techniques
for PDA platform as to which is more suitable. In order to reduce
noise and to prepare frames for the moving object detection part,
each frame is first converted to a gray-scale representation and then
smoothed using a Gaussian low pass filter. Two moving object
detection schemes i.e., BS and TFD have been analyzed. The
background frame is updated by using Infinite Impulse Response
(IIR) filter so that the background frame is adapted to the varying
illuminate conditions and geometry settings. In order to reduce the
effect of noise pixels resulting from frame differencing
morphological filters erosion and dilation are applied. In this
research, it has been found that TFD technique is more suitable for
motion detection purpose than the BS in term of speed. On average
TFD is approximately 170 ms faster than the BS technique
Abstract: Surveillance system is widely used in the traffic
monitoring. The deployment of cameras is moving toward a
ubiquitous camera (UbiCam) environment. In our previous study, a
novel service, called GPS-VT, was firstly proposed by incorporating
global positioning system (GPS) and visual tracking techniques for
the UbiCam environment. The first prototype is called GODTA
(GPS-based Moving Object Detection and Tracking Approach). For a
moving person carried GPS-enabled mobile device, he can be
tracking when he enters the field-of-view (FOV) of a camera
according to his real-time GPS coordinate. In this paper, GPS-VT
service is applied to the tracking of vehicles. The moving speed of a
vehicle is much faster than a person. It means that the time passing
through the FOV is much shorter than that of a person. Besides, the
update interval of GPS coordinate is once per second, it is
asynchronous with the frame rate of the real-time image. The above
asynchronous is worsen by the network transmission delay. These
factors are the main challenging to fulfill GPS-VT service on a
vehicle.In order to overcome the influence of the above factors, a
back-propagation neural network (BPNN) is used to predict the
possible lane before the vehicle enters the FOV of a camera. Then, a
template matching technique is used for the visual tracking of a target
vehicle. The experimental result shows that the target vehicle can be
located and tracking successfully. The success location rate of the
implemented prototype is higher than that of the previous GODTA.
Abstract: Recent developments in automotive technology are focused on economy, comfort and safety. Vehicle tracking and collision detection systems are attracting attention of many investigators focused on safety of driving in the field of automotive mechatronics. In this paper, a vision-based vehicle detection system is presented. Developed system is intended to be used in collision detection and driver alert. The system uses RGB images captured by a camera in a car driven in the highway. Images captured by the moving camera are used to detect the moving vehicles in the image. A vehicle ahead of the camera is detected in daylight conditions. The proposed method detects moving vehicles by subtracting successive images. Plate height of the vehicle is determined by using a plate recognition algorithm. Distance of the moving object is calculated by using the plate height. After determination of the distance of the moving vehicle relative speed of the vehicle and Time-to-Collision are calculated by using distances measured in successive images. Results obtained in road tests are discussed in order to validate the use of the proposed method.
Abstract: Mixed-traffic (e.g., pedestrians, bicycles, and vehicles)
data at an intersection is one of the essential factors for intersection
design and traffic control. However, some data such as pedestrian
volume cannot be directly collected by common detectors (e.g.
inductive loop, sonar and microwave sensors). In this paper, a video
based detection algorithm is proposed for mixed-traffic data collection
at intersections using surveillance cameras. The algorithm is derived
from Gaussian Mixture Model (GMM), and uses a mergence time
adjustment scheme to improve the traditional algorithm. Real-world
video data were selected to test the algorithm. The results show that
the proposed algorithm has the faster processing speed and more
accuracy than the traditional algorithm. This indicates that the
improved algorithm can be applied to detect mixed-traffic at
signalized intersection, even when conflicts occur.
Abstract: Spatial and mobile computing evolves. This paper
describes a smart modeling platform called “GeoSEMA". This
approach tends to model multidimensional GeoSpatial Evolutionary
and Mobile Agents. Instead of 3D and location-based issues, there
are some other dimensions that may characterize spatial agents, e.g.
discrete-continuous time, agent behaviors. GeoSEMA is seen as a
devoted design pattern motivating temporal geographic-based
applications; it is a firm foundation for multipurpose and
multidimensional special-based applications. It deals with
multipurpose smart objects (buildings, shapes, missiles, etc.) by
stimulating geospatial agents.
Formally, GeoSEMA refers to geospatial, spatio-evolutive and
mobile space constituents where a conceptual geospatial space model
is given in this paper. In addition to modeling and categorizing
geospatial agents, the model incorporates the concept of inter-agents
event-based protocols. Finally, a rapid software-architecture
prototyping GeoSEMA platform is also given. It will be
implemented/ validated in the next phase of our work.
Abstract: This paper presents an adaptive differentiator
of sequential data based on the adaptive control theory. The
algorithm is applied to detect moving objects by estimating a
temporal gradient of sequential data at a specified pixel. We
adopt two nonlinear intensity functions to reduce the influence
of noises. The derivatives of the nonlinear intensity functions
are estimated by an adaptive observer with σ-modification
update law.
Abstract: Optical 3D measurement of objects is meaningful in
numerous industrial applications. In various cases shape acquisition
of weak textured objects is essential. Examples are repetition parts
made of plastic or ceramic such as housing parts or ceramic bottles as
well as agricultural products like tubers. These parts are often
conveyed in a wobbling way during the automated optical inspection.
Thus, conventional 3D shape acquisition methods like laser scanning
might fail. In this paper, a novel approach for acquiring 3D shape of
weak textured and moving objects is presented. To facilitate such
measurements an active stereo vision system with structured light is
proposed. The system consists of multiple camera pairs and auxiliary
laser pattern generators. It performs the shape acquisition within one
shot and is beneficial for rapid inspection tasks. An experimental
setup including hardware and software has been developed and
implemented.
Abstract: This paper presents a novel approach for representing
the spatio-temporal topology of the camera network with overlapping
and non-overlapping fields of view (FOVs). The topology is
determined by tracking moving objects and establishing object
correspondence across multiple cameras. To track people successfully
in multiple camera views, we used the Merge-Split (MS) approach for
object occlusion in a single camera and the grid-based approach for
extracting the accurate object feature. In addition, we considered the
appearance of people and the transition time between entry and exit
zones for tracking objects across blind regions of multiple cameras
with non-overlapping FOVs. The main contribution of this paper is to
estimate transition times between various entry and exit zones, and to
graphically represent the camera topology as an undirected weighted
graph using the transition probabilities.
Abstract: In this paper a bank of velocity filters is devised to be
used for isolating a moving object with specific velocity in a sequence of frames. The approach used is a 3-D FFT based experimental procedure without applying any theoretical concept
from velocity filters. Accordingly, velocity filters are built using the
spectral signature of each separate moving object. Experimentation
reveals the capabilities of the constructed filter bank to separate moving objects as far as the amplitude as well as the direction of the
velocity are concerned.
Abstract: Intelligent traffic surveillance technology is an issue in
the field of traffic data analysis. Therefore, we need the technology to
detect moving objects in real-time while there are variations in background and natural light. In this paper, we proposed a Weighted-Center Surround Difference
method for object detection in outdoor environments. The proposed system detects objects using the saliency map that is obtained by
analyzing the weight of each layers of Gaussian pyramid. In order to validate the effectiveness of our system, we implemented the proposed
method using a digital signal processor, TMS320DM6437.
Experimental results show that blurred noisy around objects was effectively eliminated and the object detection accuracy is improved.
Abstract: Corner detection and optical flow are common techniques for feature-based video stabilization. However, these algorithms are computationally expensive and should be performed at a reasonable rate. This paper presents an algorithm for discarding irrelevant feature points and maintaining them for future use so as to improve the computational cost. The algorithm starts by initializing a maintained set. The feature points in the maintained set are examined against its accuracy for modeling. Corner detection is required only when the feature points are insufficiently accurate for future modeling. Then, optical flows are computed from the maintained feature points toward the consecutive frame. After that, a motion model is estimated based on the simplified affine motion model and least square method, with outliers belonging to moving objects presented. Studentized residuals are used to eliminate such outliers. The model estimation and elimination processes repeat until no more outliers are identified. Finally, the entire algorithm repeats along the video sequence with the points remaining from the previous iteration used as the maintained set. As a practical application, an efficient video stabilization can be achieved by exploiting the computed motion models. Our study shows that the number of times corner detection needs to perform is greatly reduced, thus significantly improving the computational cost. Moreover, optical flow vectors are computed for only the maintained feature points, not for outliers, thus also reducing the computational cost. In addition, the feature points after reduction can sufficiently be used for background objects tracking as demonstrated in the simple video stabilizer based on our proposed algorithm.
Abstract: A two-dimensional moving mesh algorithm is developed to simulate the general motion of two rotating bodies with relative translational motion. The grid includes a background grid and two sets of grids around the moving bodies. With this grid arrangement rotational and translational motions of two bodies are handled separately, with no complications. Inter-grid boundaries are determined based on their distances from two bodies. In this method, the overset concept is applied to hybrid grid, and flow variables are interpolated using a simple stencil. To evaluate this moving mesh algorithm unsteady Euler flow is solved for different cases using dual-time method of Jameson. Numerical results show excellent agreement with experimental data and other numerical results. To demonstrate the capability of present algorithm for accurate solution of flow fields around moving bodies, some benchmark problems have been defined in this paper.
Abstract: This paper introduces an intelligent system, which can be applied in the monitoring of vehicle speed using a single camera. The ability of motion tracking is extremely useful in many automation problems and the solution to this problem will open up many future applications. One of the most common problems in our daily life is the speed detection of vehicles on a highway. In this paper, a novel technique is developed to track multiple moving objects with their speeds being estimated using a sequence of video frames. Field test has been conducted to capture real-life data and the processed results were presented. Multiple object problems and noisy in data are also considered. Implementing this system in real-time is straightforward. The proposal can accurately evaluate the position and the orientation of moving objects in real-time. The transformations and calibration between the 2D image and the actual road are also considered.
Abstract: In recent years, we see an increase of interest for efficient tracking systems in surveillance applications. Many of the proposed techniques are designed for static cameras environments. When the camera is moving, tracking moving objects become more difficult and many techniques fail to detect and track the desired targets. The problem becomes more complex when we want to track a specific object in real-time using a moving Pan and Tilt camera system to keep the target within the image. This type of tracking is of high importance in surveillance applications. When a target is detected at a certain zone, the possibility of automatically tracking it continuously and keeping it within the image until action is taken is very important for security personnel working in very sensitive sites. This work presents a real-time tracking system permitting the detection and continuous tracking of targets using a Pan and Tilt camera platform. A novel and efficient approach for dealing with occlusions is presented. Also a new intelligent forget factor is introduced in order to take into account target shape variations and avoid learning non desired objects. Tests conducted in outdoor operational scenarios show the efficiency and robustness of the proposed approach.
Abstract: Speed estimation is one of the important and practical tasks in machine vision, Robotic and Mechatronic. the availability of high quality and inexpensive video cameras, and the increasing need for automated video analysis has generated a great deal of interest in machine vision algorithms. Numerous approaches for speed estimation have been proposed. So classification and survey of the proposed methods can be very useful. The goal of this paper is first to review and verify these methods. Then we will propose a novel algorithm to estimate the speed of moving object by using fuzzy concept. There is a direct relation between motion blur parameters and object speed. In our new approach we will use Radon transform to find direction of blurred image, and Fuzzy sets to estimate motion blur length. The most benefit of this algorithm is its robustness and precision in noisy images. Our method was tested on many images with different range of SNR and is satisfiable.
Abstract: This paper describes new computer vision algorithms
that have been developed to track moving objects as part of a
long-term study into the design of (semi-)autonomous vehicles. We
present the results of a study to exploit variable kernels for tracking in
video sequences. The basis of our work is the mean shift
object-tracking algorithm; for a moving target, it is usual to define a
rectangular target window in an initial frame, and then process the data
within that window to separate the tracked object from the background
by the mean shift segmentation algorithm. Rather than use the
standard, Epanechnikov kernel, we have used a kernel weighted by the
Chamfer distance transform to improve the accuracy of target
representation and localization, minimising the distance between the
two distributions in RGB color space using the Bhattacharyya
coefficient. Experimental results show the improved tracking
capability and versatility of the algorithm in comparison with results
using the standard kernel. These algorithms are incorporated as part of
a robot test-bed architecture which has been used to demonstrate their
effectiveness.
Abstract: Detecting object in video sequence is a challenging
mission for identifying, tracking moving objects. Background
removal considered as a basic step in detected moving objects tasks.
Dual static cameras placed in front and rear moving platform
gathered information which is used to detect objects. Background
change regarding with speed and direction moving platform, so
moving objects distinguished become complicated. In this paper, we
propose framework allows detection moving object with variety of
speed and direction dynamically. Object detection technique built on
two levels the first level apply background removal and edge
detection to generate moving areas. The second level apply Moving
Areas Filter (MAF) then calculate Correlation Score (CS) for
adjusted moving area. Merging moving areas with closer CS and
marked as moving object. Experiment result is prepared on real scene
acquired by dual static cameras without overlap in sense. Results
showing accuracy in detecting objects compared with optical flow
and Mixture Module Gaussian (MMG), Accurate ratio produced to
measure accurate detection moving object.
Abstract: This paper describes an algorithm to estimate realtime vehicle velocity using image processing technique from the known camera calibration parameters. The presented algorithm involves several main steps. First, the moving object is extracted by utilizing frame differencing technique. Second, the object tracking method is applied and the speed is estimated based on the displacement of the object-s centroid. Several assumptions are listed to simplify the transformation of 2D images from 3D real-world images. The results obtained from the experiment have been compared to the estimated ground truth. From this experiment, it exhibits that the proposed algorithm has achieved the velocity accuracy estimation of about ± 1.7 km/h.
Abstract: Current image-based individual human recognition
methods, such as fingerprints, face, or iris biometric modalities
generally require a cooperative subject, views from certain aspects,
and physical contact or close proximity. These methods cannot
reliably recognize non-cooperating individuals at a distance in the
real world under changing environmental conditions. Gait, which
concerns recognizing individuals by the way they walk, is a relatively
new biometric without these disadvantages. The inherent gait
characteristic of an individual makes it irreplaceable and useful in
visual surveillance.
In this paper, an efficient gait recognition system for human
identification by extracting two features namely width vector of
the binary silhouette and the MPEG-7-based region-based shape
descriptors is proposed. In the proposed method, foreground objects
i.e., human and other moving objects are extracted by estimating
background information by a Gaussian Mixture Model (GMM) and
subsequently, median filtering operation is performed for removing
noises in the background subtracted image. A moving target classification
algorithm is used to separate human being (i.e., pedestrian)
from other foreground objects (viz., vehicles). Shape and boundary
information is used in the moving target classification algorithm.
Subsequently, width vector of the outer contour of binary silhouette
and the MPEG-7 Angular Radial Transform coefficients are taken as
the feature vector. Next, the Principal Component Analysis (PCA)
is applied to the selected feature vector to reduce its dimensionality.
These extracted feature vectors are used to train an Hidden Markov
Model (HMM) for identification of some individuals. The proposed
system is evaluated using some gait sequences and the experimental
results show the efficacy of the proposed algorithm.