US20130002866A1 - Detection and Tracking of Moving Objects - Google Patents

Detection and Tracking of Moving Objects Download PDF

Info

Publication number
US20130002866A1
US20130002866A1 US13/609,393 US201213609393A US2013002866A1 US 20130002866 A1 US20130002866 A1 US 20130002866A1 US 201213609393 A US201213609393 A US 201213609393A US 2013002866 A1 US2013002866 A1 US 2013002866A1
Authority
US
United States
Prior art keywords
tracking
motion
moving objects
images
blob
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/609,393
Inventor
Arun Hampapur
Jun Li
Sharathchandra Pankanti
Charles A. Otto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/609,393 priority Critical patent/US20130002866A1/en
Publication of US20130002866A1 publication Critical patent/US20130002866A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/207Analysis of motion for motion estimation over a hierarchy of resolutions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/215Motion-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B15/00Special procedures for taking photographs; Apparatus therefor
    • G03B15/16Special procedures for taking photographs; Apparatus therefor for photographing the track of moving objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Definitions

  • Embodiments of the invention generally relate to information technology, and, more particularly, to object detection.
  • UAV Unmanned Aerial Vehicle
  • frame rate is very low (for example, 1 frame per second) so as to increase the difficulties of detecting and tracking ground moving targets, and small object size will bring another challenge for object detection and tracking.
  • a camera's strong illumination change and stripe noise can create some hard problems to separate true moving objects from the background.
  • An exemplary method for performing visual surveillance of one or more moving objects, according to one aspect of the invention, can include steps of registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames, performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images, and tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects.
  • One or more embodiments of the invention or elements thereof can be implemented in the form of a computer product including a tangible computer readable storage medium with computer useable program code for performing the method steps indicated. Furthermore, one or more embodiments of the invention or elements thereof can be implemented in the form of an apparatus including a memory and at least one processor that is coupled to the memory and operative to perform exemplary method steps.
  • one or more embodiments of the invention or elements thereof can be implemented in the form of means for carrying out one or more of the method steps described herein; the means can include (i) hardware module(s), (ii) software module(s), or (iii) a combination of hardware and software modules; any of (i)-(iii) implement the specific techniques set forth herein, and the software modules are stored in a tangible computer-readable storage medium (or multiple such media).
  • FIG. 1 is a diagram illustrating sub-pixel position estimation, according to an embodiment of the present invention.
  • FIG. 2 is a diagram illustrating sub-region selection, according to an embodiment of the present invention.
  • FIG. 3 is a diagram illustrating forward and backward geometric registration, according to an embodiment of the present invention.
  • FIG. 4 is a flow diagram illustrating forward and backward frame differencing, according to an embodiment of the present invention.
  • FIG. 5 is a flow diagram illustrating false blob filtering, according to an embodiment of the present invention.
  • FIG. 6 is a flow diagram illustrating multi-object tracking, according to an embodiment of the present invention.
  • FIG. 7 is a diagram illustrating reference plane-based registration and tracking, according to an embodiment of the present invention.
  • FIG. 8 is a flow diagram illustrating automatic urban road extraction, according to an embodiment of the present invention.
  • FIG. 9 is a block diagram illustrating architecture of an object detection and tracking system, according to an aspect of the invention.
  • FIG. 10 is a flow diagram illustrating techniques for performing visual surveillance of one or more moving objects, according to an embodiment of the invention.
  • FIG. 11 is a system diagram of an exemplary computer system on which at least one embodiment of the invention can be implemented.
  • Principles of the invention include detection, tracking, and searching of moving objects in visual surveillance.
  • one or more embodiments of the invention include motion segmentation (motion blobs versus background region), multiple object tracking (for example, consistently tracking in over-time) and reference plane-based registration and tracking.
  • one or more embodiments of the invention include using multiple cameras (for example, registered with each other) mounted, for example, on mobile platforms (for example, unmanned aerial vehicle (UAV) videos) to detect, track and search for moving objects by forming a panoramic view from the images received from the cameras based on global/local geometric registration, motion segmentation, moving object tracking, reference plane-based registration and tracking and automatic urban road extraction.
  • UAV unmanned aerial vehicle
  • the techniques described herein include recursive geometric registration, which includes region-based image registration for adjacent frames instead of for an entire frame, sub-pixel image matching techniques, and region-based geometric transformation for handling lens geometric distortion.
  • one or more embodiments of the invention include two-way motion detection and hybrid target tracking using colors and features.
  • Two-way motion detection includes forward and backward frame differencing, automatic dynamic threshold estimation based on temporary and/or spatial filtering, as well as false moving pixel removal based on independent motions of features.
  • Hybrid target tracking includes Kanade-Lucas-Tomasi feature tracker (KLT) and meanshift, auto kernel scale estimation and updating, and consistently tracking in over-time using coherent motion of feature trajectories.
  • KLT Kanade-Lucas-Tomasi feature tracker
  • the techniques detailed herein include multi-target tracking algorithms based on feature matching and distance matrices for small targets, as well as, for example, a UAV surveillance system implementation with Low frame rate (1 f/s) for detecting and tracking the targets with small size (for example, without any known shape model).
  • one or more embodiments of the invention include local/global geometric registration of videos (for example, UAV videos).
  • a frame-to-frame video registration process is implemented.
  • An accurate way to register two images can include matching every pixel in each image.
  • the high computation is not feasible.
  • An efficient way is to find a relatively small set of feature points in the image that will be easy to find again and use only those points to estimate a frame-to-frame homography.
  • 500-600 feature points can be extracted for an image of 1280 ⁇ 1280 pixels.
  • Harris corner detector can be applied to image registration and motion detection due to its invariance to scale, rotation and illumination variation.
  • Harris corner detector can be used as a feature point detector. Its algorithm can be described as follows:
  • one or more embodiments of the invention include using a normalized correlation coefficient, which is an efficient statistical method.
  • the actual feature matching is achieved by maximizing the correlation coefficient over small windows surrounding the points.
  • the correlation coefficient is given by:
  • g 1 (r,c) represents individual gray values of template matrix
  • u 1 represents average gray value of template matrix
  • g 2 (r,c) represents individual gray values of corresponding part of search matrix
  • u 2 represents average gray value of corresponding part of search matrix
  • R, C represents number of rows and columns of template matrix.
  • the block matching process can be achieved as follows. For each point in a reference frame, all points in the chosen frame are examined and its most similar point is chosen. Next, it is tested whether the achieved correlation is reasonably high. The point with maxima correlation coefficient is taken as a candidate point.
  • Video registration requires real-time implementation.
  • the block-matching algorithm is only implemented for the features. As such, the computational expense can be significantly reduced.
  • One or more embodiments of the invention also include corresponding features checking and outlier removal.
  • Feature-based block matching can sometimes cause a mismatch.
  • one or more embodiments of the invention include using forward searching to process the mismatching data which cases are one too many, keeping the candidate corresponding feature with the maximum gradient value and removing the others. Also, backward searching is employed to solve the remaining mismatching problem using the same approach.
  • a pair of features with similar attributes is accepted as a match. Nevertheless, some false matches may occur. Therefore, in one or more embodiments of the invention, a random sample consensus (RANSAC) outlier removal procedure is performed to remove incorrect matches and improve the registration precision.
  • RNSAC random sample consensus
  • the techniques detailed herein can additionally include coarse-to-fine feature matching.
  • Multi-resolution feature matching can reduce searching space and false matching.
  • feature matching is performed and the searching scope is determined.
  • the matching results at the last layer can be taken as initial results and the matching process can be performed by using equation (1) noted above.
  • a search scope is limited to 1-3 pixel(s). Further, the same operation can be repeated until the highest resolution layer is reached.
  • one or more embodiments of the invention include accurate position determination.
  • pixel level accuracy may not enough.
  • a sub-pixel position approach is considered, and a distance-based weighting interpolation is determined to the peak.
  • the horizontal and vertical locations of the peak can be separately estimated for the feature.
  • the one-dimensional horizontal and vertical correlation curves can be obtained.
  • the correlation value in x,y directions is interpolated separately, and the accurate location of the peak is computed.
  • FIG. 1 is a diagram illustrating sub-pixel position estimation, according to an embodiment of the present invention.
  • the techniques described herein also include local geometric registration.
  • a sub-region geometric registration can be selected, and the entire frame can be divided into 2 ⁇ 2 sub-regions.
  • FIG. 2 illustrates two selection models.
  • FIG. 2 is a diagram illustrating sub-region selection, according to an embodiment of the present invention.
  • FIG. 2 depicts sub-region selection model 202 and sub-region selection model 204 .
  • One or more embodiments of the invention also include an affine-based local transformation, such as, for example, the following:
  • [ x y ] [ a 0 + a 1 ⁇ u + a 2 ⁇ v b 0 + b 1 ⁇ u + b 2 ⁇ v ]
  • (x, y) is the new transformed coordinate of (u, v)
  • one or more embodiments of the invention include using a least squares technique to compute the transformation parameters.
  • One or more embodiments of the invention also include forward/backward frame-to-frame registration.
  • forward/backward frame-to-frame registration is carried out for multi-frame differencing.
  • FIG. 3 illustrates an approach.
  • FIG. 3 is a diagram illustrating forward and backward geometric registration, according to an embodiment of the present invention.
  • FIG. 3 depicts frame 302 (F i ⁇ 1 ), frame 304 (F i ) and frame 306 (F 1+1 ).
  • frame 304 (F i ) which is taken as a reference frame
  • previous frame 302 (F i ⁇ 1 ) and next frame 306 (F i+1 ) are geometrically registered to the reference frame. Motion estimation for each frame is carried out in such a fashion.
  • FIG. 4 is a flow diagram illustrating forward and backward frame differencing, according to an embodiment of the present invention.
  • forward/backward frame-to-frame images for example, frame 402 , frame 404 and frame 406
  • difference images are calculated.
  • forward/backward frame differencing in steps 412 and 414 to reduce motion noise and compensate the illustration variation such as automatic gain control.
  • Step 418 includes median filtering, which can reduce random motion noise.
  • step 422 includes performing a morphological operation to remove small isolated spots and fill holes in foreground image and step 424 includes generating motion pixels (for example, a motion map).
  • a threshold for each pixel is statistically calculated automatically in terms of statistical characteristics and spatial high frequency data of difference image. Further, a morphology step can be applied to remove small isolated spots and fill holes in the foreground image.
  • FIG. 5 is a flow diagram illustrating false blob filtering, according to an embodiment of the present invention.
  • Step 502 includes generating a motion map.
  • Step 504 includes applying a connected component process to link each blob data.
  • Step 506 includes creating a motion blob table.
  • Step 508 includes performing an optical flow estimation.
  • Step 510 includes making a displacement determination. If there is displacement, the process proceeds to step 512 , which includes performing post-processing such as, for example, data association, object tracking, trajectory maintenance and track data management. If there is no displacement, the process proceeds to step 514 , which includes filtering false blobs.
  • each blob data is verified.
  • One or more embodiments of the invention apply a KLT process to estimate the motion of each blob after forward/backward frame-to-frame registration is done. A false blob will be deleted from the blob table.
  • the process steps can include, for example, applying a connected component process to link each blob data, creating a blob table, extracting features for each blob in a previous registered frame, applying the KLT method to estimate the motion of each blob, and if no motion occurs, the blob is deleted from the blob table. Also, the above-noted steps can be repeated for all blobs.
  • FIG. 6 is a flow diagram illustrating multi-object tracking, according to an embodiment of the present invention.
  • Step 602 includes generating a motion map.
  • Step 604 includes identifying moving blobs.
  • Step 606 includes object initialization and step 608 includes object checking.
  • Step 610 includes identifying object regions.
  • Step 612 includes identifying candidate regions.
  • step 614 includes meanshift tracking and step 616 includes identifying new locations.
  • step 618 features can be extracted in step 618 .
  • moving blobs can be found as potential object candidates in step 622 .
  • KLT matching is performed in step 624 and outlier removal based on an affine transform with RANSAC is performed in step 626 .
  • a new region candidate is identified in step 628 .
  • Meanshift is applied in step 614 to compute the inter-frame translation. This yields a candidate region location in step 616 .
  • step 632 includes target model updating for solving drift issues
  • step 634 includes trajectory updating.
  • a hybrid tracking model based on the combination of KLT and Meanshift method is applied from step 618 to 630 .
  • the techniques described herein include object initialization.
  • the motion detection results from forward/backward frame differencing can contain some correct real moving objects and some false objects, and miss some true objects.
  • a moving object does not have any overlapping regions between two consecutive frames so that traditional methods for object initialization will not work.
  • one or more embodiments of the invention include combining a distance matrix with a similarity measure to initialize moving objects.
  • the processing steps can include, for example, the following.
  • a search radius is set, matching score threshold and minimum length of tracked history.
  • the distance matrix between the objects (including object candidates) and all the blobs in the table is computed. If the length of object trajectory is less than the preset value, a Kernel-based algorithm is applied to find the match between the object candidate and blobs in terms of a preset matching score. Also, if the object candidate appears in several consecutive frames, this candidate will be initialized and stored on the object table. Otherwise, the object candidate will be considered as a false object.
  • one or more embodiments of the invention include projecting the previous blob set into a current frame after geometrical registration.
  • the motion of each object according to its previous position can be estimated by a KLT tracking process.
  • affine transformation parameters can be computed from as few as four feature points. To determine these parameters, a least squares technique can be used to compute them.
  • Accuracy estimation can be performed, for example, when the number of mismatched pairs occurs.
  • One measure of tracking accuracy is the root mean square error (RMSE) between the matched points before and after the affine transformation formula. This measure is used as a criterion to eliminate the matches that are considered imprecise.
  • RMSE root mean square error
  • one or more embodiments of the invention includes performing the RANSAC algorithm to sequentially remove mis-matches in an iterative fashion until the RMSE value is lower than the desired threshold.
  • the techniques detailed herein additionally include meanshift tracking and object representation.
  • meanshift tracking and object representation By way of example, for a UAV tracking system, traditional intensity-based target representation is no longer suitable for multi-object tracking due to large scale variation and perspective geometric distortion.
  • histogram-based feature space can be chosen.
  • a metric based on the Bhattacharyya coefficient is used to define a similarity measure between a reference object and a candidate for multi-object tracking. Given an object region histogram q in the reference frame, the Bhattacharyya coefficient based objective function is given by:
  • M is the histogram dimension
  • x 0 is the 2D center
  • the candidate region histogram p u (x) at 2D center x in the current frame is defined as:
  • h 2D bandwidth vector of k(x)
  • is the Kronecker delta function and each pixel value is denoted by b(x i ).
  • the techniques described herein can additionally include object positioning.
  • one or more embodiments of the invention include applying a meanshift tracking algorithm that is based on a gradient ascent optimization rather than an exhaustive search.
  • Strengths of the meanshift method include computational effectiveness and suitability to real-time application.
  • a target can be lost, for example, due to an intrinsic limitation of exploring local maxima, especially when the tracked object moves quickly.
  • the candidate region histogram p u (x) can be obtained from the above equation.
  • the new location of the tracked object can be estimated as:
  • ⁇ i 1 n ⁇ ⁇ i ⁇ g ⁇ ( ⁇ y ⁇ 0 ⁇ X i h ⁇ 2 )
  • One or more embodiments of the invention can also include target model updating on a temporal domain.
  • a meanshift approach without target model updating can suffer from abrupt changes in target model.
  • the model updating for every frame can result in decreasing the reliability of the tracking results due to cluttered environment, occlusion, random noise, etc.
  • One way to change the target model is to periodically update the target distributions.
  • one or more embodiments of the invention include model updating that use both recent tracking results and older target model to impact a current target model for object tracking.
  • the updating procedure is formulized as:
  • the superscripts of new and old denote the newly obtained target model and the old model, respectively.
  • s represents the recent tracking result.
  • weights the contribution of the recent tracking result (normally ⁇ 0.1).
  • q and p represent the target model and the candidate model, respectively.
  • one or more embodiments of the invention include target model updating on a spatial domain.
  • meanshift based tracking hardly provides precise boundary position of the tracked object due to lack of utilizing spatial data.
  • detection results derived from KLT tracker and motion detection results can provide much more accurate information, such as the precise position and object size compared with meanshift tracker.
  • one or more embodiments of the invention use the following merging method:
  • overlapping represents the degree of overlapping region.
  • FIG. 7 is a diagram illustrating reference plane-based registration and tracking, according to an embodiment of the present invention.
  • FIG. 7 depicts a geo-reference plane 702 .
  • the first frame 704 is registered to geo-reference plane 702
  • the second frame 706 is registered to the geo-reference 702 from the first registered frame and corresponding inter-frame transformation parameters TC i (equation 712 in FIG. 7 ).
  • frames 708 and 710 are registered to the geo-reference 702 , respectively.
  • each object is projected into geo-reference 702 using navigation data.
  • FIG. 8 is a flow diagram illustrating automatic urban road extraction, according to an embodiment of the present invention.
  • Step 802 includes framing an image.
  • Step 804 includes performing a Gaussian smoothing operation.
  • step 806 includes using a canny detector and step 808 includes implementing a hough transformation.
  • Step 810 includes determining a maximum response finding.
  • Step 812 includes determining if the length of the stripe is greater than a pre-defined threshold. If the length of the stripe is not greater than the threshold, the process stops at step 814 . If the length of the stripe is greater than the threshold, the process continues to step 816 , which includes performing a straight line extraction. Further, step 818 includes performing stripe pixels removal (which can, for example, lead to a return to step 808 ).
  • step 820 includes performing frame differencing
  • step 822 includes verification via motion history images (MHI) (which can, for example, lead to a return to step 816 ).
  • MHI motion history images
  • one or more embodiments of the invention can also include extraction of road stripes via iterative hough transform.
  • one or more embodiments of the invention include recursive geometric registration with sub-pixel matching accuracy that can handle various geometrical residual errors from un-calibrated camera. Additionally, the techniques detailed herein include motion detection based on forward/backward frame differencing that can efficiently separate moving objects from background. Further, a hybrid object tracker can be implemented that uses colors, features and intensity statistical characteristics overtime to detect and track multiple small objects.
  • FIG. 9 is a block diagram illustrating architecture of an object detection and tracking system, according to an aspect of the invention.
  • An example software architecture construction for a detection and tracking system (for example, a UAV system) can be built on multiple services to provide a track database for object search and intelligent analysis.
  • the software architecture can include multiple sensor modules 904 , video streaming service modules 906 , tracking suite service modules 908 , a track database (DB) server module 910 , a user interface module 902 and a visualization console 912 .
  • a video streaming module 906 serves to capture and make available imagery from multiple sensors. The acquired images are used by a tracking suite module 908 as the basis for multi-object detection and tracking.
  • Tracking suite modules 908 includes a geometric registration sub-module 914 , a motion extraction sub-module 916 , an object tracking sub-module 918 , a tracking data sub-module 920 and a geo-coordinate mapping sub-module 922 .
  • Track DB server 910 serves track metadata management.
  • Visualization console 912 creates graphical overlays, indexes them to the imagery on the display, and presents them to a user. These overlays can be any type of graphical information that supports the higher level components, such as, for example, class types, moving directions, trajectories and object sizes.
  • User interface 902 provides data access and operation by the user.
  • FIG. 10 is a flow diagram illustrating techniques for performing visual surveillance of one or more moving objects, according to an embodiment of the present invention.
  • Step 1002 includes registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames. This step can be carried out, for example, using a geometric registration sub-module 914 in tracking suite service module 908 .
  • Registering images can include recursive global and local geometric registration of the one or more images (for example, region-based geometric transformation for handling lens geometric distortion). Registering images can also include using sub-pixel image matching techniques.
  • Step 1004 includes performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images. This step can be carried out, for example, using a motion extraction sub-module 916 in tracking suite service module 908 .
  • Performing motion segmentation of the images can include forward and backward frame differencing. Forward and backward frame differences can include, for example, automatic dynamic threshold estimation based on temporary filtering and/or spatial filtering, removing false moving pixels based on independent motions of image features, and performing a morphological operation and generating motion pixels.
  • Step 1006 includes tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects.
  • This step can be carried out, for example, using an object tracking sub-module 918 in tracking suite service module 908 .
  • Tracking the moving objects can include performing hybrid target tracking, wherein hybrid target tracking includes using a Kanade-Lucas-Tomasi feature tracker and meanshift, using auto kernel scale estimation and updating, and using feature trajectories.
  • One or more embodiments of the invention can also include using colors for tracking.
  • Tracking moving objects can additionally include using multi-target tracking algorithms based on feature matching and distance matrices for one or more (small) targets.
  • tracking moving objects can include generating a motion map, identifying one or more moving objects (blobs), performing object initialization and object checking, identifying object regions in the motion map, extracting features, setting a search region in the motion map, identifying candidate regions in the motion map, meanshift tracking, identifying moving objects in the candidate regions, performing Kanade-Lucas-Tomasi feature matching, performing an affine transform (with RANSAC), making a final regions determination via the Bhattacharyya coefficient, and updating a target model and trajectory information.
  • Tracking moving objects can additionally include reference plane-based registration and tracking.
  • the techniques depicted in FIG. 10 can also include relating each camera view with one or more other camera views, and forming a panoramic view from the images captured by one or more cameras.
  • One or more embodiments of the invention additionally include estimating motion of each camera based on video information of static objects in the panoramic view, as well as estimating one or more background (for example, road) structures in the panoramic view based on linear structure detection and statistical analysis of the moving objects over a period of time.
  • the techniques depicted in FIG. 10 include automatic feature (for example, a road) extraction, wherein automatic feature extraction includes framing an image, performing a Gaussian smoothing operation, using a canny detector to extract one or more feature (for example, road) edges, implementing a hough transformation for feature (for example, road stripe) analysis, determining a maximum response finding for reducing an influence of multiple peaks in a transform space, determining if a length of a feature (for example, a road stripe) is greater than a certain threshold, and if the length of the feature is greater than the threshold, performing feature extraction and pixel removal.
  • Automatic feature extraction can additionally include performing frame differencing and verification via motion history images.
  • One or more embodiments of the invention also include performing outlier removal to remove incorrect moving object matches (and improve the registration precision).
  • the techniques depicted in FIG. 10 can additionally include false blob filtering.
  • False blob filtering includes generating a motion map, applying a connected component process to link each blob data, creating a motion blob table, extracting features for each blob in a previously registered frame, and applying a Kanade-Lucas-Tomasi method to estimate motion of each blob, and, if no motion occurs for a blob, deleting the blob from the blob table.
  • one or more embodiments of the invention can include updating a target model on a temporal domain and/or a spatial domain, as well as creating an index (for example, a searchable index) of object appearances and object tracks in a panoramic view.
  • the object appearance and tracks template index can be stored in a template data store with a pointer to the corresponding video segments for easy retrieval.
  • one or more embodiments of the invention can include determining a similarity metric between a query and an entry in the index, which can facilitate searching for the object appearance and tracks in a template data store/index based on the similarity metric, and outputting/listing the search results for a human operator based on similarity of the query.
  • the techniques depicted in FIG. 10 can also, as described herein, include providing a system, wherein the system includes distinct software modules, each of the distinct software modules being embodied on a tangible computer-readable recordable storage medium. All the modules (or any subset thereof) can be on the same medium, or each can be on a different medium, for example.
  • the modules can include any or all of the components shown in the figures.
  • the modules include sensor modules, video streaming service modules, tracking suite service modules (including the sub-modules detailed herein), a track database (DB) server module, a user interface module and a visualization console module that can run, for example on one or more hardware processors.
  • DB track database
  • a computer program product can include a tangible computer-readable recordable storage medium with code adapted to be executed to carry out one or more method steps described herein, including the provision of the system with the distinct software modules.
  • the techniques depicted in FIG. 10 can be implemented via a computer program product that can include computer useable program code that is stored in a computer readable storage medium in a data processing system, and wherein the computer useable program code was downloaded over a network from a remote data processing system.
  • the computer program product can include computer useable program code that is stored in a computer readable storage medium in a server data processing system, and wherein the computer useable program code are downloaded over a network to a remote data processing system for use in a computer readable storage medium with the remote system.
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • One or more embodiments of the invention, or elements thereof, can be implemented in the form of an apparatus including a memory and at least one processor that is coupled to the memory and operative to perform exemplary method steps.
  • processors can make use of software running on a general purpose computer or workstation.
  • FIG. 11 such an implementation might employ, for example, a processor 1102 , a memory 1104 , and an input/output interface formed, for example, by a display 1106 and a keyboard 1108 .
  • the term “processor” as used herein is intended to include any processing device, such as, for example, one that includes a CPU (central processing unit) and/or other forms of processing circuitry. Further, the term “processor” may refer to more than one individual processor.
  • memory is intended to include memory associated with a processor or CPU, such as, for example, RAM (random access memory), ROM (read only memory), a fixed memory device (for example, hard drive), a removable memory device (for example, diskette), a flash memory and the like.
  • input/output interface is intended to include, for example, one or more mechanisms for inputting data to the processing unit (for example, mouse), and one or more mechanisms for providing results associated with the processing unit (for example, printer).
  • the processor 1102 , memory 1104 , and input/output interface such as display 1106 and keyboard 1108 can be interconnected, for example, via bus 1110 as part of a data processing unit 1112 .
  • Suitable interconnections can also be provided to a network interface 1114 , such as a network card, which can be provided to interface with a computer network, and to a media interface 1116 , such as a diskette or CD-ROM drive, which can be provided to interface with media 1118 .
  • a network interface 1114 such as a network card
  • a media interface 1116 such as a diskette or CD-ROM drive
  • computer software including instructions or code for performing the methodologies of the invention, as described herein, may be stored in one or more of the associated memory devices (for example, ROM, fixed or removable memory) and, when ready to be utilized, loaded in part or in whole (for example, into RAM) and implemented by a CPU.
  • Such software could include, but is not limited to, firmware, resident software, microcode, and the like.
  • a data processing system suitable for storing and/or executing program code will include at least one processor 1102 coupled directly or indirectly to memory elements 1104 through a system bus 1110 .
  • the memory elements can include local memory employed during actual implementation of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during implementation.
  • I/O devices including but not limited to keyboards 1108 , displays 1106 , pointing devices, and the like
  • I/O controllers can be coupled to the system either directly (such as via bus 1110 ) or through intervening I/O controllers (omitted for clarity).
  • Network adapters such as network interface 1114 may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
  • a “server” includes a physical data processing system (for example, system 1112 as shown in FIG. 11 ) running a server program. It will be understood that such a physical server may or may not include a display and keyboard.
  • aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon. Any combination of one or more computer readable medium(s) may be utilized.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • Media block 1118 is a non-limiting example.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • any appropriate medium including but not limited to wireless, wireline, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • each block in the flowchart or block diagrams may represent a module, component, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • any of the methods described herein can include an additional step of providing a system comprising distinct software modules embodied on a computer readable storage medium; the modules can include, for example, any or all of the components shown in FIG. 9 .
  • the method steps can then be carried out using the distinct software modules and/or sub-modules of the system, as described above, executing on one or more hardware processors 1102 .
  • a computer program product can include a computer-readable storage medium with code adapted to be implemented to carry out one or more method steps described herein, including the provision of the system with the distinct software modules.
  • At least one embodiment of the invention may provide one or more beneficial effects, such as, for example, automatic dynamic threshold determination based on temporary and/or spatial domain.

Abstract

Techniques for performing visual surveillance of one or more moving objects are provided. The techniques include registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames, performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images, and tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 12/972,836, filed Dec. 20, 2010, incorporated by reference herein.
  • FIELD OF THE INVENTION
  • Embodiments of the invention generally relate to information technology, and, more particularly, to object detection.
  • BACKGROUND OF THE INVENTION
  • In recent years, reconnaissance, surveillance, disaster relief, search and rescue, agriculture information gathering and fast remote sensing mapping has gained increasingly attentions in civilian and military purposes. For example, due to their small size and low-cost sensor platform, Unmanned Aerial Vehicle (UAV) can be an attractive platform for executing such operations. However, UAV introduces some significant challenges when used in surveillance systems. For an instance, the background significantly changes as the camera has a fast motion and an irregular rotation, and the motion of a UAV vehicle is usually not smooth. Further, frame rate is very low (for example, 1 frame per second) so as to increase the difficulties of detecting and tracking ground moving targets, and small object size will bring another challenge for object detection and tracking. Also, a camera's strong illumination change and stripe noise can create some hard problems to separate true moving objects from the background.
  • Existing approaches also include object initialization issues, and are additionally unable to obtain high-accuracy registration results, to handle rotation and scale variation of a target, and to deal with similar distribution between target and background.
  • SUMMARY OF THE INVENTION
  • Principles and embodiments of the invention provide techniques for detection and tracking of moving objects. An exemplary method (which may be computer-implemented) for performing visual surveillance of one or more moving objects, according to one aspect of the invention, can include steps of registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames, performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images, and tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects.
  • One or more embodiments of the invention or elements thereof can be implemented in the form of a computer product including a tangible computer readable storage medium with computer useable program code for performing the method steps indicated. Furthermore, one or more embodiments of the invention or elements thereof can be implemented in the form of an apparatus including a memory and at least one processor that is coupled to the memory and operative to perform exemplary method steps. Yet further, in another aspect, one or more embodiments of the invention or elements thereof can be implemented in the form of means for carrying out one or more of the method steps described herein; the means can include (i) hardware module(s), (ii) software module(s), or (iii) a combination of hardware and software modules; any of (i)-(iii) implement the specific techniques set forth herein, and the software modules are stored in a tangible computer-readable storage medium (or multiple such media).
  • These and other objects, features and advantages of the present invention will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram illustrating sub-pixel position estimation, according to an embodiment of the present invention;
  • FIG. 2 is a diagram illustrating sub-region selection, according to an embodiment of the present invention;
  • FIG. 3 is a diagram illustrating forward and backward geometric registration, according to an embodiment of the present invention;
  • FIG. 4 is a flow diagram illustrating forward and backward frame differencing, according to an embodiment of the present invention;
  • FIG. 5 is a flow diagram illustrating false blob filtering, according to an embodiment of the present invention;
  • FIG. 6 is a flow diagram illustrating multi-object tracking, according to an embodiment of the present invention;
  • FIG. 7 is a diagram illustrating reference plane-based registration and tracking, according to an embodiment of the present invention;
  • FIG. 8 is a flow diagram illustrating automatic urban road extraction, according to an embodiment of the present invention;
  • FIG. 9 is a block diagram illustrating architecture of an object detection and tracking system, according to an aspect of the invention;
  • FIG. 10 is a flow diagram illustrating techniques for performing visual surveillance of one or more moving objects, according to an embodiment of the invention; and
  • FIG. 11 is a system diagram of an exemplary computer system on which at least one embodiment of the invention can be implemented.
  • DETAILED DESCRIPTION OF EMBODIMENTS
  • Principles of the invention include detection, tracking, and searching of moving objects in visual surveillance. In an example setting including moving objects and one or more moving cameras, one or more embodiments of the invention include motion segmentation (motion blobs versus background region), multiple object tracking (for example, consistently tracking in over-time) and reference plane-based registration and tracking. As detailed herein, one or more embodiments of the invention include using multiple cameras (for example, registered with each other) mounted, for example, on mobile platforms (for example, unmanned aerial vehicle (UAV) videos) to detect, track and search for moving objects by forming a panoramic view from the images received from the cameras based on global/local geometric registration, motion segmentation, moving object tracking, reference plane-based registration and tracking and automatic urban road extraction.
  • The techniques described herein include recursive geometric registration, which includes region-based image registration for adjacent frames instead of for an entire frame, sub-pixel image matching techniques, and region-based geometric transformation for handling lens geometric distortion. Also, one or more embodiments of the invention include two-way motion detection and hybrid target tracking using colors and features. Two-way motion detection includes forward and backward frame differencing, automatic dynamic threshold estimation based on temporary and/or spatial filtering, as well as false moving pixel removal based on independent motions of features. Hybrid target tracking includes Kanade-Lucas-Tomasi feature tracker (KLT) and meanshift, auto kernel scale estimation and updating, and consistently tracking in over-time using coherent motion of feature trajectories.
  • Further, the techniques detailed herein include multi-target tracking algorithms based on feature matching and distance matrices for small targets, as well as, for example, a UAV surveillance system implementation with Low frame rate (1 f/s) for detecting and tracking the targets with small size (for example, without any known shape model).
  • As noted herein, one or more embodiments of the invention include local/global geometric registration of videos (for example, UAV videos). In order to reduce the camera motion effect, a frame-to-frame video registration process is implemented. An accurate way to register two images can include matching every pixel in each image. However, the high computation is not feasible. An efficient way is to find a relatively small set of feature points in the image that will be easy to find again and use only those points to estimate a frame-to-frame homography. By way of example only, 500-600 feature points can be extracted for an image of 1280×1280 pixels.
  • Harris corner detector can be applied to image registration and motion detection due to its invariance to scale, rotation and illumination variation. In one or more embodiments of the invention, Harris corner detector can be used as a feature point detector. Its algorithm can be described as follows:
  • 1. For a pixel in an image I, compute its x- and y-directional derivatives I x and I y, and I xy=I x I y.
  • 2. Apply a window function A, that is, hx=AI x, hy=AI y, hxy=AI xy.
  • 3. Compute H=hxhy−hxy 2−κ(hx+hy)2 (κ is a constant) to measure variations in both directions.
  • 4. Threshold H and find local maxima to obtain a corner.
  • To compare the windows, one or more embodiments of the invention include using a normalized correlation coefficient, which is an efficient statistical method. The actual feature matching is achieved by maximizing the correlation coefficient over small windows surrounding the points. The correlation coefficient is given by:
  • ρ = r = 1 R c = 1 C [ g 1 ( r , c ) - u 1 ] · [ g 2 ( r , c ) - u 2 ] r = 1 R c = 1 C [ g 1 ( r , c ) - u 1 ] 2 r = 1 R c = 1 C [ g 2 ( r , c ) - u 2 ] 2 ; - 1 ρ 1 ( 1 )
  • where:
    g1(r,c) represents individual gray values of template matrix;
    u1 represents average gray value of template matrix;
    g2(r,c) represents individual gray values of corresponding part of search matrix;
    u2 represents average gray value of corresponding part of search matrix; and
    R, C represents number of rows and columns of template matrix.
  • Therefore, the block matching process can be achieved as follows. For each point in a reference frame, all points in the chosen frame are examined and its most similar point is chosen. Next, it is tested whether the achieved correlation is reasonably high. The point with maxima correlation coefficient is taken as a candidate point.
  • Video registration requires real-time implementation. In one or more embodiments of the invention, the block-matching algorithm is only implemented for the features. As such, the computational expense can be significantly reduced.
  • One or more embodiments of the invention also include corresponding features checking and outlier removal. Feature-based block matching can sometimes cause a mismatch. To avoid a mismatching problem, one or more embodiments of the invention include using forward searching to process the mismatching data which cases are one too many, keeping the candidate corresponding feature with the maximum gradient value and removing the others. Also, backward searching is employed to solve the remaining mismatching problem using the same approach.
  • In many instances, a pair of features with similar attributes is accepted as a match. Nevertheless, some false matches may occur. Therefore, in one or more embodiments of the invention, a random sample consensus (RANSAC) outlier removal procedure is performed to remove incorrect matches and improve the registration precision.
  • The techniques detailed herein can additionally include coarse-to-fine feature matching. Multi-resolution feature matching can reduce searching space and false matching. At a coarsest resolution layer, feature matching is performed and the searching scope is determined. At the current resolution layer, the matching results at the last layer can be taken as initial results and the matching process can be performed by using equation (1) noted above. In one or more embodiments of the invention, a search scope is limited to 1-3 pixel(s). Further, the same operation can be repeated until the highest resolution layer is reached.
  • As additionally described herein, one or more embodiments of the invention include accurate position determination. For video registration and motion detection purposes, pixel level accuracy may not enough. In such instances, a sub-pixel position approach is considered, and a distance-based weighting interpolation is determined to the peak. The horizontal and vertical locations of the peak can be separately estimated for the feature. Also, the one-dimensional horizontal and vertical correlation curves can be obtained. Further, the correlation value in x,y directions is interpolated separately, and the accurate location of the peak is computed. By way of example, FIG. 1 is a diagram illustrating sub-pixel position estimation, according to an embodiment of the present invention.
  • The techniques described herein also include local geometric registration. By way of example, a sub-region geometric registration can be selected, and the entire frame can be divided into 2×2 sub-regions. FIG. 2 illustrates two selection models.
  • FIG. 2 is a diagram illustrating sub-region selection, according to an embodiment of the present invention. By way of illustration, FIG. 2 depicts sub-region selection model 202 and sub-region selection model 204.
  • One or more embodiments of the invention also include an affine-based local transformation, such as, for example, the following:
  • [ x y ] = [ a 0 + a 1 u + a 2 v b 0 + b 1 u + b 2 v ]
  • Where (x, y) is the new transformed coordinate of (u, v), and (aj, bk) (j, k=1, 2, 3) is the set of transformation parameters. Further, to determine the local transformation parameters for each sub-region, one or more embodiments of the invention include using a least squares technique to compute the transformation parameters.
  • One or more embodiments of the invention also include forward/backward frame-to-frame registration. For example, with instances of rapid camera motion, strong illumination variation and heavy stripe noise, to avoid residual error propagation, forward/backward frame-to-frame registration is carried out for multi-frame differencing. FIG. 3 illustrates an approach.
  • FIG. 3 is a diagram illustrating forward and backward geometric registration, according to an embodiment of the present invention. By way of illustration, FIG. 3 depicts frame 302 (Fi−1), frame 304 (Fi) and frame 306 (F1+1). To estimate object motion at frame 304 (Fi), which is taken as a reference frame, previous frame 302 (Fi−1) and next frame 306 (Fi+1) are geometrically registered to the reference frame. Motion estimation for each frame is carried out in such a fashion.
  • Forward/backward frame differencing can also be implemented for motion detection. A diagram of the approach used in one or more embodiments of the invention is illustrated in FIG. 4. FIG. 4 is a flow diagram illustrating forward and backward frame differencing, according to an embodiment of the present invention. After forward/backward frame-to-frame images (for example, frame 402, frame 404 and frame 406) are geometrically registered and aligned in steps 408 and 410, difference images are calculated. Instead of using simple subtraction between the aligned frames, one or more embodiments of the invention use forward/backward frame differencing in steps 412 and 414 to reduce motion noise and compensate the illustration variation such as automatic gain control.
  • Additionally, step 416 includes performing image arithmetic via Inew=ΔIi−1,i AND Δi,i+1. Step 418 includes median filtering, which can reduce random motion noise. To extract moving pixels of object moving objects, automatic dynamic threshold estimation based on spatial filtering in step 420 is carried out. Further, step 422 includes performing a morphological operation to remove small isolated spots and fill holes in foreground image and step 424 includes generating motion pixels (for example, a motion map).
  • To further reduce random noise and illumination variation effect, logical AND operation is implemented for forward/backward difference images to get a final difference image.
  • { D i - 1 , i ( x , y ) = F i - 1 ( x , y ) - F i ( x , y ) ; D i , i + 1 ( x , y ) = F i ( x , y ) - F i + 1 ( x , y ) ; D i ( x , y ) = D i - 1 , i ( x , y ) D i , i + 1 ( x , y ) ; i = 1 , 2 , , N
  • A threshold for each pixel is statistically calculated automatically in terms of statistical characteristics and spatial high frequency data of difference image. Further, a morphology step can be applied to remove small isolated spots and fill holes in the foreground image.
  • As described herein, one or more embodiments of the invention also include motion verification. FIG. 5 is a flow diagram illustrating false blob filtering, according to an embodiment of the present invention. Step 502 includes generating a motion map. Step 504 includes applying a connected component process to link each blob data. Step 506 includes creating a motion blob table. Step 508 includes performing an optical flow estimation. Step 510 includes making a displacement determination. If there is displacement, the process proceeds to step 512, which includes performing post-processing such as, for example, data association, object tracking, trajectory maintenance and track data management. If there is no displacement, the process proceeds to step 514, which includes filtering false blobs.
  • Accordingly, after a blob table is created, in order to remove false motion blobs from the blob table, each blob data is verified. One or more embodiments of the invention apply a KLT process to estimate the motion of each blob after forward/backward frame-to-frame registration is done. A false blob will be deleted from the blob table. The process steps can include, for example, applying a connected component process to link each blob data, creating a blob table, extracting features for each blob in a previous registered frame, applying the KLT method to estimate the motion of each blob, and if no motion occurs, the blob is deleted from the blob table. Also, the above-noted steps can be repeated for all blobs.
  • As also detailed herein, one or more embodiments of the invention include multi-object tracking. FIG. 6 is a flow diagram illustrating multi-object tracking, according to an embodiment of the present invention. Step 602 includes generating a motion map. Step 604 includes identifying moving blobs. Step 606 includes object initialization and step 608 includes object checking. Step 610 includes identifying object regions. Step 612 includes identifying candidate regions. Also, step 614 includes meanshift tracking and step 616 includes identifying new locations.
  • Additionally, after identifying object regions in step 610, features can be extracted in step 618. Once a search region is set in step 620, moving blobs can be found as potential object candidates in step 622. KLT matching is performed in step 624 and outlier removal based on an affine transform with RANSAC is performed in step 626. A new region candidate is identified in step 628. Meanwhile, Meanshift is applied in step 614 to compute the inter-frame translation. This yields a candidate region location in step 616. From steps 628 and 616, the process can proceed to step 630, which determines the final region location based on the Bhattacharyya coefficient. Also, step 632 includes target model updating for solving drift issues, and step 634 includes trajectory updating. Also, to track moving objects, a hybrid tracking model based on the combination of KLT and Meanshift method is applied from step 618 to 630.
  • As noted, the techniques described herein include object initialization. The motion detection results from forward/backward frame differencing can contain some correct real moving objects and some false objects, and miss some true objects. By way of example, for an UAV video with low frame rate (for example, 1 frame/second), a moving object does not have any overlapping regions between two consecutive frames so that traditional methods for object initialization will not work. To efficiently isolate promising moving objects among all detection results for current frame, one or more embodiments of the invention include combining a distance matrix with a similarity measure to initialize moving objects. The processing steps can include, for example, the following.
  • A search radius is set, matching score threshold and minimum length of tracked history. The distance matrix between the objects (including object candidates) and all the blobs in the table is computed. If the length of object trajectory is less than the preset value, a Kernel-based algorithm is applied to find the match between the object candidate and blobs in terms of a preset matching score. Also, if the object candidate appears in several consecutive frames, this candidate will be initialized and stored on the object table. Otherwise, the object candidate will be considered as a false object.
  • From the previous frame, one or more embodiments of the invention include projecting the previous blob set into a current frame after geometrical registration. The motion of each object according to its previous position can be estimated by a KLT tracking process. In a KLT tracking process, a motion model is approximately represented by an affine transformation, such that, Icurr(Ax+T)=Iprev(x), where A is a two-dimensional (2D) transformation matrix and T is the translation vector.
  • In one or more embodiments of the invention, affine transformation parameters can be computed from as few as four feature points. To determine these parameters, a least squares technique can be used to compute them.
  • Accuracy estimation can be performed, for example, when the number of mismatched pairs occurs. One measure of tracking accuracy is the root mean square error (RMSE) between the matched points before and after the affine transformation formula. This measure is used as a criterion to eliminate the matches that are considered imprecise.
  • Additionally, to eliminate the outliers, one or more embodiments of the invention includes performing the RANSAC algorithm to sequentially remove mis-matches in an iterative fashion until the RMSE value is lower than the desired threshold.
  • The techniques detailed herein additionally include meanshift tracking and object representation. By way of example, for a UAV tracking system, traditional intensity-based target representation is no longer suitable for multi-object tracking due to large scale variation and perspective geometric distortion. To efficiently characterize the object, histogram-based feature space can be chosen. In one or more embodiments of the invention, a metric based on the Bhattacharyya coefficient is used to define a similarity measure between a reference object and a candidate for multi-object tracking. Given an object region histogram q in the reference frame, the Bhattacharyya coefficient based objective function is given by:
  • ρ ( p , q ) = u = 1 M p u ( x ) q u ( x 0 )
  • where M is the histogram dimension, and x0 is the 2D center.
  • The candidate region histogram pu(x) at 2D center x in the current frame is defined as:
  • p u ( x ) = k ( x - x i h 2 ) δ ( b ( x i ) , u ) k ( x - x i h 2 )
  • Here, u=1, 2, . . . , M. k(x) denotes a non-negative, non-increasing and piecewise-differentiable kernel profile which weights the pixel location, h is 2D bandwidth vector of k(x), δ is the Kronecker delta function and each pixel value is denoted by b(xi).
  • Additionally, in one or more embodiments of the invention, in determining a similarity measure between distributions, the Bhattacharyya distance can include B(Ix, Iy)=√{square root over (1−ρ(px, py))}, where ρ(px, py)=∫√{square root over ({circumflex over (p)}x(u){circumflex over (p)}y(u))} du, and where ρx and py represent the target and the candidate distributions, respectively.
  • The techniques described herein can additionally include object positioning. To search the location corresponding to the object from one frame to the next, one or more embodiments of the invention include applying a meanshift tracking algorithm that is based on a gradient ascent optimization rather than an exhaustive search. Strengths of the meanshift method include computational effectiveness and suitability to real-time application. However, a target can be lost, for example, due to an intrinsic limitation of exploring local maxima, especially when the tracked object moves quickly. The candidate region histogram pu(x) can be obtained from the above equation.
  • The new location of the tracked object can be estimated as:
  • y ^ 1 = i = 1 n X i ω i g ( y ^ 0 - X i h 2 ) i = 1 n ω i g ( y ^ 0 X i h 2 )
  • where:
  • ω 1 = u = 1 m δ [ b ( X i ) - u ] q ^ u p ^ u ( y ^ 0 )
  • g(x)=−k(x), that the derivative of k(x).
  • One or more embodiments of the invention can also include target model updating on a temporal domain. In some circumstances, a meanshift approach without target model updating can suffer from abrupt changes in target model. On the other hand, the model updating for every frame can result in decreasing the reliability of the tracking results due to cluttered environment, occlusion, random noise, etc. One way to change the target model is to periodically update the target distributions.
  • To obtain a precise tracking result, the target model can be updated dynamically. Accordingly, one or more embodiments of the invention include model updating that use both recent tracking results and older target model to impact a current target model for object tracking. The updating procedure is formulized as:

  • q u new=(1−α)q u old +α·p u s
  • Here, the superscripts of new and old denote the newly obtained target model and the old model, respectively. s represents the recent tracking result. α weights the contribution of the recent tracking result (normally <0.1). q and p represent the target model and the candidate model, respectively.
  • Further, one or more embodiments of the invention include target model updating on a spatial domain. Normally, meanshift based tracking hardly provides precise boundary position of the tracked object due to lack of utilizing spatial data. Fortunately, detection results derived from KLT tracker and motion detection results can provide much more accurate information, such as the precise position and object size compared with meanshift tracker.
  • Each individual algorithm may unable to do a perfect job on multi-object tracking. Thus, fusion among their data can be used in a multi-object tracking procedure. According to the strengths of each method, one or more embodiments of the invention use the following merging method:
  • Output = { result by motion detector ; if Overlapping T KLT result ; if Outlier for MS occurs result by meanshift ; otherwise
  • where overlapping represents the degree of overlapping region.
  • FIG. 7 is a diagram illustrating reference plane-based registration and tracking, according to an embodiment of the present invention. By way of illustration, FIG. 7 depicts a geo-reference plane 702. The first frame 704 is registered to geo-reference plane 702, and the second frame 706 is registered to the geo-reference 702 from the first registered frame and corresponding inter-frame transformation parameters TCi (equation 712 in FIG. 7). In such fashion, frames 708 and 710 are registered to the geo-reference 702, respectively. Moreover, each object is projected into geo-reference 702 using navigation data.
  • FIG. 8 is a flow diagram illustrating automatic urban road extraction, according to an embodiment of the present invention. Step 802 includes framing an image. Step 804 includes performing a Gaussian smoothing operation. Also, step 806 includes using a canny detector and step 808 includes implementing a hough transformation. Step 810 includes determining a maximum response finding. Step 812 includes determining if the length of the stripe is greater than a pre-defined threshold. If the length of the stripe is not greater than the threshold, the process stops at step 814. If the length of the stripe is greater than the threshold, the process continues to step 816, which includes performing a straight line extraction. Further, step 818 includes performing stripe pixels removal (which can, for example, lead to a return to step 808).
  • As also depicted in FIG. 8, step 820 includes performing frame differencing, and step 822 includes verification via motion history images (MHI) (which can, for example, lead to a return to step 816). Additionally, one or more embodiments of the invention can also include extraction of road stripes via iterative hough transform.
  • As detailed herein, one or more embodiments of the invention include recursive geometric registration with sub-pixel matching accuracy that can handle various geometrical residual errors from un-calibrated camera. Additionally, the techniques detailed herein include motion detection based on forward/backward frame differencing that can efficiently separate moving objects from background. Further, a hybrid object tracker can be implemented that uses colors, features and intensity statistical characteristics overtime to detect and track multiple small objects.
  • FIG. 9 is a block diagram illustrating architecture of an object detection and tracking system, according to an aspect of the invention. An example software architecture construction for a detection and tracking system (for example, a UAV system) can be built on multiple services to provide a track database for object search and intelligent analysis. As illustrated in FIG. 9, the software architecture can include multiple sensor modules 904, video streaming service modules 906, tracking suite service modules 908, a track database (DB) server module 910, a user interface module 902 and a visualization console 912. A video streaming module 906 serves to capture and make available imagery from multiple sensors. The acquired images are used by a tracking suite module 908 as the basis for multi-object detection and tracking. Tracking suite modules 908 includes a geometric registration sub-module 914, a motion extraction sub-module 916, an object tracking sub-module 918, a tracking data sub-module 920 and a geo-coordinate mapping sub-module 922.
  • By processing the real-time imagery from multiple sensors, sophisticated transformation of data to track information is achieved. Track DB server 910 serves track metadata management. Visualization console 912 creates graphical overlays, indexes them to the imagery on the display, and presents them to a user. These overlays can be any type of graphical information that supports the higher level components, such as, for example, class types, moving directions, trajectories and object sizes. User interface 902 provides data access and operation by the user.
  • FIG. 10 is a flow diagram illustrating techniques for performing visual surveillance of one or more moving objects, according to an embodiment of the present invention. Step 1002 includes registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames. This step can be carried out, for example, using a geometric registration sub-module 914 in tracking suite service module 908. Registering images can include recursive global and local geometric registration of the one or more images (for example, region-based geometric transformation for handling lens geometric distortion). Registering images can also include using sub-pixel image matching techniques.
  • Step 1004 includes performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images. This step can be carried out, for example, using a motion extraction sub-module 916 in tracking suite service module 908. Performing motion segmentation of the images can include forward and backward frame differencing. Forward and backward frame differences can include, for example, automatic dynamic threshold estimation based on temporary filtering and/or spatial filtering, removing false moving pixels based on independent motions of image features, and performing a morphological operation and generating motion pixels.
  • Step 1006 includes tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects. This step can be carried out, for example, using an object tracking sub-module 918 in tracking suite service module 908. Tracking the moving objects can include performing hybrid target tracking, wherein hybrid target tracking includes using a Kanade-Lucas-Tomasi feature tracker and meanshift, using auto kernel scale estimation and updating, and using feature trajectories. One or more embodiments of the invention can also include using colors for tracking. Tracking moving objects can additionally include using multi-target tracking algorithms based on feature matching and distance matrices for one or more (small) targets.
  • Also, tracking moving objects can include generating a motion map, identifying one or more moving objects (blobs), performing object initialization and object checking, identifying object regions in the motion map, extracting features, setting a search region in the motion map, identifying candidate regions in the motion map, meanshift tracking, identifying moving objects in the candidate regions, performing Kanade-Lucas-Tomasi feature matching, performing an affine transform (with RANSAC), making a final regions determination via the Bhattacharyya coefficient, and updating a target model and trajectory information. Tracking moving objects can additionally include reference plane-based registration and tracking.
  • The techniques depicted in FIG. 10 can also include relating each camera view with one or more other camera views, and forming a panoramic view from the images captured by one or more cameras. One or more embodiments of the invention additionally include estimating motion of each camera based on video information of static objects in the panoramic view, as well as estimating one or more background (for example, road) structures in the panoramic view based on linear structure detection and statistical analysis of the moving objects over a period of time.
  • Further, the techniques depicted in FIG. 10 include automatic feature (for example, a road) extraction, wherein automatic feature extraction includes framing an image, performing a Gaussian smoothing operation, using a canny detector to extract one or more feature (for example, road) edges, implementing a hough transformation for feature (for example, road stripe) analysis, determining a maximum response finding for reducing an influence of multiple peaks in a transform space, determining if a length of a feature (for example, a road stripe) is greater than a certain threshold, and if the length of the feature is greater than the threshold, performing feature extraction and pixel removal. Automatic feature extraction can additionally include performing frame differencing and verification via motion history images.
  • One or more embodiments of the invention also include performing outlier removal to remove incorrect moving object matches (and improve the registration precision). The techniques depicted in FIG. 10 can additionally include false blob filtering. False blob filtering includes generating a motion map, applying a connected component process to link each blob data, creating a motion blob table, extracting features for each blob in a previously registered frame, and applying a Kanade-Lucas-Tomasi method to estimate motion of each blob, and, if no motion occurs for a blob, deleting the blob from the blob table.
  • Additionally, one or more embodiments of the invention can include updating a target model on a temporal domain and/or a spatial domain, as well as creating an index (for example, a searchable index) of object appearances and object tracks in a panoramic view. Also, the object appearance and tracks template index can be stored in a template data store with a pointer to the corresponding video segments for easy retrieval. Further, one or more embodiments of the invention can include determining a similarity metric between a query and an entry in the index, which can facilitate searching for the object appearance and tracks in a template data store/index based on the similarity metric, and outputting/listing the search results for a human operator based on similarity of the query.
  • The techniques depicted in FIG. 10 can also, as described herein, include providing a system, wherein the system includes distinct software modules, each of the distinct software modules being embodied on a tangible computer-readable recordable storage medium. All the modules (or any subset thereof) can be on the same medium, or each can be on a different medium, for example. The modules can include any or all of the components shown in the figures. In one or more embodiments, the modules include sensor modules, video streaming service modules, tracking suite service modules (including the sub-modules detailed herein), a track database (DB) server module, a user interface module and a visualization console module that can run, for example on one or more hardware processors. The method steps can then be carried out using the distinct software modules of the system, as described above, executing on the one or more hardware processors. Further, a computer program product can include a tangible computer-readable recordable storage medium with code adapted to be executed to carry out one or more method steps described herein, including the provision of the system with the distinct software modules.
  • Additionally, the techniques depicted in FIG. 10 can be implemented via a computer program product that can include computer useable program code that is stored in a computer readable storage medium in a data processing system, and wherein the computer useable program code was downloaded over a network from a remote data processing system. Also, in one or more embodiments of the invention, the computer program product can include computer useable program code that is stored in a computer readable storage medium in a server data processing system, and wherein the computer useable program code are downloaded over a network to a remote data processing system for use in a computer readable storage medium with the remote system.
  • As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • One or more embodiments of the invention, or elements thereof, can be implemented in the form of an apparatus including a memory and at least one processor that is coupled to the memory and operative to perform exemplary method steps.
  • One or more embodiments can make use of software running on a general purpose computer or workstation. With reference to FIG. 11, such an implementation might employ, for example, a processor 1102, a memory 1104, and an input/output interface formed, for example, by a display 1106 and a keyboard 1108. The term “processor” as used herein is intended to include any processing device, such as, for example, one that includes a CPU (central processing unit) and/or other forms of processing circuitry. Further, the term “processor” may refer to more than one individual processor. The term “memory” is intended to include memory associated with a processor or CPU, such as, for example, RAM (random access memory), ROM (read only memory), a fixed memory device (for example, hard drive), a removable memory device (for example, diskette), a flash memory and the like. In addition, the phrase “input/output interface” as used herein, is intended to include, for example, one or more mechanisms for inputting data to the processing unit (for example, mouse), and one or more mechanisms for providing results associated with the processing unit (for example, printer). The processor 1102, memory 1104, and input/output interface such as display 1106 and keyboard 1108 can be interconnected, for example, via bus 1110 as part of a data processing unit 1112. Suitable interconnections, for example via bus 1110, can also be provided to a network interface 1114, such as a network card, which can be provided to interface with a computer network, and to a media interface 1116, such as a diskette or CD-ROM drive, which can be provided to interface with media 1118.
  • Accordingly, computer software including instructions or code for performing the methodologies of the invention, as described herein, may be stored in one or more of the associated memory devices (for example, ROM, fixed or removable memory) and, when ready to be utilized, loaded in part or in whole (for example, into RAM) and implemented by a CPU. Such software could include, but is not limited to, firmware, resident software, microcode, and the like.
  • A data processing system suitable for storing and/or executing program code will include at least one processor 1102 coupled directly or indirectly to memory elements 1104 through a system bus 1110. The memory elements can include local memory employed during actual implementation of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during implementation.
  • Input/output or I/O devices (including but not limited to keyboards 1108, displays 1106, pointing devices, and the like) can be coupled to the system either directly (such as via bus 1110) or through intervening I/O controllers (omitted for clarity).
  • Network adapters such as network interface 1114 may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
  • As used herein, including the claims, a “server” includes a physical data processing system (for example, system 1112 as shown in FIG. 11) running a server program. It will be understood that such a physical server may or may not include a display and keyboard.
  • As noted, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon. Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. Media block 1118 is a non-limiting example. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, component, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
  • It should be noted that any of the methods described herein can include an additional step of providing a system comprising distinct software modules embodied on a computer readable storage medium; the modules can include, for example, any or all of the components shown in FIG. 9. The method steps can then be carried out using the distinct software modules and/or sub-modules of the system, as described above, executing on one or more hardware processors 1102. Further, a computer program product can include a computer-readable storage medium with code adapted to be implemented to carry out one or more method steps described herein, including the provision of the system with the distinct software modules.
  • In any case, it should be understood that the components illustrated herein may be implemented in various forms of hardware, software, or combinations thereof; for example, application specific integrated circuit(s) (ASICS), functional circuitry, one or more appropriately programmed general purpose digital computers with associated memory, and the like. Given the teachings of the invention provided herein, one of ordinary skill in the related art will be able to contemplate other implementations of the components of the invention.
  • The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a,” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
  • The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
  • At least one embodiment of the invention may provide one or more beneficial effects, such as, for example, automatic dynamic threshold determination based on temporary and/or spatial domain.
  • It will be appreciated and should be understood that the exemplary embodiments of the invention described above can be implemented in a number of different fashions. Given the teachings of the invention provided herein, one of ordinary skill in the related art will be able to contemplate other implementations of the invention. Indeed, although illustrative embodiments of the present invention have been described herein with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various other changes and modifications may be made by one skilled in the art.

Claims (23)

1. A method for performing visual surveillance of one or more moving objects, wherein the method comprises:
registering one or more images captured by one or more cameras, wherein registering the one or more images comprises region-based registration of the one or more images in two or more adjacent frames;
performing motion segmentation of the one or more images to detect one or more moving objects and one or more background regions in the one or more images; and
tracking the one or more moving objects to facilitate visual surveillance of the one or more moving objects.
2. The method of claim 1, wherein registering one or more images comprises recursive global and local geometric registration of the one or more images.
3. The method of claim 1, wherein registering one or more images comprises using one or more sub-pixel image matching techniques.
4. The method of claim 1, wherein performing motion segmentation of the one or more images comprises forward and backward frame differencing.
5. The method of claim 4, wherein forward and backward frame differences comprises automatic dynamic threshold estimation based on at least one of temporary filtering and spatial filtering.
6. The method of claim 4, wherein forward and backward frame differences comprises removing one or more false moving pixels based on independent motions of one or more image features.
7. The method of claim 4, wherein forward and backward frame differences comprises performing a morphological operation and generating one or more motion pixels.
8. The method of claim 1, wherein tracking the one or more moving objects comprises performing hybrid target tracking, wherein hybrid target tracking comprises using a Kanade-Lucas-Tomasi feature tracker and meanshift, using auto kernel scale estimation and updating, and using one or more feature trajectories.
9. The method of claim 1, wherein tracking the one or more moving objects comprises using one or more multi-target tracking algorithms based on feature matching and distance matrices for one or more targets.
10. The method of claim 1, wherein tracking the one or more moving objects comprises:
generating a motion map;
identifying one or more moving objects;
performing object initialization and object checking;
identifying one or more object regions in the motion map;
extracting one or more features;
setting a search region in the motion map;
identifying one or more candidate regions in the motion map;
meanshift tracking;
identifying one or more moving objects in the one or more candidate regions;
performing Kanade-Lucas-Tomasi feature matching;
performing an affine transform;
making a final regions determination via the Bhattacharyya coefficient; and
updating a target model and trajectory information.
11. The method of claim 1, wherein tracking the one or more moving objects comprises reference plane-based registration and tracking.
12. The method of claim 1, further comprising relating each camera view with one or more other camera views.
13. The method of claim 1, further comprising forming a panoramic view from the one or more images captured by one or more cameras.
14. The method of claim 13, further comprising estimating motion of each camera based on video information of one or more static objects in the panoramic view.
15. The method of claim 13, further comprising estimating one or more background structures in the panoramic view based on linear structure detection and statistical analysis of the one or more moving objects over a period of time.
16. The method of claim 1, further comprising automatic feature extraction, wherein automatic feature extraction comprises:
framing an image;
performing a Gaussian smoothing operation;
using a canny detector to extract one or more feature edges;
implementing a hough transformation for feature analysis;
determining a maximum response finding for reducing an influence of multiple peaks in a transform space;
determining if a length of a feature is greater than a certain threshold, and if the length of the feature is greater than the threshold, performing feature extraction and pixel removal.
17. The method of claim 16, wherein automatic feature extraction further comprises performing frame differencing and verification via motion history images.
18. The method of claim 1, further comprising performing outlier removal to remove one or more incorrect moving object matches.
19. The method of claim 1, further comprising false blob filtering, wherein false blob filtering comprises:
generating a motion map;
applying a connected component process to link each blob data;
creating a motion blob table;
extracting one or more features for each blob in a previously registered frame; and
applying a Kanade-Lucas-Tomasi method to estimate motion of each blob, and, if no motion occurs for a blob, deleting the blob from the blob table.
20. The method of claim 1, further comprising updating a target model on at least one of a temporal domain and a spatial domain.
21. The method of claim 1, further comprising creating an index of object appearances and object tracks in a panoramic view.
22. The method of claim 21, further comprising determining a similarity metric between a query and an entry in the index.
23. The method of claim 1, further comprising providing a system, wherein the system comprises one or more distinct software modules, each of the one or more distinct software modules being embodied on a tangible computer-readable recordable storage medium, and wherein the one or more distinct software modules comprise a geometric registration module, a motion extraction module and an object tracking module executing on a hardware processor.
US13/609,393 2010-12-20 2012-09-11 Detection and Tracking of Moving Objects Abandoned US20130002866A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/609,393 US20130002866A1 (en) 2010-12-20 2012-09-11 Detection and Tracking of Moving Objects

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/972,836 US9147260B2 (en) 2010-12-20 2010-12-20 Detection and tracking of moving objects
US13/609,393 US20130002866A1 (en) 2010-12-20 2012-09-11 Detection and Tracking of Moving Objects

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/972,836 Continuation US9147260B2 (en) 2010-12-20 2010-12-20 Detection and tracking of moving objects

Publications (1)

Publication Number Publication Date
US20130002866A1 true US20130002866A1 (en) 2013-01-03

Family

ID=45495895

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/972,836 Active 2032-12-13 US9147260B2 (en) 2010-12-20 2010-12-20 Detection and tracking of moving objects
US13/609,393 Abandoned US20130002866A1 (en) 2010-12-20 2012-09-11 Detection and Tracking of Moving Objects

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/972,836 Active 2032-12-13 US9147260B2 (en) 2010-12-20 2010-12-20 Detection and tracking of moving objects

Country Status (6)

Country Link
US (2) US9147260B2 (en)
JP (2) JP2014504410A (en)
CN (1) CN103262121A (en)
DE (1) DE112011103690T5 (en)
GB (1) GB2500839B (en)
WO (1) WO2012084703A1 (en)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110317924A1 (en) * 2010-06-28 2011-12-29 Sony Corporation Image processing apparatus, image processing method, and image processing program
US20140079320A1 (en) * 2012-09-17 2014-03-20 Gravity Jack, Inc. Feature Searching Along a Path of Increasing Similarity
CN104202559A (en) * 2014-08-11 2014-12-10 广州中大数字家庭工程技术研究中心有限公司 Intelligent monitoring system and intelligent monitoring method based on rotation invariant feature
CN104217442A (en) * 2014-08-28 2014-12-17 西北工业大学 Aerial video moving object detection method based on multiple model estimation
CN104407619A (en) * 2014-11-05 2015-03-11 沈阳航空航天大学 Method enabling multiple unmanned aerial vehicles to reach multiple targets simultaneously under uncertain environments
US9076212B2 (en) 2006-05-19 2015-07-07 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
CN104766319A (en) * 2015-04-02 2015-07-08 西安电子科技大学 Method for improving registration precision of images photographed at night
US9305365B2 (en) 2013-01-24 2016-04-05 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
CN105913459A (en) * 2016-05-10 2016-08-31 中国科学院自动化研究所 Moving object detection method based on high resolution continuous shooting images
US20160321820A1 (en) * 2015-05-01 2016-11-03 Raytheon Company Systems and methods for 3d point cloud processing
US20160358018A1 (en) * 2015-06-02 2016-12-08 SK Hynix Inc. Moving object detection device and object detection method
US9606209B2 (en) 2011-08-26 2017-03-28 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
US9717461B2 (en) 2013-01-24 2017-08-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9734589B2 (en) 2014-07-23 2017-08-15 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
CN107172341A (en) * 2016-03-07 2017-09-15 深圳市朗驰欣创科技股份有限公司 A kind of unmanned aerial vehicle (UAV) control method, unmanned plane, earth station and UAS
US9782141B2 (en) 2013-02-01 2017-10-10 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US9943247B2 (en) 2015-07-28 2018-04-17 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
US20180112979A1 (en) * 2015-06-26 2018-04-26 SZ DJI Technology Co., Ltd. System and method for measuring a displacement of a mobile platform
WO2018053341A3 (en) * 2016-09-16 2018-05-17 Interactive Intelligence Group, Inc. System and method for body language analysis
US10004462B2 (en) 2014-03-24 2018-06-26 Kineticor, Inc. Systems, methods, and devices for removing prospective motion correction from medical imaging scans
CN108848304A (en) * 2018-05-30 2018-11-20 深圳岚锋创视网络科技有限公司 A kind of method for tracking target of panoramic video, device and panorama camera
CN108885469A (en) * 2016-09-27 2018-11-23 深圳市大疆创新科技有限公司 System and method for the initialized target object in tracking system
CN108982521A (en) * 2018-08-04 2018-12-11 石修英 Visualize the horizontal detection device of soil health
CN109146862A (en) * 2018-08-04 2019-01-04 石修英 Soil remediation condition intelligent detection device
CN109215056A (en) * 2017-07-03 2019-01-15 昊翔电能运动科技(昆山)有限公司 Target tracking method and device
CN109214243A (en) * 2017-07-03 2019-01-15 昊翔电能运动科技(昆山)有限公司 Method for tracking target, device and unmanned plane
US10327708B2 (en) 2013-01-24 2019-06-25 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10636152B2 (en) * 2016-11-15 2020-04-28 Gvbb Holdings S.A.R.L. System and method of hybrid tracking for match moving
US10659684B2 (en) 2016-02-16 2020-05-19 Samsung Electronics Co., Ltd. Apparatus and method for providing dynamic panorama function
US10716515B2 (en) 2015-11-23 2020-07-21 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
CN112308880A (en) * 2019-08-30 2021-02-02 华为技术有限公司 Target user locking method and electronic equipment

Families Citing this family (115)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110123744A (en) * 2009-01-28 2011-11-15 배 시스템즈 피엘시 Detecting potential changed objects in images
US20110115913A1 (en) * 2009-11-17 2011-05-19 Werner Lang Automated vehicle surrounding area monitor and display system
US9628755B2 (en) * 2010-10-14 2017-04-18 Microsoft Technology Licensing, Llc Automatically tracking user movement in a video chat application
US8625905B2 (en) * 2011-01-28 2014-01-07 Raytheon Company Classification of target objects in motion
EP2710557B1 (en) * 2011-05-16 2020-04-29 Max-Planck-Gesellschaft zur Förderung der Wissenschaften e.V. Fast articulated motion tracking
CN102982304B (en) * 2011-09-07 2016-05-25 株式会社理光 Utilize polarized light image to detect the method and system of vehicle location
CN103196550A (en) * 2012-01-09 2013-07-10 西安智意能电子科技有限公司 Method and equipment for screening and processing imaging information of launching light source
EP2850831A1 (en) 2012-05-14 2015-03-25 Luca Rossato Encoding and decoding based on blending of sequences of samples along time
GB201213604D0 (en) 2012-07-31 2012-09-12 Bae Systems Plc Detectig moving vehicles
JP5983209B2 (en) 2012-09-07 2016-08-31 株式会社Ihi Moving object detection method
JP6094099B2 (en) * 2012-09-07 2017-03-15 株式会社Ihi Moving object detection method
JP6094100B2 (en) * 2012-09-07 2017-03-15 株式会社Ihi Moving object detection method
US9672605B2 (en) * 2012-09-27 2017-06-06 Panasonic Intellectual Property Management Co., Ltd. Image processing device and image processing method
TWI482468B (en) * 2012-11-23 2015-04-21 Inst Information Industry Device, method and computer readable storage medium thereof for detecting object
US9201958B2 (en) * 2013-10-24 2015-12-01 TCL Research America Inc. Video object retrieval system and method
US9165208B1 (en) * 2013-03-13 2015-10-20 Hrl Laboratories, Llc Robust ground-plane homography estimation using adaptive feature selection
US9367067B2 (en) * 2013-03-15 2016-06-14 Ashley A Gilmore Digital tethering for tracking with autonomous aerial robot
US9625995B2 (en) * 2013-03-15 2017-04-18 Leap Motion, Inc. Identifying an object in a field of view
US9430846B2 (en) * 2013-04-19 2016-08-30 Ge Aviation Systems Llc Method of tracking objects using hyperspectral imagery
US9317770B2 (en) * 2013-04-28 2016-04-19 Tencent Technology (Shenzhen) Co., Ltd. Method, apparatus and terminal for detecting image stability
JP6202879B2 (en) * 2013-05-20 2017-09-27 株式会社朋栄 Rolling shutter distortion correction and image stabilization processing method
US9491360B2 (en) 2013-06-06 2016-11-08 Apple Inc. Reference frame selection for still image stabilization
US9384552B2 (en) * 2013-06-06 2016-07-05 Apple Inc. Image registration methods for still image stabilization
US9277129B2 (en) * 2013-06-07 2016-03-01 Apple Inc. Robust image feature based video stabilization and smoothing
CN104424651A (en) * 2013-08-26 2015-03-18 株式会社理光 Method and system for tracking object
US20150071547A1 (en) 2013-09-09 2015-03-12 Apple Inc. Automated Selection Of Keeper Images From A Burst Photo Captured Set
CN103914850B (en) * 2014-04-22 2017-02-15 南京影迹网络科技有限公司 Automatic video labeling method and system based on motion matching
CN103957423A (en) * 2014-05-14 2014-07-30 杭州古北电子科技有限公司 Video compression and reconstruction method based on computer vision
EP3060966B1 (en) 2014-07-30 2021-05-05 SZ DJI Technology Co., Ltd. Systems and methods for target tracking
CN104243819B (en) * 2014-08-29 2018-02-23 小米科技有限责任公司 Photo acquisition methods and device
CN107004054B (en) * 2014-12-04 2022-04-15 皇家飞利浦有限公司 Calculating health parameters
US10713506B2 (en) 2014-12-18 2020-07-14 Magna Electronics Inc. Vehicle vision system with 3D registration for distance estimation
US9792664B2 (en) 2015-01-29 2017-10-17 Wipro Limited System and method for mapping object coordinates from a video to real world coordinates using perspective transformation
US10380430B2 (en) 2015-04-17 2019-08-13 Current Lighting Solutions, Llc User interfaces for parking zone creation
US10043307B2 (en) 2015-04-17 2018-08-07 General Electric Company Monitoring parking rule violations
US10762112B2 (en) * 2015-04-28 2020-09-01 Microsoft Technology Licensing, Llc Establishing search radius based on token frequency
CN110049206B (en) * 2015-04-28 2021-08-10 腾讯科技(深圳)有限公司 Image processing method, image processing apparatus, and computer-readable storage medium
US9483839B1 (en) * 2015-05-06 2016-11-01 The Boeing Company Occlusion-robust visual object fingerprinting using fusion of multiple sub-region signatures
US9936128B2 (en) 2015-05-20 2018-04-03 Google Llc Automatic detection of panoramic gestures
JP6558951B2 (en) * 2015-05-27 2019-08-14 株式会社パスコ Tunnel wall damage detection device and tunnel wall damage detection program
CN105974940B (en) * 2016-04-29 2019-03-19 优利科技有限公司 Method for tracking target suitable for aircraft
CN105023278B (en) * 2015-07-01 2019-03-05 中国矿业大学 A kind of motion target tracking method and system based on optical flow method
KR101645722B1 (en) * 2015-08-19 2016-08-05 아이디어주식회사 Unmanned aerial vehicle having Automatic Tracking and Method of the same
EP3347789B1 (en) * 2015-09-11 2021-08-04 SZ DJI Technology Co., Ltd. Systems and methods for detecting and tracking movable objects
KR101756391B1 (en) 2015-10-30 2017-07-26 이노뎁 주식회사 Object link system, intergrated control system having the same linking the object images and big data analysis sysytem, and operating method thereof
KR102410268B1 (en) * 2015-11-20 2022-06-20 한국전자통신연구원 Object tracking method and object tracking apparatus for performing the method
US10540901B2 (en) 2015-11-23 2020-01-21 Kespry Inc. Autonomous mission action alteration
WO2017091768A1 (en) 2015-11-23 2017-06-01 Kespry, Inc. Autonomous mission action alteration
US10339387B2 (en) 2016-03-03 2019-07-02 Brigham Young University Automated multiple target detection and tracking system
US10421012B2 (en) 2016-03-25 2019-09-24 Zero Latency PTY LTD System and method for tracking using multiple slave servers and a master server
US9916496B2 (en) 2016-03-25 2018-03-13 Zero Latency PTY LTD Systems and methods for operating a virtual reality environment using colored marker lights attached to game objects
US10071306B2 (en) 2016-03-25 2018-09-11 Zero Latency PTY LTD System and method for determining orientation using tracking cameras and inertial measurements
US10486061B2 (en) 2016-03-25 2019-11-26 Zero Latency Pty Ltd. Interference damping for continuous game play
US10717001B2 (en) 2016-03-25 2020-07-21 Zero Latency PTY LTD System and method for saving tracked data in the game server for replay, review and training
CN109071015B (en) 2016-04-29 2021-11-30 美国联合包裹服务公司 Unmanned aerial vehicle picks up and delivers system
US10730626B2 (en) 2016-04-29 2020-08-04 United Parcel Service Of America, Inc. Methods of photo matching and photo confirmation for parcel pickup and delivery
US10026193B2 (en) 2016-05-24 2018-07-17 Qualcomm Incorporated Methods and systems of determining costs for object tracking in video analytics
CN106127801A (en) * 2016-06-16 2016-11-16 乐视控股(北京)有限公司 A kind of method and apparatus of moving region detection
US10284875B2 (en) 2016-08-08 2019-05-07 Qualcomm Incorporated Systems and methods for determining feature point motion
US10751609B2 (en) 2016-08-12 2020-08-25 Zero Latency PTY LTD Mapping arena movements into a 3-D virtual world
CN106355606A (en) * 2016-08-30 2017-01-25 山东航天电子技术研究所 Method for tracking high-speed target image of unmanned aerial vehicle
TWI616843B (en) * 2016-09-12 2018-03-01 粉迷科技股份有限公司 Method, system for removing background of a video, and a computer-readable storage device
US10497143B2 (en) * 2016-11-14 2019-12-03 Nec Corporation Advanced driver-assistance system using accurate object proposals by tracking detections
CN106548487B (en) * 2016-11-25 2019-09-03 浙江光跃环保科技股份有限公司 Method and apparatus for detection and tracking mobile object
US10269133B2 (en) * 2017-01-03 2019-04-23 Qualcomm Incorporated Capturing images of a game by an unmanned autonomous vehicle
CN108733042B (en) * 2017-04-19 2021-11-09 上海汽车集团股份有限公司 Target tracking method and device for automatic driving vehicle
KR101876349B1 (en) * 2017-04-20 2018-07-09 사회복지법인 삼성생명공익재단 System and method for controlling medical treatment machine
US11010630B2 (en) * 2017-04-27 2021-05-18 Washington University Systems and methods for detecting landmark pairs in images
CN108876811B (en) * 2017-05-10 2024-02-02 中兴通讯股份有限公司 Image processing method, device and computer readable storage medium
US10775792B2 (en) 2017-06-13 2020-09-15 United Parcel Service Of America, Inc. Autonomously delivering items to corresponding delivery locations proximate a delivery route
EP3640887A4 (en) * 2017-06-13 2021-03-17 IHI Corporation Mobile body observation method
DE102017113794A1 (en) * 2017-06-22 2018-12-27 Connaught Electronics Ltd. Classification of static and dynamic image segments in a driver assistance device of a motor vehicle
CN107392937B (en) * 2017-07-14 2023-03-14 腾讯科技(深圳)有限公司 Target tracking method and device and electronic equipment
US10453187B2 (en) * 2017-07-21 2019-10-22 The Boeing Company Suppression of background clutter in video imagery
US10210391B1 (en) * 2017-08-07 2019-02-19 Mitsubishi Electric Research Laboratories, Inc. Method and system for detecting actions in videos using contour sequences
CN109509210B (en) 2017-09-15 2020-11-24 百度在线网络技术(北京)有限公司 Obstacle tracking method and device
US10872534B2 (en) 2017-11-01 2020-12-22 Kespry, Inc. Aerial vehicle inspection path planning
US10902615B2 (en) 2017-11-13 2021-01-26 Qualcomm Incorporated Hybrid and self-aware long-term object tracking
CN108446581B (en) * 2018-01-22 2022-07-19 北京理工雷科电子信息技术有限公司 Unmanned aerial vehicle detection method in severe environment
CN108596946A (en) * 2018-03-21 2018-09-28 中国航空工业集团公司洛阳电光设备研究所 A kind of moving target real-time detection method and system
WO2019227352A1 (en) * 2018-05-30 2019-12-05 深圳市大疆创新科技有限公司 Flight control method and aircraft
CN110610514B (en) * 2018-06-15 2023-09-19 株式会社日立制作所 Method, device and electronic equipment for realizing multi-target tracking
US10916031B2 (en) 2018-07-06 2021-02-09 Facebook Technologies, Llc Systems and methods for offloading image-based tracking operations from a general processing unit to a hardware accelerator unit
CN109118510A (en) * 2018-08-10 2019-01-01 平安科技(深圳)有限公司 A kind of monitor video processing method, device and computer-readable medium
CN109087378A (en) * 2018-09-11 2018-12-25 首都师范大学 Image processing method and system
CN109556596A (en) 2018-10-19 2019-04-02 北京极智嘉科技有限公司 Air navigation aid, device, equipment and storage medium based on ground texture image
US10839531B2 (en) * 2018-11-15 2020-11-17 Sony Corporation Object tracking based on a user-specified initialization point
CN109584266B (en) * 2018-11-15 2023-06-09 腾讯科技(深圳)有限公司 Target detection method and device
CN109522843B (en) * 2018-11-16 2021-07-02 北京市商汤科技开发有限公司 Multi-target tracking method, device, equipment and storage medium
WO2020114585A1 (en) * 2018-12-05 2020-06-11 Telefonaktiebolaget Lm Ericsson (Publ) Object location determination in frames of a video stream
WO2020128587A1 (en) * 2018-12-20 2020-06-25 Pratik Sharma Intelligent image sensor
CN109685830B (en) * 2018-12-20 2021-06-15 浙江大华技术股份有限公司 Target tracking method, device and equipment and computer storage medium
CN109685062B (en) * 2019-01-02 2023-07-25 南方科技大学 Target detection method, device, equipment and storage medium
JP7093015B2 (en) * 2019-04-24 2022-06-29 日本電信電話株式会社 Panorama video compositing device, panoramic video compositing method, and panoramic video compositing program
IT201900007815A1 (en) * 2019-06-03 2020-12-03 The Edge Company S R L METHOD FOR DETECTION OF MOVING OBJECTS
CN110223344B (en) * 2019-06-03 2023-09-29 哈尔滨工程大学 Infrared small target detection method based on morphology and visual attention mechanism
CN110349172B (en) * 2019-06-28 2022-12-16 华南理工大学 Power transmission line external damage prevention early warning method based on image processing and binocular stereo ranging
CN110675428B (en) * 2019-09-06 2023-02-28 鹏城实验室 Target tracking method and device for human-computer interaction and computer equipment
WO2021061112A1 (en) 2019-09-25 2021-04-01 Google Llc Gain control for face authentication
US10984513B1 (en) 2019-09-30 2021-04-20 Google Llc Automatic generation of all-in-focus images with a mobile camera
CN113496136A (en) * 2020-03-18 2021-10-12 中强光电股份有限公司 Unmanned aerial vehicle and image identification method thereof
KR102396830B1 (en) * 2020-10-16 2022-05-11 한양대학교 산학협력단 Moving object judgment device and method thereof
CN112330717B (en) * 2020-11-11 2023-03-10 北京市商汤科技开发有限公司 Target tracking method and device, electronic equipment and storage medium
TR202019736A2 (en) * 2020-12-04 2022-06-21 Aselsan Elektronik Sanayi Ve Ticaret As NOISE ANTI-NOISE METHOD FOR DETECTION APPLICATIONS
CN112947035B (en) * 2021-01-28 2022-03-08 四川写正智能科技有限公司 Eye-protecting posture-correcting intelligent watch ranging sensor installation and ranging method
CN112785628B (en) * 2021-02-09 2023-08-08 成都视海芯图微电子有限公司 Track prediction method and system based on panoramic view angle detection tracking
CN113361651B (en) * 2021-03-05 2022-01-04 牧今科技 Method and computing system for generating safe space list for object detection
CN113763419B (en) * 2021-04-29 2023-06-20 腾讯科技(深圳)有限公司 Target tracking method, device and computer readable storage medium
CN113642463B (en) * 2021-08-13 2023-03-10 广州赋安数字科技有限公司 Heaven and earth multi-view alignment method for video monitoring and remote sensing images
CN113766089B (en) * 2021-09-18 2023-08-18 北京百度网讯科技有限公司 Method, device, electronic equipment and storage medium for detecting video scroll bar
CN114037633B (en) * 2021-11-18 2022-07-15 南京智谱科技有限公司 Infrared image processing method and device
CN114040114A (en) * 2021-11-26 2022-02-11 重庆紫光华山智安科技有限公司 Panoramic shooting and light supplementing method, system, equipment and medium
CN114419102B (en) * 2022-01-25 2023-06-06 江南大学 Multi-target tracking detection method based on frame difference time sequence motion information
CN114972440B (en) * 2022-06-21 2024-03-08 江西省国土空间调查规划研究院 Chained tracking method for ES database pattern spot objects for homeland investigation
CN117095029A (en) * 2023-08-22 2023-11-21 中国科学院空天信息创新研究院 Method and device for detecting small target in air flight

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1996038006A1 (en) * 1995-05-24 1996-11-28 Motorola Inc. Video encoder employing motion estimation with adaptive threshold termination
US6798897B1 (en) * 1999-09-05 2004-09-28 Protrack Ltd. Real time image registration, motion detection and background replacement using discrete local motion estimation
US20050073585A1 (en) * 2003-09-19 2005-04-07 Alphatech, Inc. Tracking systems and methods
US20080037869A1 (en) * 2006-07-13 2008-02-14 Hui Zhou Method and Apparatus for Determining Motion in Images
US20090251544A1 (en) * 2008-04-03 2009-10-08 Stmicroelectronics Rousset Sas Video surveillance method and system

Family Cites Families (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6284391A (en) * 1985-10-09 1987-04-17 Fujitsu Ltd Extraction system for narrow rectangle
JPH08185521A (en) 1994-12-28 1996-07-16 Clarion Co Ltd Mobile object counter
US7015945B1 (en) 1996-07-10 2006-03-21 Visilinx Inc. Video surveillance system and method
US6757008B1 (en) 1999-09-29 2004-06-29 Spectrum San Diego, Inc. Video surveillance system
JP3540696B2 (en) 1999-12-06 2004-07-07 三洋電機株式会社 Image synthesizing method, image synthesizing device, recording medium storing image synthesizing program
IL136080A0 (en) * 2000-05-11 2001-05-20 Yeda Res & Dev Sequence-to-sequence alignment
JP2002074369A (en) 2000-08-28 2002-03-15 Ntt Data Corp System and method for monitoring based on moving image and computer readable recording medium
US7842727B2 (en) * 2001-03-27 2010-11-30 Errant Gene Therapeutics, Llc Histone deacetylase inhibitors
KR20050046822A (en) 2002-10-18 2005-05-18 사르노프 코포레이션 Method and system to allow panoramic visualization using multiple cameras
JP2004157908A (en) 2002-11-08 2004-06-03 Dainippon Pharmaceut Co Ltd Device and method for extracting movement information
US7113185B2 (en) 2002-11-14 2006-09-26 Microsoft Corporation System and method for automatically learning flexible sprites in video layers
US7599512B2 (en) 2003-01-14 2009-10-06 Tokyo Institute Of Technology Multi-parameter highly-accurate simultaneous estimation method in image sub-pixel matching and multi-parameter highly-accurate simultaneous estimation program
JP4492036B2 (en) * 2003-04-28 2010-06-30 ソニー株式会社 Image recognition apparatus and method, and robot apparatus
US20050063608A1 (en) * 2003-09-24 2005-03-24 Ian Clarke System and method for creating a panorama image from a plurality of source images
JP2005286656A (en) * 2004-03-29 2005-10-13 Fuji Photo Film Co Ltd Image pickup device, vehicle and driving support method
US7382897B2 (en) * 2004-04-27 2008-06-03 Microsoft Corporation Multi-image feature matching using multi-scale oriented patches
US7352919B2 (en) * 2004-04-28 2008-04-01 Seiko Epson Corporation Method and system of generating a high-resolution image from a set of low-resolution images
US7576767B2 (en) 2004-07-26 2009-08-18 Geo Semiconductors Inc. Panoramic vision system and method
JP4480083B2 (en) * 2005-02-23 2010-06-16 アイシン精機株式会社 Object recognition device
JP2006285403A (en) * 2005-03-31 2006-10-19 Konica Minolta Holdings Inc Image processing method, and image processing device
US7583815B2 (en) 2005-04-05 2009-09-01 Objectvideo Inc. Wide-area site-based video surveillance system
US7884849B2 (en) 2005-09-26 2011-02-08 Objectvideo, Inc. Video surveillance system with omni-directional camera
WO2007050707A2 (en) * 2005-10-27 2007-05-03 Nec Laboratories America, Inc. Video foreground segmentation method
JP4040651B2 (en) * 2005-12-02 2008-01-30 株式会社日立製作所 Camera shake correction method and image system
WO2007085950A2 (en) * 2006-01-27 2007-08-02 Imax Corporation Methods and systems for digitally re-mastering of 2d and 3d motion pictures for exhibition with enhanced visual quality
US9182228B2 (en) * 2006-02-13 2015-11-10 Sony Corporation Multi-lens array system and method
KR101195942B1 (en) * 2006-03-20 2012-10-29 삼성전자주식회사 Camera calibration method and 3D object reconstruction method using the same
JP2008203992A (en) * 2007-02-16 2008-09-04 Omron Corp Detection device, method, and program thereof
JP4668220B2 (en) * 2007-02-20 2011-04-13 ソニー株式会社 Image processing apparatus, image processing method, and program
US8855848B2 (en) * 2007-06-05 2014-10-07 GM Global Technology Operations LLC Radar, lidar and camera enhanced methods for vehicle dynamics estimation
US8417037B2 (en) * 2007-07-16 2013-04-09 Alexander Bronstein Methods and systems for representation and matching of video content
JP4636064B2 (en) * 2007-09-18 2011-02-23 ソニー株式会社 Image processing apparatus, image processing method, and program
JP2009134509A (en) * 2007-11-30 2009-06-18 Hitachi Ltd Device for and method of generating mosaic image
JP4337929B2 (en) 2007-12-25 2009-09-30 トヨタ自動車株式会社 Moving state estimation device
US8150165B2 (en) * 2008-04-11 2012-04-03 Recognition Robotics, Inc. System and method for visual recognition
FR2930211A3 (en) 2008-04-21 2009-10-23 Renault Sas Vehicle e.g. car, surrounding visualizing and object e.g. pit, detecting device, has controller managing camera function based on object detection mode and vehicle surrounding visualization mode, where modes are exclusive from one another
US8605947B2 (en) * 2008-04-24 2013-12-10 GM Global Technology Operations LLC Method for detecting a clear path of travel for a vehicle enhanced by object detection
CN101286232A (en) * 2008-05-30 2008-10-15 中国科学院上海技术物理研究所 High precision subpixel image registration method
JP5294343B2 (en) * 2008-06-10 2013-09-18 国立大学法人東京工業大学 Image alignment processing device, area expansion processing device, and image quality improvement processing device
US8345944B2 (en) * 2008-08-06 2013-01-01 Siemens Aktiengesellschaft System and method for coronary digital subtraction angiography
KR101498206B1 (en) * 2008-09-30 2015-03-06 삼성전자주식회사 Apparatus and method for obtaining high resolution image
US8340400B2 (en) * 2009-05-06 2012-12-25 Honeywell International Inc. Systems and methods for extracting planar features, matching the planar features, and estimating motion from the planar features
DE102009036200A1 (en) 2009-08-05 2010-05-06 Daimler Ag Method for monitoring surrounding area of vehicle utilized for transport of e.g. goods, involves generating entire image, where course of boundary lines is given such that lines run away from objects depending on positions of objects
US20110128385A1 (en) * 2009-12-02 2011-06-02 Honeywell International Inc. Multi camera registration for high resolution target capture
JP4911230B2 (en) * 2010-02-01 2012-04-04 カシオ計算機株式会社 Imaging apparatus, control program, and control method
CN101984463A (en) * 2010-11-02 2011-03-09 中兴通讯股份有限公司 Method and device for synthesizing panoramic image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1996038006A1 (en) * 1995-05-24 1996-11-28 Motorola Inc. Video encoder employing motion estimation with adaptive threshold termination
US6798897B1 (en) * 1999-09-05 2004-09-28 Protrack Ltd. Real time image registration, motion detection and background replacement using discrete local motion estimation
US20050073585A1 (en) * 2003-09-19 2005-04-07 Alphatech, Inc. Tracking systems and methods
US20080037869A1 (en) * 2006-07-13 2008-02-14 Hui Zhou Method and Apparatus for Determining Motion in Images
US20090251544A1 (en) * 2008-04-03 2009-10-08 Stmicroelectronics Rousset Sas Video surveillance method and system

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
A. Dore, A. Beoldo, & C.S. Regazzoni, "Multiarget Tracking with a Corner-based Particle Filter", 12th IEEE Int'l Conf. on Computer Vision Workshops 1251-1258 (Oct. 2009) *
B. Yu & A.K. Jain, "Lane Boundary Detection Using a Multiresolution Hough Transform", 2 Int'l Proc. of 1997 Int'l Conf. on Image Processing 748-751 (Oct. 1997) *
G. Jing, D. Rajan, & C.E. Siong, "Motion Detection With Adaptive Background And Dynamic Thresholds", 5 Int'l Conf. on Info., Comms. & Signal Processing 41--45 (Dec. 2005) *
I. Ahmad, W. Zheng, J. Luo, & M. Liou, "A Fast Adaptive Motion Estiamtion Algorithm", 16 IEEE Trans. on Circuits & Sys. for Video Tech. 420--438 (Mar. 2006) *
J. Canny, "A Computational Approach to Edge Detection", 6 IEEE Trans. on Pattern Analysis & Machine Intelligence 697-698 (Nov. 1986) *
T. Fei, Liang Xaio-hui, He Zhi-ying, & Hua Guo-liang, "A Registration Method Based on Nature Feature with KLT Tracking Algorithm for Wearable Computers", 2008 Int'l Conf. on Cyberworlds 416-421 (2008) *
X. Zheng, Y. Zhao, N. Li, & H. Wu, "An Automatic Moving Object Detection Algorithm for Video Surveillance Applications", 2009 Int'l Conf. on Embedded Software & Sys. 541-543 (May 2009) *

Cited By (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9138175B2 (en) 2006-05-19 2015-09-22 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US9867549B2 (en) 2006-05-19 2018-01-16 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US9076212B2 (en) 2006-05-19 2015-07-07 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US10869611B2 (en) 2006-05-19 2020-12-22 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US8675970B2 (en) * 2010-06-28 2014-03-18 Sony Corporation Image processing apparatus, image processing method, and image processing program
US20110317924A1 (en) * 2010-06-28 2011-12-29 Sony Corporation Image processing apparatus, image processing method, and image processing program
US10663553B2 (en) 2011-08-26 2020-05-26 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
US9606209B2 (en) 2011-08-26 2017-03-28 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
US9076062B2 (en) * 2012-09-17 2015-07-07 Gravity Jack, Inc. Feature searching along a path of increasing similarity
US20140079320A1 (en) * 2012-09-17 2014-03-20 Gravity Jack, Inc. Feature Searching Along a Path of Increasing Similarity
US9717461B2 (en) 2013-01-24 2017-08-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9305365B2 (en) 2013-01-24 2016-04-05 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US10339654B2 (en) 2013-01-24 2019-07-02 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US10327708B2 (en) 2013-01-24 2019-06-25 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9779502B1 (en) 2013-01-24 2017-10-03 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US9607377B2 (en) 2013-01-24 2017-03-28 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US9782141B2 (en) 2013-02-01 2017-10-10 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US10653381B2 (en) 2013-02-01 2020-05-19 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US10004462B2 (en) 2014-03-24 2018-06-26 Kineticor, Inc. Systems, methods, and devices for removing prospective motion correction from medical imaging scans
US9734589B2 (en) 2014-07-23 2017-08-15 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10438349B2 (en) 2014-07-23 2019-10-08 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US11100636B2 (en) 2014-07-23 2021-08-24 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
CN104202559A (en) * 2014-08-11 2014-12-10 广州中大数字家庭工程技术研究中心有限公司 Intelligent monitoring system and intelligent monitoring method based on rotation invariant feature
CN104217442A (en) * 2014-08-28 2014-12-17 西北工业大学 Aerial video moving object detection method based on multiple model estimation
CN104407619A (en) * 2014-11-05 2015-03-11 沈阳航空航天大学 Method enabling multiple unmanned aerial vehicles to reach multiple targets simultaneously under uncertain environments
CN104766319A (en) * 2015-04-02 2015-07-08 西安电子科技大学 Method for improving registration precision of images photographed at night
US9767572B2 (en) * 2015-05-01 2017-09-19 Raytheon Company Systems and methods for 3D point cloud processing
US20160321820A1 (en) * 2015-05-01 2016-11-03 Raytheon Company Systems and methods for 3d point cloud processing
US9984297B2 (en) * 2015-06-02 2018-05-29 SK Hynix Inc. Moving object detection device and object detection method
US20160358018A1 (en) * 2015-06-02 2016-12-08 SK Hynix Inc. Moving object detection device and object detection method
US11346666B2 (en) 2015-06-26 2022-05-31 SZ DJI Technology Co., Ltd. System and method for measuring a displacement of a mobile platform
US10760907B2 (en) * 2015-06-26 2020-09-01 SZ DJI Technology Co., Ltd. System and method for measuring a displacement of a mobile platform
US20180112979A1 (en) * 2015-06-26 2018-04-26 SZ DJI Technology Co., Ltd. System and method for measuring a displacement of a mobile platform
US10527416B2 (en) * 2015-06-26 2020-01-07 SZ DJI Technology Co., Ltd. System and method for measuring a displacement of a mobile platform
US9943247B2 (en) 2015-07-28 2018-04-17 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
US10660541B2 (en) 2015-07-28 2020-05-26 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
US10716515B2 (en) 2015-11-23 2020-07-21 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10659684B2 (en) 2016-02-16 2020-05-19 Samsung Electronics Co., Ltd. Apparatus and method for providing dynamic panorama function
CN107172341A (en) * 2016-03-07 2017-09-15 深圳市朗驰欣创科技股份有限公司 A kind of unmanned aerial vehicle (UAV) control method, unmanned plane, earth station and UAS
CN105913459A (en) * 2016-05-10 2016-08-31 中国科学院自动化研究所 Moving object detection method based on high resolution continuous shooting images
US10289900B2 (en) 2016-09-16 2019-05-14 Interactive Intelligence Group, Inc. System and method for body language analysis
WO2018053341A3 (en) * 2016-09-16 2018-05-17 Interactive Intelligence Group, Inc. System and method for body language analysis
CN108885469A (en) * 2016-09-27 2018-11-23 深圳市大疆创新科技有限公司 System and method for the initialized target object in tracking system
US10636152B2 (en) * 2016-11-15 2020-04-28 Gvbb Holdings S.A.R.L. System and method of hybrid tracking for match moving
CN109214243A (en) * 2017-07-03 2019-01-15 昊翔电能运动科技(昆山)有限公司 Method for tracking target, device and unmanned plane
CN109215056A (en) * 2017-07-03 2019-01-15 昊翔电能运动科技(昆山)有限公司 Target tracking method and device
CN108848304A (en) * 2018-05-30 2018-11-20 深圳岚锋创视网络科技有限公司 A kind of method for tracking target of panoramic video, device and panorama camera
US11509824B2 (en) 2018-05-30 2022-11-22 Arashi Vision Inc. Method for tracking target in panoramic video, and panoramic camera
CN109146862A (en) * 2018-08-04 2019-01-04 石修英 Soil remediation condition intelligent detection device
CN108982521A (en) * 2018-08-04 2018-12-11 石修英 Visualize the horizontal detection device of soil health
CN112308880A (en) * 2019-08-30 2021-02-02 华为技术有限公司 Target user locking method and electronic equipment
WO2021036717A1 (en) * 2019-08-30 2021-03-04 华为技术有限公司 Target user locking method and electronic device

Also Published As

Publication number Publication date
JP2015181042A (en) 2015-10-15
GB201311821D0 (en) 2013-08-14
JP2014504410A (en) 2014-02-20
WO2012084703A1 (en) 2012-06-28
US20120154579A1 (en) 2012-06-21
DE112011103690T5 (en) 2013-09-05
GB2500839A (en) 2013-10-02
GB2500839B (en) 2017-04-05
JP6095018B2 (en) 2017-03-15
CN103262121A (en) 2013-08-21
US9147260B2 (en) 2015-09-29

Similar Documents

Publication Publication Date Title
US9147260B2 (en) Detection and tracking of moving objects
CN111462200B (en) Cross-video pedestrian positioning and tracking method, system and equipment
US10964054B2 (en) Method and device for positioning
US8446468B1 (en) Moving object detection using a mobile infrared camera
US9969337B2 (en) Methods and systems for mobile-agent navigation
US10339389B2 (en) Methods and systems for vision-based motion estimation
US7187809B2 (en) Method and apparatus for aligning video to three-dimensional point clouds
US8073196B2 (en) Detection and tracking of moving objects from a moving platform in presence of strong parallax
US20150138310A1 (en) Automatic scene parsing
CN104204721B (en) Single camera distance estimations
US8498472B2 (en) Image-based localization feature point registration apparatus, method and computer-readable medium
Bhattacharya et al. Moving object detection and tracking in forward looking infra-red aerial imagery
Lin et al. Efficient detection and tracking of moving objects in geo-coordinates
Ji et al. RGB-D SLAM using vanishing point and door plate information in corridor environment
CN111383252B (en) Multi-camera target tracking method, system, device and storage medium
GB2541153A (en) Processing a series of images to identify at least a portion of an object
CN113256731A (en) Target detection method and device based on monocular vision
CN113673288B (en) Idle parking space detection method and device, computer equipment and storage medium
Tanathong et al. Using GPS/INS data to enhance image matching for real-time aerial triangulation
Su et al. Moving objects detecting and tracking for unmanned aerial vehicle
Pudchuen et al. Visual odometry based on k-nearest neighbor matching and robust motion estimation
Koutlemanis et al. Foreground detection with a moving RGBD camera
Zhou et al. UAV Based Indoor Localization and Objection Detection
Hamza et al. Runway positioning and moving object detection prior to landing
Chien et al. Multi-frame feature integration for multi-camera visual odometry

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION