CA2132515C - An object monitoring system - Google Patents

An object monitoring system Download PDF

Info

Publication number
CA2132515C
CA2132515C CA002132515A CA2132515A CA2132515C CA 2132515 C CA2132515 C CA 2132515C CA 002132515 A CA002132515 A CA 002132515A CA 2132515 A CA2132515 A CA 2132515A CA 2132515 C CA2132515 C CA 2132515C
Authority
CA
Canada
Prior art keywords
monitoring system
image
images
regions
clusters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
CA002132515A
Other languages
French (fr)
Other versions
CA2132515A1 (en
Inventor
Glen William Auty
Peter Ian Corke
Paul Alexander Dunn
Ian Barry Macintyre
Dennis Charles Mills
Benjamin Francis Simons
Murray John Jensen
Rodney Lavis Knight
David Stuart Pierce
Ponnampalam Balakumar
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Commonwealth Scientific and Industrial Research Organization CSIRO
Original Assignee
Commonwealth Scientific and Industrial Research Organization CSIRO
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Commonwealth Scientific and Industrial Research Organization CSIRO filed Critical Commonwealth Scientific and Industrial Research Organization CSIRO
Publication of CA2132515A1 publication Critical patent/CA2132515A1/en
Application granted granted Critical
Publication of CA2132515C publication Critical patent/CA2132515C/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01PMEASURING LINEAR OR ANGULAR SPEED, ACCELERATION, DECELERATION, OR SHOCK; INDICATING PRESENCE, ABSENCE, OR DIRECTION, OF MOVEMENT
    • G01P3/00Measuring linear or angular speed; Measuring differences of linear or angular speeds
    • G01P3/36Devices characterised by the use of optical means, e.g. using infrared, visible, or ultraviolet light
    • G01P3/38Devices characterised by the use of optical means, e.g. using infrared, visible, or ultraviolet light using photographic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/174Segmentation; Edge detection involving the use of two or more images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/255Detecting or recognising potential candidate objects based on visual cues, e.g. shapes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/54Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/017Detecting movement of traffic to be counted or controlled identifying vehicles
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/04Detecting movement of traffic to be counted or controlled using optical or ultrasonic detectors
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/052Detecting movement of traffic to be counted or controlled with provision for determining speed or overspeed
    • G08G1/054Detecting movement of traffic to be counted or controlled with provision for determining speed or overspeed photographing overspeeding vehicles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle
    • G06T2207/30261Obstacle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/625License plates

Abstract

An object monitoring system includes a camera node (2) for monitoring movement of an object (18) to determine as acquisition time when an image of the object (18) is to be acquired and acquiring the image at the predetermined time. The system includes a camera (6) which is able to monitor moving objects (18), and image processing circuitry (10), responsive to the camera (6), which is able to detect a predetermined moving object (18) from other moving and static objects. From the image acquired, information identifying the object (18) can be automatically extracted. The system is particularly suited to monitoring and discriminating large vehicles (18) from other vehicles over a multi-lane roadway, and acquiring high resolution images of the large vehicles (18) at a predetermined acquisition point (22). Image data acquired by a plurality of camera nodes (2) can be sent over a digital telecommunications network (45) to a central image processing system (42) which can exact extract vehicle identifying data, such as licence plate details, and obtain information on vehicle travel between nodes (2).

Description

WO 93/19441_ _ _ ~ ~ ~ ~ ~ ~ ~ p~'/Al,'93/0p1 i5 -I-Arr oatECr MarrcroRarrc sY~rr~.rr i The prrsenc invention relates to as object monitoring system sad, its particular, to a system for monitoring vehicles.
Authorities responsible for traffic maaagcment and the iayus which govern the use of vehicles require systems which can monitor t:affc continuously and detect breach of the law, withour requiting the eapease of having personnel present at rise scene of cye iaftiagemeac. syscecas which are able to monitor a large number of locations, detect in5ria~meats and issue iatrissgement sodas are particularly advantageous as they relieve personnel; such as police, from the task of traffic management and allow there co pursue ocher tasks. By coatiauously moaiaoring a location the systems also as as a deterrent to is in5iagers and may asai:t in redtcciag acadeats which cause toad fatalities sad casualties.
It would also be advto be able to monitor road wage is order to make decisions on toad damage by heavy veiudes. .
A number of traffic management systems atr preaaatly in use, such as speed 30 caraecas sad red light amerss for road t:'all~c. The known sy:rcems employ cameras which else triggered wl~n sa ia~tlagement is detected, optical :enaors placed oa the side of the toad, due aea:ora placed underneath the road sad radar signals reflected from vehicles are wed to detect the p:eseace of a vehicle and deta~mine ia>xio~g~neat. The semoa sad ruler dgoals art used to generate a trigger signal to a4tivace a camera to take r5 a picture of vehicle whidb include details isom which the vehicle can be idled, such as a cat licence plate. Use of toad based kneels is disadvantageous as they require the road to be altered ~ excavated for iascallation or, when placed o~ the side of du road, can be easily det~ed and daaa~aged. Also elearieal cabling aceds to be installed and conaecteil between the aensots and the camera. The use of dtic aigasts which 30 are transmitted to sad reaeaed from a vehicle. suds as aadu signal:, to decea p and ia6~iagamaat is also d>sadwntageous as these signals can be~detaaed by deteai~
r units planed is a vehids to glen the driver as to their .
i,.

H'0 93/14441 PC?/Ah93l0pt t c It is advantageous therefore to provide a system which can detect vehicle presence anti in$ingement without transmitting any electromagnetic signals or using road based sensOn.
'Ihe cameras pteseathy is use also use photographic film which has the disadvantage that it needs to be ooatiauaily replaced at the I location of the camera.
Accordiapty, a number of red light cameras is metropolitan areas do not always include film and do not continuously monitor the corresponding intersection.
I
t0 Speed detection systems which ux only cameras ate bed is a number of publications. T'he systems are able to monitor tragic flow and dared instaataaeous speed iafriagemeats but the systems an relatively limited with respect to the iaformatioa they can obtain on a vehicle whilst it is being monitored, and the systems are also tenable to selectively acquire inforrrnation on specified vehicle types.
is .
T3e pintoes or image: aoqnired by the caaxra also norm~lty need to be examined by personnel to eucract the informacioa to identify the vehicle slid de:ermiaa elx persaa responsible for it, which is a time ~ming process. If the 'ale could be prae~ed within a relatively short time of acquisition rhea it could be task as a bats for alerting 20 authorities is the region to seek tad hold the vehicle, for example, if the infonoaation identifies it a: being uolea. A~ooordiagly, it would be advantaget~'ous to provide s system which cat process images is real time to obtain detailed iafon~atioa on a vehicle and ' i issue alert iaFocmacion tad ia>yiagament notices without reqtdti~ human internatioa.
:S Whoa trsvallins a long distance, vehicle user's, is partia~r truck drivers, read to traasgrass apoed timiu so as ro shorten the lima in travelling to t»
deaiaation and bring the journey their speed may vary from a range which is the limit to one which extxeds the limit. 'the known systems for deteaiag speed ia~ringement coaoentrate oa derocxiag the inataataaeous speed of a vehicle at a pastiarlar ytoeation tad thesefoee 30 don the location a which the deletion unit is pLaoed, it may not detect user's who infri~e sporadically over a long diaraace.. Also tntek and bus drivers wlm exceed a reoammended time of travel by avoiding rest slaps tad iasaurately oomptae log books f~
I

WO 93/ l9aa 1 pGT/A L'93/00114
2~j~~~~ ' _3_ I
may not be detected. Heave, it would be advantageous to provide a syseem which can detect the average speed of a vehicle over a relatively long distaece. It is also advantageous to provide a system which can a~tonitor vehicles-in more than one lane of a mufti-lane carriageway. j i s .
The present invention provides an object monitoring system comprising camen means for monitoring movement of as object to determine an acquisition tame when as imsge of said object is to lx acquired and acquiring said image at said predetermaaed time.
~
The pseseat invention also provides as object monitoring systeaa oonmp:~ing camera means for monitoring moving objects, sad image processing means, responsive to said camera mesas, for detetxiag a predetermined moving object from other moving and static objects.
~s ~ 1 The present inveatian further provides an object monitoring system comprising camera rnea~ for sacking sad acquiring as image of a mo~!ing object from which information idattifying :aid object can be automatically extracted.
30 Preferably said system includes mesas for transmitting said image over a digital telocommunicatiaas network.
The pre:aut invention also provicks a vehicle moaitorlag system, comprising camaca n~ for coatinuaa:ty detactiag sad tnckiag moving velrielea over a mufti-lane 25 carlageway, sad ~quiilaa images of predetermined vehicles at, m aalttisition area oa said catr6a~way from wlriich idrntifyiag information on said vehicles can be exttaaed.
The present iavencion fortf~a provides a vehicle monitoring system comprising a plurality of casters means for trackins sad acquiring images of predetarminod moving 30 vehicles for a reapoctive ales, sad means for processing the image data obuined from said areas ro i~ntify acid vdycle: and obtain iafo:matian oa the travd of said vehicles beeween said areas. i The pr~aent invention also provides a vehicle monitoring system comprising camera means for monitoring; moving vehicles to determine if' said vehicle is of a predetermined type and, in response thereto, capturing respective images of vehicles of said predetermined type.
The present inventi,m further provides a vehicle monitoring system comprising camera means for monitoring a vehicle to detect a law infringement and determine a predetermined time to acquire an image c~f said vehicle, and for capturing an image of said vehicle at said predetermined time in response to detecting said i o infringement.
The present invention also provides a. vehicle monitoring system comprising camera means for monitoring vehicles on a roadway, discriminating between large vehicles, such as trucks and buses, and small vehicles, such as cars, on said roadway 15 so as to acquire images of only the large vehicles from which vehicle information can be obtained.
In accordance with one aspect of the present invention an object monitoring system comprising camera means characterised in that the camera means is adapted to 2o monitor movement of an object to predetermine, based on the monitored movement of the object an acquisition limy at which an image can be acquired at a predetermined position of said object relative to said camera means, and to acquire an image at the predetermined acquisition time and the predetermined position.
25 In accordance with a~zother aspect of the present invention an object monitoring system comprising:
camera means for generating images of an area and for acquiring an image of a prc;determined object, and image procc;s;sing means including:
3o means for subtracting a background image of said area from said images of said area to generate difference images representative of moving objects in said area;

_ L~a_ segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area;
classification means for processing and classifying said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of raid valid regions, clusters corresponding to respective ones of said rr~oving object , and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said predetermined object; and tracking means for tracking said one of said clusters corresponding to t5 said predetermined object to trigger said camera means to acquire said image of said predetermined object.
In accordance with a further aspect of the present invention an object monitoring system compri~~ing:
2o camera means for generating images of an area and for acquiring an image of a predetermined object;
image processing means including:
means for subtracting a background image of said area from said images of said area to gencyrate difference irr~ages representative of moving objects in 25 said area, segmentation means for processing said difference images to generate region image; representative of regions corresponding to parts of said moving objects in said area, classificatic:>n means for processing said region images, said 3o classification means inclu~:Iing:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valii:l regions and invalid regions, -4b-clustering means for rejecting said invalid regions and generating, on the basis of the: geometry of said valid regions, clusters corresponding to respective ones of said moving objects, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters correspond:; t~ said predetermined object, and tracking means for tracking said one of said clusters corresponding to said predetermined object to trigger said camera means to acquire said image of said predetermined object; and t o extraction means for processing said image of said predetermined object to extract information identifying said predetermined object.
In accordance with one aspect of the present invention a vehicle monitoring system comprising:
t 5 camera means for generating images of a carriageway and for acquiring imal;es o1' predetermined vehicles, and image processing means including:
means for s~,ibtracting a background image of said carriageway from said images of said carriageway to generate difference images representative of moving vehicles on said carriageway;
2o segmentation rr~eans for processing said difference images to generate region images representati~s;e of regions corresponding to parts of said moving vehicles on sand carriagew;ay;;
classification means for processing said region images, said classification means including:
25 means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and 3o means for cuassifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said predetermined veihicles; and -~C-tracking means for tracking said clusters corresponding to said predetermined vehicles to trigger said camera means to acquire said images of said predetermined vehicles.
In accordance with another aspect of the present invention a vehicle monitoring system comprising:
a plurality c~f c;amera means for generating images of respective areas and for acquiring images o l-' predetermined vf:hicles, said areas being remote with respect to one another; and a plurality of image processing means including:
means fox sxabtracting background images of said areas from said images of said areas to generate difference images representative of moving vehicles in said areas;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area;
classification means for processing said region images, said classification rneans including means for analyzing the shape of said regions and, on the basis of the: analysis, determining valid regions and invalid regions, clustering means for rejecting said insealid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for cla;ssiif;ying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters corresponds to said predetermined vehicles;
tracking means for tracking said clusters corresponding to said predetermined vehicles to trigger said camera means to acquire said images of said predetermined vehicles; and recognition means for processing said images of said predetermined vehicles to obtain information identifying said predetermined vehicles' In accordance with ,:mother aspect of the present invention a vehicle monitoring system comprising:

-4~d-camera means for generating images of an area and for acquiring an image of a vehicle associated with a law infringement, and image processing means including:
means for subtracting a background image of said area from said images of said area to gencvrate difference images representative of moving vehicles in said area;
segmentati<m means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area;
1 o classificati<an means for processing said region images, said classification means inclucling:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valic:l regions and invalid regions, clustering means for rejecting said invalid regions and generating, on 15 the basis of the geometry <~f said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for detecting said law infringement by comparing at least one characteristic of said clusters, t:o classification data of said system to determine if one of said clusters corresponds i:o said vehicle; and 2o tracking me:ar~s for tracking said one of said clusters corresponding to said vehicle to trigger said camera means to acquire said image of said vehicle.
In accordance with a further aspect of the present invention a vehicle monitoring system comprising camera means for generating images of a carriageway 25 and for acquiring high resc:~lution images of large vehicles, such as trucks and buses, and image processing means including:
means for subtracting a background image of said carriageway from said images cof said carriageway to generate difference images representative of moving vehicles of said carriiageway;
3o segmentation means for processing said difference images to generate region images. representative of regions corresponding to parts of said moving vehicles on said carriageway;
classificatic:m means for processing said region images, said classification means including:

~.e~
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of thc: geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing of at least one characteristic of said clusters to classification. data of said system to determine if said clusters correspond to said large vehicles; and tracking means for tracking said clusters corresponding to said large 1 o vehicles to trigger said camera means to acquire said high resolution images of said large vehicles.
In accordance with one aspect of the present invention an object monitoring system comprising:
I 5 video camera means for generating images of an area to monitor moving objects in said area:, image capture camera means for acquiring a high resolution image of a predetermined object; and image processing means including:
2o means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving objects in said area;
segmentatio~u means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects 25 in said area;
classification. means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, 3o clustering means for rejecting said invalid regions and generating, on the basis of the geometry oi~ said valid regions, clusters corresponding to respective ones of said moving objects, and - 4f~
means for classifying said clusters by comparing at least one characteristic of said clust~:rs to classification data of said system to determine if one of said clusters corresponds to said predetermined object; and tracking means for tracking said one of said clusters corresponding to said predetermined object I:o trigger said image capture means to acquire said high resolution image of said predetermined ob~eca.
In accordance with another aspect of the present invention an object monitoring system comprisin.g.:
1 o video camera :means for generating images of an area to monitor moving objects in said area;
image capture camera means for acquiring a high resolution image of a predetermined object; and image processing means including:
15 means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving objects in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects 2o in said area;
classification means for processing said region images, said classification means including means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, 25 clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving objects, and means for clr:~ssifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one 30 of said clusters corresponds to said predetermined object;
tracking means for tracking said one of'said clusters corresponding to said predeternai.ned object to trigger said image capture camera means to acquire said high resolution image of said predetermined abject; and _~.g_ extraction means for processing said image of said predetermined object to extract information identifying said predetermined object.
In accordance with a further aspect of the present invention a vehicle monitoring system comprising:
video camera means for generating images of a carriageway to monitor moving vehicles in said carriageway;
image capture camera means fox acquiring a high resolution image of a predetermined vehicle; anct 1o image processing means including:
means for sg;~btracting a background image of said carriageway from said images of'said caxriagcway to generate difference images representative of said moving vehicles on said carriageway;
segmentation means for processing said difference images to generate 15 region images representative of regions corresponding to parts o:f said moving vehicles on said carriageway;
classification means for processing said region images, said classification means including;:
means for analyzing the shape of said regions and, on the basis of the 2o analysis, determining valid regions and invalid regions, clustering means far rejecting said invalid regions and generating, on the basis of the: geometry ol's,aid valid regions, clusters corresponding to respective ones of said moving vehiclw°,s, and means for classifying said clusters by comparing at least one 25 characteristic of said clusters to classification data of said system to determine if said clusters correspond to said r~ra:determined vehicle; and tracking mesan;> for tracking said clusters corresponding to said predetermined vehicle to trigger said image capture camera means to acquire said high resolution image of said predetermined vehicle.
In accordance with ono aspect of the present invention a vehicle monitoring system comprising:

-4h-a plurality of 'video camera means for generating images of respective areas to monitor moving vehicles in said area, said areas being remote with respect to one another;
a plurality c~f image capture camera means for acquiring a high resolution image of one or more predetermined vehicles; and a plurality c;~f iimage processing means including:
means for subtracting background images of said areas from said images of said. areas to generate difference images representative of said moving vehicles in said areas;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said areas;
classification uneans for processing said region images, said classification means including:
15 means for analyzing the shapes of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis ofthc; geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles., and 20 means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said predetermined vehicle;
tracking mean's for tracking said clusters corresponding to said predetermined vehicle to trigl;er said camera means to acquiring said image of said 25 predetermined vehicle; and recognition means for processing said images of said predetermined vehicle to obtain intormatiom identifying said predetermined vehicle.
In accordance with another aspect of the present invention a vehicle 3o monitoring system comprisi.n;~:
video camera means for generating images of a carriageway to monitor moving vehiclc;s in said area;
image captua°e camera means for acquiring a high resolution image of a large vehicle, such as a truck and a bus; and - ~r ~
image proces sing means including:
means for subtracting a background image of said carnageway from said images ojE said carriageway to generate difference images representative of said moving vehicl'aes on said carriageway;
segmentaticyn means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles on said carriagew;:~y;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid relgions and invalid regions, clustering nnerrns for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and t5 means for cl.as;sifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said large vehicle; and tracking me;:~ns for tracking said clusters corresponding to said large vehicle to trigl;er said image c<rpture camera means to acquire said high resolution 2o image of said large vehicle.
In accordance with ,;t further aspect a vehicle monitoring system comprising:
video camera means for generating images of an area to monitor moving vehicles in said area;
2s image capture camera means for acquiring a high resolution image of a vehicle associated with a law infringement; and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving 3o vehicles in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area:

_ ra.l _ classification means for processing said region images, said classification :means including means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of ;>aid valid regions, clusters corresponding to respective ones of said moving vehicies, and means fur detecting said law infringement by comparing at least one characteristic of said clustf~rs to classification data of said system to determine if one to of said clusters corresponds to said vehicle; and tracking means for tracking said one of said clusters corresponding to said vehicle to trigger said image capture camera means to acquire said high resolution image of said vehicle.
t 5 A prefE;rred embodiment of the present invention is hereinafter described, by way of example only, with reference to the accompanying drawings wherein:
Figures 1 to 3 are side views illustrating use of a preferred system for monitoring vehicles;
Figurca 4 is a front perspective view illustrating use of a preferred system for 2o monitoring vehicles;
Figure 5 is a block diagram of a preferred embodiment of the vehicle monitoring system;
Figure 6 is a block diagram of connection across a digital telecommunications network of two nodes and a central server of the vehicle monitoring system;
25 Figure '7 is a view illustrating connection of a large number of nodes of the vehicle monitoring system;
Figure ~ is a block diagram of vehicle detection and image capture circuitry of the vehicle monitoring system;
Figure'3 is a digitised image produced by the vehicle detection circuitry from 3o an image generated by a detection camera of the system;

W9 93119441 ~ ~ ~j ~ ~ ~ ~ I Pf'T/AL:93/0p11:
_s_ Figure IO is a block dial of the control of the circuit boards of the vehicle detection circuitry to perform a se~entation process;
Figure 11 is a static baek~ound image stored in the vehicle detection circuitry:
Figure 12 is a di~etence image generated by the vehicle detection circuitry;
Figure 13 is an image illustrating regions of shadow which are faltered from the image obtained by the detection cataeta:
I"sgs::~-~~ is w ~ae~lxd ::'.»~ de:i~~o~ b~ ;.L.a :~as~..°~k.
,ioexe:o:~-.-.soia:;r~;_ ...
Figure is is a histogram of pixel grey levels;
Figure 16 is a real time status display generated by the system;
Figure 17 is a flow diagram illustrating flow between the software tasks of the system;
Figure 1B is a diagram of the formation of "black triangles" in a processing window of the system;
Figure I9 is a diagram illustrating meaaureiment of cawerage of blob regions produced by the system;
Figure 20 is a diagram itlusttating vertical exte:xsion of ,'blob Legions to perform Blusters;
Figure 21 is a graph of extension amounts which are stored in a look-up table of the system;
30 Figure Z2 is a disgtam illusttitiag extaasioa based on blob region width;
Figure 23 is a diagtsam of overlap detection for clusters produced by the system;
Figure 24 is a diagram illustrating a labelling method performed by the systear:
Figure 25 is a daagtam of the roadway coordinates used,'by the system;
Figure 26 is a gnpls of the trajoaory of clusters; ;:
Figure 27 is a graph of the trajectory of clusters transformed to the roadway coordinates;
Figure 28 is a diag:am of data values obtained by tr~jecxory software of the System; I
Figuae 29 is a block diagram of a timing control board bf the system;
Figure 30 is a graph of tl3e operating chaaaaecistics of the aequiaition camera and infrared flash of tire vehicle monitoring system;
Figures 31 sad 32 are image: acquired by the system;

WO 93/ l9d~t t ~ ~ ~ ~ ~ ~ ~ PCT/A 1:93/0011:

Figure 33 is a block diagram of components of the acquisition ratnera, and intorface components for the camera of the image capture circuitry;
Figure 34 is a block diagram of comtnuttications components of nodes of the system, and the eotnponents of an acquisition imago processing system of the system S eotmected over the digital telecommunications network;
Figure 35 is a diagram of the memory layout for a buffer board of the image capture cuaaitry;
Figure 36 is a flow diagram illusreatiag software modules of the acquisition image processing system and communications modules of the ;
Figure 37 is a block diagram of a lice. plate reco~ieian system of the vehicle monitoring system;
Figure 38 is a flow diagram of an image a~uisition p:oc~tte of the liceaex plate recognition system;
Figure 39 is a flow diagram of the software module of the licence plate recognition system;
Figure 40 is a flow diagram of a locate piste module of the liccaa piste recognition system; and Figure 41 is a flow diagram of an optical cbaracter recognition module of the license plate recognition system.
i A vebicie monito~iag system, as shown is Fisures 1 to 7; includes a camera node 2 which is mounted on a bridge or pylon 4 above vehicle trsffi~, as shown in Figures 1 to 3. 'The ca~eru node 2 includes a vehicle detection camera i6, as image ae quisiaoa camera 8 and a tads ooatrol unit 10. Both canuras 6 and 8~ are moaocitrome Ct'D
?3 cameras, with the vehicle detection camera 6 being a wide angle video camera of medium resolution, and the image acquisition camera being a high raso~udon camera.
The detection camera 6 has a wide; field of view 1~ of part of a vehicle cartiageway 16 whic3t is to be monitored by the node 2. '~'he detaxitm camera b 30 monitors vehicles in the fleid of view 12 sad the coat:oi unit 10 pthe images acquired by the detection camera 10 to detect and disa~imisiate vehicles from other objects in the field of view 12. As a vehicle lg eattxa the 5e1~ of view 12 and moves i i WO 93/ 1944a ~ ~ ~ ~ ~ ~ ~ , P~'f / A l'93/40 t 13 _?
towards the node 2, the node 2 analyses the images produced by the detection careers 6 to first detect the vehicle 18 as bring a moving object, which is different from other moving objects or the still background in the view 12, and determines whether the vehicle 18 constieutes an object for which a high resolution image thereof should be obtained by the image acquisition camera 8. The image acquisition cavtaera 8 is mounted on the bridge or pylon 4 so as to have a limited ~cld of view 20 which will include the front of a vehicle 18 when it rcach~ a predetermined Iocaiion 22 oa a carriageway 16. TThe location 22 and the field of view 20 are chosen to bs near the point where movinE
vehicles will leave the 5eld of view 12 of the detection camera 6, as shown in Figure 3.
On determining that the vehicle 18 reprexats as object for which an image is to be acquired, the node 2 estimates the time when the vehicle 8 will enter the 5eld of view of the acquisition camera 8, on the basis of the movement of the vehicle which has been monitored by the detection camera 6. The nod 2 provides trigger ia~ormation to control circuitry associated with the aoquisitioa camera 8 so as to trigger the camera 8 15 at the estimated time. A high resolution image of the $oat of the vehicle I8 is obtained from which considerable identifying information can be derived, such as vehicle type sad licence plate details, by subxquent digital elecxronic processia~~of the image.
i In addition to identifying the vehicle 18 and estimating the time for triggering the 20 acquisition camera 8 the node 2 is able to ux the iafrom ~ the detection camera 6 to discriminate between vehicles on a number of charaete 'rrktica, such as size, to determine those for which high resolution images are t~ be acqtt~red. For example, the system is able to distinguish between large vehicles such as hurk~ sad coaches, and other moving objects within the field of view 12, sub as cars and motor bicycles.
'The soda : is also able to determine from the images obtains by the ~deteccion camera 6 the current speed of tho vehicle 18 sad whether the dtwer is oommi~t~ say crafi;ic or other offences, such as tailgating or illegal lane cheagi~ag. The system can also be used to detect stolen vehicles.
'1"he detection camera 6 sad the control wait 10 are ab6e to monitor all of the moving vehicles 18 and Z2 within the held of view 12 whilst aoquiriag the images of selected vehicles at the location 22. For a mufti-lane carriageway 21, as shown is Figure wp 93i19~4t ~ ~ ~ ~ ~ ~ ~ p~ T/At;~93/ppt 1:
_8_ 4, the field of view 12 of the detection caatacra 6 extends over all of the lanes ~3 and 25 of the carziageway and an image acquisition camera 8 is provided for each Lane ~3 and 35. The node 2 is therefore able to monitor the moving vehicle 18 to determine in which lane it will be when it roaches the image caprute location 22 and activates.
as required.
the acquisition camera 8 corresponding to that lane 23 or 23. .
i The control unit I0, as shown in Figure S, includes vehicle detection circuitry 30 for processing the images generated by the detetxion camera 6 so as to provide nigger signals oa a bus 32 to the image acquisition camera 8. ~r sele~d camera 8 is triggered to acquire an image in aaordaner. with the timing iafotmation determined by the detection circuitry 30, and the camera 8 provides a trigger sigmil on a line 36 to a flash ttiggtring circuit 38, of a corresponding infrared flash 40 mouatod adjacent the scteaed acquisition camera 8. The image obtained by the trigger acquisition camera 8 is received by as image acquisition circuit 34. The detection circuit 3fl deteriniaes the light intensity within the field of view 12 of the detection camera b so as to da~ermiae the correct revel of exposure for the acquisition emcee 8, and is turn the correct level of erJergy to be discharged by the flash 40 to achieve the desired level of expostue. The use of as flt flash is advantageous as activation is difficult to detect visuall~r. Visible wavelengths produced by the flash are removal by 1R band pass filters.
?0 The vehicle monitoring system includes as acquisition '~mage processing system 42 connected to the control unit 10 for receiving and processing'the images acquired by the caaaara $ to extract vehicle information therefrom. 'Zbe aequ~sition image processing system 42 may form pt~rt of the node 2 of be pocitionad remote from the node and r5 connected to the control unit by a telecammuttications iiae 44 from the acquisition circuit 34. The system 42 comprises a processing station 43 eoaf'tgused ~o automatically extract the required information $om the image. such as licence plats Is 50.
'The acquisition intagc processing system ~2 when impleedentod at a remote cxntrai 30 site, a: shown in Figure 6, include: somrauaicatio~ ouan~ol3ers 55 ~eaed to a public digital telecommunications network 45, and a 1 oomputa server 47 which serves a Local area aetverork (l.c~d) connecting computer3 which implement as acquisition image ACT/ A 1.93/001 ! 5 W~93/t94~1 ~~~~515 _g_ database 49, a iicence piste recognition system 51 and a remote site user interface 53.
T'he communications controllers 55 are provided for each ~e ? which sends itmages to the processing system 42. T'he nodes ~. each ineiude an image buyer and communications controller 57 for storing images ob~ined by the acquisition circuit and communicating with the communications coneroilers 55 of the central image processing system 42 to send the images over the integrated services digital networlt (ISDN) 45 to the central server 47. The eonataunications controller 55 manage the high speed image transfers over the ISDN 45, and handle houxkeeping, error detection grad correction for image transfers between the nodes 2 arid the cxntral ses~ier 4'~. The central server 47 commuaicatcs with the controllers 55 sa the nodes 2 as asextensions of the LAN
maintained by the server 47. Image processing can also lx performed at each of the nodes r, for example, the nodes : may each include a liceasme~ plats recognition system S1 which performs optical character recognition (~) on tile acquired images to e~ctract vehicle information, such as licence plate details.
The vehicle monitoring system, as shown in Figure 7,' comprises a plurality of camera n~s 2 mounted at a number o$ locations 52 to 60 oa dehicie caraiageways. The nodes 2 may be connected by telecommunications lines of the tSDN 45 to communicate with another as~idlor connected to a central coatroi station. 62, so as to compare information collected at each of the avdas :. The control; station 62 includes the acquisition image processing system 42. ?he nodes 2 sad the coactol station 62 are able to monitor a vehicle's progress along the carriageways 16, 64 usi~rg information collected by the nodes 2, which iarludes, in addition to vehicle identifying information, the ante, time and iocation at which as image is acquired. This ix paaticvlarly advantageous as the 35 information eau be used to determine the avetage speed at which a vehicle has travelled between two nodes 2. If the average speed indicate that the vehicle has exceeded the speed limit fn travelling between the codes, then authorities cacn be coata~ed so as to intercept the vehicle. Alternativeiy, the centre! station b2 iss>ses as iafrittgemeat notice to the 'registered owner of the vehicle. 'The station 62 aad/dr the nodes 2 may also captain information on stole$ vehicles sad the authorities are ~wbea a stoics vehicle is detected. Vehicle drivels negotiating long distaa~ would be relucxast to instantaneously exceed the speed limit at chosen iocatiotas, if they are aware that they will i ~1'O 93/ 1944 i r P~T/AL~93~0011 ~
~~j~~l~ ' -10- ' be intercepted or issued with an infringement notice by tsavellirsg between two locations 5. and 54 of two nodes, too quickly. The distance bctwecn the nodes would be relatively large and an allowable time for travel between the nods would be established corresponding to a permitted average speed. The ability to rnoaitor average speeds by the system represents a significant development which can be used to deter excessive speeding by large vehicles, such as tructcs and busts, on major loads, and further scabies deaection of drivers who fail to take scheduled rest stops.
The detection camera 6 produces video herds of 312 and X13 horizontal scan liaaes respectively which are each duplicated to producx a complete 6~5 lice vido frame:. The fields are converted into S12 x 512 pixel 8 bit quaattised digitall images which oecttr st a video field period of 20 ms. The vertical resolution of the ~tection camera 6 is dependent on the ve~ical field line resolution which is appzo~imately 300 elements, digitixd i~oto 512 piurels, for a maximum. distance which the i~amera 6 can view on a horiaontal roadway. The maximum distance D is gives by:
h - titan I,~b? ~ ~1 i t1) where D = distance along road covered by camera view h = height of sera above road D,r = distance of closest position of camera vii along roadway 4f = lenr field of view eagle The !lord of view across the roadway is given by:
~V =~~~
where W ~ held of view across the roadway w = width of the shot ' f ~ lens focal length L ~ object distance from camera The camera 6 includes a 12 mm lane aced as 8.8 mm x 6.6 mm BCD sensor to WO 93119441- ~ ~ J ~ ~ ~ ~ P~'T/r1193/001 t=
1~ _ optimise vehicle image size and maintain a four lane c~verage, 3.5 metres per lane, at the image acquisition points 2Z. An antiblooming and antismear sensor is included to prevent blooming or smearing of an image by vehicle tights. The in~~d filter of the camera permits a infrared wavelengths up to 450 ram, which allows the detection camera 6 to receive the infrared component of vehicle lights, thereby providing more image information to detect and monitor vehicles. The detection camera 6 bas a +~0 dB gain range, and the exposure time is iced at the field period, 20 tns.
The exposurt eontrol of the detention camera 6 controls the intensity of light falling on the camera sensor so as to maintain consistent video s~gZtal quality and obtain a predictable repre:eatation of a vehicle. Acceptable ~cposu~e of the sensor can be maintained through the appropriate match of sensor sensitivity arid control of the intensity or power of the elec~tmtnagaetic wavele:agth failing oa tht ~seasor, as shows with F
retetence to equation 3.
E ac (HA)T (3) 1 s where E = exposure of light on sensor H R isadent e.m.r. power per cma (uradiance) A = area of pixel site in cm=
T s time in seconds that light or e.m.r. falls on tensor The time T light falls on the trigger camera is held const~at at the video field rate of 20 ms. Thin is autficieatiy short to "freeze" tix motion of tha~wehicle in the nlativcly large field of view 12 of a mufti-lane carriageway. A shuttei is not included is the deteaioa camera 6 as elscaonic abutters or short duration ex~trre control produced adveme affects from either image smear or blooming from sunlight reflections or vehicle ZS headlights, as exposure times ware sho:<ened. The incident lighf iaradiaace, H, required to provide su~ciaat expvattre of a sensor pixel is dependent' oa the sensitivity to a particular wavelength of light. Sensor pixels also have a miaia~um light sensitivity to produce a satisfactory signal to noise ratio in the video signal, and a maximum light level before the senSdr pixels bet:4me saturated. The range of tight in~adis~e that can be imaged in a single exposure for the sensor is approximately 100:1. The range of light t WO 93! i 94A a i'CT/ ~ 1.93/001 l ~
w _ ~. e) ~ ~ ~ c~
irradianse which can be presented to the camera 6 duavag a 2~ hour period c~a be varied by as much as 10s:1. Accordingly, the exposure control system litai~ l~
sufficiently to maizltain is within the dynamic taa~,e of the sensor to prevent setzsor saturation from the illumination levels typically ptese:rt durita~ a 24 hour period. 'The exposure control is a f1.8 to f1000 auto iris lens systtm which is designed to provide exposure adjusttraeat based on leas aperture and progressive neutral density fllterin$ of light as the tens aperture decreases. The rate of change of the exposure control, or the rate that H
changes. is restricted as atoviag vehiefes are located by di~ereneiatg images obtained by the camera 6 tom a slowly changing background image, as d~cxibed hereinafter.
'The rate of change is restricted to ensure chat>ga in exposure of the ~ season are sot mistakes for changes in the background image, which would adver~ly affect detection and nvnitoring of vehicles. The auto iris reaction time is set to t~atc~ the ratio at which background images are subtracted from the current ice. 'f~e~~iow rate of cbaage also prevents the leis responding too fast to transient c6aages in lig~, for example, reflected off roofs of vehicles as they peas close to the camera 6. The rate of change is restricted to 10 seconds for a halvir>s or doubling of light irradiance H.
i, The exposure comrol system e»taaes that traasirnt e~cr~sly bri~6t reflections yr headlights do not saturate the sensor pixels by limiting the exposiue on the season to keep 30 is below the sensor's saturation level for the peavk intensity of light received in the field of view lr. The peak video level obtained fmm the camera 6 is; monitored. as discussed hereinafter, sad usexi as a basis for controlling the setting of the diaphragm ~f the iris.
i The sensor sensitivity is selected in order to psodttce video sills which allow 35 the subtraction of the background for vehicles not using headlights during dusk and dawn illumination levels. The sensor is also respot~ive to near 6afra-red light to maximise the sisal from large vehicle side and perimeter lights, yet the respr~ase must be still below a threshold where blooming may occur from vehiclt headlights. ~ 'I~e lei of the camera 6 can be controlled fully to provide sufficient exposure for the ae~sor for vehicles without 30 headlights during the dav~ta sad dusk periods. The maximum lenns aperture is held at f4 for a lurais>aace value of about 10 r~llmi reflecting ~ the ~ , y. ice the c~tiageway luminance level fall below ap iy ~~'o' off this level, vehicle t~.T/Al'93/00. s~
w0 93/ t 944 z ~ 3. ~ ~ ~ ~, 5 segmentation, as discussed hereinafter, is based on vehicle headla~,hts.
Cotnrol si8nals representative of the illumination levels are derived from an illumination hisiogrartt of video signet levels for the pixels, described herei»after.
The control unit 10 of a camera nods 2, as shown .in Figure 8, includes a Motorola 68030 CPU 64 and a detection and trigger sub-system b6 connected to receive images frown the detection camera b, and as acquisition sub-system b8 eoctaseted to receive images from the acquisition camera 8. The sub-systems 66 and 68 include a number of Dataeube pipelined pixel rate video processing circuit boards which are controlled by the C'PU 64. The boards and the CPU 64 are mounted on and i~atcrluoiked by a VM>r (Veaa Module Europe) bus. The CPU 64 and the boards of the sub-systeaas 66 and 68 run a software operrrttaag system knowta as VxWor~s, which is a real time mufti-tasking system. The detection sub-system 66, the fPU 64 and controlling software form the detection circuit 30, and the acquisition sub-~ysttm 68, the and the controlling software form the acquisition cir~it 34. ~ The image buffer and communications controller 57 caw be connected to the acquisition circuit to provide access to she ISD~I 4s.
~i i The detection sub-system 66 the 512 x 512 piatel images of each video field obtained by the detection camera 6 and is dcsigaed to achi~,we low latency between change: in the field of view 12, by using pipelined processing off' the image data with rto intermediate storage. The data rate through tlse video data paths of the pipeline, known as MA7~US, is 10 million pixels per second. 'Processing the vadeo fields individtsaliy, as two consecutive frtmes of half vertical resolution, achieves a ply rate of 50 HZ and :.5 eliminates the deiaterIacing latency required for full frame pro i essing.
The deeectioa sub-sysseas b6 includes a video digitiser ,beard 74 which reoeivts the Eeids output via the detection cannery 6 and converts them into the 512 x 512 pixel representation. The digitiser board 74 is a Dataaxtbe Digima~ board wad produus a greyscale image representation with each pixel having a value within the 2's complement positive range of 0 to 127~ 0 representing black and 127 ra:presentisg white.
'I~ac 313 x 512 pixels era able to produce a live image display as shohvn in Figure 9. The ~'O 93/t9d4;, ~ ~~ ~ ~ ~ ~ ~ I PCC/AL'93/0011;:
_ 1~ _ image produced by the digitiser board 74 is input t~ a background diffezeneer board 76 which, as shown in Figurc 10, subtras;ts a background image, as shown io Figure 11, from the cutreut of rive image to produce a pretiminary diffcrencc raga, shown in Figure 12.
'il~e difference image contprises a grey lever of representation of the moving objects within the ejeld of view 12. ~y virtue of the imaage subtraction the pixci image raage for the difference image extends from -128 to 127. The background differences board 76 is a Datacube MaxSP board.
The background image represents the static backgaound 'viewed by the d~te~ion camera b and is stored in one of two framGStores 71 of a background image score board 70, being a Datacube Fra~mestore board. The ~~ ~~ is coutisualty updated by a background update board 72. which is aaotber l~atacube ISP beard that ensuecs i one of the framestores 71 holds an image correctly representative. of the static d within the ~sld of view 12 of the deteexion camera 6. 'The updatb board 72 then receaves the curnnt background image from one of the faamestorec 71b ~amd is combifled with a filtered form of the preliminary difference image to produce a; new ad image which is outputted by the update board 72 t~ the other framestriae 7ia. The cxmtrotling software thcat switches to the other framestore 71a for submission of the bac~eour~d image to the differettcer board 76, ~ ettsuses the next updated '~tnage is submitted to the ?0 first framestore 71b. 'The background update board i:,lters thtprelim6nary difference image is accordance with a filter characteristic 73, as shown in Figure I0, which is brad in RAI~t aad perfoans a limiting function vn the grey level p3xe9g of the pnlimiaary difference image so as to restrict them between a programmable range, for example -3 and +2 pixel ran;e. The timitins functicm te~accs the ion made to the current 35 background image when' it is combined with the dif~e~ 'u, after having been subject to a delay 74 to allow for the time taken to apply the liittititag filter function 73.
The limiting fuaetion ensue the correction made to the baokgr~und image per frame ~
only slight so that traastent dtffere:xes, such as those produ~d~ by moving ~b~eocsy are not allowed to signafirantly alter the stored background inaageheld in the image store 30 board 7Q. "f be shape of the altar function 73 that greet level differences added to the background image are to a level t for all ' ~ levels ~t and -t for all difference levels <-t, where t is s low tjtreshold such as 2< The state of the ii PCT/ A 1,'93/0011 - L~ -bacleground update board 72 can also be changed to disable update of t$te background image. The rate of change in the background image is xt so as to be faster theta the rate of change of scenic exposure due to variation in the lens aperture of the detection camera 6. The rate change governed by the limiting function is impoata>9t because if tht rate is too slow fighting changes tin produce incorrect difference images, and if the rate is too fast then moving objects may appear in the background image as a blur.
The preliminary difference image produced by the backgtouad differencer board 76 is outpusted to a third Datacube MaxSP board, a shadow eliaxination board 77. The shadows produced by vehielas which appear in the di~ereace ieaage, shown in Figure ~2, pose a significant problem for the images processed to determiated the type of vehicle.
T7te shadows can mistakenly represent the vehicle as being larger,; than its actual site, and if a discrimination is being made between the large vehicles, such as trucks and buses, and small vehicles, such as cars and motorcycles, then the shadow cast by a cat cast lead to it being classified as a large vehicle. Therefore the shadow elimination board 77 is employed to eliminate all grey levels is the difference imaged which caould represent shadows. This is done by defining a grey level window range 79 is RAM, a shown in Figure 10, wheml~y ~ preliminary difference image is proceed so as to set to zero all pixels having a gaey level within the window 79. The result is; then u~d to mask the preliminary difference image so that the elimiaatioa board 77 outputs a shadow 5ltered difference image having ail of the pixels with grey levels withi~a the window range 79 removed. Figure I3 illustrates a Iive image with all of the pixels having a grey level within the range of the window 79 shown as given. T'he range defined by the window 79 is adjusted depending on the light conditions within the ~ald of view 12 of the ~5 detection camera 6, as discussed hereinafter.
The shadow filtered difference image is inputted to a threshold and mediaa 5lter board 78, which is a Dacacube Snap board. The f iter board 78 ~petforms b inary image processing on the difference image so as to convert the grey level representation of the moving objes~ to a binary repraentadoa, which oorrcsponds' . to white or black, for further pr~oeessing by the deteaioa sub-system 66. 3be 5lter board 78 tries a threshold value to convert all of the pixels, with grey level values within t~ range -i28 to *1Z7.

w~ ~3W a4t_ pCWA~,~93ioot t:

to pixels having values of either 0 or ?55. Accordingly, the faatal difference image produced by the filter board 78, when viewed by a real time display, shows the m~vi~g objects within the field of view 12 as a collection of white pixel blobs, as illustrated in Figure 14. The blobs may correspond to parts of moving vehicles which reflect sunlight grad, at night, may correspond to light produced by a vehicle's external lights. Noise regions of one or more pixels in sin ate eliminated by the board 78 which performs binary median filtering on 3 by 3 pixel neighbout5.
The light conditions within the field of view 12 of the detection eamera 6 are determined with reference to a histogram 150, as shown is Figural 15, of pixel grey levels produced by the CPU 64. Tbc CPU 64 processes a window o~ the stored background image which is appeo~dmately 300 x 400 pixels every 10 seconds. The CPU 64 calculates the number of pixels its the window having each grey level and sabulate5 the results as the histogram 150, with the number of pixels on the vertical axis 152 and the grey level values on the horizontal axis 154. The lusto150 can be displayed to provide a real time representation of the light within the field of view 12.
From the grey level value which represents the position of the median 136; one of three lighting conditions, day, dusk, or ttigbt, can be instantaneously detetmin~ed. Dawn is considered to be the same lighting a9ndition as dusk. The positiooa of thd peak i55, median 156 and the minimum 15g art used to determine the range of the 79 used its the shadow elimination board 77. For daytime conditions, the ,shadow window 79 is determined as being from the values a.peak to (peak + mediea)~, where a is typically 0.5. For dusk conditions, the shadow window 79 is from minimu~a to (peak +
~edian)/2.
Shadow pixels of a~urse, do not need to be eliminated duri~ night conditions.
r5 Estimation of the shadow pixel range is as approximate techeiqu~ which is aided if areas of Permanent shadow are in the i9eld of view 12, such as cast fby trees or an overpass bridge.
't i The segmented im~ss produced by the fitter board 78 ale submitted to ~ Area Perimeter Aareletator (APA) board 80, which is an APA 512 boaitd produced by Atlantek Micsosystetns, of Adelaide Australia, d~i~ed to acxaelermte ' the pressing of axes parameters of objects in a video scene, The hosed 80 v~ith concrollin=
software ~'O 93/ 19~d1 ~CT/A L'93/pOt t o ~~.Jj~~S

to perform analysis of the white piacel blobs within a 3~ x 4p0 pixel window corresponding to the window on which the histogram 150 is produced. The APA
board Rti and the software perfor>?t a classification and feature extraction process iat teal tints on the blobs so as to facSlitate the fomaation of clustezs of blobs which correspond to a moving vehicle. The APA board 8t? computes features of the white pixel blobs and the ftatures are used by the clustering software to determine, on the basis of rules aztd classification code, whether the blobs can be combined to form a cluster. Unce formed, the size of a cluster indicates whether it corresponds to a large vehicle, such as a truck or bug, Or a small vehacl~, such as a csr. Labelling software is used t0 monitor movement of clustezs over successive fields so as to detetaaine ~srhich clusters are to be assigned a unique label and which clusters are to share a label, a~ they are considered to relate to the same vehicle.
Different considerations apply in respect to whether the c~iageway 15 is being viewed by the detection camera 6 at night or during the day, and the :ales and cia$sifications used are adjusted, on the basis of the data provided by the histogram 150, to account for night conditions, rain and inctement weather, w~sioh result in a moving vehicle producing different corresponding pixel blobs. For rxsanple, the :ales and classification code needs to be adjusted to account for refleetidn produced by vehicle '?0 lights on the road during night conditions.
Once a cluster his been formed, its movement is mot~tored to detern~ine its insiaataneous speed and its position with rcspeci to a point on th~ edge of the road using KaAmaa filter techniques. Corrections are made for per~ctive as the cluster moves 35 towards the cameras b and 8, The information obtained from mibnitoring the movement of the cluster is used by this C'CpU 64 to predict when the cluster will cater the field of view 20 of the acquisition camera 8, aid in particular when a vehicle scathes a position '-d which an image of the vehicle is t0 acquired. The prrrdieted liras estimate is updated for overt' field generated by the detection samara 6, 50 tunes peg second, The predicted 30 time is continually corrected as the Cpt3 b4 ~onito~ m~verne~t of a cluster until it is satisfied the cluster will enter the 5eid off view within 10 to 20 ~uts. A CPU
64 predicts the time by specifying the number of scan li~aes wlti~ net t~ ~~ sped by the camera ~O 93/19441 PCT/At,'9310011c ~13~~1J

6 before the clusters within the field of ~~iow ~0.
Performance of the control utsit 10 can be monitored and'controiled by peripheral devices, such as a printer 94 for error and event lagging, a real came seaeus display 98, and a control workstation 100, which may all be cormected to the CPU 64 and the boards of the control unit 10 directly or by a local area network 102. A display of the rest time I
status display 98 is illustrated in Figtue 16 w~eJt is the live imago produced by the digitiser board y4 superimposed with cluster markings and other data. The histogr~ 150 is displayed at the left of the screen and the box around the vehicles are cl,~t~ which have been formed. The label number for each duster is sbowm me the lower right hand i comer of oath cluster, and the estimated speed of the vehicle, obt~iaed by monitoring the cluster, is displayed directly below the label cumber. TI~ large box around the vehicles represents the processing window, on which the clustering, ~ labelling and freckles sofsware operate, in addition to the Iaistogram software. The I'>~ across the window is an acquisition line which cxlTesponds to the position 22 at which,! high resolution images are to be acquired by the acquisition camera 8. A diagnostic glSaphics botttd 82, which is a ~atacubc Maxgraph board, is used to queue sad configure graphic irna=es for the real i time status display 98.
i The image processing performed by the ~7 64 and the APA hosed 80 for vehicle classification is baadled by feature extraction, clustatan~. labelling and erackirrg software. The operation of the software a largely controlled Iby parameter variables, which msy bo altered via an interactive shell of the software or byroarrote procedure calls ~ a graphical interactive command tool runaiag under Xv~hado~rs oa the control ZS workshtion 140.
The AF'A hood 80 roduads the binary image pixels ~to a stream of feature vectors representing the blobs, or regions, in the imates. Only; a small sub-set of the features which can be computed by the APA are requared, being! the area, perimeter and 30 bounding box for each blob, or region. A region is tepresermd ,by raw data of ib bytes and for a field of view I2 which includes 20 r~egio~, ehe dam ate a I6 kbytesls which is less then 0.2~r of the data rate for binary images, and is ~bte for software i WO 93Jt9441- ~ ~ ~ ~ ~ ~ ~ p~lpD.'93/00115 P~~sin~ by the CPti 64.
The raw seed parameters are read from the ,~pe~ h~~,are by the A,Qt~Task 170.
as shown in Figure 17. A timC stamp is givers to each blob, nerd some initial screening is performed, why regions such as "black triangles" described hereinafter, are located and removed. Time stamping, inter alia, allows any latency in the system to be mcasyred and compensated for. The seeds which ~~~nd to wee blobs within certain area constraints are passed via a VxVWorks message pipe to the aeedTaSk 17:. The setdTask unpacks the raw seed parameters, or structures, and perftnms classification of regions based on each regions height to width ratio, "circularity", (urea sad "coverage". as described hereinafter. Umvanted regions such as headlight end road reflections are renooved and then each classified region is passed via aaot6er message pipe to the clusterTask 174. i i 1s The clustering task is divided fnco five ~bs~tions 1'16, region classification.
region extension, clustering, region unextenaion and cluster el~tsification.
One the regions have been clustered into clusters wlueb have been classified as corrtspondia~; to separate vehicles, the coordinates of the eluste~ are gassed onto a label task 178, once again by a message pipe. The label cask monitors each ch~stdr over a given period of 30 time and if a cluster appears in rnugltly the same place as did ~ cluster from a previous video flame, then the label task considers them to be the same ~,~~ter. In this case, the new cluster inherits t,'be label from ehe previous cluster.
ache if no match can be made, the new cittster is given a new label. The elustea's label, l: they panned via .m dies, along with its allege pipe to a trajotxory 180. The trajectory task I80 determines the tiara to tr9Bger the acquisition c~aaexa 8 for cluster of a selected cFass, such as large vehicles. The put cluster box task 182, move cluster box task 184, put label task 186, remove label task 188 and the histogna~ t~k 190 are tai used to generate graphics overlaid on the video image, as shown in ir'iguure 16, for diagnostic purposes, ?he blob shape analysis performed by t~ ,~sATssk 174 sad seedlask I72 is not extensive during daytime s=ensation, as all blobs are ooa~idered valid.
However, i ~'O 93119491 ~° ~, ~ ,~ ~ ~ ~ 1 ~~I r~ lr'93/p~p 1 f ~
-r0-during dusk and night time segate>ztation, blobs tine occur due to vehicle headlight reflection, and if these blobs are clustered in with tnae vehicle blobs, then the front-of-vehicle coordinates, which are taken from the bottom of the cluster box, wilt bs incorrect. Itt order to correctly locate each cluster box at the front of each vehicle, blobs which are recognised as being due to headlight reflections are identified and removed fxfore blobs are clustered. ether problem blobs are those which cc,~gsp~d to road lane markets. These appear when the mount for the detection ca~anera 6 shakes.
During camera shake, the iacon~ing video israage ao longer precisely coarespoatds to the stored static back~ound image, and therefore the result frorai the backgcntand image t0 subtraction is that the road tnukers appear to have moved. Ate, the blobs that result from camera shake aae identi~cd snd filtered out before cdusteri~ commes. A
further problem is "black triangles". The APA board 80 posse3ses a hardware fault which causes the polarity of blobs to be specified iacorzectly. If a black region finishes at the right head side of the pixel processing window, it can be ~ ~ tly labelled as a white IS region by the APA board 80. These white regions can rhea' become eandidates for clustering unless filtered out by the seedl"ask 19Z. Typically, ,~whm a lane marker 190 appears on the right side of the pixel processing window 19Z, ~ s~~ Figure 18, it pmduees a black triangular blob 194, a "bt~~e", which is iaadvtrtently represented by white pixels, in the top right heard corner. 'I9~ triangular blob 1~4 is ?0 identified and removed. A canvenient side effect of the polarity fault is that the toad lane line~matker 1!~0, which usually mast be identified and tetnoved by other shape characteristics, is la~'belled by the APA beard gp ~ blue, ~ ~ therefore automatically filtered nut. ~ .
Regions are classified into one of the following types;
(i) Headlight refleaioas;
(iij Road artefacts; such as road lane markers, which ~e to carnets i shake, (iiij Laghts; attd (iv) Other; dutilag daytime segraeatatioa staost of ~he regions that are not classified as road artefacts are classified "other".

''~'O 93/19441 ~ .~ e9 :v ;) ~ e7 PCTI~1,~3/0011s I
During day and dusk conditions, illumiaaated headlig' do not appear segmented from other segmented parts of a moving vehicle, and so ~. :. ~ts are trot classified. At night, however, cfrcular regions are elassified as cithcr "headli~t" or "stnallalight".
depending on the area and position within the field of view lm: Distant headlight pairs which are typically segmented tom the baekgtound image as a'siagle joined region, are classified as "joined headlights". To obtain coma initial ~ Blusters, distant joined headlights need to be distinguishod horn the small perimeter lights of large vehicles.
The main shape measure that is used duaing dusk and bight time processing is "circularity". This is a measure which co~iskrs bow dose each blob is to the shape of the circle by comparing the blob's area to its perimeter. Ia tht case of a circle:
i ~a = errs (4D
= 2Rr (~
The radius team eaa be elimiaatexi, since it is only relevant for circles, by squaring the perimeter equation sad taking the quotient of the two terms. For a drcle, this produces a constant:
I
><r= = 1 (paimeca~~ (2aa~ 4rt (6) To make a circularity measurement equal to 1 for a cirCls, equatioa 6 is simply multiplied by the iaverse of the constant. Tlvs provides a circularity measure which can be sppiiexi to blobs whereby a cirarlar blob will have a m ~ea~urement value of 1, as followvs:
4~ ' i.o ~ c~>
i For a square blob of unit area, Area = 1. Perimeter = 4, ~ the circularity measurers ZO is as follows:

WO93/194~t1 I P(:T/A~.'93/mpll~
Circularity = 4~ _ 'e = p.7g5 (g) (4)=
i-For an c~uilatcral triangle with sides of unit length, Atca = X314, lyeaimeter s 3.
the circuiariey treasures is as follows;
= 0.6 (9) A further measurement employed, that is particularly u~ful in detecting road laad/line markings, is "coverage". Coverase is the measured ratio between the arcs of s a blob to the area of ice bounding box. The bounding box 200, as shown is Figure I9, is aligned with the ApA board coordinate axes, which arc the sates of the APA
processing window. The APA axes ate not ne~ssariiy aligned with the anajor axis of the blob itself.
For inssaace, a rectangular blob X02 which is aligned with she APA coordinate axes would have a high coverage value, whereat a rectangul,~ blob ZOt which is not aligned with the axes nsay have a medium coverage value. A concave ape 20b would produced a medium coverage value, sad a line 208, diagonal to the A19A eaoordinate axis r01 would produce a low coverage value. Road lace markings can be simply detected betwause they have a low coverage value. If the lane markings are sat diagqnrtl, but vertical, then the measure is insufficient and is such cases a measure of the ratio of the blob's major axis length to it's minor axis length can be used instead. ' During night time segmentation the coagulated blobs ~ of joined headlights are identified by their height to width ratio as they Mead to be twice the expected area of one headlight. Joined headlights need to be detected sd that a headlight count maintained for 30 each cluster is correct. i Headlight reflections appear as large elongated blob:, acid are detected initially on the basis of their size and chara~cristie shape. The blobs art; fed as relating td headlight reflections by extending the blobs ver~ric~lly to deter~ine whether try extend r5 from a headlight region.
As the vehicle moaatoriag system is capable of ~riairous automatic operative.
I

~'O 93/19441 ~ ~ ~ ~ ~ ~ ~ PCT/AC,'93/OOt 1~
- r3 clustering of regions takes into account different lighting conditions. 'the technique of static background subtraction. described previously, segments moving objects froth the .
~~ideo image obtained by the detection oarncra 6, but the regions that result from the scgrmentation process depend on the ambient lighting conditions at the time of day.
Dining daytime scgmeniatiaa> large regions typically result, whereas during night time only headlights and the smaller sidelights on trucks are segaieated. lauring dusk, lit headlights do not appear segmented from the other visible parts of moving vehicles, however, reflections upon the surface of the road caused by the headlights need to be removed, as discussed above.
The clustering process operates on the segmented regions or blobs sad each vehicle is typically segmented into several separate regions, as ~hovYn is Figure 12. For instance, a car will often appear split by its wi»dscxeen into a roof-region and a bonnet-region. Large vehicles typically segment into more regions. The cluster task groups these regions into "logical vehicles" so that they can be backed.
Distant vehicles tend to be segmented together into one region due to vehicle, oociusioa at the image horizon. Whale the segtaeneed regions at this distance car ba~ tracked, they cannot be reliably clustered into separate vshiclcs. Emphasis is planed on cbrrcctly clustering lower regions that art: closer to the acquisition line 22, and con:equeatly the clustering process scans from lower regions to higher regions in each image.
i I~ two vehicles are sagrneated into the,same region, tha~ they will be clustered together. The cluster task does not separate vehicles that have been segmented together into a single region. The coordinates of each cluster are seat to label task 178 which ?5 matches tend separates clusters over consecutive video fields. The cluster task and the label task classify clusters on the basis of classi~cstioa data. The coordinates passed to the trajectory task 180 Correspond to as estimation as to the fr ,~t of the vehicle, at the road surface level. Cluster information oa all vehicles is provid~d to the trajectory task, which tracks the clusters and selects far which vehicles itaagcsue to be acquired.
clustering is achieved as a middle paint batw~a "over Clustering,. sad "under clustering". dot the over clustering extreme, all rcgioat sae clustered into one ~'O 93/19441 ~ ~ ~ ~' ~ ~ ~ PCT~Atr'9310~11:~
single cluster and then only the lowest vehicle in the cluster is tracked.
'I°his is because the lowest point of each cluster is passed by the label task t~ the trajectory task. The classification of the cluster, which is based on its height and width will be iacotrect. At the under ciusteria~~ extreme, if no regions are clustered together, that is each region obtains its own unique cluster and label, then the trajectory task is over-burdened in an attempt to track every region, vehieie classification will fail is a taun~ber of instances, and images will be inadvertently acquired sand missed. For the purposes of vehicle image acquisition, it is better to mistake a vehicle-roof for a vehieie-fi~nt and begin to track it than it is to mistake a vehicle-front for vehicle-roof and so, by addiacg it to the beak of another cluster, not track the vehicle-front, Therefore the cluster task has been written to use an optimal middle point which ties on the side of under clustering rather than over clustering. i The cluster task performs clustering essentially by extending the boundary of each segmented region by a certain ataouttt, and then joining asiy region$ that overlap.
Regions that overlap are "clustered". The cluster task, howeverl determines correctly the amount of cutension which should be applied to each region. I?uring daytime segmentation, very little region extension is required, yet ~duting night time, the segmentation process produces small sparse region that require large amounts of ?0 extension in order to achieve overlap.
i An important aspect is the construction of a cluster is these the bottom region of each cluster should be the front of a vehicle. Invalid rcgxons~ such as regions due to headlight reflections, must not tx clustered, and are thus not extended. After every valid ~5 vehicle region is the image is extead~ by a oertaan aaaount. the clustering process begins with the lowest region is the image, The lowest is considered' first which is the region most likely to cause triggering of the acquisition camera g. ~' The coordinates of the lowest region are used to initaalix a Bluster strueture.
30 'rhea al! exteacdcd regions above the initial region are tied fdr overlap.
If any region does not overlap with the coordinates of the clctster, then t~ cluster coosdinatcs are updated to include the region and the region is a~arkod as sh~tertd.
V6~hsnever a new wc~ ~~~,~~ot ~ ~. ~ ~' S 1 ~ , ~cr~,~LV3~oozt:
~~s~
region is added to a cluster, all remaining unclustered regions bccorne possible cluster candidates again. Thus the list of regions is traversed agasa fiom the bottotit of the image. Although the regions in the list which have already beets snacked as cltastered cats be skipped, this is considered sub«-optimal. t?nce the entire rtst of regions have beet, traversed without any overlap detected, the next cluster is begun with the lowtst remaining region. The clustering pcontinues in this saaaner until no regions are left unciustcred. The list of clttstets are then unextended and passed to the label task.
1rt perfvrrning region extension, regions are extended by a variable aa~otsat in the vertical direction, but extended by a standard amount in the horizontal direction, with reference to the APA coordinate axis. Horizontal extessioti is unnecessary during daytime segmentation, as a vehicle blobs tend to be coasecte~ aaoss the fuQ
width of the vehicle. It is in the vertical disectioa that blobs due to ~tlae same vehicle appear disc~nected. For example, two blobs that typis~liy re t ~ car might be due to its bonnet and its roof. 'Ihcss two blobs wall streteh over the full dvidth of the vehicle, and appear one above the other. Furthermore, so long as ono blob ~or each vehicle st~ches the full width, the cluster croordiaates will be wide enough to '~corpocate stay blobs that might otherwise need horizontal extension to be clustered tog,at~ter. The full width blob provides the extension. with reference to the example iltusuated in Figure 20, tl~ region 30 r10 becomes added to the region 212 on the right, from which;the cluster 214 is begun, only because the full width region 21b above was added to tb~ region '13 to form the cluster 214. It the repon list wag not researched from t ~be beginning of the list I
continuously, the overlap of she previously tested region 2I0 wo!uid oot have been found.
It is for this reason that the clustering task, as di~ussrd above, r9es~oasiders all uaclustered '?5 aegioas after addia~g a region.
The cluster task is able to perform one of ~o eeatttens~ioa methods. 'I~e fist method takes the vertical or Y coordinate of the region as as input to a loak~up table that speeiSes the amount of extension to be applied. 'IVs amount oi~tbe extension, and hence 30 the degree of cltastering, is they modified according to Gghtiag itions.
.~s the outside light level de~asss, sad regions reduce is sizes, the anaouat of extension applied to regions can be gradually inasascd. Fut2hermote, the ~'ve is the image can be I

WO 93119441 PCT/A193/0011:
~~j~~$~
_ 26 -compensated for by adjusting the values stored in the iook~up t2~ble accordiztgly, i.e.
distant regions high in the canasta irz~age earl be extended less than near regions which are low in the image. An example of the extension values stored in the look-up table is illustrated by the graph 2I8 shown in Figure 21 of extensioa~ar~ount v. Y
coordinates.
S All extension amounts arc pxovidcd in pixel numbers. The ixcond extensioat method extends each region by an amount proportional to its width. 'IVs method is largely based on an observation of the shapes of legions obtained during day;ime segmentation. Small regions, which arc typically far away, are n~inimaily extended, large vehicle body regions.
which are typically close, square and occur one per vehicle, axe~rninimally extended, and wide short regions, which are often vehicle fxon~, ace greatly txtended.
Essendaliy, as illustrated in Figure 22, this results is every region bai~ary 220 and 2',w', being approximately square. In Figute 22, the boe~ies 220 and 22$ of both regions 224 and 2~6 have been extended vetticaily to equal at least their width. ~'Thcrefore the wide short region 224 has been extended a great deal more ihaa the large s~uara region 226. Region IS 224 would tx a vehicle front portion disposed under the vehicle body region '. a6.
Therefore, the two regions 224 and 226 can be matched without too much extension. If i the large region 226 is ovtr extended, then it may overlap with a succeeding vehicle fi~ant. In tht preferred embodiment, this atethod is only t ~ ploycd during daytime segmentation as nighi time processing requires a large amount of region extension, 30 although it is envisaged the extetASion factor used in the extension calculation can be enlarged foe night time use.
During night time clustering all of the regions to be elus~ered arc essentially small circles, asd s truck cluster, for example, is aonsaucted by eoa~idering the possibility of 25 whether each light could feasibly fit into a stored wck template. For the first region in a cluster, to fit within the template, there is a maximum distance of light separation which cannot be cxaeeded. :i Overlap of regions is detected by competing the coordinates of regions and 30 clusters, wherein the top_left (xt,Y') and bottom-right (xyY~ coordinates for both regions and clusters are known. For the image pleas coordinates, x ' from left to right and y increases from top to bottom. Considering first the boaixontal, x coordinate.
i w0 93/t944t PCT/,~L'93/OOt 1~
overlap for the regions R" Ra, R" R" Rs and R6 illustrated is Figure ?3 the test for overlap with the cluster C~ i~:
R~(xt) < C~(x~ (10) Cat'tt) ' Ra(~ (11) If both of the two equations a~tte true, then there is overlap is the horiaontal direction. Therefore, horizontal overlap is tnae for R~, ltr, R, and ~ but region R= fails 5 the test as equation 10 is not true and region R6 fails the scat a equation 8 is not true.
A similar test is performed is the vertical direciioa as follows:
I~(yz) a C~(yi) (i2) There is no need to perform the complimsatary test for R,(y~ because the regions are outputted from the APA board 80 is order tom top to bottorm and as the cluster task processes all regions in a list fiom the bottom up. the complimentary test, C,(Y~' R,(Y~), is unnecessary as it will always be true.
Clustering during day ti~htir~' conditions is based oa the overlap test discussed above, yet during dusk and night oonditioas clusterir~ involves soasideratioa of additional soles, primarily due to she increased ambiguity and greater separation between regions of the same vehicle. Certain regions should also txvet be clustered, such as headlight reflections and aoiae from baeitgmund image areas due to vibration of the detection camera 6 discussed previously. Clustetzag therefore also involves consideration of a series of rules based oa the various re&ioa cl~sificxtiona 'dzs~sed previously. The rules include: ;
30 (i) An e~cteaded region must spatially overlap a cluster t~ be added to that cluster.
(ii) If a region overlaps more that one cluster, then it' is added to the lowest cluster.
(iii) A region to be clustered caaaat already .have bees added to the cluster.
(iv) A "joined headlights" rewoaa ratumt be added to an existing cluster.
Retiomms of this type cps only initiate a cluster.

F°C1'I,~L'93/001 IS

_2g_ (v) f?nly a predetermined number of "headlight" regions can be added to a cluster, the predetermined number being a system parameter which can ~
adjusted froth the user interface.
(vi) As many "other" and "sasall light" regions as is spatially allowed teas be added to a clustet.
(v ii) A region which touches or includes part of the cop of the processing window can initiate a cluster but cannot be added to a cluster.
(viii) A further "headlight" region to be added to a cluster must be horiaontally aligned with another "headlight" region in that cluster, which is determined an the basis of the difference between the regions lower y ordinates.
(ix) "Reflection" and "toad artefact" rtgioas are not added to any cluster.
For monitoring a roadway, clusters are classi~od into one of three claosses:
car, ute (a small fiat-bed utility truck) or truck. Therefore all large vehicles, such as buses and artiarlated vehicles, are classified as a crock. Cluster ciass~fi~tion is based on the height and width of each cluster box, sand the number of lights within the cluster dur9ag night conditions. The height sad width data for each classification is modified via procedure calls to the histogram tart 190 a~ the lighting conditions change faom day to dusk and eo night, ate. The cluster width is as important as th i cluster height because, ?0 for example, a large four wheel drive vehicle towing a trailer might product a cluster which exceeds ttu truck height throshoid but is unlikely to be atwide as a uvck or bus.
A histogram of cluster heights and widths of motor vehicles iachides distinct peaks which correspond to vanious vehicle class, and . is used to set the stored classification i thresholds automatically, The height and width histogram is in the display of ~5 Ffgure 16, For example, a cluster is classified as a truck ~f' one of the following co:Dditions is true:
(i) The cluster height sad width excxed the truck threshold.
(ii) The lighting condition is night and the cluster .exceeds the truck width threshold.
~0 (iii) The lighting condition is night sad the number of fail lights in the cluster exceeds the small light thick threshold.
(iv) The chsster height is within a predc r~oge of the ttu~k height i r W~ 93/19441 /A4'93/ti011c - ?9 r threshold and the number of small light regions in the ~luster~ exceeds the truck small light threshold.
~,s ambient lighting drops, the six of the track c ~ stars are reduced, and S consequently the height and width thresholds decxease, dtpgndin~ on the lighting conditions, as determined by the histogram task 190. The classification for eaeh Bluster is scored in a clustered data strut~turc, together with the cluster's coordinates and time stamp. The clustered data is then passed to the label task 1713.
The label task 1'18 assigns a label to each unique clusreri and tracks clusters over time by matching an array of previously seen clustcas to each subsequent video field of clusters. If a cluster appears in roughly the same place as a cluster from a previotes field, then the label task 178 consideas them to be the sane cltasmr.~ Where a mateb cats be made, the new cluster inherits the ttaique label of the previously s~en duster. If a cluster cannot be matched, then a txw label is created for that olusser. I Clusters may disappear for a few fields, and it is an objective of the label task 178 to ~iierntine whether a cluster is indeed new or whether It has just appeared again after a parsed of absence.
i The matching of clusters is bases: on location. C,'h~te~, size can be used as as emra match parameter but the current location heuristic has been found suffieacat. It oar be assumed the clusters will not move very far from their position in the previous held, and if a cluster moves so far that its boundary coordinates is the present frame do not overlap with its boundary ooordinatea from the previous frame, then the previous label will not lx traasferted. Cttuters can split and join, both vereie~liy and horizontally, as ~5 they an tracked from held to field. Two labelling methods heave been developed, with the second being the preferred method which is presently used.
i The first labelling method involves two seoipxo~l tests whiob are used to determine whether a new cluster should inherit an old clusters label.
°t'hha fiast test is to determine whether the carne of a ~vv cluster 230 lien witlltin the boundary of any clusters 232 and 234, as shows m 1 igurc 24, on a list of pf~evidusly aeon elttstets, oall~
the label list. For the ehrster 230, the tsst fails, but for the ~v citasters ~6 and ?3$

~'~'O 93/ 194A 1 p~ T/ A L' 93~ 0011 ~1j~~1~
_3p_ their centres fall within the older cluster 240 so the lowest new cluster 238 inherits the tabet of the old cluster 240, and the upper new cluster ?36 is assi~ed a new label. The second test, whieh is executed when the fizst test fails, determines whether the centres of any of the clusters on the label List lie within the boundaries ~f the clusters from the S current video field. Therefore as the centres of the old clusters 232 and 234 fall within the boundaries of the new cluster 230, a match is detected, sled the new cluster 230 inherits the label of the lower old Bluster 234. Applying the .second cast to the new clusters 236 and 238 results in failure as the centre of the old cluster 240 does not Iie within any of the sew ciusters 236 and 238, and therefore applying this test to these clusters would result in the new clusters 236 and 238 both being assigened sew labels.
The second labelling method is based ors the dustetilsg overlap technique described previously. Essentially, the bounding box of tech duster from the current field is tested for an overlap with dustem in the cluster list. The duster list is search from bosom to top, in a similar mauaer to the search method described for detecting overlapping regions. In this way, if two clusters merje into a single duster, then the 5rsa overlap found will be an overlap with the lower duster. Once a match is found, the search is terminated, and the label which is tasnsferred is marked as applied to a new cluster. Therefore a label censor be transferred twice within one search of a new video frame. 'ihe second method is preferred as it requires half the nu~nbcr of tests as the first method, and a cluster can move further between successive frrlmes yet still inherit its label, Ia the fsnt method, where ceatres are matched to the boundaries, the maxilmum displseament allowed between fields is half the width (or height) ~of the clusters, wherta~s in the second method, where boundaries are checked for overlap, the maximum displacement is the entire width (or height) of the diner. '17>e~fore the second method allows a cluster to move ewice the distance of the first method:
As clusters travel successive fields in tune, they tend to split or join, and if a cluster splits, then the lalxl is taaasferrred eo the lower of the two dusters, and the upper cluster, which would typically be another vehide behiad t~ idwcr duster, is provided with a new label. Alternatively, if two dustet3 join, then the old lower duster's label is transfersed to the sew combiaed duster, and the other duster's label is allowed to expire.

~~ 93119441, ~ ~ ~ ~ ~ ~ ~ ~~'T/Al.'931p01 t i _31_ The label of the lower of two clusteas is uansfcrred after a split or join because the lowest cluster is most likely to include the front of a vehicle, an~ is therefore gi~~en priority with regard to maintaining t:luster labels.
A record of the bounding box coordinates is maintained for each cluster in the cluster list, together with its label, the labels age, and when the cluster was last soon.
whenever a lalx! is inherited, its age increases, aed its last scene value is reset. If a label is not transferred in the couur~e of one bald, its last scene value is incremented. A
label is removed from the cluster list if its last scene value exceeds a label tenure lU chacshold. Cluster labels, coordinates and classifications are passed to the trajectory task 1 gl0. i The trajectory task 180 uses the received cluster data to track the elustets over successive video ~clds. Tine coordinates used for tracking ~ cluster box are the coordinates of the cenm of the base of the box, and the eoo~dinatc system for the roadway 16 which is adopted is illustrated in Figure 25. 'I"ltc datum 300 of the roadway ~ .
coordinate system is an arbitrary poina on the roadway, which ~as been chorea as the centre of the left hand fog line underneath the edge of as overpass brid=e holding the cameras b and 8. Vehicles 302 travel in the positive Y axis direction on the roadway lb, 30 staving at a negative value is the distance. The trajectory of a cluster box in image plane coordinates (xi, y~, as shown in the graph of Figure 26 is not linear with time due to the effect of perspective. Therefore a camera transformation is e~plied so as to covert image plane coordinates to real world 3-D coordinates. In m~atri~c forge, the coverall camera transformation is as follows:

~ 0 J~ 0 ~ ar ~0 0 0 I 0 0 a,I, ~y (13) a 0 0 -l~f 1 0 0 1 ~ ~ 0 0 I I I
2~ where ax X-axis scaling factor is pixelslmm (iatr~asic) c~, Y-axis scalang factor in pixelslmm (in~ias'sc) ' WO 93/19441 PCT/AL'93/Opl 1:~
~~j~~l~
_ Xa image place offset in pixels (intrinsic) Ya image plane offset is pixels (intrinsic) focal length (intrinsic) °f~~, detection camera 6 position in world coordinates-(extrinsic) 'I~e intrinsic parameters are incite characteristics of the sera and sensor, while ttte extrinsic parameters are characteristics only of the position and orientation of the camera. The principle point of the image plane is the intsrststioa of the optical axis and that plane, at coordinates (Xo,Yo). Equation 13 can be writtenas:
x x yt o ~ i x; 1 where C is the camera calibration matrix, a 3 x 4 homogeneous transform which performs scaling, translation and perspective corteaion. The image pl~e coordinates are then cxgressed in terms of homogeneous coordinate= as:
X~ s si ~' iii s y i ~ .'~~ ~ , !1~
z The general petspectlve transform maps a ray in three ~ dimensional space to a point on the image plane. For vehicle coordinates in the imaue plane as seen by the detection camera 6, a unique three dimensional location of the vehicle cannot be detesmiaed so the bottorg of a cluster box received ft~ the label task is considered to v be on the roadway, i,e. z * 0, and therefore the box can be with reference to the i roadway x and y coordinates. Tire equations 14, 15 and lh,; given the image pleas coordinates and z, can be solved simultaneously for the roadway co~d~ates x and y to specify the position of a vehicle. The equations have been s0~ved asiag the computer algebra package ARAPLE, and the solution, is C notation, ~ a~s~.follows:
den * (-Xi'C31'C22+Xi'4'3.'.'C~lt(Yi'1-X21)°C12+(-Y~'C32+C~)'Cll);

~'O 93/ 19~d4 d PCT/de L'93/00 t !:

y = -(-Xi'C31°C"24+~Ci°(:34'C21+(Yi°C?1-C'~1)°Cl~
b (Xi'C33°L'? 1-Xi'C31 °CZ3)°z+(Yi°C31-C'1 ~'z'C: ;ø
{-Yi'C34+~4+(-xi°C33+C,~3)sz)'Cll ) / dcn;
~ x = (-C'r4°Xi°C;3.',+C'=''Xl°C34+(Yi°C3'?-C~')'C14.+
{CZ3'Xi'C33-C23°Xi'C3Z)°z+
(Yi°C32-C'2Z)'z°CI3+{-'Yi~C~4+C'24+(-Yi°C33+G"23)°z)°C1s ) I den; _ .
'The solution explicitly includes height above the rosdw~y, z, which can be set at zero for daytime operation or some marginal distance above the roadway, whereas at ttigbt, the bottom of the cluster box geaeraily corresponds to the height of the headlights above the road, and therefore z is set to a notional headlight heist. Figure M7 illustrates a graph of the same vehicle trajeetory as in Fib 2b, after the trajectory bas been mapped to the roadway cootdinatea x and y. 'tee trajectory illuStratcs the vehicle is moving at a constant speed, and is the left hand lane.
The time at which the vehicle 302 wilt reach the acq:iisition line Zo, and the future location of the vehicle 302, need to be predicted, due tb latency is ehe systeia.
Considerable latency exists between a taigger request and 'ea~age aaquisitlon via the r0 acquisition camera 8, and additional latency is caused by pixel transfer, image processing pipeline delay and software processing delay. The iaformatiaa obtaaned off the basis of the iazagea required by the detection eamera 6 provide a delayed representation of the actual vehicle position, and therefore it is neerxsary estimate the ~utuze position and speed of the vehicle 302.
The position estimates of a vehicle obtained by the inverse perspective discussed above are quite noisy due to quantisation effects, gartieularly when vehicles are in the distance, therefore simple di~ersncing c~onot be used to ate velocity of a vehicle and therefore the software uses a Kahn filter to reco~ruot Vibe vehicle's lateral and 30 longitudinal position and velocity states, based oa flea noisy observations of the vehicle position. °1'he vehi~a s~ for each of the loagicudiaal a~ lateral axes comprises position m and speed w of the vehicle, represented as follows;

t~~ 93/19441 ~ PC."'f~~193/QQI1;
~~.~ ~1~ i K=t~~lr (1~
Ire space state for8n, assuming constant va6ocity motion, ~h~e vehicle dynamics arc I-K = d~K ~ (li) 3C ° Chi ( 19) I
where Y is the observable output of the system, bei~~ the vebacle's lateral or longitudinal position, ~ is the state-transition mataix, and C ~ the observation matrix.
For constant velocity motion the asatriees ate as follows:
~ $ 1 °t' ' (20) 0 1~
C m (I ~1 !Zl) where T is the sa~pliag interval, beta' equal to the video field interval which is ms. The Kalman filter e~uativns for one axis are I
IS K m ~PCr(~PC'r + I~"~ i (Z~) $ = dd$ + K(a -1C~ i (x3) r .
P ~ ~P~ + ItyI= - K' (Z~) ?.4 The filter is predictive, and ~ is the predictive valve of the vehicle state for the next sample interval. K is a gala, P is the error co-variance' ~a~ix, sad Ia is a 3 x 3 identity matrix. R, and Its are iapest and output co-variance etes, and a~ee used to adjust the dynamics of the filter.
' i The Kalmaa filter aquatio~ 22. 23 and 24 are c~plex acd time ping to execute in matriu form, and the computer algebra ' was iced to redoes i ~'~ 93119441 ~ ~ ~ ~ ~_ ~ ~ II p~/,4 L'93/pp t 1=

the equations to ss~lar form, as follows, in C notation:
;' compute the filter gain "/ ' data = kp-apl l + 'R2; _ kI = (kp-apil + T ' kp-aplZ) I den k°' = kp-apl2 / den:
I' update the sate vector 'I
xl = kp->xi + T " kp-~~x2 +ki " (°y - kp-axl);
i0 x2 = kp->x2 + k2 ' ('y - kp->xl);
kp-axl = xI;
kP-a~ _ ~;
r /" update the eovar matrix (symmetric so keep only 3 eletaen:t) "/
pll = 'Rl + kp-apl l + 2.0 ' T " kp->p12 + T ' T ' kp->p22 -kl °' kp-apl l - kl ' kp-apl2 " T;
p12 = kp->pI2 + T ° kp->p22 - kl ' kp-apl2;
p2 : ~ °Rl + kp-ap~"2 - I~c2 ~ kp-apl2;
!
kp~->pil = pli;
kp->p1Z = p12; ' kP'>PZ2 = P22 The estimated values for the state of the vehicle ammd error wariaace for the her are calculated using the equations and are stored iu a data a ~kp. Optis~al values for Rt and R2 are detetmi~ed empirically. Figure 2~ ill gasphs which can be plotted from the estimated values for one axis, being the t~l position and estimated speed of the vehicle, aad rhea estimated error ed'wlth the camera alter calculations, as tech vidso ~eid as received. The estimate acq~isitio~ rims is calculated by using the estimated vehicle state data. ~ the potion ~ at which acquisition co ocs~r is , the estimated acqui;zitiott time is calculated ~y taking the difference a H'O 93! 19~! 1 Pt_°t'/ A t.' 93/00 t t ~

bctweea the estimated position and the acquisition position, and dividing the result by the estimated velocity of the vehicle. 'den the estimated acquisition tune falls below a vaiue whip indicates acquisition is to occur within the time of the next vide~
field then the estimated time information is provided to a trigger board g4: -The estimated c~ehicle state coordinate fez the x direction indicates which camera g of a mufti-lane cartaageway is to be triggered.
The scaling matrix C of equation 14 muy be calibtat~ usiap road markers or preferabiy telescopic stakes which are placid at predeter~'ua~d positions along the roadway 16. The stakes are surveyed with respect to the r0aduv~y datum 300 to obtain the x, y sad z coordiaates for di~eteat positions o3o the stapes, sad rhea removed.
Equation 14 can be expanded as follows:
Csix * C~ * C~Z * C~ - CayX'x - C~C'y - Cx,X'Z - Cue' a p (Zgj C= x * C *
mY ~ - Ct~ - C-s~Y'x ° ~~'Y - ~Y'~ - Ca.y' ~ 0 t26) i which relate as image pleas eoordiaate (X'.Y°) to a real world coordinate (x,y,z).
For n observations this can be expressed in math form as folldws:
xt yt zt 1 0 0 0 0 -X'txt -X'tyt -X'~zt ~ x'I
0 0 0 0 xt yt zt 1 -Y'txt -Y'tyt ..lntzt ~ Ctt Y't . . . . . . . . . . , I
~ . . . . . . . a . . ; .
xa ya Zs E6 ~ 0 0 ~ -X''~x~ -X'~s -X'ata ~ ~a 0 0 0 0 x' y~ z' a -~'.~ -Y'~'. -Y'az. ~ Z"s The nations are hem I g e9 ogeaeous and therefore the overall sin of the C matrix is simply chorea so that Ca, ~ 1, sad this parameter is not idettified.
Equation 27 has 11 unknowns sad for a solution reqstires at least 5.5 ob~r,ratiooos, being pairs of (X','~'°) ~ (x~Y~)~ '~ system of equation: is generally ova deter~iaed, and a least square solution is obtained using s singular value d~poiitioa tee~aique. For solution the H'O 93/19441 ~ ~ ~ ~ ~ ~ ~ P~/A~.'93/m(111a --3?~
calibration points x,y.z must not lie in a common plane. The real world coordinates are obtained from the survey results, and the image plane coordiasaces (X'.Y°) are obtained from a display of the detection camera image of the survey stakes using a cursor plotting software package.
To achieve correct triggering of the acquisition camera g, the timing of the systetn needs to cake into account the follawin=:
{i) The system timing: the system must have suffiCimt temporal resolution to facilitate accurate image capture, i.e. the system ~ttst have a sufficient '~ehicle aoquisltion rate, such ,as two vehicles gex second, to avoid omi~ion of vehicles on the roadway.
(ii) Prediction: determining the time of w$ich an image of a vehicle is to ix acquired, and thus initiate image acquisition. ' (iii) Acquisition data flow: timing to perform the physical ineetfaciag between 13 the acquisition camera and the acquisition sub-system 6g res~oasibte for each image capture and storage.
The system timing is resolved at two levels, a coarse leveg,considered to start from periods greater than 24 hours, and a high resolution, one level. xhe coarse level timing 34 is maintained by a real time master clock 354 of the trigger board 84, as shown in Figtue ?9. 'The geometry of the acquisition camera 8 is chorea ~to limit tl:e effects of perspective, limit image blur and take into account other constraints imposed by limitations in the depdt of field available, and for an overpass bridge mounting, the image acquisition point 22 is between 17 and 20 metres fraza the camera 8, and tbs camera is '?5 at as eagle greater rhea 15° and approximately 24° t~ the roadv~ay. A target vehicle traverses the acquisition point 22 within the field of view 20 in approRimately 40 ms, being the acquisition window, at a nominal vehicle speed of 1~0 . 'I~e real time clock 354 provides timing down to 20 ms intervals. Due to the uncertsinties in the position of.tarset vehicle accumulated dutaag the segmentation, ,rlustett°ing and trajectory 30 tasks, one or more timing events duria: t~ a~t~isition window are not sufficient to allow reliable image capture, therefore the high resolution tiaoiag is resolved to hori$ontal video line scan times, being apprr»tamatcly G4,sd.

~'O 93/19~AI
PC°T~ A L' 83/0011 _ 38 -The CPtJ 64, as described above. is able to classify v~hieles during the region analysis and clusterirsg procedures and, in particular, is able to distinguish large vehicles and small vehicles on the basis of the size of a cluster, if theLCPU 64 determines that a cluster represents a vehicle for which an image is to be acquired, the final estimated acduisltion time determined by the trajectory cask is supplied t~ the erir board 86, as shown in Figure 27 via the VIA bras interface 350. The CP1J 64 supplies the estimated time as a 3I bit value, which regeesents the number of the hori~o~atal scan line during the next field, which when reached indicates the acquisition camera 8 is to be triggered. The vI~IE bus can be used to writs the number of tire scatsning lin~ at which uisitioa is ZO ao occur into an acquisition line register 352 of the board X84, For a mufti-la>se carriageway, the CPL 6a also provides data to indicate the cor~ct acquisition cataera 8 to be activated, as determined by the horizontal position data of the vehicle, yn addition eo the acquisition line register 352, and the master clock 354, the tsi~ts board 84 ~includas a comparator 336, and a scanning line counter 38 which also includes a count reg5ster to store the value of the line s~unt. The master clock bas a battery back-up 360 and is synchronised to the horizontal sync of the detection camera 6 so ~ to atxurately keep traaDc of video fields, reset the line counter 38 at the end of each ~aeld and be used as a basis on which timE scamp infot~ation can be generated 9oe allocation to the raw seed parameters processed by the APA b~uard 80. After the mtmber of the acquisition line 30 has been read into the acquisition line register 35:, the line ~ counter 358 counts the horizontal scan iiaas on the basis of pulses provided from the digitiser board 74. T'hs fins coast of the countex 338 dad the number held in the acquisiti~n line register 352 are compared by the ccmoparator dad when the two numbers are tt~e same, abe coaaparator issues as aoquisitioa pulse on the lint: 32 for the acquisition damera 8.
Providing the 35 trigger board 84 to trigger the acquisition camera 8 as also more acataate than relying on software eorataol as the CPU 64 is open co intemcpts and theref ~ose be relied on co accurately control c~ signals of real time events. I
The image acquisition caanera 8 bas been developed to acquired detailed electronic 30 stills or images of vehicle: travelling head-oa to"rards the cato~esa 8 at speeds up t~ if0 km/h. 'tee ~amc rate is at least two pitaures per using a inoa-interlaced mode. Stand~d camera architeccums suffertd limitation of insuf~tdent resolution. image ~'O 93/ 19441 pCfIAL'93/001 I:
~~~~~~a smear and unwanted effects caused by blooming of the image sensor whey vehicle driving lights or sun reflections shone into the camera 8. Hiooming is considered an unacceptable image anomaly, particularly if it appears in the licence plate region of an image which can severely restrict the automatic or even manual reading of the licence piste characters. Another unacceptable image anomaly is image smear, which tended to occur for standard camera architectures in areas of high contrast, which may include the licence plate region as most large ve>ucles have bght~ meted in the vicinity of the licence plate. The effect of image smear tended to iae as sensor exposure decreases, sad for standard camera architectures, image smear was uaptably detrimental at < xstrre times of 1 ms.
The image acquisition camera ~ is a high resolution. t~.illumivated full-a camera architecture having a 1280 x 1024 pixel ~o'M~ moatnatic silicon charge coupled device (CCD) serssor. To prevent smeaaag or bloomang across as image, the IS earners 8 includes a lateral overflow drain sensor architecture which provides 10008 antiblooming characteristics. The architecture provides a site to drain exce$s electrons for each pixel, and Eastman Kodak Co. has developed one suchv:emsor incor~ratiag this architecture. 'This combined with extended infrared aensitiv~ty to 1.I
micrometres, enables near infrared imaging of vehicles and reduces bioomia~ to as accxptable image without degrading the clarity of the vehicle licence plate in the images.
The pixels of the camera 8 arc I6 miarometres equate w~th a 7096 frli factor and have a qwntum efficiency of 0.25 e-lpboton at the image exposure wavelengeb of 800-g00 am. This makes the camera suitable to operation st ;exposure times of 1 ms, d5 which is required to fete the moving vehicle:. The seasoe has low light i~o0aging capability at 1 millisecond exposure time, but is pFacxice the i~a$r:red flash 40 is required to provide SII-in illurrniaatioa as duaiag m~ openti>ag condition extreane lighting ratios wore experienced. This occurs, for example, when shadows iaipiage on the vehicle oar when imaging is performed at eight. ~Synrluonous shuttering of the CCD sensor is achieved with a mechanical shutter, a camera butter made bye Robot, Germany, which is elsetronicslly triggered for I millisaecond. The shutter also pr~vi~s a basis for sYr~~Onaaataon of the olectronic flash 40, ate desexibsd below.

w'O 93/l9aai ~~T/mtr'93/OOt 1~
~~~~~~
i The analogue output from the image sensor is directly cot9verted to digital data by the camera 8, and the digital irrtage data is capable of being outputted in either an 8 bit grsy level format or in a compressed format, using stastda3d ll~IrG image compression.
The flash 40 has a flash head which includes an six-cooed Xenon short-duration (~500 acs) flash rube anounted behind a mirror reflector. The aniaror reflector products a narrow beam width for the illumination of one lane 35. 'The power pack for the flash consists of an air-cooled 100 to 15~ Joule variable outpaat povubr capacitor pack which has a cycle time of two flashes per second. The flash 40 bss aj wavelength range of 695 nm to 1300 nrn. An infrared band pass filter is placed on tb~ front of the flash tube which trangrnits electroma~etic wavelengths primarily outside the human visible range, thereby preventing "flash dazzle" of oncoming drivers ~ pm~ly eiaating delectability of the flash 40, is The wavelettgtb at which the 5lter allows transmission is selected so as to balancx elimination of driver "flash dazzle" and still obtain an .ptabie contrast ran;e for retro- 'reflective licence plates. Licence plates with both the cl~a~cters staid background having rctxo-.reflective properties are relatively ~di~cult to image, and the sefecte~d balance between the CCIy adasor spectral sensitivity, the flat and pass filter and the ~0 lens filter for the camera 8 is illustrated in the graph of Figure 3t9.
,fin expt~ure control circuit is connceted to the Robot shutter and the iris aperture rdec~nist» of the lens of the camera 8. The C;irwit controls the aperture p~ition in accordance with the level of asnbiont light sensed by t~ circuit. T'he circuit provides a k sigh ~ line 36 to control the power and triggering of the infrared flash ~0. As tie a~uisition camera 8 'S aperture closes with imxeased ambient ildumiaarioa, the flasf t power is increased to maintain an optimum balatnae between ambient light and flash "fill-in"
illumination. T'he circuit also includes a delay element io maintain the average gash power dosing large transient fluetuation of light received that can be why white tsucks pass of sunlight is directly reflected faom vehicle windscreens onto the carntra 8.
The circuit is 30 based ort standard expo:ore control circasits, and, in ~dicio~ to the delay element, includes an infraaed sensor to measum the ambient light. The ~~ power is opntrolled by adjusting the capacitanot= of the power pack for the flash 40.

evp 93/d944r d'C'T/AL'93/ppd d:
~1~~~~~ ' _41_ 'I~a infrared flash 40 is mountctt at an angle of 9.5° With respect to the optical axis of the acquisition camera 8, ancJ at an angle of grCater Chars 15' to the roadway I6.
The field of view 39 of the flash 40 is sitniIar to the field of vaew ?0 of the acquisition camera 8. The geometry of the flash 40 is important so as to reduce any retro-reflective effect from the exposed vehicle, in particular its licence plato. The rctro-reflective progenies of the paint used on licence plates is such that the maxinsum reflected light is back along the axis of the flash illuminating beans. The angle of illumination and the illumination energy is seieaed to rafts into at:count the diverse Mange of retro-reflective and non-retroreflective paint colours and formulations used on Licence plates.
Examples IO of the images which can be obtained by the acquisition ca~era 8 of the vehicle monitoring system are illustrated in Figures 31 and 32. ' The acquisition camera 8 is connected to the deteetie~f sub-system 66 aid acquisition sub-system 58 by an interface board 359, as sho own in Figure 33.
'T>ae I3 interface board 359 provides power to the camera 8, can issud data iatertupts for the ptncessor 360 of the camera 8, and, is connected to an image buffer 361 and trigger interfaces 363 of the r.~era 8 by optical isolators 365. The iateifaoe board 359 provides communications to the control unit 10 via differential RS422 communications interfaces 367 which are coaneaed by communications cables 369. The tr~gg~r signal is provided 2t) from the trigger board 84 to the trigger interfacx 363 of the sa~deta 8 by the RS4.°'.'-interconneet. Image data produced by the CCD sensor 371 i~ available in the image bu~'er 361 of the can;rera 8 approximately 300 ms after the r~me~a 8 receives the trigger signal. Ac that time a dace interrupt signal is , sent fiom the cx>ntroi unit 10 to request traasfsr of the image data from the camera 8. ?be image data' is read from the image ?5 buyer 361 as 16 bit vvords at a rate of 1 I~Iword/s, where each word represents two 8 bit pixel values. A strobe clock edge is also included in each ~16 bit word for tirauag purposes. The 16 bit data stream is coeverted to 8 bit data st ~ logic levels by the CPU 64, and the async)uonous imaas data is rhea ,~ by a frame grabber 86 of the acquisition sub-system 68, which is a Datacube A~axsauu~ board. 'T'he image data 30 is then clocked into as acquisition image buffer board 88 where et is held until transferred by a bus d~epeater 89 to the image buffer and eommunfrations e~natroller 57 or a lioeaa~s plate recognition system 51, as shovrn in Figure 6.
i w0 93!19441 PCT/A,~.'93/0~1 t' The images captured by the acquisition carnets 8 possess the following eharactetistics:
(i) A full lane width of 3.~ metres is imaged.
(ii) The pixel tesoiutiotts of each licence piste character. for character sets of 40 x 80 mm, were at least 10 x 20 pixels for W, Z and 9, and a nqiaimum of four pixels for a eltaracter stroke, such as the letters I, L etc. Pixel resole;io8s of up to 15 x 30 were achieved on characters for a full lane field of view 20.
i (ill) The average grey level of a character sttolee;is at least .'.0 grey levels higher thaw the background of the grey level of Bhe licence pate.
(iv) Both the ficxnce plate region and the vehicle; body work are imaged adequately to enable identification and verification of vehicle ~ype.
(v) The quality of the licencx plate image is nlati~ely constane throughout a 24 hour period for all vehicle aad lidettct plate types.
13 The image buffer and communications controi57 include a Silicon Caraphics Personal IRIS 4DI355 msshane as a buffer box 381 for handling intermediate storage of images on disk 383, a CISC4 Internet Psotoso! (IP) roster 385 and a Summit Technologies 52000 ISDId bandwidth manager 387, as shown in iFigure 34. The remaining description relates to image transfer between the repel ter 89 and the buffer box 30 381, but the description also applies to amaac transfer between the repeater 89 and a licence plate recog~ution system 51 located at the node 2, as shown in Figure 6.
The data tsruosfer by the bus repeater.89 to the buffer; box 38i is made by a
3 M~/s digital line. The repeater 89, which is a ~tT3 Model 413!x-bus repe:..er, with '$ D1~1A capability, ettabies the buffer box 381 to copy data due~tly from the buffer 88 in the acquisition subsystem 68. To coordinate i~ga transfer between the buffer box 381 and the system 68, an image header scruauro is established fob storage of the images in the buffer 88, sad messages are allowed to be pasted back and forth between the buffer box 381 sad the system 68 via incenqtpts in a mail box location; 'The memory layout for 30 the image buffer 88 is shown in Figure 35 and the highca armory locations are used to store acquired images in bu~ea segments 370 with a ~72 for each image buffer being stored in the lower mesaory locations, The ' header 372 includes date of I

w093/19441. I PCT/AL'93/pOtt>
21j~J15 image acquisition, a base address far the image in the buffer 88, a busy flag to indicate whether the image is presently being read, and information on the size of the image. A
memory header 374 at the lowest location in the buffer is shared'with the buffer box 381.
and inciudes the following 5elds:
1. ha-hostintr: used by the buffer box 381 to specify which type of intenvpts they are sending.
I
2. ha-imagenum: used to tell the buffer box 381 which image to read after an image available irnermpt is sent.
i 3. ha-numbufs: the number of image buffers allocated in ithe buffer 88.
4. ha-height, ha-width: the orgaafsation of the image within the buffer 88.
is S. ha-bufsize: the siu of the buffer, which is s multiple of ~6 bytes.
i i The architecture of tht software modules usod by the buffer box and the acquisition image proceaaittg system 42 is illustrated in Figura l3fi. '1"he buffer box runs '_0 a capture module 401 which is responsible for communication ,between the acquisition sub-system 68 vla the BIT3 interface board 4$9. Ti:e modtt~e poles the acquisition sub-ayacem 68 for images, stores them in a memory buffer, aid rhea stores them in a dimctory GprureQ as a 51e with a unique name. 'The came is trade up of the 5rst 5ve cbaraaers of the name of the buffer brnc 381 eu~d a ten digit number. A CRC
essor :5 chec~ng value is generated and image data is stored in a header of the 51e, including the name of the remote site or nook 2, the tiara the image was capet~ied by the camera 8, the image header length, the CRC value and the image width and height. The CaptureQ is able to score S00 images, the data of which each occupy approximately 1.5 Mbytes. 1~
the CaptuseQ overflows, the images are discarded, and the file names of the lost image 30 is recorded in an error fife together with the time the images ueidiaasded.
Overflow of the CaptureQ may ocau if the acquisition sub-system 68 acquits imps at a high rate for a long perm of time, or the lick thsvugh the ISDN 45 to the,' ventral server 47 is cue ~O 93/19441 PCT~AI'93/Opl 1?
of servict for as cxcended period of cimc. The communications link to the eentsai server 47 fzom each remote site : is prov ided by the scoter 385 connected to the buffer box 381.
and a ~G? 1 protoeol little 389 between the muter 385 afld the bandwidth manager 387.
which prov ides a Macrolink'" 391 to the ISDN 45. The central server 47 in turn is also connected to the ISDN 45 for each remote site 2 by a CISCO IP scoter 385, a Summit Technologies S2000 bandwidth manager 387, an X21 link 389 between the manager and the muter 385, aad a Macroliak 391 to the iSDN 45. The bandwidth manager and the muter 385 form the communications controller 42 of the acquisition image proceaain~ system 42. The X21 links 389 are scaadard H-ISI~N communications Iink governed by CCITT standards. The Macroliaks 391 are Primary Rate Access Iitzlcs provided by the second applicant aad are based on the CCITT s for Primary Rate Access in the B-ISDN. The X21 Iirrks operate at 768 KRJs ~d the Macroliaks provide two virtual links operaciag at 384 KHJs. The bandwidth manage 381 is esaetatially a multiplexes which uses a data aggregation protocol sad provide aocxss to the ISDN 45.
t5 The remote sites 2 are each represented at the central server 47, which is a Silicoa i Gtaphia Crimson machine, by a retrieve module 403, as shown in Figure 36, which makes a socket connection to the respective remote site 2 and polls for an image from the remote site 2. The FTP protocol, which is a Uaix fate traa~fer protocol, is used to retrieve images, including their associated dsts, from the xemotc;site 2 and when received the ima~,e is checked for integrity of the imago data on the basis of the CRC
value, aad stored oa a Ret:;avaliQ directory 405 of the carver 47 which has a c~aeity of 720 images.
'Ihe images arc atost~d on the RetrievalQ 405 with the time w~ea the image was first r~equeued and the rites when the image was finsliy received. Aa SDistributor module 35 40'1 is responsible for distributing the image 81e names to store modules 409 of the central server 47. The store modules 409 retrieve images fmmi the R~etrievalQ
405 sad archive them in respective imags store: 4i1 which have the ~pacity to store images acquired over a week from each site 2. 'Ihe image store 4i1 ate Exabyte IOI
tape storage systems which can each hold up to ten tapes that each hive a capacity set at 3000 images. The store module 409 cotamunicates with a tape drivtf for each fore 411 which based on a tape driver developed by Crene Daonsk of Vulcan 'haboratories; U.S.
The driver controls laadiag sad ualoadiag of a tape from a stare 4~ 1 by a robot arm. The W093/19441 ~ ~ r9 ~ ~ ~ ~ ~Pt=1'lAl.'93/0411~
I

driver on initialisation dctcrmincs the number of tapes in the store 411, and for a Bold start formats each tape and loads the 5rst tape, For a warm start the drivel simply xlects the tape last used. When a tape reaches its 3000 image capacity it is returned to its storage bay and the next tape is selected.
The SDistributor module 407 has a list of the names of the Sles in the RetrievalQ
405 called store list and another list of the names of files which are in the process of being stored on an istage store 411 rxlled InlProgress list. When a store module 409 requests a file name, the SDistributot module 407 returns a file name from the store list and moves that same to the InProgress list. If a f1e names is not available, the module 407 a«xsses names fmm the RetrievalQ, adds them to the store Kist and then returns file aaaus. When the module 407 receives a storage ackaowledgmevt from the store module 409, rhea the file name is removed from the IaProg:esa list. T'ht s:are module 409 poles the SDistributor module 407 for a 5ie name, sad on recxiviag the file name retrieves the corresponding 51e from the RetrievalQ and copies it onto the imaiSe store. The same files is also copied onto a directory of the server 47, IPQ 413 whi~h can hold 750 ieaages.
It IPQ 413 is thll, the file is discarded and the hander of the 51e is copied onto a further directory DatabaseQ. Act acknowledgment message is rhea seat to the SD~stributor module 407. A dare staaop is plsnd on all files indicating whoa the file is archived.
'0 An IPDfatributor module 417 distributes images to a licence plate rccogoitinn aystam 51 connected to the Ethernef LAN 419 of the ocntral ter~er 47. The module 417 maintain: a ties of 51e names, called Image list, which its the 5les paid is IPQ
413. Whoa the liceacx plate ceoogaition system S1 poles for a file name, the module 417 ?5 returns a file name from image liar sad moves that Sle acme td another list, IP~'rogzess list. When the sy:tem 51 acknowledges that it bas received the oocreaponding 5ie, then the file name is deleted from the IPI?rogzess list, together with the file froze IPQ 413.
If file namts ace not available is Image list, the names are obtrihed from the IPQ by the module 417, sad added to the liar. The module 4I7 rommuni~ates with the system 30 via a socket coattaxion. Licentx plate dataits cxRraaed by the >iecogaitioa system 51 art stored on a DatabaseQ 415 of the aervcr 47 together with othsr image data details, such as image acquisition time, and ins:aaaaaeous speed of vehicle which have already been I .

W093119~4t ' 'y ~' PC"t/A14310011~
~1~~5~
provided with the image from the remote sites 3. A database module 419 poles foe files placed on DatabaseQ 415, and then stores the fibs on an imag i database 421.
The licence plats recognition system 51 has been implemented using a Silicon ~ Graphics workstation 400 which is connected to the LAN 419, as shown in Figure 37.
but can also be connected directly to the repeater board 89 at a remote site 3. A Pixar II
image computer 402 is connected to the workstation 400 and acts as an image co-processor. The system 51 also includes a monitor 404, keyboud 406, disc storage of 600 MB 408 and optfcal disc storage of 1.2 GB 410 oonnecte~ to the worksution 400.
The workstation 4p0 usrs, inter alia, VIVID (Vehicle Identification by Video Image Detection) software owned by the State of Victoeia~ which is abio~to locate a m;mberplate its a vehicle image, and then perform optical character recogtution (OCR) on the located numberplate to extract the liaace plate cha:acters. 'Ihs parameter settings of the VIVID
software have been adjusted to handle the images provided by theiscquisitioa sub-system 68, according to the size and contrast of the images. To accept tie images at a peak rate of ? per second, a real time image handling procedure 412, as shown in Figure 38 is used. The procedure begin at step 414 by requesting as image 51e name from the IP
distributor model 417. If a name is not received at step 416, the IP
distributor module 417 is polled again, ot6ervvi:e the received namo is used to sooe~ the If'Q
414 and store the image 51e o,n the disk 408, at step 418.
Images aro saxsaed from the disk 408 and procr~scd b~y four separate software modules of the workstation 400, a locate plate ~ moduie 420, a glyph extraction module 422, and OCR motiule 424 ttnd s plate recognition module 426, ac shown is Figure 39.
23 The iocate plate module 420, as shown in Figure 40, begirt at step 430 by preparing the 1280 x 1024 pixel image for processing as a number of pusl windows for the Pixar co-processor 402. At step 432, the system 51 attempts to dote~t an edge of a ehafacter sire object, and when deceetcd the object's location is determined at step 434. An object assembler is used at stop 436 to group adjsoent objects together, and the groups ate processed by a plate clauifier 438 to determine whethca the objecx groups could constitute a licence plate. If an object group is classed as a ølate according to a plate template, a bounding box is formed, sad its coordinates to the gtypb extraction ~~~~~1~
WO 93/ 19441 pCT/ A L:93/001 !
_4'7-module 423. The glyph extraction rnoG~:.e 4?3 processes each bounding box to binarise and extract individual characters in a pounding box and then pass the "glyphs", i.e.
licer:cc plate letters and numbers, to the OCR module 424. 'Fhe OCR module 424, as shown in Figure 41 begins at step 4ø8 by building a typological graphical representation of a glyph from the glyph bitmap provided by the glyph extraction module 42'.', for each glyph. The graphical representation is analysed at step 440 so as to detect any characteristic features, such as holes, arcs and vertical and hotizontai lines. From the results of step 440 an 81 bit string representing the charasxeristic features of the glyph is created at step 442. A bayesian statistical analysis is then performed at step 444 on the feature string to try and match the features against a set of prcviouely determined features characteristic of known ASCII characias. The ASCII value of the snatch with the highest probably of being correrx is retutaad to the plate recognition module 426.
The plate recognition module 426 detesmiaes whethest the glyphs in a bounding box constitute a valid lictace plate. The module 42b effe~ively coatmls the other image processing modules as it has the ability to override the results bf the OCR
module 424 of to force the glyph extraction module 422 to use a bounding box othex thaw that found by the locate module 420. The majority of vehicle !leaner plates in Australia have six characters and fall into one of two classes, Federal plates or non=Federal plates. Federal ..0 plates comprise two alphabetic characters, two digits and two alphabetic cha:aacra.
whereas non-Feder<t place comprise three alphabetic cbaraccoa aad are followed by three digits. The plate recognition module 426 is able to determine whether a valid licencx plate has been found oa the basis of this information, aad.~other inf0n~atioa, such as the spacing of characters and the specific charaetcristic alphanumeric sequetxxs used .'.5 by the non-Federal plates. The OCR module, for example, may not be able to distinguish between capital B and 8, sad fps many plate foots; there: is no difference between a 0 sad O or a 1 sad as i. Therefore: the plate caoog~itiob module 426 tray need to override the results obtained by the OCdt raodula 424 The plate re~gnitiota module 426 is also able to ithe glyph extraction module 424 to procGa6 an altered 30 bounding box if the module 426 de;terrmines that there tray be ~ additional glyph to the left or right of an original bounding box returned by the locate module 420.
The: licence plate details obtained by the plate reso~ition metduie 426 art; shed oa l~atabaaeQ 415 ~'~'O 93/ 19441 PCT/ A ir'93/o0 i i i _48~
of the server 47, arad archived on the optical disk 410. The opticisl disk 410 alSO archives image files which the system 51 is nnablc to process when received.
The database on the optical disc 410 stores for each processed image as does DatabaseQ 415, data concerning the position, size and characters of the aumberplate located in the image, and other details such as time and date df aequisitiota.
It is also structured with data pointers whieh facilitate access to the stored data by the workstation 400. The workstation 400 iz:cludes graphical ttxr interface software which enables as operator to review the results of the procedures 412 and 414, anti perform further optical character recognition or mmbe~late regions. as xlected. Any further dCR
proc~3iingg performed on a plate region xlecced by the operator of the woritatatiota 400 is normally used to aaalyse the performaece Of the proctrdures 412 and 4I4 abd not to alter cht integrity of the data held im the optical disc 410.
The image data stared on database 421 is procaud by aiatchiag software which looks for matches amongst the Licence plate details fields of the image data so as to locate ocxurrences of detection of the sasae iiatxe plate at different remote sites or nodes 2. Once a match has been located, the acquisition time field: era be used to determine whether speed or time violatiom~ have oaurred in travel betvrie,~n trmdte sites 2, as distance between the; sites 2 is known. The m~atchiaig aofflvare is run 4a a Sun Miarosyatems worknstion 450 cotusecced to the 1 AN 4I9, or aloernstively, the matching software is run on a systr~m of a road traffic authority, with the image data being sent by the cemral carver 47 over the iSDN 45 to the road traffic auttiority. The road traff'ae authotfty is able to c~~mmttaicate with the central server 47 via the ISDN 45 to obtain :S archived images, s~ required. ' To avoid sending all images to the central server 47, a lame cumber of which may not be of interest, images can be archived at the nods 2, aid licetaoe plate details extracted at the remote nodes 2 by respective licence plate seca~itioa systems conaeaed directly co the BITS npsatrr 89 of a ttode'a aoquisitioa stab-sy~em 6B.
The server 47 then only recsives the ext:saed licence plotte details, and other data on the image, such as acquisition time, the remote site, and insnmta:sestss spetd, sad not the WO 93/39441 ~ ~ j ro ~ ~ ~ p~,"i'/e~L'93/pOtle image itself. Images archived at the remote sites 2 can be retrieved by the central stwct 47 when required.
Control of the remote nodes ? is performed by the remote site user interface which tuns on the Sua workstation 450 connected to the LAN 419 of the central server 47. The interface 53 includes a user tool which communicates vvitb a super task of each eemote site 2 using a Sun Miaosystetns Remote Procedure CaII (RPC'~
communications protocol. The super cask provides a set of paocedural functions wlrieh caa be called by the user tool using the RPC protocol, regardless of the locations of the workstation 450.
The RPC protocol handlca data type conversions and atignmoat. 'ihe procedures provided by the super task perform various actions which together allow ramplete control of the software of a node 2. For exempla, a parameter Tilt maintains a fiat of all vatiablas u;s~
by the sof:'vare of the node= 2, to;cthrer with their initial values. ~ The form of the values indicates the variable type, which may be a decimal integer, ~ bexadaimal integer, a floating point value, a character string or booleaa value. The vsr$ables can be aiteted by adjustins tae parameter file, sad location of the variables Ustod ~ in the parameter file is done vis a VxWorka system: table which contains adl global symbols. The user tool, in addition to changing system parameters, can accxss the super task to obtain status and configuratiota information on each node 2.
The super task accepu ItPC transaction via both tba~ Transnaissios Control Protocol (TCP) and the User Datagram Protocol (UDP), both of which use the Interest protocol (IP) for otansmission of data~raa~ betvvcan composer systems. UDP fs eaaneaioale:s proton>l which prlnuuily involve: multiplexing of datagtams, wheres~s TCP
:5 is a connection orientated protocol which seeks to rnsure dats~integsity is maintained.
The user tool pnsandy uses TCP/iP which, together with the RPC protocol, is provided with Sun Miaosystam's SunOs oreration system sad the VxWorkt rea! time operating system. To proterx against different central stations accessing a remote node sad making conil9cting changes to system parameters, tho user tool ptoviber infoamation on the current state of the node software before any alteration can be .
The master clocks 354 of the remote sites 2 era synchto>aiaed to the clock of the Vo~~ 93/19a4i pCT/A1,93~0~11~
_so_ central ser~~ez 47, and the systems S 1 and 450 connected to the L~tl~1491 using a network time protocol (wv°l'P), which is a staazdard UNIX utility norsnally used to synchronise the clocks of stations on a LAN. Tht: N'IT polls the remote sites 3 and on the basis of information received from the sites 2 eoncstning r~uaunisatidns between the sites and the server 47, the hTP applies offsets to the remote sites 2 so as to synchronise the sites 2 and accourn for network propagation delays, including transient network problems such as 6ink congestion.
The vehicle monitoring system is particularly advantageous as it is able to detect and discrimiztate moving vehicles from other objects, and enquire an image of selected vehicles from which they can be identified, using only elactroaie cbmeras and processing circuitry and software housed at a relents sits 2. 1"ha system eatbles automatic e~ttraction of licence plate details and does not serlttin road based equipment or markings, the emission of electromagnetic sigaala or the aeplaament of filet at the node 2.
The system is able to simultaneously track a number of vehicles on mufti-lane carriageways sad classify them by vehicle type. A high rasoludon image of a vehicle teat be obtained over a full traffic lane, the resolution sad clarity of the invention being sufficient to enable eatrxtion of the licence plate detail:. "L'l:re system can operate ~0 continuously in all ccmditions where visibility is greater than 100 metres.
using in»ared imaging techniyue~t. The high maolucion camera iaco:porata~ antibloomiag txbnOlogy to prevent pixel saturation due to vehicle haadJights, and the in~ted flash used is oonflgured so as to t;~e substantially undetectabia and inhibit flash dale.
i YS The system Gas also be controlled and initialised frond a remote ctataal station.
with images and dew being traastnitted ova a digital oommuaicatioats network.

I
The system can further be usod for a somber of puepioets. such as moaitoritng tailgating offences, road toll collertioa, sad transit lane monitoring. It ran also be 30 adapted for red light intersation monitoring.

WO 93/19441 ~ ~ ~ ~' ~ ~ ~ PC'T/At'9310~1I~
'Phe system caa also be adapted to monitor aaad acquire images of other moving objects, such as the movement of shipping containers within transport depots, and the mov emeat of objects oa as assembly tine, s

Claims (141)

CLAIMS:
1.~An object monitoring system comprising camera means characterised in that the camera means is adapted to monitor movement of an object to predetermine, based on the monitored movement of the object an acquisition time at which an image can be acquired at a predetermined position of said object relative to said camera means, and to acquire an image at the predetermined acquisition time and the predetermined position.
2. ~An object monitoring system as claimed in claim 1. wherein said camera means is adapted to detect said object and discriminate said object from static objects and other moving objects.
3. ~An object monitoring system as claimed in claim 2, wherein the camera means includes video camera means for monitoring a respective area in which objects move, and image processing means for subtracting a background image of said area from images of said area generated by said video camera means so as to produce difference images representative of moving objects in said area.
4. ~An object monitoring system as claimed in claim 3, wherein said image processing means includes classification means for forming clusters from parts of said difference images which correspond to the same moving object.
5. ~An object monitoring system as claimed in claim 4, wherein said image processing means processes each cluster to determine if it corresponds to said object and determines said acquisition time.
6. ~An object monitoring system as claimed in claim 5, wherein said image processing means filters said difference images to disregard pixels within a predetermined level range.
7. ~An object monitoring system as claimed in claim 6, wherein said image processing means includes segmentation means for processing said difference images to generate segmented images which include regions corresponding to parts of moving objects in said area that produce at least a predetermined light level at said camera means.
8. ~An object monitoring system is claimed in claim 7, wherein said classification means analyses and generates measurements of the shape of said regions and on the basis of the analysis and measurements determines valid regions and invalid regions to be disregarded.
9. ~An object monitoring system as claimed in claim 8, wherein said classification means includes clustering means for generating said clusters, each cluster comprising the valid regions which are considered to correspond to an object, said regions being extended to determine if regions overlap with and have to be combined with another to form a cluster.
10. An object monitoring system as claimed in claim 9, wherein said classification means includes labelling means for assigning a label to each cluster for each image to identify respective clusters and for matching and separating clusters over consecutive images to determine if labels are to be inherited or new labels assigned.
11. An object monitoring system as claimed in claim 10, wherein said classification means is adapted to classify said clusters as corresponding to predetermined objects by comparing characteristics of said clusters with classification data of said system, such that said classification means is adapted to identify a cluster corresponding to said object.
12. ~An object monitoring system as claimed in claim 11, including means for maintaining a histogram of said characteristics for objects monitored by said camera means and adjusting said classification data on the basis of said histogram.
13. ~An object monitoring system as claimed in claim 12, including light intensity means for monitoring a lighting level of said area, and wherein said predetermined level range, said analysis of said regions, the extension applied to said regions by said clustering means, and said classification data are adjusted depending on said lighting level.
14. ~An object monitoring system as claimed in claim 13, wherein said image processing means includes means for tracking the cluster corresponding to said object over consecutive images, comprising transformation means for transforming coordinates of said cluster to compensate for a perspective view of said camera means, and means for predicting the speed and position of said cluster for each succeeding image.
15. ~An object monitoring system as claimed in claim 14, wherein said tracking means determines said acquisition time an the basis of said predetermined position and the predicted speed and position of said cluster.
16. ~An object monitoring system as claimed in claim 15, wherein said camera-means includes image capture camera means to acquire said image of said object at said acquisition time.
17. ~An object monitoring system as claimed in claim 16, wherein the image capture: camera means is adapted to acquire a high resolution image of said object.
18. ~An object monitoring system as claimed in claim 17, wherein said video camera means has a wide field view relative to said image capture camera means, which has a limited field of view.
19. ~An object monitoring system as claimed in claim 18, including an infrared flash which is synchronised with said image capture camera means, the energy level of said flash being dependent on said lighting level.
20. ~An object monitoring system as claimed in claim 19, wherein said image capture camera means includes image sensor means and exposure control means for inhibiting saturation of said image sensor means in response to said lighting level.
21. ~An object monitoring system as claimed in claim 20, wherein said flash includes means for inhibiting the emission of visible light therefrom.
22. ~An object monitoring system as claimed in claim 21, wherein said extension applied by said clustering means is increased when said lighting level corresponds to a night condition.
23. ~An object monitoring system as claimed in claim 22, wherein said extension is less for regions corresponding to objects distant from said camera means.
24. ~An object monitoring system as claimed in claim 23, wherein said labelling means performs said matching and separating on the basis of comparing boundaries or centres of said clusters for consecutive images.
25. ~An object monitoring system as claimed in claim 24, wherein said difference images are filtered and used to update said background image.
26. ~An object monitoring system as claimed in claim 25, including means for triggering said image capture camera means at said acquisition time, comprising means for receiving and storing a number of a scan line corresponding to said acquisition time from said tracking means, means for counting scan lines of said images, and means for generating a trigger signal for said image capture camera means when said count reaches said number.
27. ~An abject monitoring system as claimed in claim 26, wherein said light intensity means generates a histogram of pixel grey levels for said images generated by said video camera means, and determines a day, night or twilight light condition on the basis of the median of said histogram.
28. ~An object monitoring system as claimed in claim 27, wherein said predetermined level range is determined on the basis of the minimum, median and peak of said histogram.
29. ~An object monitoring system as claimed in claim 28, wherein said measurements comprise circularity and coverage of said regions.
30. ~An object monitoring system as claimed in any one of claims 1 to 29, including recognition means for processing the acquired image to obtain information identifying said object.
31. ~An object monitoring system as claimed in claim 30, including a plurality of said camera means for monitoring respective areas and adapted to communicate with one another so as to transfer information on said object.
32. ~An object monitoring system as claimed in claim 30 or 31, including a plurality of said camera means for monitoring respective areas, and adapted to communicate with a central station so as to transfer information on said object.
33. ~An object monitoring system as claimed in claim 32, wherein said information on said object is acquired by at least two of said camera means and said information can be used to determine the time which said object took to travel between at least two of said areas.
34. ~An object monitoring system as claimed in claim 33, wherein said central station includes remote control means for controlling said camera means from said central station.
35. ~An object monitoring system as claimed in claim 34, wherein said central station and said plurality of camera means include respective telecommunications controllers and communicate using a digital telecommunications network.
36. ~An object monitoring system as claimed in claim 35, including means for archiving said information and allowing subsequent access thereto.
37. ~An object monitoring system as claimed in claim 36, wherein said information includes acquired images of said object and the times of acquisition, and said central station includes said recognition means.
38. ~An object monitoring system as claimed in claim 36, wherein said information includes said identifying information and the times of acquisition of acquired images of said object, and a plurality of said recognition means are connected to said plurality of camera means, respectively, at the sites of said plurality of camera means.
39. ~An object monitoring system as claimed in claim 37 or 38, wherein said recognition means is adapted to process said acquired image to locate pixels representative of characteristic pixels identifying said object.
40. ~An object monitoring system as claimed in any one of claims 1 to 39, wherein said object is a vehicle.
41. ~An object monitoring system as claimed in claim 40 when dependent on claim 30, wherein said recognition means comprises means for locating a licence plate in said image and means for determining the characters of said licence plate, said characters comprising said identifying information.
42. ~An object monitoring system as claimed in claim 41 when dependent on claim 6 wherein said predetermined level range covers pixel levels produced by shadows of said vehicle.
43. ~An object monitoring system as claimed in claim 42 when dependent on claim 8, wherein said invalid regions correspond to headlight reflections produced by said vehicles or road lane markings within said area.
44. An object monitoring system as claimed in any one of claims 40 to 43, wherein said vehicle is a large vehicle, such as a bus or truck.
45. A monitoring system as claimed in any of claims 1 to 44, wherein said object is a vehicle and wherein said camera means is adapted to monitor said vehicle to detect a law infringement acid captures an image of said vehicle at said predetermined time in response to detecting said infringement.
46. A monitoring system as claimed in claim 45, including recognition means for processing said images to obtain information identifying said vehicle.
47. A monitoring system according to claim 1, wherein the camera means generates images of an area and acquires an image of a predetermined object, including image processing means having:
means for subtracting a background image of said area from said images of said area to generate difference images representative of moving objects in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area;
classification means for processing said region images, said classification means including means for analysing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for generating, on the basis of said valid regions, clusters corresponding to respective ones of said moving objects, and means for classifying said clusters. by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said predetermined object;
and tracking means for tracking said one of said clusters corresponding to said predetermined object to determine an image acquisition time for acquiring said image of said predetermined object.
48. ~An object monitoring system as claimed in claim 47, wherein said image processing means filters said difference images to disregard pixels within a predetermined intensity range.
49. ~An object monitoring system as claimed in claim 48, wherein said parts of moving objects correspond to at least a predetermined light level received at said camera means.
50. ~An object monitoring system as claimed in claim 49, wherein said classification means extends said valid regions to determine if said regions have to be combined to form said clusters.
51. ~An object monitoring system comprising:
camera means for generating images of an area and for acquiring an image of a predetermined object, and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of moving objects in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area;

classification means for processing and classifying said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving objects, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said predetermined object; and tracking means for tracking said one of'said clusters corresponding to said predetermined object to trigger said camera means to acquire said image of said predetermined object.
52. ~An object monitoring system as claimed in claim 51, wherein said image processing means filters said difference images to disregard pixels within a predetermined intensity range.
53. ~An object monitoring system as claimed in claim 52, wherein said parts of moving objects correspond to at least a predetermined light level received at said camera means.
54. ~An object monitoring system as claimed in claim 53, wherein said classification means extends said valid regions to determine if said valid regions have to be combined to form said clusters.
55. ~An object monitoring system as claimed in claim 54, wherein said classification means includes labeling means for assigning labels to clusters, respectively, for each one of said images of said area to identify said clusters, and for matching clusters over consecutive ones of said images of said area to determine if labels are to be inherited or new labels assigned.
56. ~~An object monitoring system as claimed in claim 55, including means for maintaining a histogram of said at least one characteristic of said clusters, and adjusting said classification data on the basis of said histogram.
57. ~~An object monitoring system as claimed in claim 56, including light intensity means for determining a lighting level of said area, and wherein said predetermined intensity range, said analysis of said regions, extension applied to said valid regions by said classification means, and said classification data are adjusted depending on said lighting level.
58. ~~An object monitoring system as claimed in claim 57, wherein said tracking means includes transformation means for transforming coordinates of said one of said clusters to compensate for a perspective view of said camera means, and means for predicting the speed and position of said one of said clusters for each succeeding image of said images of said area.
59. ~~An object monitoring system as claimed in claim 57, wherein said tracking means determines an image acquisition time on the basis of an image capture position and the position of said one of said clusters.
60. ~~An object monitoring system as claimed in claim 59, wherein said camera means includes video camera means for monitoring said moving objects and image capture camera means to acquire said image of said predetermined object at said acquisition time, said image being a high resolution image of said predetermined object.
61. ~~An object monitoring system as claimed in claim 60, wherein said video camera means has a wide field view relative to said image capture camera means, which has a limited field of view.
62. ~~An object monitoring system as claimed in claim 61, including an infrared flash which is synchronized with said image capture camera means, the energy level of said flash being dependent on said lighting level.
63. ~~An object monitoring system as claimed in claim 62, wherein said image capture camera means includes image sensor means and exposure control means for inhibiting saturation of said image sensor means in response to said lighting level.
64. ~~An object monitoring system as claimed in claim 63, wherein said flash includes means for attenuating the emission of visible light therefrom.
65. ~~An object monitoring system as claimed in claim 64, wherein said extension applied by said valid regions is increased when said lighting level corresponds to a night condition.
66. ~~An object monitoring system as claimed in claim 65, wherein said extersion is less for valid regions corresponding to said moving objects distant from said camera means.
67. ~~An object monitoring system as claimed in claim 66, wherein said labeling means performs said matching on the basis of comparing boundaries of said clusters for said consecutive ones of said images of said area.
68. ~~An object monitoring system as claimed in claim 67, wherein said difference images are filtered and used to update said background image.
69. ~~An object monitoring system as claimed in claim 68, including means for triggering said image capture camera means at said acquisition time, comprising means for receiving and storing a number of scan lines corresponding to said acquisition time from said tracking means, means for counting scan lines of said images of said area, and means for generating a trigger signal for said image capture camera means when said count reaches said number.
70. ~~An object monitoring system as claimed in claim 68, wherein said light intensity means generates a histogram of pixel grey levels for said images generated by said video camera means, and determines a day, night or twilight light condition on the basis of the median of said histogram.
71. ~~An object monitoring system as claimed in claim 70, wherein said predetermined. intensity range is determined on the basis of the minimum, median and peak of said histogram.
72. ~~An object monitoring system as claimed in claim 71, wherein said analysis includes determining circularity and coverage of said valid and invalid regions.
73. ~~An object monitoring system as claimed in claim 66, wherein said labeling means performs said matching on the basis of comparing centres of said clusters for said consecutive: one of said images of said area.
74. ~~An object monitoring system as claimed in claim 57, wherein said camera means includes exposure control means for adjusting camera exposure on the basis of said lighting level.
75. ~~An object monitoring system as claimed in claim 51, including recognition means for processing said image of said predetermined object to obtain information identifying said predetermined object.
76. ~~An object monitoring system as claimed in claim 7.5, including a plurality of said camera means and image processing means for monitoring a plurality of areas, respectively, and being adapted to communicate with one another so as to transfer information on said predetermined object, said areas being remote with respect to one another.
77. ~~An object monitoring system as claimed in claim 75, including a plurality of said camera means and image processing means for monitoring a plurality of areas, respectively, and being adapted to communicate with a central station so as to transfer information on said predetermined object, said areas being remote with respect to one another.
78. ~~An object monitoring system as claimed in claim 77, wherein said information on said predetermined abject is acquired by at least two of said camera means and image processing means and said information can be used to determine the time which said predetermined object took to travel between at least two of said areas.
79. ~~An object monitoring system as claimed in claim 78, wherein said central station includes remote control means for controlling said plurality of said camera means and image processing means from said central station.
80. ~An abject monitoring system as claimed in claim 79, wherein said central station and said plurality of camera means and image processing means include respective telecommunications controllers and communicate using a digital telecommunications network.
81. ~An object monitoring system as claimed in claim 80, including means for archiving said information and allowing subsequent access thereto.
82. ~An object monitoring system as claimed in claim 81, wherein said information includes acquired images of said predetermined object and the times of acquisition, and said central station includes said recognition means.
83. ~An object monitoring system as claimed in claim 82, wherein said recognition means is adapted to process said image of said predetermined object to locate pixels representative of'characteristic pixels identifying said object.
84. ~An object monitoring system as claimed in claim 81, wherein said information includes said identifying information and the times of acquisition of acquired images of said predetermined abject, and a plurality of said recognition means are connected to said plurality of camera means and said image processing means, respectively, at the sites of said plurality of camera means and image processing means.
85. ~An object monitoring system as claimed in claim 78, wherein the objects are vehicles.
86. ~An object monitoring system as claimed in claim 85, wherein said recognition means comprises means for locating a licence plate in said image of said predetermined object and means for determining the characters of said licence plate, said characters comprising said identifying information.
87. ~~An object monitoring system as claimed in claim 86, wherein said predetermined intensity range covers pixel intensities produced by shadows of said vehicles.
88. ~~An object monitoring system as claimed in claim 87, wherein said invalid regions correspond to headlight reflections produced by said vehicles.
89. ~~An object monitoring system as claimed in claim 88, wherein said predetermined object is a large vehicle, such as a bus or truck.
90. ~~An object monitoring system as claimed in claim 89, wherein said central station includes remote control means for controlling said plurality of camera means and image processing means from said central station.
91. ~~An object monitoring system as claimed in claim 90, wherein said central station and said plurality of camera means and image processing means include respective telecommunications controllers and communicate using a digital telecommunications network.
92. ~~An object monitoring system as claimed in claim 91, including means for archiving said information and allowing subsequent access thereto.
93. ~~An object monitoring system as claimed in claim 92, wherein said information includes said high resolution image and the time of acquisition, and said central station includes said recognition means.
94. ~~An object monitoring system as claimed in claim 92, wherein said information includes said identifying information and the time of acquisition of said high resolution image, and a plurality of said recognition means are connected to said plurality of camera means and image processing means, respectively, at the sites of said plurality of camera means and image processing means.
95. ~~An object monitoring system as claimed in claim 93, wherein said recognition means is adapted to process said high resolution image to locate pixels representative of characteristic pixels identifying said predetermined object.
96. ~~An object monitoring system as claimed in claim 87, wherein said invalid regions correspond to road lane markings within said area.
97. ~~An object monitoring system as claimed in claim 51, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
98. ~~An object monitoring system comprising:
camera means for generating images of an area and for acquiring an image of a predetermined object;
image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of moving objects in said area, segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area, classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving object, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said predetermined object, and tracking means for tracking said one of said clusters corresponding to said predetermined object to trigger said camera means to acquire said image of said predetermined object; and extraction means for processing said image of said predetermined object to extract information identifying said predetermined object.
99. ~~An object monitoring system as claimed in claim 98, including means for transmitting said image of said predetermined object over a digital telecommunications network.
100. ~An object monitoring system as claimed in claim 98, including a plurality of said camera means and image processing means for monitoring a plurality of areas, respectively, said areas being remote with respect to one another, and means for comparing said information respectively obtained at said areas.
101. ~An object monitoring system as claimed in claim 98, wherein the objects are vehicles.
102. ~An object monitoring system as claimed in claim 98, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
103. ~A vehicle monitoring system comprising:
camera means for generating images of a carriageway and for acquiring images of predetermined vehicles, and image processing means including:
means for subtracting a background image of said carriageway from said images of said carriageway to generate difference images representative of moving vehicles on said carriageway;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles on said carriageway;
classification means for processing said region images, said classification means including:
means for analysing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing at least one characteristic. of said clusters to classification data of said system to determine if said clusters correspond to said predetermined vehicles; and tracking means for tracking said clusters corresponding to said predetermined vehicles to trigger said camera means to acquire said images of said predetermined vehicles.
104. ~~A vehicle monitoring system as claimed in claim 103, wherein said camera means includes video camera means for monitoring said carriageway and a plurality of image capture camera means for acquiring said images of said predetermined vehicle for respective lanes of said carriageway.
105. A vehicle monitoring system as claimed in claim 104, wherein said images of said predetermined vehicles are high resolution images covering the width of a lane of said carriageway and enable optical character recognition means to extract licence plate characters of said predetermined vehicles.
106. A vehicle monitoring system as claimed in claim 103, including optical character recognition means for processing said images of said predetermined vehicles to extract licence plate characters identifying said predetermined vehicles.
107. A vehicle monitoring system as claimed in claim 103, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
108. A vehicle monitoring system as claimed in claim 103, wherein said images of said predetermined vehicles are high resolution images covering the width of a lane of said carriageway and enable optical character recognition means to extract licence plate characters of said predetermined vehicles.
109. A vehicle monitoring system comprising:
a plurality of camera means for generating images of respective areas and for acquiring images of predetermined vehicles, said areas being remote with respect to one another; and a plurality of image processing means including:

means for subtracting background images of said areas from said images of said areas to generate difference images representative of moving vehicles in said areas;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area;
classification means for processing said region images, said classification means including means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving;
vehicles, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters corresponds to said predetermined vehicles;
tracking means for tracking said clusters corresponding to said predetermined vehicles to trigger said camera means to acquire said images of said predetermined vehicles; and recognition means for processing said images of said predetermined vehicles to obtain information identifying said predetermined vehicles.
110. A vehicle monitoring system as claimed in claim 109, including means for comparing said information obtained to determine the average speed between at least two of said areas of at least one of said predetermined vehicles.
111. A vehicle monitoring system as claimed in claim 109, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
112. A vehicle monitoring system comprising:
camera means for generating images of an area and for acquiring an image of a vehicle associated with a law infringement, and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of moving vehicles in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for detecting said law infringement by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said vehicle; and tracking means for tracking said one of said clusters corresponding to said vehicle to trigger said camera means to acquire said image of said vehicle.
113. A vehicle monitoring system as claimed in claim 112, including recognition means for processing said image of said vehicle to obtain information identifying said vehicle.
114. A vehicle monitoring system as claimed in claim 112, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
115. A vehicle monitoring system comprising camera means for generating images of a carriageway and for acquiring high resolution images of large vehicles, such as trucks and buses, and image processing means including:
means for subtracting a background image of said carriageway from said images of said carriageway to generate difference images representative of moving vehicles of said carriageway;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles on said carriageway;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry oil said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing of at least one;
characteristic of said clusters to classification data of said system to determine if said clusters correspond to said large vehicles; and tracking means for tracking said clusters corresponding to said large vehicles to trigger said camera means to acquire said high resolution images of said large vehicles.
116. A vehicle monitoring system as claimed in claim 115, including recognition means for automatically extracting information on said large vehicles, such as licence plate characters, from said high resolution images.
117. A vehicle monitoring system as claimed in claim 115, wherein said classification means generates and operates on the basis of region feature vectors representative of said regions and cluster feature vectors representative of said clusters.
118. An object monitoring system comprising:
video camera means for generating images of an area to monitor moving objects in said area;
image capture camera means for acquiring a high resolution image of a predetermined object; and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving objects in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving objects, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said predetermined object; and tracking means for tracking said one of said clusters corresponding to said predetermined object to trigger said image capture means to acquire said high resolution image of said predetermined object.
119. The object monitoring system of claim 118 further including a plurality of said video camera means and image processing means for monitoring a plurality of said areas, respectively, said areas being remote with respect to one another, and means for comparing said information respectively obtained at said areas.
120. The object monitoring system of claim 118, wherein the predetermined object is a vehicle and said area is a carriageway.
121. The object monitoring system of claim 120 further including a plurality of image capture camera means for acquiring said image of said vehicle in respective lanes of said carriageway.
122. The object monitoring system of claim 118 further including recognition means for processing said image of said predetermined object to obtain information identifying said object.
123. An object monitoring system comprising:
video camera means for generating images of an area to monitor moving objects in said area;
image capture camera means for acquiring a high resolution image of a predetermined object; and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving objects in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving objects in said area;

classification means for processing said region images, said classification means including means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving objects, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine; if one of said clusters corresponds to said predetermined object;
tracking means for tracking said one of said clusters corresponding to said predetermined object to trigger said image capture camera means to acquire said high resolution image of said predetermined object; and extraction means for processing said image of said predetermined object to extract information identifying said predetermined object.
124. The object monitoring system of claim 123, including a plurality of said video camera means and image processing means for monitoring a plurality of areas, respectively, said areas being remote with respect to one another, and means for comparing said information respectively obtained at said areas.
125. The object monitoring system of claim 123, wherein the predetermined object is a vehicle and said area is a carriageway.
126. The object monitoring system of claim 125 further including a plurality of image capture camera means for acquiring said image of said vehicle in respective lanes of said carriageway.
127. The object monitoring system of claim 123 further including recognition means for processing said image of said predetermined object to obtain information identifying said object.
128. A vehicle monitoring system comprising:
video camera means for generating images of a carriageway to monitor moving vehicles in said carriageway:
image capture camera means for acquiring a high resolution image of a predetermined vehicle; and image processing means including:
means for subtracting a background image of said carriageway from said images of said carriageway to generate difference images representative of said moving vehicles on said carriageway;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving;
vehicles on said carriageway;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said predetermined vehicle; and tracking means for tracking said clusters corresponding to said predetermined vehicle to trigger said image capture camera means to acquire said high resolution image of said predetermined vehicle.
129. The vehicle monitoring system of claim 128, including a plurality of said video camera means and image processing means for monitoring a plurality of carriageways, respectively, said carriageways being remote with respect to one another, and means for comparing said information respectively obtained at said carriageways.
130. The vehicle monitoring system of claim 128 further including a plurality of image capture camera means for acquiring said image of said predetermined vehicle in respective lanes of said carriageway.
131. The vehicle monitoring system of claim 128 further including recognition means for processing said image of said predetermined vehicle to obtain information identifying said vehicle.
132. A vehicle monitoring system comprising:
a plurality of video camera means for generating images of respective areas to monitor moving vehicles in said area, said areas being remote with respect to one another;
a plurality of image capture camera means for acquiring a high resolution image of one or more predetermined vehicles; and a plurality of image processing means including:
means for subtracting background images of said areas from said images of said areas to generate difference images representative of said moving vehicles in said areas;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said areas;
classification means for processing said region images, said classification means including:

means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said predetermined vehicle;
tracking means for tracking said clusters corresponding to said predetermined vehicle to trigger said camera means to acquiring said image of said predetermined vehicle; and recognition means for processing said images of said predetermined vehicle to obtain information identifying said predetermined vehicle.
133. The vehicle monitoring system of claim 132, wherein said area is a carriageway, and said image capture camera means acquire said images of said predetermined vehicle in respective lanes of said carriageway.
134. A vehicle monitoring system comprising:
video camera means for generating images of an area to monitor moving vehicles in said area;

image capture camera means for acquiring a high resolution image of a vehicle associated with a law infringement; and image processing means including:
means for subtracting a background image of said area from said images of said area to generate difference images representative of said moving vehicles in said area;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles in said area;

classification means for processing said region images, said classification means including means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for detecting said law infringement by comparing at least one characteristic of said clusters to classification data of said system to determine if one of said clusters corresponds to said vehicle; and tracking means for tracking said one of said clusters corresponding to said vehicle to trigger said image capture camera means to acquire said high resolution image of said vehicle.
135. The vehicle monitoring system of claim 134, including a plurality of said video camera means and image processing means for monitoring a plurality of areas, respectively, said areas being remote with respect to one another, and means for comparing said information respectively obtained at said areas.
136. The vehicle monitoring system of claim 134 further including a plurality of image capture camera means for acquiring said image of said vehicle in respective lanes of said carriageway.
137. The vehicle monitoring system of claim 134 further including recognition means for processing said image of said vehicle associated with said law infringement to obtain information identifying said vehicle.
138. A vehicle monitoring system comprising:

video camera means for generating images of a carriageway to monitor moving vehicles in said area;
image capture camera means for acquiring a high resolution image of a large vehicle, such as a truck and a bus; and image processing means including:
means for subtracting a background image of said carriageway from said images of said carriageway to generate difference images representative of said moving vehicles on said carriageway;
segmentation means for processing said difference images to generate region images representative of regions corresponding to parts of said moving vehicles on said carriageway;
classification means for processing said region images, said classification means including:
means for analyzing the shape of said regions and, on the basis of the analysis, determining valid regions and invalid regions, clustering means for rejecting said invalid regions and generating, on the basis of the geometry of said valid regions, clusters corresponding to respective ones of said moving vehicles, and means for classifying said clusters by comparing at least one characteristic of said clusters to classification data of said system to determine if said clusters correspond to said large vehicle; and tracking means for tracking said clusters corresponding to said large vehicle to trigger said image capture camera means to acquire said high resolution image of said large vehicle.
139. The vehicle monitoring system of claim 138, including a plurality of said video camera means and image processing means for monitoring a plurality of carriageways, respectively, said carriageways being remote with respect to one another, and means for comparing said information respectively obtained at said carriageways.
140. The vehicle monitoring system of claim 138 further including a plurality of image capture camera means for acquiring said image of said large vehicle in respective lanes of said carriageway.
141. The vehicle monitoring system of claim 138 further including recognition means for processing said image of said large vehicle to obtain information identifying said vehicle.
CA002132515A 1992-03-20 1993-03-22 An object monitoring system Expired - Lifetime CA2132515C (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
AUPL146592 1992-03-20
AUPL1465 1992-03-20
AUPL146492 1992-03-20
AUPL1464 1992-03-20
PCT/AU1993/000115 WO1993019441A1 (en) 1992-03-20 1993-03-22 An object monitoring system

Publications (2)

Publication Number Publication Date
CA2132515A1 CA2132515A1 (en) 1993-09-21
CA2132515C true CA2132515C (en) 2006-01-31

Family

ID=25644212

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002132515A Expired - Lifetime CA2132515C (en) 1992-03-20 1993-03-22 An object monitoring system

Country Status (8)

Country Link
US (1) US5809161A (en)
EP (1) EP0631683B1 (en)
JP (1) JPH07505966A (en)
AT (1) ATE203844T1 (en)
CA (1) CA2132515C (en)
DE (1) DE69330513D1 (en)
NZ (1) NZ249799A (en)
WO (1) WO1993019441A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210183042A1 (en) * 2018-12-03 2021-06-17 Mistras Group, Inc. Systems and methods for inspecting pipelines using a robotic imaging system
US11635391B2 (en) 2018-12-03 2023-04-25 Mistras Group, Inc. Systems and methods for inspecting pipelines using a pipeline inspection robot

Families Citing this family (385)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3288474B2 (en) * 1993-03-31 2002-06-04 富士通株式会社 Image processing device
US6430303B1 (en) 1993-03-31 2002-08-06 Fujitsu Limited Image processing apparatus
US5917405A (en) 1993-06-08 1999-06-29 Joao; Raymond Anthony Control apparatus and methods for vehicles
US6542077B2 (en) 1993-06-08 2003-04-01 Raymond Anthony Joao Monitoring apparatus for a vehicle and/or a premises
US5434927A (en) * 1993-12-08 1995-07-18 Minnesota Mining And Manufacturing Company Method and apparatus for machine vision classification and tracking
US6028626A (en) 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
US5666157A (en) * 1995-01-03 1997-09-09 Arc Incorporated Abnormality detection and surveillance system
AU5027796A (en) 1995-03-07 1996-09-23 Interval Research Corporation System and method for selective recording of information
DE19516083A1 (en) * 1995-05-03 1996-11-07 Albrecht Buchner Road traffic monitoring system
EP0804774B1 (en) * 1995-09-15 2003-05-21 Interval Research Corporation A method of compressing a plurality of video images
US5938717A (en) * 1996-03-04 1999-08-17 Laser Technology, Inc. Speed detection and image capture system for moving vehicles
US10152876B2 (en) 1996-03-27 2018-12-11 Gtj Ventures, Llc Control, monitoring, and/or security apparatus and method
US6587046B2 (en) 1996-03-27 2003-07-01 Raymond Anthony Joao Monitoring apparatus and method
US10011247B2 (en) 1996-03-27 2018-07-03 Gtj Ventures, Llc Control, monitoring and/or security apparatus and method
US7253731B2 (en) 2001-01-23 2007-08-07 Raymond Anthony Joao Apparatus and method for providing shipment information
US5974158A (en) * 1996-03-29 1999-10-26 The Commonwealth Of Australia Commonwealth Scientific And Industrial Research Organization Aircraft detection system
AUPN903296A0 (en) * 1996-03-29 1996-04-26 Commonwealth Scientific And Industrial Research Organisation An aircraft detection system
AU720315B2 (en) * 1996-03-29 2000-05-25 Commonwealth Scientific And Industrial Research Organisation An aircraft detection system
AU760788B2 (en) * 1996-03-29 2003-05-22 Commonwealth Scientific And Industrial Research Organisation An image acquisition system
US5889505A (en) * 1996-04-04 1999-03-30 Yale University Vision-based six-degree-of-freedom computer input device
US7613801B2 (en) * 1996-07-12 2009-11-03 Microsoft Corporation System and method for monitoring server performance using a server
EP1890483A3 (en) * 1996-07-23 2011-07-06 Canon Kabushiki Kaisha Camera control apparatus and method
IT1286684B1 (en) * 1996-07-26 1998-07-15 Paolo Sodi DEVICE AND METHOD FOR DETECTION OF ROAD INFRINGEMENTS WITH DYNAMIC POINTING SYSTEMS
JP2998791B2 (en) * 1996-10-31 2000-01-11 日本電気株式会社 3D structure estimation device
US6263507B1 (en) 1996-12-05 2001-07-17 Interval Research Corporation Browser for use in navigating a body of information, with particular application to browsing information represented by audiovisual data
US5893062A (en) 1996-12-05 1999-04-06 Interval Research Corporation Variable rate video playback with synchronized audio
CZ266699A3 (en) * 1997-01-13 1999-12-15 John Overton Automated system for archiving pictures
US5963253A (en) * 1997-01-17 1999-10-05 Raytheon Company Light sensor and thresholding method for minimizing transmission of redundant data
DE19701484A1 (en) * 1997-01-17 1998-07-23 Bosch Gmbh Robert Process for controlling the exposure of video cameras
GB9702849D0 (en) * 1997-02-12 1997-04-02 Trafficmaster Plc Traffic monitoring
US6611610B1 (en) * 1997-04-02 2003-08-26 Gentex Corporation Vehicle lamp control
US6587573B1 (en) 2000-03-20 2003-07-01 Gentex Corporation System for controlling exterior vehicle lights
NL1005785C2 (en) * 1997-04-10 1998-10-14 Traffic Test B V System for detecting vehicles overtaking in zone where overtaking is forbidden
US6760061B1 (en) * 1997-04-14 2004-07-06 Nestor Traffic Systems, Inc. Traffic sensor
US6256046B1 (en) * 1997-04-18 2001-07-03 Compaq Computer Corporation Method and apparatus for visual sensing of humans for active public interfaces
US6263088B1 (en) 1997-06-19 2001-07-17 Ncr Corporation System and method for tracking movement of objects in a scene
US6295367B1 (en) * 1997-06-19 2001-09-25 Emtera Corporation System and method for tracking movement of objects in a scene using correspondence graphs
GB9719694D0 (en) * 1997-09-16 1997-11-19 Canon Kk Image processing apparatus
GB2330679B (en) * 1997-10-21 2002-04-24 911 Emergency Products Inc Warning signal light
WO1999022336A1 (en) * 1997-10-24 1999-05-06 Magic Circle Media, Inc. Objet specification in a bit mapped image
US6121898A (en) * 1997-10-28 2000-09-19 Moetteli; John B. Traffic law enforcement system
NZ332266A (en) 1997-10-28 2000-12-22 John B Moetteli Speed camera system having license plate readers connected to central computer which calculates vehicle's speed
EP0930585B1 (en) 1998-01-14 2004-03-31 Canon Kabushiki Kaisha Image processing apparatus
US6061088A (en) * 1998-01-20 2000-05-09 Ncr Corporation System and method for multi-resolution background adaptation
US6154133A (en) * 1998-01-22 2000-11-28 Ross & Baruzzini, Inc. Exit guard system
WO1999040554A2 (en) * 1998-02-06 1999-08-12 Laser Technology, Inc. Portable motor vehicle speed awareness and surveillance system
US6546119B2 (en) * 1998-02-24 2003-04-08 Redflex Traffic Systems Automated traffic violation monitoring and reporting system
US9075136B1 (en) 1998-03-04 2015-07-07 Gtj Ventures, Llc Vehicle operator and/or occupant information apparatus and method
DE19810302A1 (en) * 1998-03-10 1999-09-16 Wienand Hans Theo Method for measuring speed of motor vehicle
US6697103B1 (en) * 1998-03-19 2004-02-24 Dennis Sunga Fernandez Integrated network for monitoring remote objects
US6400831B2 (en) * 1998-04-02 2002-06-04 Microsoft Corporation Semantic video object segmentation and tracking
US8543372B1 (en) 1998-05-18 2013-09-24 Dennis S. Fernandez System design rights management
JP3511892B2 (en) * 1998-05-25 2004-03-29 日産自動車株式会社 Ambient monitoring device for vehicles
US6675189B2 (en) 1998-05-28 2004-01-06 Hewlett-Packard Development Company, L.P. System for learning and applying integrated task and data parallel strategies in dynamic applications
US7103640B1 (en) * 1999-09-14 2006-09-05 Econnectix, Llc Network distributed tracking wire transfer protocol
US7233978B2 (en) * 1998-07-08 2007-06-19 Econnectix, Llc Method and apparatus for managing location information in a network separate from the data to which the location information pertains
EP0978811A3 (en) * 1998-08-07 2000-08-16 Siemens Aktiengesellschaft Method and device to obtain travel times of vehicles
US6711278B1 (en) 1998-09-10 2004-03-23 Microsoft Corporation Tracking semantic objects in vector image sequences
US6606411B1 (en) * 1998-09-30 2003-08-12 Eastman Kodak Company Method for automatically classifying images into events
WO2000031707A1 (en) * 1998-11-23 2000-06-02 Nestor, Inc. Non-violation event filtering for a traffic light violation detection system
US6754663B1 (en) 1998-11-23 2004-06-22 Nestor, Inc. Video-file based citation generation system for traffic light violations
AUPP839199A0 (en) * 1999-02-01 1999-02-25 Traffic Pro Pty Ltd Object recognition & tracking system
US6396535B1 (en) 1999-02-16 2002-05-28 Mitsubishi Electric Research Laboratories, Inc. Situation awareness system
DE60037360T2 (en) * 1999-05-28 2008-12-04 Nippon Telegraph And Telephone Corp. Method and device for measuring the speed of vehicles with an image processing system
DE19926559A1 (en) * 1999-06-11 2000-12-21 Daimler Chrysler Ag Method and device for detecting objects in the vicinity of a road vehicle up to a great distance
JP2001036801A (en) * 1999-07-23 2001-02-09 Sharp Corp Image pickup device
GB9918248D0 (en) * 1999-08-04 1999-10-06 Matra Bae Dynamics Uk Ltd Improvements in and relating to surveillance systems
US7468677B2 (en) * 1999-08-04 2008-12-23 911Ep, Inc. End cap warning signal assembly
US7783507B2 (en) * 1999-08-23 2010-08-24 General Electric Company System and method for managing a fleet of remote assets
US6263265B1 (en) 1999-10-01 2001-07-17 General Electric Company Web information vault
AUPQ281299A0 (en) * 1999-09-14 1999-10-07 Locktronic Systems Pty. Ltd. Improvements in image recording apparatus
US6553131B1 (en) * 1999-09-15 2003-04-22 Siemens Corporate Research, Inc. License plate recognition with an intelligent camera
US6473517B1 (en) * 1999-09-15 2002-10-29 Siemens Corporate Research, Inc. Character segmentation method for vehicle license plate recognition
JP3352655B2 (en) * 1999-09-22 2002-12-03 富士重工業株式会社 Lane recognition device
US7155735B1 (en) 1999-10-08 2006-12-26 Vulcan Patents Llc System and method for the broadcast dissemination of time-ordered data
US6338152B1 (en) 1999-10-28 2002-01-08 General Electric Company Method and system for remotely managing communication of data used for predicting malfunctions in a plurality of machines
US6795935B1 (en) 1999-10-28 2004-09-21 General Electric Company Diagnosis of faults in a complex system
US6615367B1 (en) 1999-10-28 2003-09-02 General Electric Company Method and apparatus for diagnosing difficult to diagnose faults in a complex system
US6625589B1 (en) 1999-10-28 2003-09-23 General Electric Company Method for adaptive threshold computation for time and frequency based anomalous feature identification in fault log data
US6959235B1 (en) * 1999-10-28 2005-10-25 General Electric Company Diagnosis and repair system and method
US6651034B1 (en) * 1999-10-28 2003-11-18 General Electric Company Apparatus and method for performance and fault data analysis
US6704621B1 (en) * 1999-11-26 2004-03-09 Gideon P. Stein System and method for estimating ego-motion of a moving vehicle using successive images recorded along the vehicle's path of motion
DE19960887A1 (en) * 1999-12-17 2001-06-21 Robot Foto Electr Kg Location information for file compression
US6522889B1 (en) * 1999-12-23 2003-02-18 Nokia Corporation Method and apparatus for providing precise location information through a communications network
US6757682B1 (en) 2000-01-28 2004-06-29 Interval Research Corporation Alerting users to items of current interest
US6940998B2 (en) 2000-02-04 2005-09-06 Cernium, Inc. System for automated screening of security cameras
KR100800040B1 (en) * 2000-03-08 2008-01-31 기븐 이미징 리미티드 A capsule for in vivo imaging
US20010043721A1 (en) * 2000-03-21 2001-11-22 Sarnoff Corporation Method and apparatus for performing motion analysis on an image sequence
US6681195B1 (en) 2000-03-22 2004-01-20 Laser Technology, Inc. Compact speed measurement system with onsite digital image capture, processing, and portable display
JP3987264B2 (en) * 2000-03-24 2007-10-03 富士通株式会社 License plate reader and method
AU2001253619A1 (en) * 2000-04-14 2001-10-30 Mobileye, Inc. Generating a model of the path of a roadway from an image recorded by a camera
US6734896B2 (en) * 2000-04-28 2004-05-11 Matsushita Electric Industrial Co., Ltd. Image processor and monitoring system
US7010788B1 (en) 2000-05-19 2006-03-07 Hewlett-Packard Development Company, L.P. System for computing the optimal static schedule using the stored task execution costs with recent schedule execution costs
US20080005275A1 (en) * 2000-06-02 2008-01-03 Econnectix, Llc Method and apparatus for managing location information in a network separate from the data to which the location information pertains
DE10030894A1 (en) * 2000-06-23 2002-01-03 Bosch Gmbh Robert Traffic information method uses analysis of images provided by cameras at defined locations in traffic network for providing traffic data
US7319479B1 (en) 2000-09-22 2008-01-15 Brickstream Corporation System and method for multi-camera linking and analysis
US6965324B1 (en) * 2000-10-21 2005-11-15 Robert W. Suggs, Sr. Automotive picture and data acquisition center and method
US8564661B2 (en) 2000-10-24 2013-10-22 Objectvideo, Inc. Video analytic rule detection system and method
US8188878B2 (en) 2000-11-15 2012-05-29 Federal Law Enforcement Development Services, Inc. LED light communication system
US7439847B2 (en) 2002-08-23 2008-10-21 John C. Pederson Intelligent observation and identification database system
US7113867B1 (en) 2000-11-26 2006-09-26 Mobileye Technologies Limited System and method for detecting obstacles to vehicle motion and determining time to contact therewith using sequences of images
JP3390426B2 (en) * 2000-11-29 2003-03-24 技研トラステム株式会社 Object image identification method
US7015949B1 (en) 2001-04-12 2006-03-21 Ipix Corporation Method and apparatus for hosting a network camera with refresh degradation
US8026944B1 (en) 2001-04-12 2011-09-27 Sony Corporation Method and apparatus for hosting a network camera with image degradation
US7024488B1 (en) 2001-04-12 2006-04-04 Ipix Corporation Method and apparatus for hosting a network camera
US7076085B1 (en) 2001-04-12 2006-07-11 Ipix Corp. Method and apparatus for hosting a network camera including a heartbeat mechanism
US7177448B1 (en) 2001-04-12 2007-02-13 Ipix Corporation System and method for selecting and transmitting images of interest to a user
US8028249B2 (en) * 2001-05-23 2011-09-27 Eastman Kodak Company Method and system for browsing large digital multimedia object collections
US6996782B2 (en) 2001-05-23 2006-02-07 Eastman Kodak Company Using digital objects organized according to a histogram timeline
US6828920B2 (en) 2001-06-04 2004-12-07 Lockheed Martin Orincon Corporation System and method for classifying vehicles
ATE404114T1 (en) 2001-06-18 2008-08-15 Given Imaging Ltd SWALLOWABLE IN-VIVO CAPSULE WITH A CIRCUIT BOARD HAVING RIGID AND FLEXIBLE SECTIONS
US7283672B1 (en) * 2001-06-29 2007-10-16 Harris Scott C Image compression by object segregation
US9149175B2 (en) 2001-07-26 2015-10-06 Given Imaging Ltd. Apparatus and method for light control in an in-vivo imaging device
US20060184039A1 (en) * 2001-07-26 2006-08-17 Dov Avni Apparatus and method for light control in an in-vivo imaging device
US8493370B2 (en) * 2001-08-29 2013-07-23 Palm, Inc. Dynamic brightness range for portable computer displays based on ambient conditions
DE10148289A1 (en) * 2001-09-29 2003-04-24 Vitronic Dr Ing Stein Bildvera Detector for classifying moving vehicles has camera mounted on motorway bridge with computer to determine vehicle speed and path
DE60228744D1 (en) * 2001-10-09 2008-10-16 Sirf Tech Inc METHOD AND SYSTEM FOR SENDING POSITION-CODED IMAGES ON A WIRELESS NETWORK
US8331621B1 (en) * 2001-10-17 2012-12-11 United Toll Systems, Inc. Vehicle image capture system
US7734500B1 (en) 2001-10-17 2010-06-08 United Toll Systems, Inc. Multiple RF read zone system
US7725348B1 (en) * 2001-10-17 2010-05-25 United Toll Systems, Inc. Multilane vehicle information capture system
DE10160719B4 (en) * 2001-12-11 2011-06-16 Deutsches Zentrum für Luft- und Raumfahrt e.V. Method and device for detecting and recognizing moving objects
JP3817174B2 (en) * 2001-12-28 2006-08-30 矢崎総業株式会社 Vehicle image correction device and night driving visibility support device
US9092841B2 (en) 2004-06-09 2015-07-28 Cognex Technology And Investment Llc Method and apparatus for visual detection and inspection of objects
JP3747866B2 (en) * 2002-03-05 2006-02-22 日産自動車株式会社 Image processing apparatus for vehicle
US7016518B2 (en) * 2002-03-15 2006-03-21 Extreme Cctv Inc. Vehicle license plate imaging and reading system for day and night
KR100459475B1 (en) * 2002-04-04 2004-12-03 엘지산전 주식회사 System and method for judge the kind of vehicle
US10562492B2 (en) 2002-05-01 2020-02-18 Gtj Ventures, Llc Control, monitoring and/or security apparatus and method
US6904341B2 (en) * 2002-06-12 2005-06-07 Sea-Watch Technologies, Inc. Integrated vessel monitoring and control system
AU2003247643A1 (en) * 2002-06-25 2004-01-06 Robert G. Combs Data logging and digital video recording/playback system
US20040001149A1 (en) * 2002-06-28 2004-01-01 Smith Steven Winn Dual-mode surveillance system
JP4287647B2 (en) * 2002-12-27 2009-07-01 株式会社Ihi Environmental status monitoring device
US7171028B2 (en) * 2002-11-22 2007-01-30 The Boeing Company Method and apparatus for covertly determining the rate of relative motion between two objects
JP4080316B2 (en) * 2002-12-10 2008-04-23 富士フイルム株式会社 Imaging device, terminal device, and imaging system
DE10301898A1 (en) * 2003-01-17 2004-08-05 Robert Bosch Gmbh Image sensor adjustment method
SE531334C2 (en) * 2003-01-30 2009-02-24 Facility Parking Europ Ab Vehicle ID
US7382277B2 (en) 2003-02-12 2008-06-03 Edward D. Ioli Trust System for tracking suspicious vehicular activity
EP1463301A1 (en) * 2003-03-19 2004-09-29 Thomson Licensing S.A. Method for identification of tokens in video sequences
EP1460835B1 (en) * 2003-03-19 2018-11-14 InterDigital Madison Patent Holdings Method for identification of tokens in video sequences
US20040201460A1 (en) * 2003-04-09 2004-10-14 Bucholz Andrew J. System and method of vehicle surveillance
US6865297B2 (en) * 2003-04-15 2005-03-08 Eastman Kodak Company Method for automatically classifying images into events in a multimedia authoring application
US7298895B2 (en) * 2003-04-15 2007-11-20 Eastman Kodak Company Method for automatically classifying images into events
US20100002070A1 (en) 2004-04-30 2010-01-07 Grandeye Ltd. Method and System of Simultaneously Displaying Multiple Views for Video Surveillance
US7528881B2 (en) * 2003-05-02 2009-05-05 Grandeye, Ltd. Multiple object processing in wide-angle video camera
US6970102B2 (en) * 2003-05-05 2005-11-29 Transol Pty Ltd Traffic violation detection, recording and evidence processing system
US7986339B2 (en) * 2003-06-12 2011-07-26 Redflex Traffic Systems Pty Ltd Automated traffic violation monitoring and reporting system with combined video and still-image data
US7263209B2 (en) * 2003-06-13 2007-08-28 Sarnoff Corporation Vehicular vision system
US7711150B2 (en) * 2003-07-10 2010-05-04 James Simon Autonomous wide-angle license plate recognition
US7747041B2 (en) * 2003-09-24 2010-06-29 Brigham Young University Automated estimation of average stopped delay at signalized intersections
US7171024B2 (en) * 2003-12-01 2007-01-30 Brickstream Corporation Systems and methods for determining if objects are in a queue
EP1542186A1 (en) * 2003-12-12 2005-06-15 Chao-Hung Chang Active searching and identifying system and method
EP1738198A4 (en) * 2004-02-13 2011-05-25 Technion Res & Dev Foundation Enhanced underwater imaging
WO2005086078A1 (en) * 2004-03-02 2005-09-15 Sarnoff Corporation Method and apparatus for classifying an object
WO2005086079A1 (en) * 2004-03-02 2005-09-15 Sarnoff Corporation Method and apparatus for differentiating pedestrians, vehicles, and other objects
WO2005098782A1 (en) 2004-04-08 2005-10-20 Mobileye Technologies Limited Collision warning system
US8427538B2 (en) * 2004-04-30 2013-04-23 Oncam Grandeye Multiple view and multiple object processing in wide-angle video camera
EP1751970A4 (en) * 2004-05-01 2010-04-21 Eliezer Jacob Digital camera with non-uniform image resolution
US7605852B2 (en) 2004-05-17 2009-10-20 Micron Technology, Inc. Real-time exposure control for automatic light control
US20050276445A1 (en) * 2004-06-09 2005-12-15 Silver William M Method and apparatus for automatic visual detection, recording, and retrieval of events
US8243986B2 (en) * 2004-06-09 2012-08-14 Cognex Technology And Investment Corporation Method and apparatus for automatic visual event detection
US8891852B2 (en) * 2004-06-09 2014-11-18 Cognex Technology And Investment Corporation Method and apparatus for configuring and testing a machine vision detector
US8127247B2 (en) 2004-06-09 2012-02-28 Cognex Corporation Human-machine-interface and method for manipulating data in a machine vision system
JP2006025007A (en) * 2004-07-06 2006-01-26 Fuji Photo Film Co Ltd Image processor and image processing program
IL162921A0 (en) * 2004-07-08 2005-11-20 Hi Tech Solutions Ltd Character recognition system and method
US7659906B2 (en) * 2004-07-29 2010-02-09 The United States Of America As Represented By The Secretary Of The Navy Airborne real time image exploitation system (ARIES)
US20060088188A1 (en) * 2004-10-11 2006-04-27 Alexander Ioffe Method for the detection of an obstacle
US8081214B2 (en) 2004-10-12 2011-12-20 Enforcement Video, Llc Method of and system for mobile surveillance and event recording
US20070092110A1 (en) * 2004-11-08 2007-04-26 Li-Qun Xu Object tracking within video images
US7720315B2 (en) 2004-11-12 2010-05-18 Cognex Technology And Investment Corporation System and method for displaying and using non-numeric graphic elements to control and monitor a vision system
US9292187B2 (en) 2004-11-12 2016-03-22 Cognex Corporation System, method and graphical user interface for displaying and controlling vision system operating parameters
US7636449B2 (en) 2004-11-12 2009-12-22 Cognex Technology And Investment Corporation System and method for assigning analysis parameters to vision detector using a graphical interface
EP1659551A1 (en) * 2004-11-23 2006-05-24 S.A.V.E. - Sistemi Avanzati di Elettronica di Domenico Ing. Monopoli & C. S.A.S. Electronic registration system for detecting traffic violations
US7512250B2 (en) * 2004-11-24 2009-03-31 Siemens Corporate Research, Inc. System and method for fast illumination-invariant background subtraction using two views
US7688999B2 (en) * 2004-12-08 2010-03-30 Electronics And Telecommunications Research Institute Target detecting system and method
US7623676B2 (en) * 2004-12-21 2009-11-24 Sarnoff Corporation Method and apparatus for tracking objects over a wide area using a network of stereo sensors
US7457433B2 (en) * 2005-01-20 2008-11-25 International Business Machines Corporation System and method for analyzing video from non-static camera
US9489717B2 (en) 2005-01-31 2016-11-08 Invention Science Fund I, Llc Shared image device
US9124729B2 (en) 2005-01-31 2015-09-01 The Invention Science Fund I, Llc Shared image device synchronization or designation
US8902320B2 (en) 2005-01-31 2014-12-02 The Invention Science Fund I, Llc Shared image device synchronization or designation
US20060170956A1 (en) 2005-01-31 2006-08-03 Jung Edward K Shared image devices
US9910341B2 (en) 2005-01-31 2018-03-06 The Invention Science Fund I, Llc Shared image device designation
US9082456B2 (en) 2005-01-31 2015-07-14 The Invention Science Fund I Llc Shared image device designation
US8606383B2 (en) 2005-01-31 2013-12-10 The Invention Science Fund I, Llc Audio sharing
US8502870B2 (en) * 2005-02-02 2013-08-06 Pima Electronic Systems Ltd. Device, system, and method of rapid image acquisition
WO2006082644A1 (en) * 2005-02-03 2006-08-10 Fujitsu Limited Vehicle image data generation program and vehicle image data generation device
US7568628B2 (en) 2005-03-11 2009-08-04 Hand Held Products, Inc. Bar code reading device with global electronic shutter control
US7710452B1 (en) 2005-03-16 2010-05-04 Eric Lindberg Remote video monitoring of non-urban outdoor sites
US7050908B1 (en) * 2005-03-22 2006-05-23 Delphi Technologies, Inc. Lane marker projection method for a motor vehicle vision system
GB2425385B (en) 2005-04-18 2007-08-01 Pips Technology Ltd Vehicle speed monitoring system
US9167195B2 (en) 2005-10-31 2015-10-20 Invention Science Fund I, Llc Preservation/degradation of video/audio aspects of a data stream
US8964054B2 (en) 2006-08-18 2015-02-24 The Invention Science Fund I, Llc Capturing selected image objects
US9942511B2 (en) 2005-10-31 2018-04-10 Invention Science Fund I, Llc Preservation/degradation of video/audio aspects of a data stream
US9819490B2 (en) 2005-05-04 2017-11-14 Invention Science Fund I, Llc Regional proximity for shared image device(s)
US9621749B2 (en) 2005-06-02 2017-04-11 Invention Science Fund I, Llc Capturing selected image objects
US20070139529A1 (en) * 2005-06-02 2007-06-21 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Dual mode image capture technique
US9001215B2 (en) 2005-06-02 2015-04-07 The Invention Science Fund I, Llc Estimating shared image device operational capabilities or resources
US8681225B2 (en) 2005-06-02 2014-03-25 Royce A. Levien Storage access technique for captured data
US9191611B2 (en) 2005-06-02 2015-11-17 Invention Science Fund I, Llc Conditional alteration of a saved image
US20070222865A1 (en) 2006-03-15 2007-09-27 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Enhanced video/still image correlation
US9076208B2 (en) 2006-02-28 2015-07-07 The Invention Science Fund I, Llc Imagery processing
US10003762B2 (en) 2005-04-26 2018-06-19 Invention Science Fund I, Llc Shared image devices
US9451200B2 (en) 2005-06-02 2016-09-20 Invention Science Fund I, Llc Storage access technique for captured data
US9967424B2 (en) 2005-06-02 2018-05-08 Invention Science Fund I, Llc Data storage usage protocol
US7920959B1 (en) 2005-05-01 2011-04-05 Christopher Reed Williams Method and apparatus for estimating the velocity vector of multiple vehicles on non-level and curved roads using a single camera
JP4772115B2 (en) * 2005-05-31 2011-09-14 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Method and system for detecting roads at night
US7770799B2 (en) 2005-06-03 2010-08-10 Hand Held Products, Inc. Optical reader having reduced specular reflection read failures
NL1031867C1 (en) * 2005-07-08 2007-01-09 Everhardus Fransiscu Weijdeven Method for determining vehicle data.
CN100458817C (en) * 2005-07-21 2009-02-04 欧姆龙株式会社 Monitoring apparatus
JP2007052770A (en) * 2005-07-21 2007-03-01 Omron Corp Monitoring apparatus
US7504965B1 (en) 2005-08-05 2009-03-17 Elsag North America, Llc Portable covert license plate reader
US20070040065A1 (en) * 2005-08-19 2007-02-22 Von Thal German Flexible refueling boom extendable tube
US20070046484A1 (en) * 2005-08-22 2007-03-01 Andrew Bucholz Electronic parking identification and surveillance system and method
DE102005045017A1 (en) * 2005-09-21 2007-03-22 Robert Bosch Gmbh Method and driver assistance system for sensor-based approach control of a motor vehicle
US8982944B2 (en) * 2005-10-12 2015-03-17 Enforcement Video, Llc Method and system for categorized event recording of images in multiple resolution levels
US20070120980A1 (en) * 2005-10-31 2007-05-31 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Preservation/degradation of video/audio aspects of a data stream
US9041851B2 (en) * 2005-11-15 2015-05-26 The Trustees Of Princeton University Organic electronic detectors and methods of fabrication
ATE429627T1 (en) * 2005-11-22 2009-05-15 Yarayan Ali DEVICE FOR CHECKING TIRE TREAD DEPTH AND TYPE, SPEED AND GROUND CLEARANCE ON VEHICLES WHILE DRIVING
US7733357B2 (en) * 2006-01-13 2010-06-08 Hewlett-Packard Development Company, L.P. Display system
US8078618B2 (en) 2006-01-30 2011-12-13 Eastman Kodak Company Automatic multimode system for organizing and retrieving content data files
US7579965B2 (en) * 2006-03-03 2009-08-25 Andrew Bucholz Vehicle data collection and processing system
ITTO20060214A1 (en) * 2006-03-22 2007-09-23 Kria S R L VEHICLE DETECTION SYSTEM
US20070250898A1 (en) * 2006-03-28 2007-10-25 Object Video, Inc. Automatic extraction of secondary video streams
US20130163822A1 (en) * 2006-04-04 2013-06-27 Cyclops Technologies, Inc. Airborne Image Capture and Recognition System
US20130163823A1 (en) * 2006-04-04 2013-06-27 Cyclops Technologies, Inc. Image Capture and Recognition System Having Real-Time Secure Communication
TWI302879B (en) * 2006-05-12 2008-11-11 Univ Nat Chiao Tung Real-time nighttime vehicle detection and recognition system based on computer vision
US7956735B2 (en) 2006-05-15 2011-06-07 Cernium Corporation Automated, remotely-verified alarm system with intrusion and video surveillance and digital video recording
WO2007135488A1 (en) * 2006-05-22 2007-11-29 Nokia Corporation Detection system for detecing use of a mobile telecommunications device
US7786898B2 (en) 2006-05-31 2010-08-31 Mobileye Technologies Ltd. Fusion of far infrared and visible images in enhanced obstacle detection in automotive applications
US7724962B2 (en) * 2006-07-07 2010-05-25 Siemens Corporation Context adaptive approach in vehicle detection under various visibility conditions
US20080062009A1 (en) * 2006-08-30 2008-03-13 Marton Keith J Method and system to improve traffic flow
FR2906920A1 (en) * 2006-10-04 2008-04-11 Nikos Ioannou AUTONOMOUS SYSTEM FOR DETECTING THE EXCEEDING BY A VEHICLE OF A LIMITED SPEED.
US7580547B2 (en) 2006-10-24 2009-08-25 Iteris, Inc. Electronic traffic monitor
DK1870868T3 (en) * 2006-11-10 2009-01-19 Engine Srl System and method for detecting the average speed of vehicles in connection with speed control
JP4454681B2 (en) * 2006-12-05 2010-04-21 富士通株式会社 Traffic condition display method, traffic condition display system, in-vehicle device, and computer program
WO2008088409A2 (en) * 2006-12-19 2008-07-24 Indiana University Research & Technology Corporation Real-time dynamic content based vehicle tracking, traffic monitoring, and classification system
US8189962B2 (en) * 2006-12-19 2012-05-29 Hitachi Kokusai Electric Inc. Image processing apparatus
CN101211407B (en) * 2006-12-29 2011-10-26 东软集团股份有限公司 Diurnal image recognition method and device
US8213685B2 (en) * 2007-01-05 2012-07-03 American Traffic Solutions, Inc. Video speed detection system
EP2168079B1 (en) * 2007-01-23 2015-01-14 Valeo Schalter und Sensoren GmbH Method and system for universal lane boundary detection
US8132759B2 (en) * 2007-03-21 2012-03-13 The Boeing Company System and method for facilitating aerial refueling
US7959110B2 (en) * 2007-04-11 2011-06-14 The Boeing Company Methods and apparatus for resisting torsional loads in aerial refueling booms
US8144214B2 (en) 2007-04-18 2012-03-27 Panasonic Corporation Imaging apparatus, imaging method, integrated circuit, and storage medium
JP5083841B2 (en) * 2007-04-27 2012-11-28 本田技研工業株式会社 Vehicle periphery monitoring device, vehicle periphery monitoring program, and vehicle periphery monitoring method
US7952021B2 (en) 2007-05-03 2011-05-31 United Toll Systems, Inc. System and method for loop detector installation
US9414458B2 (en) 2007-05-24 2016-08-09 Federal Law Enforcement Development Services, Inc. LED light control assembly and system
US9294198B2 (en) 2007-05-24 2016-03-22 Federal Law Enforcement Development Services, Inc. Pulsed light communication key
US9258864B2 (en) 2007-05-24 2016-02-09 Federal Law Enforcement Development Services, Inc. LED light control and management system
US9455783B2 (en) 2013-05-06 2016-09-27 Federal Law Enforcement Development Services, Inc. Network security and variable pulse wave form with continuous communication
US11265082B2 (en) 2007-05-24 2022-03-01 Federal Law Enforcement Development Services, Inc. LED light control assembly and system
WO2008148050A1 (en) 2007-05-24 2008-12-04 Federal Law Enforcement Development Services, Inc. Led light interior room and building communication system
US9100124B2 (en) 2007-05-24 2015-08-04 Federal Law Enforcement Development Services, Inc. LED Light Fixture
US7869621B1 (en) * 2007-06-07 2011-01-11 Aydin Arpa Method and apparatus for interpreting images in temporal or spatial domains
US8237099B2 (en) * 2007-06-15 2012-08-07 Cognex Corporation Method and system for optoelectronic detection and location of objects
US8718319B2 (en) * 2007-06-15 2014-05-06 Cognex Corporation Method and system for optoelectronic detection and location of objects
WO2009012289A1 (en) 2007-07-16 2009-01-22 Cernium Corporation Apparatus and methods for video alarm verification
US20120002045A1 (en) * 2007-08-08 2012-01-05 Mayer Tony Non-retro-reflective license plate imaging system
US8599368B1 (en) * 2008-01-29 2013-12-03 Enforcement Video, Llc Laser-based speed determination device for use in a moving vehicle
US20090046157A1 (en) * 2007-08-13 2009-02-19 Andrew Cilia Combined wide-angle/zoom camera for license plate identification
US8306268B2 (en) * 2007-09-12 2012-11-06 General Electric Company Method and system for image integrity determination
KR101104609B1 (en) * 2007-10-26 2012-01-12 주식회사 만도 Method and System for Recognizing Target Parking Location
EP2219166B1 (en) 2007-11-01 2012-07-11 Igor Yurievich Matsur Traffic monitoring system
US8204273B2 (en) * 2007-11-29 2012-06-19 Cernium Corporation Systems and methods for analysis of video content, event notification, and video content provision
US8108055B2 (en) * 2007-12-28 2012-01-31 Larry Wong Method, system and apparatus for controlling an electrical device
WO2009097449A1 (en) 2008-01-29 2009-08-06 Enforcement Video, Llc Omnidirectional camera for use in police car event recording
US20090213218A1 (en) * 2008-02-15 2009-08-27 Andrew Cilia System and method for multi-resolution storage of images
US8229168B2 (en) * 2008-02-20 2012-07-24 International Business Machines Corporation Fast license plate verifier
DE102008017137A1 (en) 2008-03-19 2009-10-01 Siemens Aktiengesellschaft Method for recording image from vehicle i.e. police car, to provide evidence to court proceedings about accident, involves decoding marking by evaluating marked image, and storing code of marking in portable data memory
US8208022B2 (en) * 2008-06-05 2012-06-26 Lg Electronics Inc. Apparatus and method for removing blooming of camera image
US7786850B1 (en) 2008-06-16 2010-08-31 Gianni Arcaini Method and apparatus for bridge collision alert monitoring and impact analysis
DE102008002560A1 (en) * 2008-06-20 2009-12-24 Robert Bosch Gmbh Image data visualization
US9405968B2 (en) * 2008-07-21 2016-08-02 Facefirst, Inc Managed notification system
US8320674B2 (en) * 2008-09-03 2012-11-27 Sony Corporation Text localization for image and video OCR
JP5183392B2 (en) * 2008-09-25 2013-04-17 キヤノン株式会社 Image processing apparatus, image processing method, and program
NL2002080C (en) * 2008-10-10 2010-04-13 Lucas Christiaan Fretz Information system and method of providing information.
US8611590B2 (en) 2008-12-23 2013-12-17 Canon Kabushiki Kaisha Video object fragmentation detection and management
US7974314B2 (en) * 2009-01-16 2011-07-05 Microsoft Corporation Synchronization of multiple data source to a common time base
US20100246890A1 (en) * 2009-03-26 2010-09-30 Microsoft Corporation Detection of objects in images
US8890773B1 (en) 2009-04-01 2014-11-18 Federal Law Enforcement Development Services, Inc. Visible light transceiver glasses
US8571261B2 (en) * 2009-04-22 2013-10-29 Checkvideo Llc System and method for motion detection in a surveillance video
US20100276962A1 (en) * 2009-04-29 2010-11-04 Gm Global Technology Operations, Inc. Active face shade detection in auto sun-shade system
US20100302371A1 (en) * 2009-05-27 2010-12-02 Mark Abrams Vehicle tailgating detection system
US8310377B2 (en) * 2009-08-24 2012-11-13 Optotraffic, Llc Mobile automated system for traffic monitoring
EP2306428B1 (en) * 2009-10-01 2011-11-09 Kapsch TrafficCom AG Device and method for determining the direction, speed and/or distance of vehicles
PT2306425E (en) * 2009-10-01 2011-11-10 Kapsch Trafficcom Ag Device and method for detecting wheel axles
US20110234829A1 (en) * 2009-10-06 2011-09-29 Nikhil Gagvani Methods, systems and apparatus to configure an imaging device
DE102009046362A1 (en) * 2009-11-03 2011-05-05 Tesa Se Pressure-sensitive adhesive made of a crosslinkable polyolefin and an adhesive resin
CN101739829B (en) * 2009-12-03 2014-04-23 北京中星微电子有限公司 Video-based vehicle overspeed monitoring method and system
US20110133510A1 (en) * 2009-12-07 2011-06-09 Gm Global Technology Operations, Inc. Saturation-based shade-line detection
EP2535881B1 (en) * 2010-02-08 2015-10-28 Obshchestvo s ogranichennoj otvetstvennost'ju "Korporazija "Stroy Invest Proekt M" Method and device for determining the speed of travel and coordinates of vehicles and subsequently identifying same and automatically recording road traffic offences
AU2011201189B2 (en) * 2010-03-16 2014-08-28 The University Of Sydney Vehicle localization in open-pit mining using GPS and monocular camera
US8736680B1 (en) 2010-05-18 2014-05-27 Enforcement Video, Llc Method and system for split-screen video display
US8922656B2 (en) 2010-06-16 2014-12-30 Ndi Technologies, Inc. USB video interface for ALPR cameras and associated method
US9626696B2 (en) 2010-06-17 2017-04-18 Microsoft Technology Licensing, Llc Techniques to verify location for location based services
TWI452540B (en) * 2010-12-09 2014-09-11 Ind Tech Res Inst Image based detecting system and method for traffic parameters and computer program product thereof
US20120162431A1 (en) * 2010-12-23 2012-06-28 Scott Riesebosch Methods and systems for monitoring traffic flow
US20120179518A1 (en) * 2011-01-06 2012-07-12 Joshua Timothy Jaipaul System and method for intersection monitoring
WO2012097291A1 (en) 2011-01-14 2012-07-19 Federal Law Enforcement Development Services, Inc. Method of providing lumens and tracking of lumen consumption
US9373142B2 (en) 2011-03-04 2016-06-21 Digital Recognition Network, Inc. Method and system for locating a mobile asset
WO2012125687A2 (en) * 2011-03-14 2012-09-20 The Regents Of The University Of California Method and system for vehicle classification
US8565481B1 (en) * 2011-05-26 2013-10-22 Google Inc. System and method for tracking objects
DK2538239T3 (en) * 2011-06-21 2014-02-24 Kapsch Trafficcom Ag Devices and method for detection of wheels
US9031948B1 (en) * 2011-07-06 2015-05-12 Shawn B. Smith Vehicle prediction and association tool based on license plate recognition
US8781187B2 (en) * 2011-07-13 2014-07-15 Mckesson Financial Holdings Methods, apparatuses, and computer program products for identifying a region of interest within a mammogram image
US8768009B1 (en) * 2011-07-26 2014-07-01 Shawn B. Smith Locating persons of interest based on license plate recognition information
DE102011053284B8 (en) * 2011-09-05 2012-09-06 Jenoptik Robot Gmbh Method for verifying the speed of an adequate vehicle by means of a camera
US8903127B2 (en) * 2011-09-16 2014-12-02 Harman International (China) Holdings Co., Ltd. Egomotion estimation system and method
WO2013041123A1 (en) 2011-09-20 2013-03-28 Friedrich-Alexander-Universität Erlangen-Nürnberg System and method for supporting an exercise movement
WO2013052773A1 (en) * 2011-10-05 2013-04-11 L-3 Communications Mobilevision Inc. Multiple resolution camera system for automated license plate recognition and event recording
US9651499B2 (en) 2011-12-20 2017-05-16 Cognex Corporation Configurable image trigger for a vision system and method for using the same
US8934020B2 (en) * 2011-12-22 2015-01-13 Pelco, Inc. Integrated video quantization
US8736463B1 (en) * 2012-01-30 2014-05-27 Google Inc. Object bounding box estimation
US9569680B2 (en) * 2012-02-02 2017-02-14 Xerox Corporation Automated running-engine detection in stationary motor vehicles
JP6051608B2 (en) * 2012-06-19 2016-12-27 市光工業株式会社 Vehicle surrounding object detection device
US9465997B2 (en) 2012-09-26 2016-10-11 General Electric Company System and method for detection and tracking of moving objects
US20140133753A1 (en) * 2012-11-09 2014-05-15 Ge Aviation Systems Llc Spectral scene simplification through background subtraction
US9641806B2 (en) * 2013-03-12 2017-05-02 3M Innovative Properties Company Average speed detection with flash illumination
WO2014160096A1 (en) 2013-03-13 2014-10-02 Federal Law Enforcement Development Services, Inc. Led light control and management system
JP6178099B2 (en) 2013-04-05 2017-08-09 ソニー株式会社 Intermediate unit and camera system
US10546441B2 (en) 2013-06-04 2020-01-28 Raymond Anthony Joao Control, monitoring, and/or security, apparatus and method for premises, vehicles, and/or articles
US9336450B2 (en) * 2013-06-05 2016-05-10 Xerox Corporation Methods and systems for selecting target vehicles for occupancy detection
MA36097B1 (en) * 2013-07-08 2016-01-29 Mascir Morrocan Foundation For Advanced Science Innovation & Res Mono-camera system for estimating vehicle speed by video processing with multi-camera calibration by stereoscopic effect
CA2856896A1 (en) * 2013-07-18 2015-01-18 Spo Systems Inc. Limited Virtual video patrol system and components therefor
JP6110256B2 (en) * 2013-08-21 2017-04-05 株式会社日本自動車部品総合研究所 Object estimation apparatus and object estimation method
US9736374B2 (en) * 2013-09-19 2017-08-15 Conduent Business Services, Llc Video/vision based access control method and system for parking occupancy determination, which is robust against camera shake
US9412031B2 (en) 2013-10-16 2016-08-09 Xerox Corporation Delayed vehicle identification for privacy enforcement
KR101507242B1 (en) * 2013-10-21 2015-03-31 포항공과대학교 산학협력단 Apparatus and method for providing motion haptic effect using video analysis
JP6351238B2 (en) * 2013-11-15 2018-07-04 キヤノン株式会社 Image processing apparatus, imaging apparatus, and distance correction method
KR102149273B1 (en) * 2013-12-10 2020-08-28 한화테크윈 주식회사 Method and apparatus for recognizing number-plate
FR3014553A1 (en) * 2013-12-11 2015-06-12 Parrot METHOD FOR ANGULAR CALIBRATION OF THE POSITION OF AN ON-BOARD VIDEO CAMERA IN A MOTOR VEHICLE
US9779284B2 (en) 2013-12-17 2017-10-03 Conduent Business Services, Llc Privacy-preserving evidence in ALPR applications
US20150198941A1 (en) 2014-01-15 2015-07-16 John C. Pederson Cyber Life Electronic Networking and Commerce Operating Exchange
US11244171B2 (en) * 2014-01-22 2022-02-08 Conduent Business Services Llc Video-based system for automated detection of double parking violations
US9478030B1 (en) * 2014-03-19 2016-10-25 Amazon Technologies, Inc. Automatic visual fact extraction
FR3018940B1 (en) * 2014-03-24 2018-03-09 Survision AUTOMATIC CLASSIFICATION SYSTEM FOR MOTOR VEHICLES
US9626476B2 (en) 2014-03-27 2017-04-18 Change Healthcare Llc Apparatus, method and computer-readable storage medium for transforming digital images
US10207193B2 (en) 2014-05-21 2019-02-19 Universal City Studios Llc Optical tracking system for automation of amusement park elements
US10061058B2 (en) 2014-05-21 2018-08-28 Universal City Studios Llc Tracking system and method for use in surveying amusement park equipment
US9433870B2 (en) 2014-05-21 2016-09-06 Universal City Studios Llc Ride vehicle tracking and control system using passive tracking elements
US9429398B2 (en) 2014-05-21 2016-08-30 Universal City Studios Llc Optical tracking for controlling pyrotechnic show elements
US9616350B2 (en) 2014-05-21 2017-04-11 Universal City Studios Llc Enhanced interactivity in an amusement park environment using passive tracking elements
US9600999B2 (en) 2014-05-21 2017-03-21 Universal City Studios Llc Amusement park element tracking system
US10025990B2 (en) * 2014-05-21 2018-07-17 Universal City Studios Llc System and method for tracking vehicles in parking structures and intersections
US9443158B1 (en) * 2014-06-22 2016-09-13 Kristopher Haskins Method for computer vision to recognize objects marked for identification with a bigram of glyphs, and devices utilizing the method for practical purposes
WO2016032524A1 (en) * 2014-08-29 2016-03-03 Hewlett-Packard Development Company, Lp Image processing
CN104574981B (en) * 2014-12-01 2017-06-06 山东易华录信息技术有限公司 Do not give way the system and method for behavior at pass through crossing of giving way of record motor vehicle
US9607245B2 (en) 2014-12-02 2017-03-28 Xerox Corporation Adapted vocabularies for matching image signatures with fisher vectors
US10217034B2 (en) 2015-01-08 2019-02-26 Sony Semiconductor Solutions Corporation Image processing device, imaging device, and image processing method
US9984544B2 (en) 2015-02-17 2018-05-29 Sap Se Device layout optimization for surveillance devices
GB2538502B (en) * 2015-05-18 2018-04-04 Roadmetric Ltd Detection and documentation of tailgating and speeding violations
US20170046950A1 (en) 2015-08-11 2017-02-16 Federal Law Enforcement Development Services, Inc. Function disabler device and system
EP3136635B1 (en) * 2015-08-25 2020-05-06 Toll Collect GmbH Method and system for capturing an image of a motor vehicle
EP3136636B1 (en) * 2015-08-25 2020-04-22 Toll Collect GmbH Method and system for capturing an image of a motor vehicle
RU2628916C2 (en) * 2015-10-14 2017-08-22 Общество с ограниченной ответственностью "АВТОДОРИЯ" (ООО "АВТОДОРИЯ") Method and system of controlling stationary camera
JP6709283B2 (en) * 2015-11-16 2020-06-10 オービタル インサイト インコーポレイテッド Detection and analysis of moving vehicles using low resolution remote sensing images
US20170186317A1 (en) 2015-12-29 2017-06-29 Tannery Creek Systems Inc. System and Method for Determining Parking Infraction
US9965696B2 (en) 2015-12-31 2018-05-08 James Alves Digital camera control system
PT3203447T (en) * 2016-02-04 2019-08-07 Holding Assessoria I Lideratge S L Hal Sl Detection of fraudulent access at control gates
JP6658054B2 (en) * 2016-02-16 2020-03-04 トヨタ車体株式会社 Body part construction site determination system and construction site determination method
US10341605B1 (en) 2016-04-07 2019-07-02 WatchGuard, Inc. Systems and methods for multiple-resolution storage of media streams
JP6789682B2 (en) * 2016-06-13 2020-11-25 キヤノン株式会社 Imaging device, its control method, and program
US11092689B2 (en) * 2016-09-20 2021-08-17 Apple Inc. Enabling lidar detection
US10629072B2 (en) * 2016-10-20 2020-04-21 Ford Global Technologies, Llc LIDAR and vision vehicle sensing
US11017243B2 (en) 2016-12-05 2021-05-25 Intel Corporation Automotive system with motion detection capabilities
EP3566078A1 (en) * 2017-01-03 2019-11-13 Innoviz Technologies Ltd. Lidar systems and methods for detection and classification of objects
US10474906B2 (en) 2017-03-24 2019-11-12 Echelon Corporation High dynamic range video of fast moving objects without blur
BR102017006860A2 (en) * 2017-04-03 2018-10-30 Compsis Computadores E Sist Industria E Comercio Ltda system for automatic detection of vehicle categories by longitudinal profile image analysis
EP3506153A1 (en) * 2017-12-28 2019-07-03 Canon Kabushiki Kaisha Information processing apparatus, system, method, and non-transitory computer-readable storage medium
EP3769295A1 (en) * 2018-03-19 2021-01-27 Kistler Holding AG Traffic monitoring system
US10836309B1 (en) * 2018-06-18 2020-11-17 Alarm.Com Incorporated Distracted driver detection and alert system
US10298910B1 (en) * 2018-06-29 2019-05-21 Zoox, Inc. Infrastructure free intrinsic calibration
US10733761B2 (en) 2018-06-29 2020-08-04 Zoox, Inc. Sensor calibration
EP3824454A4 (en) 2018-07-19 2021-09-01 Acusensus IP Pty Ltd Infringement detection method, device and system
WO2020036603A1 (en) * 2018-08-17 2020-02-20 Hewlett-Packard Development Company, L.P. Exposure control of captured images
GB2576538B (en) 2018-08-23 2020-09-02 Roadmetric Ltd Detection and documentation of speeding violations
US10967751B2 (en) * 2019-05-10 2021-04-06 Gm Global Technology Operations, Llc Method to detect the proper connection of a vehicle charging cable
CN113841188B (en) * 2019-05-13 2024-02-20 日本电信电话株式会社 Traffic flow estimating device, traffic flow estimating method, and storage medium
US20220227396A1 (en) * 2019-05-23 2022-07-21 Hitachi Astemo, Ltd. Vehicle control system and vehicle control method
US11267464B2 (en) 2019-07-24 2022-03-08 Pony Ai Inc. System and method to distinguish between moving and static objects
CN111340877B (en) * 2020-03-25 2023-10-27 北京爱笔科技有限公司 Vehicle positioning method and device
CN112309126B (en) * 2020-10-30 2022-07-05 杭州海康威视数字技术股份有限公司 License plate detection method and device, electronic equipment and computer readable storage medium
CN112926590B (en) * 2021-03-18 2023-12-01 上海晨兴希姆通电子科技有限公司 Segmentation recognition method and system for characters on cable
CN113674329A (en) * 2021-08-13 2021-11-19 上海同温层智能科技有限公司 Vehicle driving behavior detection method and system
EP4199495A1 (en) * 2021-12-15 2023-06-21 Nokia Solutions and Networks Oy Regulating frame processing rates
CN115019502A (en) * 2022-04-14 2022-09-06 安徽超远信息技术有限公司 Road vehicle superelevation early warning method
US11715305B1 (en) * 2022-11-30 2023-08-01 Amitha Nandini Mandava Traffic detection system using machine vision
CN117553695A (en) * 2024-01-11 2024-02-13 摩斯智联科技有限公司 Method, apparatus and computer storage medium for calculating vehicle height based on aspect ratio suppression noise

Family Cites Families (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE1054757B (en) * 1957-08-02 1959-04-09 Robot Berning & Co Method and device for the photographic monitoring of the vehicle speed by means of stationary recording cameras
US3685012A (en) * 1970-04-16 1972-08-15 Sperry Rand Corp Apparatus for determining data associated with objects
US3788201A (en) * 1970-12-14 1974-01-29 F Abell Method for establishing vehicle identification, speed and conditions of visibility
US3890462A (en) * 1974-04-17 1975-06-17 Bell Telephone Labor Inc Speed and direction indicator for video systems
FR2315730A1 (en) * 1975-06-25 1977-01-21 Dassault Electronique Vehicle number plate reading unit - using television camera connected to threshold circuit
SE394146B (en) * 1975-10-16 1977-06-06 L Olesen SATURATION DEVICE RESP CONTROL OF A FOREMAL, IN ESPECIALLY THE SPEED OF A VEHICLE.
US4122521A (en) * 1975-10-20 1978-10-24 Northrop Corporation Correlation system for multiple-sensor reconnaissance vehicle
US4433325A (en) * 1980-09-30 1984-02-21 Omron Tateisi Electronics, Co. Optical vehicle detection system
JPS58140900A (en) * 1982-02-17 1983-08-20 三菱電機株式会社 Photographic apparatus for vehicle number plate
US4567609A (en) * 1983-03-28 1986-01-28 The United States Of America As Represented By The Secretary Of The Navy Automatic character recognition system
US4555618A (en) * 1983-06-02 1985-11-26 R L Associates Method and means for collecting highway tolls
SU1235002A1 (en) * 1983-07-13 1986-05-30 Предприятие П/Я А-1772 Television device for monitoring vehicle traffic
US4603390A (en) * 1984-03-05 1986-07-29 Soft Plus Corp. Computerized parking system
SU1193716A1 (en) * 1984-05-04 1985-11-23 Специальное Проектное Конструкторское Бюро Промышленной Автоматики "Промавтоматика" Device for determining velocities and types of objects
JPS6161573A (en) * 1984-08-31 1986-03-29 Mitsubishi Electric Corp Traffic monitor system
SU1234957A1 (en) * 1984-09-12 1986-05-30 Предприятие П/Я В-2769 Digital pulse-time-position discriminator
FR2583882B1 (en) * 1985-06-25 1989-02-03 Renault DEVICE FOR MEASURING THE SPEED AND THE POSITION OF A MOBILE IN RELATION TO THE GROUND
JPS6287947A (en) * 1985-10-14 1987-04-22 Matsushita Electric Ind Co Ltd Image pickup device
US4817166A (en) * 1986-05-05 1989-03-28 Perceptics Corporation Apparatus for reading a license plate
DD249783A1 (en) * 1986-06-05 1987-09-16 Wismar Ing Hochschule METHOD OF COUNTING, CLASSIFICATION AND VEHICLE SPEED DETERMINATION
FR2609566B1 (en) * 1987-01-14 1990-04-13 Armine METHOD FOR DETERMINING THE TRAJECTORY OF A BODY CAPABLE OF MOVING ON A TRACK AND DEVICE FOR IMPLEMENTING THE METHOD
US4847772A (en) * 1987-02-17 1989-07-11 Regents Of The University Of Minnesota Vehicle detection through image processing for traffic surveillance and control
FR2617309B1 (en) * 1987-06-29 1993-07-16 Cga Hbs SYSTEM FOR THE AUTOMATIC READING OF IDENTIFICATION DATA, APPOINTED ON A VEHICLE
GB2217498B (en) * 1988-04-18 1992-03-18 Ind Tech Res Inst Method and apparatus for automatically recognising licence plate characters
IL86202A (en) * 1988-04-27 1992-01-15 Driver Safety Systems Ltd Traffic safety monitoring apparatus
DE58902538D1 (en) * 1988-05-19 1992-12-03 Siemens Ag METHOD FOR OBSERVING A SCENE AND DEVICE FOR IMPLEMENTING THE METHOD.
GB2219881A (en) * 1988-06-15 1989-12-20 English Electric Valve Co Ltd Vehicle monitoring system
JPH01319644A (en) * 1988-06-18 1989-12-25 Furukawa Alum Co Ltd Heat-resistant aluminum alloy material and its manufacture
JPH01319898A (en) * 1988-06-21 1989-12-26 Mitsubishi Electric Corp Toll collecting device
US4855770A (en) * 1988-09-09 1989-08-08 Polaroid Corporation Vehicle identification camera
GB2227866B (en) * 1988-12-13 1993-05-05 Matsushita Electric Ind Co Ltd Automatic vehicle identification system
US4958064A (en) * 1989-01-30 1990-09-18 Image Recognition Equipment Corporation Bar code locator for video scanner/reader system
US5046111A (en) * 1989-02-09 1991-09-03 Philip Morris Incorporated Methods and apparatus for optically determining the acceptability of products
US5146510A (en) * 1989-02-09 1992-09-08 Philip Morris Incorporated Methods and apparatus for optically determining the acceptability of products
GB2231753A (en) * 1989-05-05 1990-11-21 Golden River Ltd Traffic incident monitoring with vido cameras
JPH0752480B2 (en) * 1989-06-09 1995-06-05 名古屋電機工業株式会社 Vehicle detection device
JPH0322099A (en) * 1989-06-19 1991-01-30 Matsushita Electric Ind Co Ltd Automatic recognizing device for number plate
DE3937603A1 (en) * 1989-11-11 1991-06-27 Wannke Bodo Vehicle license number reading system - has lens system mounted on adjustable carriage for training license numbers of passing vehicle
US5262871A (en) * 1989-11-13 1993-11-16 Rutgers, The State University Multiple resolution image sensor
JP2862605B2 (en) * 1989-12-08 1999-03-03 名古屋電機工業株式会社 Vehicle detecting device and detecting method thereof
JP2712844B2 (en) * 1990-04-27 1998-02-16 株式会社日立製作所 Traffic flow measurement device and traffic flow measurement control device
JPH04147400A (en) * 1990-10-11 1992-05-20 Matsushita Electric Ind Co Ltd Vehicle detecting apparatus
US5161107A (en) * 1990-10-25 1992-11-03 Mestech Creation Corporation Traffic surveillance system
US5243418A (en) * 1990-11-27 1993-09-07 Kabushiki Kaisha Toshiba Display monitoring system for detecting and tracking an intruder in a monitor area
FR2671653B1 (en) * 1991-01-11 1995-05-24 Renault MOTOR VEHICLE TRAFFIC MEASUREMENT SYSTEM.
FR2672144B1 (en) * 1991-01-30 1993-04-23 Angenieux P Ets DEVICE FOR INSPECTING VEHICLES AT THE LOCATION OF AN INSPECTION STATION OR THE LIKE.
US5301239A (en) * 1991-02-18 1994-04-05 Matsushita Electric Industrial Co., Ltd. Apparatus for measuring the dynamic state of traffic
US5296852A (en) * 1991-02-27 1994-03-22 Rathi Rajendra P Method and apparatus for monitoring traffic flow
CA2132346A1 (en) * 1992-03-18 1993-09-30 Robert Ciolli Vision apparatus
JP3011500U (en) 1994-06-06 1995-05-30 金四郎 伴 Window glass cleaning equipment

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210183042A1 (en) * 2018-12-03 2021-06-17 Mistras Group, Inc. Systems and methods for inspecting pipelines using a robotic imaging system
US11587217B2 (en) * 2018-12-03 2023-02-21 Mistras Group, Inc. Systems and methods for inspecting pipelines using a robotic imaging system
US11635391B2 (en) 2018-12-03 2023-04-25 Mistras Group, Inc. Systems and methods for inspecting pipelines using a pipeline inspection robot
US11887291B2 (en) 2018-12-03 2024-01-30 Mistras Group, Inc. Systems and methods for inspecting pipelines using a robotic imaging system
US11946882B2 (en) 2018-12-03 2024-04-02 Mistras Group, Inc. Systems and methods for inspecting pipelines using a pipeline inspection robot

Also Published As

Publication number Publication date
AU671106B2 (en) 1996-08-15
EP0631683A1 (en) 1995-01-04
JPH07505966A (en) 1995-06-29
CA2132515A1 (en) 1993-09-21
ATE203844T1 (en) 2001-08-15
AU3740293A (en) 1993-10-21
NZ249799A (en) 1996-11-26
WO1993019441A1 (en) 1993-09-30
US5809161A (en) 1998-09-15
EP0631683A4 (en) 1996-04-17
DE69330513D1 (en) 2001-09-06
EP0631683B1 (en) 2001-08-01

Similar Documents

Publication Publication Date Title
CA2132515C (en) An object monitoring system
US8284996B2 (en) Multiple object speed tracking system
CN106611512B (en) Method, device and system for processing starting of front vehicle
Tseng et al. Real-time video surveillance for traffic monitoring using virtual line analysis
CN109817014B (en) Roadside inspection parking charging method based on mobile video and high-precision positioning
US20110164789A1 (en) Detection of vehicles in images of a night time scene
Atkočiūnas et al. Image processing in road traffic analysis
WO2005036371A2 (en) Moving object detection using low illumination depth capable computer vision
CN103021175A (en) Pedestrian red light running video detection method and device based on Davinci architecture
CN110738150B (en) Camera linkage snapshot method and device and computer storage medium
CN110033622A (en) Violation snap-shooting based on unmanned plane aerial photography technology occupies Emergency Vehicle Lane method
CN106875696A (en) The system and method and drive recorder of automatic identification traffic violations illegal activities
CN111368612A (en) Overman detection system, personnel detection method and electronic equipment
CN115331191B (en) Vehicle type recognition method, device, system and storage medium
KR102306789B1 (en) License Plate Recognition Method and Apparatus for roads
CN113936257A (en) Detection method and detection device for vehicle violation behaviors and vehicle-mounted electronic equipment
Sharma et al. Automatic vehicle detection using spatial time frame and object based classification
JP2002190013A (en) System and method for detecting congestion by image recognition
WO2008088409A2 (en) Real-time dynamic content based vehicle tracking, traffic monitoring, and classification system
Auty et al. Image acquisition system for traffic monitoring applications
JP3649277B2 (en) Speed measurement system and speed measurement method by image recognition
CN115965636A (en) Vehicle side view generating method and device and terminal equipment
CN115761668A (en) Camera stain recognition method and device, vehicle and storage medium
Jie et al. Computer vision based real-time information acquisition for transport traffic
AU671106C (en) An object monitoring system

Legal Events

Date Code Title Description
EEER Examination request