EP1061487A1 - A method and device for automatically controlling a region in space - Google Patents

A method and device for automatically controlling a region in space Download PDF

Info

Publication number
EP1061487A1
EP1061487A1 EP99830376A EP99830376A EP1061487A1 EP 1061487 A1 EP1061487 A1 EP 1061487A1 EP 99830376 A EP99830376 A EP 99830376A EP 99830376 A EP99830376 A EP 99830376A EP 1061487 A1 EP1061487 A1 EP 1061487A1
Authority
EP
European Patent Office
Prior art keywords
region
volumetric map
volumetric
map
parameter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP99830376A
Other languages
German (de)
French (fr)
Inventor
Marco Aste
Massimo Boninsegna
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Istituto Trentino di Cultura
Original Assignee
Istituto Trentino di Cultura
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Istituto Trentino di Cultura filed Critical Istituto Trentino di Cultura
Priority to EP99830376A priority Critical patent/EP1061487A1/en
Publication of EP1061487A1 publication Critical patent/EP1061487A1/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19639Details of the system layout
    • G08B13/19641Multiple cameras having overlapping views on a single scene
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19639Details of the system layout
    • G08B13/19652Systems using zones in a single scene defined for different treatment, e.g. outer zone gives pre-alarm, inner zone gives alarm

Definitions

  • the present invention relates in a general way to the automatic monitoring of a region of space, particularly as regards the detection and location of bodies present within this region.
  • the ability to detect and locate bodies present within a region of space is useful when it is necessary to monitor the presence of people or objects in particular regions or subregions.
  • the presence of people in defined areas may be judged to endanger the people themselves, or the objects present within that area.
  • a volume of space in the vicinity of a machine that may produce chips or liquids dangerous to man; or the volume represented by the radius of action of a mechanical arm; or regions of space close to equipment operating at high voltages; or more generally any region in which machines dangerous to man are operating may be considered dangerous.
  • the desirability of being able to automatically monitor and report the presence of people in a region of space regarded as dangerous for human activity may be considered dangerous.
  • Another example of an application is the protection of objects of artistic value in museum environments or in any situation in which the presence of people within a certain region can be a source of danger for the objects present in that area.
  • regions of space are monitored automatically by devices such as optical barriers of transmission and/or reflection type (typically based on infrared technology), physical barriers, pressure-sensitive mats, movement detectors based on microwaves, passive infrared or ultrasound, radar systems, and devices that use laser beams to detect the presence and position of objects.
  • devices such as optical barriers of transmission and/or reflection type (typically based on infrared technology), physical barriers, pressure-sensitive mats, movement detectors based on microwaves, passive infrared or ultrasound, radar systems, and devices that use laser beams to detect the presence and position of objects.
  • the object of this invention is therefore to provide a solution for the automatic detection of bodies within a defined region that is simple, reliable, easily set up and capable of discriminating between objects for shape and volume while also considering the movement and relative path of the objects within the monitored area.
  • the solution according to the invention is capable of automatically detecting, locating and reporting the presence of bodies within a monitored region of space.
  • the invention is also capable of discriminating with a high degree of reliability between objects of different shapes and sizes and is therefore able to select, for monitoring purposes, only one particular type of body, e.g. only people.
  • the solution according to the invention is capable of detecting with a high degree of reliability the simultaneous presence of bodies, not necessarily of the same shape, in the monitored region, and selecting for monitoring purposes only those bodies that present certain distinctive features, for example only people or objects above a certain height.
  • the quantity of information obtainable with the solution according to the invention is very much greater than could be obtained by means of conventional techniques of automatic detection, and makes possible more reliable and robust location of bodies within the monitored area, overcoming the limitations from which known systems usually suffer and so enabling it to be adapted more successfully to different working conditions and hence giving it greater generality of use.
  • the solution according to the invention is able to carry out a volumetric analysis and so extract the characteristics of form, volume and dimensions which distinguish an object or person which it is wished to pick out from other artefacts or objects that should be inside the monitored region.
  • a volumetric analysis is able to carry out a volumetric analysis and so extract the characteristics of form, volume and dimensions which distinguish an object or person which it is wished to pick out from other artefacts or objects that should be inside the monitored region.
  • volumetric map means any representation of occupied volumes due to the presence of a body, in other words a representation of a three-dimensional map in which the regions of volumetric occupation introduced by the presence of bodies are indicated.
  • Such a map is obtained after image analysis procedures have been carried out using automatic methods known per se or according to the embodiments of the invention described below.
  • the reference S indicates a region of space in which it is wished to detect the presence of people A or objects B.
  • the region S may be bounded by physical walls , as for example in the case of a room or cage, or may consist simply of a portion of space bounded by an imaginary closed surface that separates a generic space in two regions, or it may be bounded partly by physical barriers, for example the floor, and partly by an imaginary surface.
  • the monitored region has however the feature of a volume, may be of any shape and can be defined simply and flexibly according to need.
  • the volumes occupied by the bodies (such as bodies A, B visible in Figures 1 and 3) that are present in the monitored region S are found by using two or more video cameras (acting as image signal generating means) installed in such a way that the region S is in the visual field of at least two video cameras 2, as shown for example in Figures 1 and 3 (the latter figure referring to a solution in which four video cameras 2 are used). It is advisable for the video cameras 2 to be so positioned as to avoid occlusions due to the movement of objects on the same plane; for example, for bodies of different heights standing on the floor and moving about, it is preferable to have views from above.
  • the signals (of analogue type or already directly converted into digital form) output by the video cameras 2 are sent to a processing unit 1 which may be a specialized processor or, in the currently preferred embodiment of the invention, a computer such as a programmed personal computer (known per se) in order to extract from the images the shapes of the bodies A and B present within the region S to be monitored.
  • a processing unit 1 which may be a specialized processor or, in the currently preferred embodiment of the invention, a computer such as a programmed personal computer (known per se) in order to extract from the images the shapes of the bodies A and B present within the region S to be monitored.
  • the object here is to check for the possible presence of bodies not inherently belonging to the monitored region.
  • Figures al and bl included in Figure 2 and Figures al, bl, cl and dl included in Figure 4 show the images produced by the two video cameras depicted in Figure 1, on the one hand, and by the four video cameras depicted in Figure 3, on the other.
  • the unit 1 is able, in accordance with known principles (typically by using known image processing algorithms), to extract respective sets of data representing the abovementioned shapes of the bodies present within the region S.
  • Figures a2 and b2 of Figure 2 and Figures a2, b2, c2 and d2 of Figure 3 show the shapes, marked C of the body A present within the region S corresponding to images a1 and b1, or a1, b1, c1 and d1, produced, from their respective points of observation, by the video cameras 2.
  • the abovementioned shapes correspond only to bodies possessed of movement in themselves, thus eliminating - for the purposes of the subsequent processing - information corresponding to fixed parts of the image, such as the outlines of the region S represented by broken lines in parts a2 and b2, and a2, b2, c2 and d2 in Figures 2 and 4, or to the object B.
  • the region S is a room in a museum
  • the body A is the body of a visitor moving about in the room
  • the body B is a bench situated - in a fixed position of course - in the centre of the room.
  • objects having their own movement can be distinguished from fixed objects/items in such a way as to avoid, on the one hand, deception relating to the presence of an intruder body attempting to evade detection by moving very slowly and, on the other hand, the generation of false objects connected with, for example, vibratory movements, draughts, etc.
  • the processing unit 1 can be programmed to find matches on the images between projections of the same real-world points belonging to the detected bodies. In this way, information is extracted from each image frame about the positions of the projections of the same real-world points onto the different image planes corresponding to the different observation points identified by the video cameras 2 (or equivalent image signal generating means).
  • the availability of a larger number of observation points, and hence of a larger number of images gives a certain degree of redundancy which can be used to increase the accuracy and reliability of the detection action.
  • Figure 5 shows the currently preferred method of producing the volumetric map of bodies in the monitored region S.
  • the processing unit 1 is programmed, for the purposes of processing the signals generated by the video cameras 2, to divide up the entire volume of the region S into volumetric cells of fixed dimensions (some of these are shown diagrammatically at D1 and D2 in Figure 5), each cell corresponding to a portion of the real space inside the region S.
  • the region S is a room in a museum
  • this is not of course a limiting option as it has to do with the spatial resolution with which it is wished to detect and locate the objects: the smaller the cells, the greater the resolution and vice versa.
  • each three-dimensional cell is projected onto the image plane of each video camera.
  • each cell there is a certain area E on each image, as shown at the top of Figure 5, the bottom part of which meanwhile shows the location of the cells D1 and D2 within the three-dimensional Cartesian reference system used for locating the cells in question within the region S. Cells corresponding to regions of volume that are not covered by at least two video cameras are ignored.
  • the volumetric map F (see Figure 6) is obtained by checking each cell to see whether the areas corresponding to that cell's projection on the different images represent some portion of the objects present within the monitored region. If the outcome of the check is positive, that cell is judged to be occupied (for example, D1 is an example of this); otherwise it is judged to be unoccupied, as in the case of the cells marked D2. In this way the set of all occupied cells in each frame provides information about the volumetric occupation of the monitored region. From this information a volumetric map of occupation can be constructed almost immediately by checking which and how many of the cells into which the monitored space has been divided are occupied by objects. The aim of all this is to obtain, as a result, the representation seen in Figure 6.
  • the volumetric map approximates to the volume of actual occupation with a resolution based on the dimensions of the cells D1, D2 and on the spatial resolution of the means employed to generate the image signals (in particular, in the case of video cameras composed of a matrix of sensitive points, the resolution in a particular direction is given by the ratio of the dimension of the observed region to the number of sensitive elements in that direction).
  • the dimensions of the cells are defined having regard not only to the resolution but also to the processing capacity of the unit 1 and of the frequency with which the surveillance information is to be updated. Also borne in mind is the fact that the overall degree of approximation can be enhanced, as already stated, by using more video cameras in suitable positions.
  • the processing unit 1 can be programmed, again in a known manner, to carry out a volumetric analysis of bodies for the purpose of recognizing the distinctive characteristics of shape and volume of the objects analysed.
  • the programming can be done by conventional algorithmic approaches, thus coding ab initio the characteristics of shape and volume of the bodies to be detected into the processing system, or using statistical learning techniques such as for example neural networks.
  • a step 101 the unit 1 examines the data set corresponding to the images generated by the video cameras 2 (optionally already processed to refer only to moving objects) and in a second step 102 commences an action of scanning the region S in such a way as to scan the cells D into which the region S has been theoretically divided up.
  • each of these cells will be identified by three co-ordinates x i , y i , z i within the system x, y and z to which the bottom part of Figure 5 refers.
  • the unit 1 is also capable of detecting, for example, the entry into the region S of a body not previously present, the aim being to extend the scanning action to those cells (previously not included in the scanning action) which the body subsequently occupies.
  • the steps marked 1031, 1041; 1032, 1042; ...; 103n, 104n indicate successive processing stages, here shown as carried out in parallel, though in fact they can be performed serially, and therefore sequentially in time.
  • n is equal to 2, and to 4, in the illustrative embodiments shown in Figures 1 and 3, respectively
  • the unit 1 checks to see whether the cells corresponding to their respective images generated by the video cameras 2 can be regarded as occupied or unoccupied.
  • the results of the comparisons carried out in steps 1041, 1042, ..., 104n are processed in order to decide whether, on the basis of the image data, the scanned cell is to be regarded as occupied or unoccupied for the purposes of constructing the map of volumetric occupation.
  • attribution of the "occupied" value may be based on a different number of decision processes relating to the individual images than the number of images taken into consideration in attributing the "occupied" logic value to other cells.
  • the criterion used in attributing the logic value in question may be of unanimous type (the cell is judged to be occupied for the purposes of the construction of the map of volumetric occupation if and only if all the video cameras 2 whose images are taken into account produce data corresponding to occupation in the relevant image), majority type (the cell is judged to be occupied if the majority of video cameras 2 give data indicating occupation in the respective images), or correlation with the values attributed to adjacent cells (so that uncertainty in the attribution of the "occupied" value to a cell is resolved on the basis of confident values attributed to spatially adjacent cells) or different again, according to well-known criteria in the image processing field.
  • the step 106 in Figure 7 represents simply the selection step where it is decided whether or not the scan of the region S (or of the scanned subregion thereof) can be said to be complete.
  • step 106 If the result of the comparison in step 106 is positive, this indicates that the map of volumetric occupation is complete.
  • the map itself which can be represented as illustrated diagrammatically in Figure 6 (which must of course be understood to be a perspective representation of a data set which in reality is three-dimensional), is subjected to a processing step 107 for the extraction of a set of parameters P which represent in compact form the shape, volume, position and/or dimensions of the detected bodies.
  • a processing step 107 for the extraction of a set of parameters P which represent in compact form the shape, volume, position and/or dimensions of the detected bodies.
  • the map of volumetric occupation that is the set of occupied cells and their positions in space
  • the position of the centre of mass of each volumetric map can be used to represent the position of the body
  • the dimensions of width, length and height of the smallest parallelepiped which inscribes the occupied volume can be used as dimensional parameters.
  • Requicha A.G. "Representation of rigid solids: theory, methods and systems", Comp. Surveys, vol. 12, pp. 437-464, 1980
  • Requicha A.G. and Rossignac J.R. “Solid modeling and beyond”, IEEE Computer Graphics and Applications, vol. 12, pp. 31-44, 1992
  • Aggarwaal J.K. and Cai Q. "Human motion analysis: a review", Proceedings of IEEE Computer Society Workshop on Motion of Non-Rigid and Articulated Objects, pp. 90-102, 1997.
  • step 108 at least one of the parameters P obtained in this way is compared with a predetermined "model".
  • the purpose of this is to establish whether or not the map F, corresponding to the position, size and shape of a body such as the body C, is "compatible" with the criteria of monitoring or surveillance which the system according to the invention has to follow.
  • One possible model for comparison may correspond to a defined part of the region of space S in which the body C must come no closer than a limiting distance.
  • compatibility is checked by using, for example, the parameters P relating to the position and dimensions of the detected bodies.
  • the volume D corresponding to the region of space S which the body C must not enter may be an area that must be respected around a work of art exhibited in a museum (e.g. a picture hanging on a wall).
  • the volume D may be a zone that must be respected around a machine with moving parts or with exposed parts at a high temperature and/or voltage.
  • step 108 the unit checks (by applying known criteria) that, for example, none of the cells contained within the map of volumetric occupation F falls inside the volume D or is at a distance less than a minimum safety distance from the volume D.
  • the unit 1 prepares itself to repeat the monitoring action with reference to the next set of images taken by the video camera 2. The processing action thus returns upstream of step 101.
  • the processing action moves on from step 108 to a new step 109 corresponding to the emission of a warning signal.
  • a warning signal This may be represented by e.g. an acoustic or visual alarm signal (optionally at a distance, aimed at a manned remote control station) emitted by a corresponding device 3.
  • the device 3 must be understood to be of known type, depending on the alarm signal which it is wished to produce: it may for example be a siren, an acoustic indicator, a remote warning system, etc., connected to the unit 1.
  • the volumetric map of each body detected inside the monitored region and/or the manner in which it changes while the bodies are present in the monitored region can be compared with models of volumetric maps for other bodies, encoded in a similar manner and previously stored in the processing unit (take for example the model marked D in Figure 6) in order to recognize those bodies which must be detected from among all the bodies present inside the monitored region.
  • the bodies may for example be people only.
  • the solution described is highly robust and overcomes the functional limitations of currently used systems. Thus, it is capable of detecting the presence and at the same time determining the position of people or objects within a defined region of space, discriminate between objects and people, between objects or people close to each other, and between objects and people that move into the monitored region following different paths or more generally with behaviours which could easily deceive other types of sensor.
  • the method according to the invention can be carried out using, at least in part, a computer program capable of being run on a computer in such a way that the system comprising the program and the computer carries out the method according to the invention.
  • the invention therefore extends also to such a program capable of being loaded into a computer which has the means of or is capable of carrying out the method according to the invention, as well as to the corresponding information technology product comprising a means readable by a computer containing codes for a computer program which, when the program is loaded into the computer, cause the computer to carry out the method according to the invention.

Abstract

The monitored region (S) is monitored by image signal generating means such as video cameras (2) in order to obtain a succession of images of the bodies (A, B) present in the monitored region, each image corresponding to a defined instant. The images are processed in such a way as to obtain, for each instant considered, a volumetric map of each body present in the region (S). This map, which identifies characteristics of shape, position, volume and dimensions of the body to which it refers, is processed in order to extract from it at least one parameter selected from the following group: descriptors of shape and volume, such as the volumetric map itself, the co-ordinates of position and the dimensions of each body to which the volumetric map refers. The parameter or a succession of values of the parameter obtained in this way is then compared with at least one model of these characteristics stored in a processing unit (1). Depending on the outcome of this comparison operation, a procedure of surveillance and/or reporting may be selectively activated. The solution is applicable, for example, to the automatic monitoring of museum environments, e.g. to ensure that visitors do not come too close to an exhibited work, or to the monitoring of industrial environments, e.g. to ensure that an operator does not come too close to a dangerous machine or process.

Description

  • The present invention relates in a general way to the automatic monitoring of a region of space, particularly as regards the detection and location of bodies present within this region.
  • The ability to detect and locate bodies present within a region of space is useful when it is necessary to monitor the presence of people or objects in particular regions or subregions.
  • For example, for reasons of security, the presence of people in defined areas may be judged to endanger the people themselves, or the objects present within that area. To take a particular case, in an industrial environment a volume of space in the vicinity of a machine that may produce chips or liquids dangerous to man; or the volume represented by the radius of action of a mechanical arm; or regions of space close to equipment operating at high voltages; or more generally any region in which machines dangerous to man are operating, may be considered dangerous. Hence the desirability of being able to automatically monitor and report the presence of people in a region of space regarded as dangerous for human activity.
  • Another example of an application is the protection of objects of artistic value in museum environments or in any situation in which the presence of people within a certain region can be a source of danger for the objects present in that area.
  • At the present time, regions of space are monitored automatically by devices such as optical barriers of transmission and/or reflection type (typically based on infrared technology), physical barriers, pressure-sensitive mats, movement detectors based on microwaves, passive infrared or ultrasound, radar systems, and devices that use laser beams to detect the presence and position of objects.
  • In many cases it is difficult to use these techniques for reasons of practicality, layout and reliability or environmental compatibility. Known systems, in fact, have intrinsic limitations which make their use difficult. For example, many known systems are sensitive to noise, dust and dirt and are therefore unsuitable for use in industrial working environments. Other limitations have to do with the difficulty of discriminating the size of the detected object and/or the inability to analyse the behaviour and motion of bodies in the vicinity of and within the monitored region of space. Again, some systems are unsuitable as being too invasive in environments which should be respected such as museums, or more generally places of great historical and artistic worth. Moreover, many of the known systems can easily be deceived, while others, such as physical barriers, may be unacceptable for reasons of safety and/or practicality; while others, such as those based on the emission of electromagnetic radiation, may not be tolerated because of their interference with other equipment, the difficulty of setting them up and in some cases the danger which they may present to biological organisms.
  • Many of the problems cited above can be overcome by monitoring the observed region with image signal generating means represented - in commonly used surveillance systems - by video cameras, sometimes of the type often known as "slow video". These systems however have the drawback that they require the constant presence of a human operator if they are to be of any real benefit.
  • The object of this invention is therefore to provide a solution for the automatic detection of bodies within a defined region that is simple, reliable, easily set up and capable of discriminating between objects for shape and volume while also considering the movement and relative path of the objects within the monitored area.
  • According to this invention, this object is achieved with a method having the characteristics claimed specifically in the following claims. The invention also relates to apparatus for carrying out this method.
  • In summary, the solution according to the invention is capable of automatically detecting, locating and reporting the presence of bodies within a monitored region of space. The invention is also capable of discriminating with a high degree of reliability between objects of different shapes and sizes and is therefore able to select, for monitoring purposes, only one particular type of body, e.g. only people.
  • In particular, the solution according to the invention is capable of detecting with a high degree of reliability the simultaneous presence of bodies, not necessarily of the same shape, in the monitored region, and selecting for monitoring purposes only those bodies that present certain distinctive features, for example only people or objects above a certain height. In addition, the quantity of information obtainable with the solution according to the invention is very much greater than could be obtained by means of conventional techniques of automatic detection, and makes possible more reliable and robust location of bodies within the monitored area, overcoming the limitations from which known systems usually suffer and so enabling it to be adapted more successfully to different working conditions and hence giving it greater generality of use.
  • In specific terms, the solution according to the invention is able to carry out a volumetric analysis and so extract the characteristics of form, volume and dimensions which distinguish an object or person which it is wished to pick out from other artefacts or objects that should be inside the monitored region. For example, in order to recognize the presence of a person, it is possible to use the a priori knowledge that a person possesses a certain shape and therefore produces a characteristic occupied volume.
  • With the invention it is therefore possible to discriminate between objects and people regardless of how they are moving and to use the information obtainable by the method in order to define different degrees of danger and/or alarm as a consequence of the presence of bodies within defined subregions of space.
  • The invention will now be described, purely by way of non-restrictive example, with reference to the attached drawings, in which:
    • Figure 1 shows diagrammatically the characteristics of the system according to the invention used for the monitoring and automatic surveillance of a defined region of space,
    • Figure 2, comprising four parts respectively labelled a1-a2 and b1-b2, illustrates the generation of image signals within the context of the solution illustrated in Figure 1,
    • Figures 3 and 4 illustrate, in ways basically identical to those of Figures 1 and 2, another possible embodiment of the solution according to the invention; in particular, Figure 4 is composed of eight parts respectively labelled a1-a2, b1-b2, c1-c2 and d1-d2,
    • Figure 5 illustrates the methods adopted for calculating a so-called map of volumetric occupation,
    • Figure 6 illustrates schematically one of these maps capable of being obtained within the context of the invention, and
    • Figure 7 is a flow diagram relating to the generation and use of such a map.
  • In particular the expression "volumetric map" as used here means any representation of occupied volumes due to the presence of a body, in other words a representation of a three-dimensional map in which the regions of volumetric occupation introduced by the presence of bodies are indicated. Such a map is obtained after image analysis procedures have been carried out using automatic methods known per se or according to the embodiments of the invention described below. For a summary of some of these methods the following may usefully be referred to: Marr D., "Vision", Freeman, 1982; Ballard D.H. and Brown C.M. "Computer Vision", Prentice Hall, 1982; Martin W.N. and Aggarwaal J.K., "Volumetric description of objects from multiple views", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 5, pp.150-158, 1983. From this map, by means of the volumetric analysis carried out using automatic methods known per se it is possible to derive the characteristics of shape, volume, dimensions and position of the bodies present in a defined region of space in such a way that they can easily be compared with similar representations obtained from the volumetric maps of other bodies.
  • In both Figure 1 and Figure 3, the reference S indicates a region of space in which it is wished to detect the presence of people A or objects B.
  • The region S may be bounded by physical walls , as for example in the case of a room or cage, or may consist simply of a portion of space bounded by an imaginary closed surface that separates a generic space in two regions, or it may be bounded partly by physical barriers, for example the floor, and partly by an imaginary surface. The monitored region has however the feature of a volume, may be of any shape and can be defined simply and flexibly according to need.
  • In the currently preferred embodiment of the invention, the volumes occupied by the bodies (such as bodies A, B visible in Figures 1 and 3) that are present in the monitored region S are found by using two or more video cameras (acting as image signal generating means) installed in such a way that the region S is in the visual field of at least two video cameras 2, as shown for example in Figures 1 and 3 (the latter figure referring to a solution in which four video cameras 2 are used). It is advisable for the video cameras 2 to be so positioned as to avoid occlusions due to the movement of objects on the same plane; for example, for bodies of different heights standing on the floor and moving about, it is preferable to have views from above.
  • The signals (of analogue type or already directly converted into digital form) output by the video cameras 2 are sent to a processing unit 1 which may be a specialized processor or, in the currently preferred embodiment of the invention, a computer such as a programmed personal computer (known per se) in order to extract from the images the shapes of the bodies A and B present within the region S to be monitored. The object here is to check for the possible presence of bodies not inherently belonging to the monitored region.
  • In particular, Figures al and bl included in Figure 2, and Figures al, bl, cl and dl included in Figure 4 show the images produced by the two video cameras depicted in Figure 1, on the one hand, and by the four video cameras depicted in Figure 3, on the other.
  • Using the signals corresponding to the abovementioned images, the unit 1 is able, in accordance with known principles (typically by using known image processing algorithms), to extract respective sets of data representing the abovementioned shapes of the bodies present within the region S. For example, Figures a2 and b2 of Figure 2, and Figures a2, b2, c2 and d2 of Figure 3 show the shapes, marked C of the body A present within the region S corresponding to images a1 and b1, or a1, b1, c1 and d1, produced, from their respective points of observation, by the video cameras 2.
  • For an overview of these algorithms, the following may usefully be referred to: Huang T.S., "Image sequence processing and dynamic scene analysis", Springer-Verlag, 1982; Jain A., "Fundamentals of digital image processing", Prentice Hall, 1989; Jain J.R. and Martin W.N. and Aggarwaal J.K., "Segmentation through the detection of changes due to motion", Computer Graphics and Image Processing, vol. 11, pp.13-34, 1979; Debuisson M.-P., "Contour extraction of moving objects in complex outdoor scenes", Int. Journal of Computer Vision, vol. 14, pp.83-105, 1995; Bichsel M., "Segmenting Simply Connected Moving Objects in a Static Scene", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 16, n. 11, pp.1138-1142, 1994, as well as to US-A-5,212,547 or US-A-5,877,804.
  • In particular it is possible to have the abovementioned shapes correspond only to bodies possessed of movement in themselves, thus eliminating - for the purposes of the subsequent processing - information corresponding to fixed parts of the image, such as the outlines of the region S represented by broken lines in parts a2 and b2, and a2, b2, c2 and d2 in Figures 2 and 4, or to the object B. As an example, it may be imagined that the region S is a room in a museum, the body A is the body of a visitor moving about in the room, and the body B is a bench situated - in a fixed position of course - in the centre of the room.
  • In particular, it is known that objects having their own movement can be distinguished from fixed objects/items in such a way as to avoid, on the one hand, deception relating to the presence of an intruder body attempting to evade detection by moving very slowly and, on the other hand, the generation of false objects connected with, for example, vibratory movements, draughts, etc.
  • In one embodiment of the invention, the processing unit 1 can be programmed to find matches on the images between projections of the same real-world points belonging to the detected bodies. In this way, information is extracted from each image frame about the positions of the projections of the same real-world points onto the different image planes corresponding to the different observation points identified by the video cameras 2 (or equivalent image signal generating means). In particular it will be realized that, other parameters being equal, the availability of a larger number of observation points, and hence of a larger number of images (e.g. four rather than two) gives a certain degree of redundancy which can be used to increase the accuracy and reliability of the detection action.
  • Given a knowledge of the intrinsic and extrinsic parameters of the video cameras 2 - i.e. the focal distances of each lens fitted to the video cameras, the resolution of the sensor of each, the alignment of the optical axis with the centre of the active sensor of the video camera, the spatial co-ordinates relative to the origin of a reference system and the inclinations with respect to the axes of the reference system - the three-dimensional spatial positions of all observed points are unambiguously determined.
  • From these positions it is possible, with known methods, to extract the shape of the external surface of the objects present within the region S or an approximation thereto and from this to derive the volumetric map.
  • For an overview of the abovementioned methods the following may usefully be referred to: Boll R.M. and Vemuri B.C., "On three dimensional surface reconstruction methods", IEEE Transactions on Pattern Analysis and Computer Intelligence, vol. 13, no. 1, pp. 1-13, 1991; Besl P. and Jain R., "Three dimensional object recognition", Comp. Surveys, vol. 17, pp. 75-145, 1985; Aggarwaal J.K. et al, "Survey: representation methods of three dimensional objects", Progress in Pattern Recognition, vol. 1, North Holland, 1981; Morasso P. and Sandini G., "3D reconstruction from multiple stereo views", Proceedings 3rd International Conference on Image Analysis and Processing, 1985.
  • Using these techniques there may be uncertainty in the identification of surface parts of an object that are not seen simultaneously by at least two video cameras. This problem can be corrected by selecting a suitable location for another video camera or drawing on a priori information about the shape of the observed objects.
  • In all cases the characteristics of volume, shape, position and dimensions of the bodies in question can be represented by values associated with a finite set of parameters P.
  • Figure 5 shows the currently preferred method of producing the volumetric map of bodies in the monitored region S.
  • More specifically, the processing unit 1 is programmed, for the purposes of processing the signals generated by the video cameras 2, to divide up the entire volume of the region S into volumetric cells of fixed dimensions (some of these are shown diagrammatically at D1 and D2 in Figure 5), each cell corresponding to a portion of the real space inside the region S. For example, if the region S is a room in a museum, it may be decided to define the cells in question as cubic volumes with sides of, for example, ten centimetres. However, this is not of course a limiting option as it has to do with the spatial resolution with which it is wished to detect and locate the objects: the smaller the cells, the greater the resolution and vice versa.
  • Using a perspective projection function, each three-dimensional cell is projected onto the image plane of each video camera. In general, therefore, for each cell there is a certain area E on each image, as shown at the top of Figure 5, the bottom part of which meanwhile shows the location of the cells D1 and D2 within the three-dimensional Cartesian reference system used for locating the cells in question within the region S. Cells corresponding to regions of volume that are not covered by at least two video cameras are ignored.
  • The volumetric map F (see Figure 6) is obtained by checking each cell to see whether the areas corresponding to that cell's projection on the different images represent some portion of the objects present within the monitored region. If the outcome of the check is positive, that cell is judged to be occupied (for example, D1 is an example of this); otherwise it is judged to be unoccupied, as in the case of the cells marked D2. In this way the set of all occupied cells in each frame provides information about the volumetric occupation of the monitored region. From this information a volumetric map of occupation can be constructed almost immediately by checking which and how many of the cells into which the monitored space has been divided are occupied by objects. The aim of all this is to obtain, as a result, the representation seen in Figure 6. Using this technique, the volumetric map approximates to the volume of actual occupation with a resolution based on the dimensions of the cells D1, D2 and on the spatial resolution of the means employed to generate the image signals (in particular, in the case of video cameras composed of a matrix of sensitive points, the resolution in a particular direction is given by the ratio of the dimension of the observed region to the number of sensitive elements in that direction). The dimensions of the cells are defined having regard not only to the resolution but also to the processing capacity of the unit 1 and of the frequency with which the surveillance information is to be updated. Also borne in mind is the fact that the overall degree of approximation can be enhanced, as already stated, by using more video cameras in suitable positions.
  • In particular, the processing unit 1 can be programmed, again in a known manner, to carry out a volumetric analysis of bodies for the purpose of recognizing the distinctive characteristics of shape and volume of the objects analysed. The programming can be done by conventional algorithmic approaches, thus coding ab initio the characteristics of shape and volume of the bodies to be detected into the processing system, or using statistical learning techniques such as for example neural networks. It is also possible to design the unit 1 such that it is able to evaluate the way the positions of the bodies examined within the region S are changing in space and time and deduce the dynamics of their movements, in particular the line, and the direction along this line, of their displacements.
  • This point will become clearer on referring to the flow diagram given in Figure 7, which shows, in a deliberately schematic way, for ease of comprehension, the principles by which the functions of automatic monitoring are carried out in the unit 1.
  • Assuming the process to start at a starting step 100, in a step 101 the unit 1 examines the data set corresponding to the images generated by the video cameras 2 (optionally already processed to refer only to moving objects) and in a second step 102 commences an action of scanning the region S in such a way as to scan the cells D into which the region S has been theoretically divided up. As a general rule, each of these cells will be identified by three co-ordinates xi, yi, zi within the system x, y and z to which the bottom part of Figure 5 refers.
  • From now on it will be assumed, for simplicity, that this scanning operation applies, on each successive detection of the images of the region S, to all the cells contained within the region S scanned on a "matrix" principle, for example in successive lines (co-ordinate x), columns (co-ordinate y) and planes (co-ordinate z).
  • Those skilled in the art of image processing will have realized that it is possible (e.g. in order to reduce the processing cost and/or speed up the processing) to adopt different scanning systems, such as predictive-type scanning systems which, once initialized with reference to a map of initial volumetric occupation, perform subsequent scans only on cells where there exists some degree of likelihood inherent in the fact that these cells may be significant in the generation of subsequent maps, the aim being to avoid the need to perform exhaustive scanning of the entire region S for each updating operation.
  • In this context it is also known that it is possible to intervene in such a way that, when operating on the abovementioned principles, the unit 1 is also capable of detecting, for example, the entry into the region S of a body not previously present, the aim being to extend the scanning action to those cells (previously not included in the scanning action) which the body subsequently occupies.
  • The steps marked 1031, 1041; 1032, 1042; ...; 103n, 104n indicate successive processing stages, here shown as carried out in parallel, though in fact they can be performed serially, and therefore sequentially in time. In the course of these steps, for each video camera 1, ..., n (n is equal to 2, and to 4, in the illustrative embodiments shown in Figures 1 and 3, respectively) and for each cell D(xi, yi, zi) that is scanned, the unit 1 checks to see whether the cells corresponding to their respective images generated by the video cameras 2 can be regarded as occupied or unoccupied.
  • In the next step, indicated by the general reference 105, the results of the comparisons carried out in steps 1041, 1042, ..., 104n are processed in order to decide whether, on the basis of the image data, the scanned cell is to be regarded as occupied or unoccupied for the purposes of constructing the map of volumetric occupation.
  • The relevant criteria for attributing the "occupied" or "unoccupied" logic value may differ.
  • On this subject it should be remembered that the cells of the region S are not necessarily all covered by all of the video cameras 2. As a consequence, in the case of certain cells, attribution of the "occupied" value may be based on a different number of decision processes relating to the individual images than the number of images taken into consideration in attributing the "occupied" logic value to other cells.
  • The criterion used in attributing the logic value in question may be of unanimous type (the cell is judged to be occupied for the purposes of the construction of the map of volumetric occupation if and only if all the video cameras 2 whose images are taken into account produce data corresponding to occupation in the relevant image), majority type (the cell is judged to be occupied if the majority of video cameras 2 give data indicating occupation in the respective images), or correlation with the values attributed to adjacent cells (so that uncertainty in the attribution of the "occupied" value to a cell is resolved on the basis of confident values attributed to spatially adjacent cells) or different again, according to well-known criteria in the image processing field.
  • The step 106 in Figure 7 represents simply the selection step where it is decided whether or not the scan of the region S (or of the scanned subregion thereof) can be said to be complete.
  • If the answer to this is negative, the process returns upstream of the step 102 and another cell is analysed.
  • If the result of the comparison in step 106 is positive, this indicates that the map of volumetric occupation is complete. At this point the map itself, which can be represented as illustrated diagrammatically in Figure 6 (which must of course be understood to be a perspective representation of a data set which in reality is three-dimensional), is subjected to a processing step 107 for the extraction of a set of parameters P which represent in compact form the shape, volume, position and/or dimensions of the detected bodies. By using algorithms that search for connected regions, it is possible to separate out elements of the volumetric map which are associated with different bodies; by this means it is possible to derive a volumetric map for every detected body. As a rule, for each volumetric map the characteristic parameters can be found by using automatic methods that are known in themselves. For instance, the map of volumetric occupation, that is the set of occupied cells and their positions in space, can be used directly as the volume parameter, the position of the centre of mass of each volumetric map can be used to represent the position of the body, and the dimensions of width, length and height of the smallest parallelepiped which inscribes the occupied volume can be used as dimensional parameters.
  • For an overview of some of the abovementioned methods, the following may usefully be referred to: Requicha A.G., "Representation of rigid solids: theory, methods and systems", Comp. Surveys, vol. 12, pp. 437-464, 1980; Requicha A.G. and Rossignac J.R., "Solid modeling and beyond", IEEE Computer Graphics and Applications, vol. 12, pp. 31-44, 1992; Aggarwaal J.K. and Cai Q., "Human motion analysis: a review", Proceedings of IEEE Computer Society Workshop on Motion of Non-Rigid and Articulated Objects, pp. 90-102, 1997.
  • In the next step 108 at least one of the parameters P obtained in this way is compared with a predetermined "model". The purpose of this is to establish whether or not the map F, corresponding to the position, size and shape of a body such as the body C, is "compatible" with the criteria of monitoring or surveillance which the system according to the invention has to follow.
  • One possible model for comparison may correspond to a defined part of the region of space S in which the body C must come no closer than a limiting distance. In this case compatibility is checked by using, for example, the parameters P relating to the position and dimensions of the detected bodies.
  • To take a concrete example, in Figure 6 the volume D corresponding to the region of space S which the body C must not enter may be an area that must be respected around a work of art exhibited in a museum (e.g. a picture hanging on a wall). To take another example, such as industrial equipment, the volume D may be a zone that must be respected around a machine with moving parts or with exposed parts at a high temperature and/or voltage.
  • In practice, in step 108 the unit checks (by applying known criteria) that, for example, none of the cells contained within the map of volumetric occupation F falls inside the volume D or is at a distance less than a minimum safety distance from the volume D.
  • If this condition is not found, so that the map F is compatible with the abovementioned model (to refer to the examples discussed above: the visitor has kept away from the picture hanging on the wall or the machine operator has kept a safe distance from the dangerous machine), the unit 1 prepares itself to repeat the monitoring action with reference to the next set of images taken by the video camera 2. The processing action thus returns upstream of step 101.
  • If, however, the map F is found to be incompatible with the model (for example because the visitor is found to have moved too close to the picture, or the machine operator has moved too close to the dangerous machine), the processing action moves on from step 108 to a new step 109 corresponding to the emission of a warning signal. This may be represented by e.g. an acoustic or visual alarm signal (optionally at a distance, aimed at a manned remote control station) emitted by a corresponding device 3. The device 3 must be understood to be of known type, depending on the alarm signal which it is wished to produce: it may for example be a siren, an acoustic indicator, a remote warning system, etc., connected to the unit 1.
  • From the above description it will be clear that by using the volumetric map describing the objects present in the monitored region S and obtained for example by the means described above or by the equivalent methods, and the manner in which it changes over time, it is possible to derive a description of the shapes of the objects and of their movements within the region S by encoding the information in numerical strings which describe at least one of the values of the characteristics of position, shape, dimensions and volume. The volumetric map of each body detected inside the monitored region and/or the manner in which it changes while the bodies are present in the monitored region can be compared with models of volumetric maps for other bodies, encoded in a similar manner and previously stored in the processing unit (take for example the model marked D in Figure 6) in order to recognize those bodies which must be detected from among all the bodies present inside the monitored region. The bodies may for example be people only.
  • Furthermore, it is possible to detect the simultaneous presence of several bodies, even if of different kinds, in the monitored region. The manner in which the position of the bodies change within the monitored region can be used to detect violation of predefined subregions. It is thus possible to monitor, as has already been seen, the presence of a movement of people in the vicinity of a machine in an industrial environment and activate an alarm signalling procedure whenever at least one person comes within a certain distance of that machine.
  • The solution described is highly robust and overcomes the functional limitations of currently used systems. Thus, it is capable of detecting the presence and at the same time determining the position of people or objects within a defined region of space, discriminate between objects and people, between objects or people close to each other, and between objects and people that move into the monitored region following different paths or more generally with behaviours which could easily deceive other types of sensor.
  • Those skilled in the art will recognize that the method according to the invention can be carried out using, at least in part, a computer program capable of being run on a computer in such a way that the system comprising the program and the computer carries out the method according to the invention. The invention therefore extends also to such a program capable of being loaded into a computer which has the means of or is capable of carrying out the method according to the invention, as well as to the corresponding information technology product comprising a means readable by a computer containing codes for a computer program which, when the program is loaded into the computer, cause the computer to carry out the method according to the invention.
  • Clearly, without affecting the principle of the invention, the constructional details and the embodiments may be greatly altered compared to what has been described and illustrated, without thereby departing from the scope of the present invention, as defined in the accompanying claims.

Claims (13)

  1. Method for the detection and location of bodies (A, B) in a defined region of space (S), comprising the operations of generating (2) image signals capable of representing a succession of images of at least one body present in the said region (S), each image corresponding to a defined instant, characterized in that it comprises the following operations:
    processing (101 to 106) the said image signals in such a way as to obtain for each instant taken into consideration a volumetric map (F) of the said at least one body present in the said region (S), the said volumetric map (F) representing the shape, position, volume and dimensions of the body to which the said volumetric map (F) refers,
    extracting (107) from the said volumetric map (F) at least one parameter (P) taken from the following group: descriptors of shape and volume, such as the volumetric map (F) itself, the position co-ordinates and the dimensions of the said at least one body to which the said volumetric map (F) refers,
    comparing (108) the said at least one parameter (P) with at least one model (D) for compatibility of the said volumetric map (F) with predetermined conditions of occupation of the said region (S), and
    selectively generating a warning signal (109) depending on the outcome of the said comparison (108).
  2. Method according to Claim 1, characterized in that it comprises the operations of storing volumetric maps (F) or successions of the said at least one parameter (P) obtained from image signals relating to images of the said succession corresponding to successive instants, in order to detect changes in time in the said volumetric map (F) or the said at least one parameter (P), and in that the said model is itself generated as a model of changes over time.
  3. Method according to Claim 1 or Claim 2, characterized in that it comprises the operation of comparing (108) successions of the at least one parameter (P) obtained from image signals relating to images of the said succession corresponding to successive instants with at least one model for compatibility of said successions with predetermined conditions of occupation of the said region (S).
  4. Method according to Claim 1 or Claim 2, characterized in that it comprises the following operations:
    generating image signals relating to the view of the said region (S) from separate observation points (2) so as to generate at least two separate image signals relating respectively to the projections of the same points of the said region (S) viewed from separate observation points,
    processing (1) the said separate image signals by finding the match between the projections of the same real-world points onto separate images with a view to finding its position in space,
    obtaining the said volumetric map (F) from the said positions in space.
  5. Method according to Claim 1 or Claim 2, characterized in that it comprises the following operations:
    subjecting at least part of the said region (S) to scanning of cells (D(xi, yi, zi)),
    detecting (1041, 1042, ..., 104n) for each of the said cells the said separate signals (1031, 1032, ..., 103n),
    generating (105) for each scanned cell and from the values of the said separate signals detected for the said cell, an occupation signal whose value can identify the occupation of the scanned cell by the said at least one body present in the said region of space (S), the said volumetric map (F) being identified by the set of values attributed to the said occupation signal corresponding to the scanned cells.
  6. Apparatus for the detection and location of bodies (A, B) in a defined region of space (S) comprising image signal generating means (2) capable of monitoring the said region (S) in order to produce an image or succession of images of at least one body present in the said region (S), each image corresponding to a defined instant, the apparatus being characterized in that it comprises processing means (1) configured (101 to 106) so as to:
    process the said image signals in such a way as to obtain for each instant taken into consideration a volumetric map (F) of the said at least one body present in the said region (S), the said volumetric map (F) representing the shape, position, volume and dimensions of the body to which the said volumetric map (F) refers,
    extract (107) from the said volumetric map (F) at least one parameter (P) taken from the following group: descriptors of shape and volume, such as the volumetric map (F) itself, the position co-ordinates and the dimensions of the said at least one body to which the said volumetric map (F) refers,
    compare (108) the said at least one parameter (P) with at least one model (D) for compatibility of the said volumetric map (F) with predetermined conditions of occupation of the said region (S),
    and in that it also comprises warning means (3) connected to the said processing means (1) for selectively generating a warning signal (109) depending on the outcome of the said comparison (108).
  7. Apparatus according to Claim 6, characterized in that the said processing means (1) are configured so as to store volumetric maps (F) or successions of the said at least one parameter (P) obtained from image signals relating to images of the said succession corresponding to successive instants, in order to detect changes in time in the said volumetric map (F) or the said at least one parameter (P), and in that the said model is itself generated as a model of changes over time.
  8. Apparatus according to Claim 6 or Claim 7, characterized in that the said processing means (1) are configured in such a way as to compare (108) successions of the at least one parameter (P) obtained from image signals relating to images of the said succession corresponding to successive instants with at least one model for compatibility of said successions with predetermined conditions of occupation of the said region (S).
  9. Apparatus according to Claim 6 or Claim 7, characterized in that:
    a plurality of means (2) are provided for the generation of image signals relating to views of the said region (S) from separate observation points (2) so as to generate at least two separate image signals relating respectively to the projections of the same points of the said region (S) viewed from separate observation points,
    the said processing means (1) are configured so as to process the said separate image signals by finding the match between the projections of the same real-world points onto separate images in order to find its position in space and obtain the said volumetric map (F) from the said positions in space.
  10. Apparatus according to Claim 6 or Claim 7, characterized in that the said processing means (1) are configured so as to:
    subject at least part of the said region (S) to scanning of cells (D(xi, yi, zi)),
    detect (1041, 1042, ..., 104n) for each of the said cells the said separate signals (1031, 1032, ..., 103n),
    generate (105) for each scanned cell and from the values of the said separate signals detected for the said cell, an occupation signal whose value can identify the occupation of the scanned cell by the said at least one body present in the said region of space (S), the said volumetric map (F) being identified by the set of values attributed to the said occupation signal corresponding to the scanned cells.
  11. Computer program capable of being run on a computer in such a way that the system comprising the program and the computer carries out the method according to any one of Claims 1 to 5.
  12. Computer program capable of being loaded into a computer which has the means of or is capable of carrying out the method according to any one of Claims 1 to 5.
  13. Information technology product comprising a means readable by a computer containing codes for a computer program which, when the program is loaded into the computer, cause the computer to carry out the method according to any one of Claims 1 to 5.
EP99830376A 1999-06-17 1999-06-17 A method and device for automatically controlling a region in space Withdrawn EP1061487A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP99830376A EP1061487A1 (en) 1999-06-17 1999-06-17 A method and device for automatically controlling a region in space

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
EP99830376A EP1061487A1 (en) 1999-06-17 1999-06-17 A method and device for automatically controlling a region in space

Publications (1)

Publication Number Publication Date
EP1061487A1 true EP1061487A1 (en) 2000-12-20

Family

ID=8243455

Family Applications (1)

Application Number Title Priority Date Filing Date
EP99830376A Withdrawn EP1061487A1 (en) 1999-06-17 1999-06-17 A method and device for automatically controlling a region in space

Country Status (1)

Country Link
EP (1) EP1061487A1 (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003003310A1 (en) * 2001-06-29 2003-01-09 Honeywell International Inc. Moving object assessment system and method
WO2003003721A1 (en) * 2001-06-29 2003-01-09 Honeywell International, Inc. Surveillance system and methods regarding same
WO2003005315A1 (en) * 2001-07-02 2003-01-16 Koninklijke Philips Electronics N.V. Vision based method and apparatus for detecting an event requiring assistance or documentation
DE10138960A1 (en) * 2001-08-03 2003-02-27 Pilz Gmbh & Co Surveillance method for monitoring surveillance space uses different brightness values provided by different image data sets for providing distance information for given point
WO2003019490A1 (en) * 2001-08-22 2003-03-06 Koninklijke Philips Electronics N.V. Vision-based method and apparatus for detecting fraudulent events in a retail environment
WO2003036581A1 (en) * 2001-10-23 2003-05-01 Vistar Telecommunications Inc. Method of monitoring an enclosed space over a low data rate channel
WO2003042946A1 (en) * 2001-11-13 2003-05-22 Koninklijke Philips Electronics N.V. Vision-based method and apparatus for automatically activating a child safety feature
DE10245720A1 (en) * 2002-09-24 2004-04-01 Pilz Gmbh & Co. Safety method for protecting automatic machine danger area with scene analysis of different images of danger area via 2 different analysis algorithms for reliable detection of foreign object in danger area
WO2004029502A1 (en) 2002-09-24 2004-04-08 Pilz Gmbh & Co. Kg Method and device for making a hazardous area safe
EP1422022A1 (en) * 2002-11-12 2004-05-26 Makita Corporation Power tools
EP1482238A2 (en) 2003-05-29 2004-12-01 CASAGRANDE SpA Safety device for operating machines, particularly drilling machines or suchlike, and method to recognize the presence of persons, using such safety device
BE1015588A3 (en) * 2003-07-02 2005-06-07 Robosoft Nv Monitoring for presence of unauthorized personnel or foreign objects in dangerous machinery, involves comparing actual images of danger area with images stored in memory
EP1586805A1 (en) * 2004-04-14 2005-10-19 Sick Ag Method for monitoring a surveillance area
EP1635108A1 (en) * 2004-09-08 2006-03-15 Sick Ag Method and apparatus to detect an object
US7084779B2 (en) 2003-09-12 2006-08-01 Makita Corporation Power tool
WO2008028911A1 (en) * 2006-09-04 2008-03-13 Robert Bosch Gmbh Machine tool monitoring device
WO2008028871A1 (en) * 2006-09-04 2008-03-13 Robert Bosch Gmbh Machine tool monitoring device
WO2008014831A3 (en) * 2006-08-02 2008-04-03 Daimler Ag Method for observation of a person in an industrial environment
WO2008098831A1 (en) * 2007-02-15 2008-08-21 Kuka Roboter Gmbh Method and device for securing a work space
WO2009062770A1 (en) * 2007-11-12 2009-05-22 Robert Bosch Gmbh Configuration module for a video surveillance system, surveillance system comprising the configuration module, method for configuring a video surveillance system and computer program
WO2009083297A1 (en) * 2007-12-21 2009-07-09 Robert Bosch Gmbh Machine tool device
US7684590B2 (en) 2004-04-19 2010-03-23 Ibeo Automobile Sensor Gmbh Method of recognizing and/or tracking objects
US7729511B2 (en) 2002-09-24 2010-06-01 Pilz Gmbh & Co. Kg Method and device for safeguarding a hazardous area
US7960676B2 (en) * 2007-09-28 2011-06-14 Casio Computer Co., Ltd. Image capture device to obtain an image with a bright subject, even when the background is dark, and recording medium
EP2650850A1 (en) * 2012-04-12 2013-10-16 Steinel GmbH Device for controlling a building unit
EP2685421A1 (en) 2012-07-13 2014-01-15 ABB Research Ltd. Determining objects present in a process control system
DE102004044973B4 (en) * 2004-09-16 2014-12-04 Sick Ag Control of a surveillance area
CN106163713A (en) * 2014-02-14 2016-11-23 豪仕马裁板锯技术有限公司 For to equipment, method that especially plate splitting equipment operates
EP3112900A1 (en) * 2015-07-03 2017-01-04 Soilmec S.p.A. Safety system and method to detect a risk condition in a region to be monitored placed close to an operating machine, such as a drilling machine or the like
IT201900021108A1 (en) * 2019-11-13 2021-05-13 Gamma System S R L Safety system for an industrial machinery

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0356734A2 (en) * 1988-08-02 1990-03-07 Siemens Aktiengesellschaft Intruder detection device with television cameras

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0356734A2 (en) * 1988-08-02 1990-03-07 Siemens Aktiengesellschaft Intruder detection device with television cameras

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CARLSON J ET AL: "Real-time 3D visualization of volumetric video motion sensor data", SURVEILLANCE AND ASSESSMENT TECHNOLOGIES FOR LAW ENFORCEMENT, BOSTON, MA, USA, 19-20 NOV. 1996, vol. 2935, Proceedings of the SPIE - The International Society for Optical Engineering, 1997, SPIE-Int. Soc. Opt. Eng, USA, pages 69 - 79, XP000863382, ISSN: 0277-786X *
TARBOX G H ET AL: "VOLUMETRIC BASED INSPECTION*", PROCEEDINGS OF THE IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS,US,NEW YORK, IEEE, vol. -, pages 1239-1246, XP000334081, ISBN: 0-7803-0738-0 *

Cited By (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003003721A1 (en) * 2001-06-29 2003-01-09 Honeywell International, Inc. Surveillance system and methods regarding same
WO2003003310A1 (en) * 2001-06-29 2003-01-09 Honeywell International Inc. Moving object assessment system and method
WO2003005315A1 (en) * 2001-07-02 2003-01-16 Koninklijke Philips Electronics N.V. Vision based method and apparatus for detecting an event requiring assistance or documentation
DE10138960A1 (en) * 2001-08-03 2003-02-27 Pilz Gmbh & Co Surveillance method for monitoring surveillance space uses different brightness values provided by different image data sets for providing distance information for given point
WO2003019490A1 (en) * 2001-08-22 2003-03-06 Koninklijke Philips Electronics N.V. Vision-based method and apparatus for detecting fraudulent events in a retail environment
WO2003036581A1 (en) * 2001-10-23 2003-05-01 Vistar Telecommunications Inc. Method of monitoring an enclosed space over a low data rate channel
US6720880B2 (en) 2001-11-13 2004-04-13 Koninklijke Philips Electronics N.V. Vision-based method and apparatus for automatically activating a child safety feature
WO2003042946A1 (en) * 2001-11-13 2003-05-22 Koninklijke Philips Electronics N.V. Vision-based method and apparatus for automatically activating a child safety feature
WO2004029502A1 (en) 2002-09-24 2004-04-08 Pilz Gmbh & Co. Kg Method and device for making a hazardous area safe
US7729511B2 (en) 2002-09-24 2010-06-01 Pilz Gmbh & Co. Kg Method and device for safeguarding a hazardous area
DE10245720A1 (en) * 2002-09-24 2004-04-01 Pilz Gmbh & Co. Safety method for protecting automatic machine danger area with scene analysis of different images of danger area via 2 different analysis algorithms for reliable detection of foreign object in danger area
CN100389939C (en) * 2002-11-12 2008-05-28 株式会社牧田 Power tools
EP1422022A1 (en) * 2002-11-12 2004-05-26 Makita Corporation Power tools
US6959631B2 (en) 2002-11-12 2005-11-01 Makita Corporation Power tools
US7047854B2 (en) 2002-11-12 2006-05-23 Makita Corporation Power tools
EP1482238A2 (en) 2003-05-29 2004-12-01 CASAGRANDE SpA Safety device for operating machines, particularly drilling machines or suchlike, and method to recognize the presence of persons, using such safety device
EP1482238A3 (en) * 2003-05-29 2005-08-17 CASAGRANDE SpA Safety device for operating machines, particularly drilling machines or suchlike, and method to recognize the presence of persons, using such safety device
BE1015588A3 (en) * 2003-07-02 2005-06-07 Robosoft Nv Monitoring for presence of unauthorized personnel or foreign objects in dangerous machinery, involves comparing actual images of danger area with images stored in memory
US7084779B2 (en) 2003-09-12 2006-08-01 Makita Corporation Power tool
US7505620B2 (en) 2004-04-14 2009-03-17 Sick Ag Method for the monitoring of a monitored zone
EP1586805A1 (en) * 2004-04-14 2005-10-19 Sick Ag Method for monitoring a surveillance area
US7684590B2 (en) 2004-04-19 2010-03-23 Ibeo Automobile Sensor Gmbh Method of recognizing and/or tracking objects
EP1635108A1 (en) * 2004-09-08 2006-03-15 Sick Ag Method and apparatus to detect an object
US7652238B2 (en) 2004-09-08 2010-01-26 Sick Ag Method and apparatus for detecting an object through the use of multiple sensors
DE102004044973B4 (en) * 2004-09-16 2014-12-04 Sick Ag Control of a surveillance area
WO2008014831A3 (en) * 2006-08-02 2008-04-03 Daimler Ag Method for observation of a person in an industrial environment
CN101511550B (en) * 2006-08-02 2013-12-18 皮尔茨公司 Method for observation of person in industrial environment
US8154590B2 (en) 2006-08-02 2012-04-10 Pilz Gmbh & Co. Kg Method for observation of a person in an industrial environment
US8311661B2 (en) 2006-09-04 2012-11-13 Robert Bosch Gmbh Machine tool use situation monitoring device using reflected signal
US8615320B2 (en) 2006-09-04 2013-12-24 Robert Bosch Gmbh Machine tool monitoring device
WO2008028911A1 (en) * 2006-09-04 2008-03-13 Robert Bosch Gmbh Machine tool monitoring device
WO2008028871A1 (en) * 2006-09-04 2008-03-13 Robert Bosch Gmbh Machine tool monitoring device
RU2459138C2 (en) * 2006-09-04 2012-08-20 Роберт Бош Гмбх Processing machine control device
WO2008098831A1 (en) * 2007-02-15 2008-08-21 Kuka Roboter Gmbh Method and device for securing a work space
US8227738B2 (en) 2007-09-28 2012-07-24 Casio Computer Co., Ltd. Image capture device for creating image data from a plurality of image capture data, and recording medium therefor
US7960676B2 (en) * 2007-09-28 2011-06-14 Casio Computer Co., Ltd. Image capture device to obtain an image with a bright subject, even when the background is dark, and recording medium
US9549155B2 (en) 2007-11-12 2017-01-17 Robert Bosch Gmbh Configuration module for a video surveillance system, surveillance system comprising the configuration module, method for configuring a video surveillance system, and computer program
WO2009062770A1 (en) * 2007-11-12 2009-05-22 Robert Bosch Gmbh Configuration module for a video surveillance system, surveillance system comprising the configuration module, method for configuring a video surveillance system and computer program
CN101855906B (en) * 2007-11-12 2017-11-24 罗伯特·博世有限公司 The configuration module of video monitoring system, the monitoring system with configuration module and the method for configuring video monitoring system
WO2009083297A1 (en) * 2007-12-21 2009-07-09 Robert Bosch Gmbh Machine tool device
US8948903B2 (en) 2007-12-21 2015-02-03 Robert Bosch Gmbh Machine tool device having a computing unit adapted to distinguish at least two motions
EP2650850A1 (en) * 2012-04-12 2013-10-16 Steinel GmbH Device for controlling a building unit
WO2014009087A1 (en) 2012-07-13 2014-01-16 Abb Research Ltd Presenting process data of a process control object on a mobile terminal
EP2685421A1 (en) 2012-07-13 2014-01-15 ABB Research Ltd. Determining objects present in a process control system
CN106163713A (en) * 2014-02-14 2016-11-23 豪仕马裁板锯技术有限公司 For to equipment, method that especially plate splitting equipment operates
CN106163713B (en) * 2014-02-14 2018-04-13 豪迈裁板锯技术有限公司 For to equipment, the method that especially plate splitting equipment is operated
EP3112900A1 (en) * 2015-07-03 2017-01-04 Soilmec S.p.A. Safety system and method to detect a risk condition in a region to be monitored placed close to an operating machine, such as a drilling machine or the like
IT201900021108A1 (en) * 2019-11-13 2021-05-13 Gamma System S R L Safety system for an industrial machinery

Similar Documents

Publication Publication Date Title
EP1061487A1 (en) A method and device for automatically controlling a region in space
Rozsa et al. Obstacle prediction for automated guided vehicles based on point clouds measured by a tilted LIDAR sensor
RU2251739C2 (en) Objects recognition and tracking system
CN108241870B (en) Method for assigning a specific category of interest within measurement data
JP3785456B2 (en) Safety monitoring device at station platform
Zivkovic et al. Hierarchical map building using visual landmarks and geometric constraints
CN104935879B (en) For the monitoring system of the view-based access control model of activity command verification
Anousaki et al. Simultaneous localization and map building for mobile robot navigation
Cameron et al. A Bayesian approach to optimal sensor placement
US5845048A (en) Applicable recognition system for estimating object conditions
US20050012817A1 (en) Selective surveillance system with active sensor management policies
KR101877294B1 (en) Smart cctv system for crime prevention capable of setting multi situation and recognizing automatic situation by defining several basic behaviors based on organic relation between object, area and object's events
Arbuckle et al. Temporal occupancy grids: a method for classifying the spatio-temporal properties of the environment
EP3680813A1 (en) Method and system for detecting objects installed within a building
Chakravarty et al. Panoramic vision and laser range finder fusion for multiple person tracking
Kim et al. Robotic sensing and object recognition from thermal-mapped point clouds
McGreavy et al. Next best view planning for object recognition in mobile robotics
Arsic et al. Applying multi layer homography for multi camera person tracking
Apolloni et al. Machine learning and robot perception
Foresti et al. Event classification for automatic visual-based surveillance of parking lots
Nickerson et al. An autonomous mobile robot for known industrial environments
JP3630703B2 (en) Object monitoring device
KR102597692B1 (en) Method, apparatus, and computer program for measuring volume of objects by using image
Özbilge Experiments in online expectation-based novelty-detection using 3D shape and colour perceptions for mobile robot inspection
Beibei et al. Crowd analysis: a survey

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

AKX Designation fees paid
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20010621

REG Reference to a national code

Ref country code: DE

Ref legal event code: 8566