US20070164202A1 - Large depth of field line scan camera - Google Patents
Large depth of field line scan camera Download PDFInfo
- Publication number
- US20070164202A1 US20070164202A1 US11/559,669 US55966906A US2007164202A1 US 20070164202 A1 US20070164202 A1 US 20070164202A1 US 55966906 A US55966906 A US 55966906A US 2007164202 A1 US2007164202 A1 US 2007164202A1
- Authority
- US
- United States
- Prior art keywords
- imaging
- sensors
- independent linear
- imaging sensors
- linear imaging
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/04—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/04—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa
- H04N1/12—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa using the sheet-feed movement or the medium-advance or the drum-rotation movement as the slow scanning component, e.g. arrangements for the main-scanning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/04—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa
- H04N1/19—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa using multi-element arrays
- H04N1/195—Scanning arrangements, i.e. arrangements for the displacement of active reading or reproducing elements relative to the original or reproducing medium, or vice versa using multi-element arrays the array comprising a two-dimensional array or a combination of two-dimensional arrays
Definitions
- the present invention relates generally to optical scanning systems, More particularly, this invention relates to a scanning system having a large depth of field which is capable of focusing on and scanning a plurality of targets over a wide range of distances.
- optical scanning systems have been developed for reading and decoding coded symbologies.
- Such optical scanning systems typically incorporate a light source and a light detecting means, such as a CMOS- or CCD-based image sensor.
- a portion of the light which is reflected from the coded symbology is detected by the image sensor and converted into an electrical signal, which is the basis for the digital image of the coded symbology that has been scanned.
- the digital image is then processed and decoded according to the specific type of coded symbology.
- CMOS- and CCD-based scanning systems require a mechanical focusing apparatus in order to focus the light reflected from the coded symbology onto the image sensor. Since a mechanical focusing apparatus typically operates slower than the image sensor and other electronics within the scanner, the focusing apparatus tends to be the limiting factor in the speed of operation of the scanning system. Additionally, focusing parts are subject to higher wear and maintenance requirements.
- U.S. Pat. No. 6,783,068 discloses a large depth of field camera which was developed by Accu-Sort Systems, Inc., the assignee of the present invention. This reference discloses an area array sensor used for scanning a large depth of field. An object height sensor is used to determine an object height so that the correct row of pixels is read. While the system described in U.S. Pat. No. 6,783,068 is effective for scanning a large depth of field, suitable area array sensors are currently available only at a very high cost. Further, such sensors typically transmit large quantities of information that can burden internal or external processing systems.
- the present invention recognizes and addresses the foregoing disadvantages, and others, of prior art constructions and methods.
- the present invention provides a system for imaging an object comprising a surface for receiving an object to be imaged and an imaging device.
- the imaging device comprising a housing, a fixed focusing optical device mounted to the housing and comprising a lens for focusing light reflected from the object on the surface, a plurality of independent linear imaging sensors are positioned in the housing at different distances from the fixed focusing lens so that a face of each of the plurality of independent linear imaging sensors is aligned parallel to the surface and a processor coupled to the plurality of independent linear imaging sensors for receiving an output signal from each of the plurality of independent linear imaging sensors representative of the reflected light.
- the lens defines a plane parallel to the surface, and the plurality of independent linear imaging sensors are adapted to receive reflected light from the object on the surface.
- the plurality of independent linear imaging sensors are located at different distances from the lens such that a plane intersecting each linear row of pixels for each of the plurality of independent linear imaging sensors intersects the lens plane at the Sheimpflug angle such that the focal point of each of the plurality of independent linear imaging sensors differ with respect to the surface.
- a height measuring device detects a height of the object on the surface prior to reaching the imaging device, wherein the detected height is used by the processor for selectively reading the output signal from one of the plurality of independent linear imaging sensors.
- the processor forms a two-dimensional image from the output signals of the plurality of independent linear imaging sensors.
- the processor is adapted to read an output signal from individual pixels located in a different one of the plurality of independent linear imaging sensors based on an intensity of the output of the individual pixels.
- the surface is moveable in a direction perpendicular to an axis of each of the plurality of independent linear imaging sensors.
- a system for imaging an object comprises a surface for receiving an object to be imaged and an imaging device comprising a housing, a plurality of independent linear imaging sensors positioned in the housing so that a face of each of the plurality of independent linear imaging sensors is parallel to the surface, where the plurality of independent linear imaging sensors is adapted to receive reflected light from the object on the surface, a fixed focusing optical device mounted to the housing and comprising a single lens for focusing the reflected light from the object on the surface onto each of the plurality of independent linear imaging sensors, and a processor coupled to the plurality of independent linear imaging sensors for reading an output signal from each of the plurality of independent linear imaging sensors.
- the single lens may be an irregular lens that provides a plurality of focal lengths each corresponding to a respective one of the plurality of independent linear imaging sensors.
- the system may also comprise a plurality of linear actuators for independently moving each of the plurality of independent linear imaging sensors with respect to the fixed focusing optical device.
- a height measuring device may be used for detecting a height of the object on the surface prior to reaching the imaging device, wherein the detected height is used by each of the plurality of linear actuators for moving each of the plurality of independent linear imaging sensors with respect to the fixed focusing device.
- a feedback loop from the processor to each of the plurality of linear actuators may be used to assist in positioning each of the plurality of independent linear imaging sensors.
- a system for imaging an object comprising a surface for receiving an object to be imaged and an imaging device comprising.
- the imaging device comprises a housing, a fixed focusing optical device mounted to the housing and comprising a lens for focusing light reflected from the object on the surface, the fixed focusing optical device lens defining a plane parallel to the surface, a linear imaging sensor positioned in the housing radially outward from an axis normal to the lens plane, first and second mirrors rotatably coupled to a pivot axis and positioned so as to direct light entering the fixed focusing optical device onto the linear imaging sensor, and a processor coupled to the linear imaging sensor for selectively reading an output signal from individual pixels in the linear imaging sensor based on an intensity of the output signal from each of the individual pixels.
- Each of the first and the second mirrors are independently moveable about the pivot axis along a path generally perpendicular to the linear image sensor with respect to each other so that different portions of the surface can be focused onto different portions of the linear sensor.
- the system may also comprise a first and second actuator for respectively rotating the first and second mirrors about the pivot axis with respect to the linear imaging sensor.
- the system may comprise third and fourth actuators for respectively translating the first and the second mirrors along the pivot axis with respect to the lineal imaging sensor.
- a third mirror may be rotatably coupled to the pivot axis and positioned so as to direct light entering the fixed focusing optical device onto the linear imaging sensor.
- a height measuring device may be used to detect a height of the object on the surface, wherein the detected height information is used by the first and the second actuators for moving each of the first and second mirrors to focus on the object on the surface.
- FIG. 1A is a top perspective view of a coded symbology scanning system in accordance with preferred embodiments of the present invention
- FIG. 1B is bottom perspective view of the coded symbology scanning system of FIG. 1A ;
- FIG. 2 is a camera according to a first preferred embodiment of the present invention for use in the scanning system of FIG. 1A ;
- FIG. 3A illustrates a relationship between an object plane, an alternate camera with camera optics, and an image plane defined by a second preferred embodiment of the present invention for use in the scanning system of FIG. 1A ;
- FIG. 3B illustrates a 3-D spatial relationship between the object plane, camera optics, and image plane of FIG. 3A ;
- FIG. 3C shows the object plane of FIG. 3B in greater detail
- FIG. 3D shows the image plane of FIG. 3B in greater detail
- FIGS. 4A and 4B illustrate a 3-D spatial relationship between an object plane, camera optics, and an image plane defined by a third preferred embodiment of the present invention
- FIG. 5 is a flow diagram of a preferred symbol reading and decoding method
- FIG. 6 is another alternate camera according to a fourth preferred embodiment of the present invention for use in the scanning system of FIG. 1A ;
- FIG. 7 illustrates a relationship between an object plane, camera optics and an image plane defined by an embodiment of the scanning system of FIG. 1A including the alternate camera of FIG. 6 ;
- FIG. 8 is a flow diagram of another alternative preferred symbol reading and decoding method
- FIG. 9A is a camera according to a fifth preferred embodiment of the present invention for use in the scanning system of FIG. 9B ;
- FIGS. 9B is a top perspective view of a coded symbology scanning system in accordance with the fifth preferred embodiment of the present invention.
- FIG. 10 is a flow diagram of another alternative preferred symbol reading and decoding method.
- Scanning system 10 is able to scan any type of object. For simplicity, however, scanning system 10 will be explained with reference to scanning a coded symbology, and in particular a bar code symbol.
- Scanning system 10 includes a light source 12 and a camera 17 located within a housing 14 .
- camera 17 includes a plurality of independent linear image sensors 26 a - 26 c arranged side by side along a plane ( FIG. 2 ).
- any suitable number of independent linear image sensors greater than 1 can be used.
- any line scan image sensor may be utilized without departing from the spirit and scope of the present invention, for example either a CMOS or a CCD linear image sensor.
- Light source 12 preferably includes a plurality of LEDs 13 mounted in a focusing container 15 for directing light emitted by LEDs 13 .
- Light source 12 facilitates detection of subject bar code symbols 18 a and 18 b by illuminating bar code symbols 18 a and 18 b located on objects 8 and 9 .
- objects 8 and 9 are supported by a moving transport system 7 .
- the scanning system 10 is aligned such that light reflected from bar code symbols 18 a and 18 b passes through a front of camera housing 14 to be detected by linear image sensors 26 a - 26 c .
- linear image sensors 26 a - 26 c are optically aligned parallel to the surface of the transport system 7 and perpendicular to the motion of transport system 7 . It should be recognized by those skilled in the art that FIGS. 1A and 1B are illustrative only and are not drawn to scale.
- a tachometer 19 monitors the speed of the surface of transport system 7 and objects 8 and 9 and provides this information to a processor 32 (shown in FIG. 2 ) to control the read-out rate of the plurality of linear image sensors 26 a - 26 c .
- processor 32 shown in FIG. 2
- FIG. 2 A tachometer 19 monitors the speed of the surface of transport system 7 and objects 8 and 9 and provides this information to a processor 32 (shown in FIG. 2 ) to control the read-out rate of the plurality of linear image sensors 26 a - 26 c .
- camera 17 includes fixed focusing optics 31 , a plurality of independent linear image sensors 26 a - 26 c , a processor 32 and an output means 34 .
- Linear image sensors 26 a - 26 c include a column of pixels 28 a - 28 n .
- fixed focusing optics 31 includes at least one custom optic having more than one focal length, for example, an irregularly shaped lens.
- the custom optic provides a unique focal length for each of linear image sensors 26 a 26 c such that objects of different heights positioned on the transport system 7 , corresponding to different depths of field, are in focus (or nearly in focus) on different ones of linear image sensors 26 a - 26 c .
- This configuration allows two or more adjacent or side-by side items having different heights, for example objects 8 and 9 positioned on transport system 7 in FIGS. 1A and 1B , to be viewed in focus (or nearly in focus) simultaneously by different ones of linear image sensors 26 a 26 c.
- the focusing optics 31 focuses reflected light 29 onto the appropriate one of linear image sensors 26 a 26 c based on a depth of field to an object being scanned.
- Linear image sensors 26 a 26 c detect reflected light 29 and each pixel within each of linear image sensors 26 a 26 c converts the incident light energy into a stored electrical charge and ultimately into an electrical signal.
- the electrical signal comprises a value that represents the amount of light detected by a pixel. If the pixel's analog signal is digitized to have eight-bit resolution, the number of gray scale levels will be 256 .
- the resolution of the linear scan image is dependent on the number of pixels 28 a 28 n in each of linear image sensors 26 a 26 c .
- Linear image sensors 26 a 26 c are preferably aligned parallel to each other and transverse to the direction of travel of transport system 7 .
- Each of the linear image sensors 26 a 26 c can be considered one row in a series of parallel lines when projected into the object plane.
- independent linear image sensors 26 a 26 c are preferably positioned at different distances from optics 31 so that an image plane I intersects each row of pixels in each of the linear image sensors, where the image plane is disposed at the Scheimpflug angle with respect to a median plane 33 of optics 31 , as shown in FIG. 3A .
- each pixel in any specific column 28 a 28 n will have a corresponding unique focus point in an object plane 0 that also intersects with median plane 33 and image plane I along a line 35 .
- the system 10 of the present invention utilizes the selective positioning of linear image sensors 26 a 26 c in order to read out information pertaining to the objects within the depth of field that are in focus or nearly in focus on each linear image sensor.
- a processor 32 accesses the linear image sensors 26 a 26 c to obtain the desired information.
- the processor receives the data from all independent linear imaging sensors and may (1) process and store the output of one linear image sensor that is receiving data in focus or (2) process and store all data from each of the linear image sensors.
- the processor uses the processed data to form one or more 2-D images from the individual line scans detected by each linear image sensor. That is, if the data from one linear image scanner is selected, then a single 2-D image is formed from the plurality of line scans captured by the selected linear imager. If, on the other hand, two or more independent linear image sensors are selected, than an independent 2-D image is formed for each selected linear image sensor using that sensors respective data.
- a tachometer 19 can be used to determine a scan rate for sensors 26 a 26 c . This information is then output to output means 34 for further processing.
- Output means 34 may be a digital display, an image processing system, a positioning system, or any other type of system that will accept an output from an imaging camera.
- linear image sensors 26 a 26 c are positioned at different distances from optics 31 so that image plane I intersects each row of pixels in each of the linear image sensors, where image plane I is disposed at an angle with respect to median plane 33 through optics 31 .
- the angle (hereinafter, the Scheimpflug angle), is determined using the Scheimpflug condition.
- the Scheimpflug angle is dependent on the focal length of optics 31 and a desired working range defining the depth of field (DoF) along object plane O.
- Each horizontal line in object plane O has a corresponding horizontal line in image plane I that is in focus.
- each of linear image sensors 26 a 26 c in image plane I corresponds to a specific distance in object plane O; and each such distance in object plane O corresponds to a line parallel to a surface of transport system 7 and perpendicular to the direction of motion of transport system 7 .
- FIG. 3B the spatial relationship between object plane O, optics 31 and image plane I is shown.
- FIG. 3C shows object plane O of FIG. 3B in greater detail
- FIG. 3D shows image plane I of FIG. 3B in greater detail.
- an object O 1 in object plane O at a distance D 1 from optics 31 will be in focus or close to in focus in image plane I on linear image sensor 26 a on its string of pixels I 1 .
- an object O 2 in object plane O at distance D 2 will be in focus or close to in focus in image plane I on linear image sensor 26 b at its string of pixels I 2
- an object O 3 in object plane O at distance D 3 will be in focus or close to in focus in image plane I on the linear image sensor 26 c at its string of pixels I 3 .
- linear image sensors 26 a 26 c having an object plane focal point distance, for example one of distances D 1 , D 2 , D 3 that best corresponds to an object height, for example a height represented by objects O 1 , 02 , 03 , will be in optimal focus.
- each of the sensors 26 a 26 c will receive some light regardless of the height of object O 1 , O 2 , O 3 .
- the linear image sensor in optimal focus will receive the most light.
- processor 32 selects the data received from one of sensor 26 a 26 c that receives the most light, indicating optimal focus, to be processed and stored. Through comparing the intensity of the signals read by each linear image sensors 26 a 26 c , the appropriate in focus linear image sensor 26 can be read and the data further processed.
- the data from linear image sensor 26 that is selected by processor 32 is sent to output means 34 for further processing, decoding, storage and/or display as is well known by those skilled in the art. Data is periodically read out of linear image sensors 26 a 26 c as will be described in greater detail hereinafter in order to read coded symbology on objects carried past the scanner.
- a third preferred embodiment of the present invention is shown in which a system 110 includes a compact camera 117 having only two linear image sensors 126 a and 126 b and optics 131 .
- Linear image sensors 126 a and 126 b are positioned at different distances from optics 131 so that a common image plane I intersects each row of pixels in each of the linear image sensors, where image plane I is disposed at an angle with respect to a median plane 133 through optics 131 at the Scheimpflug angle.
- an object plane O intersects with image plane I and median plane 133 along line 135 so that each horizontal line in an object plane O has a corresponding horizontal line in image plane I that is in focus.
- sensor 126 b is generally in focus between O 1 and O 2 and sensor 126 a is generally in focus between O 2 and O 3 , thereby increasing the overall depth of field of camera 117 .
- system 110 includes a conveyor for transporting objects where symbology to be read is within the depth of field (DoF) of the camera.
- linear image sensors 126 a and 126 b have a lower scan rate than the linear image sensors described in the other embodiments. As a result, each line scan image is processed separately by the processor and a 2-D image is not created from the captured data.
- the first, second and third embodiments of the invention allow two or more singulated or non-singulated objects having substantially different heights positioned on transport system 7 in FIG. 1 , for example objects 8 and 9 , to be viewed in focus (or nearly in focus) by different linear image sensors 26 a 26 c or 126 a - 126 b without a mechanical focusing or height detection device. Accordingly, information received simultaneously from adjacent objects can be processed simultaneously by the system recognizing and reading in focus data from more than one linear image sensor.
- scanning systems 10 and 110 execute the bar code symbol reading and decoding procedure 400 shown in FIG. 5 .
- a light source 12 illuminates a subject bar code symbol 18 (step 410 ).
- Light is reflected from bar code symbol 18 toward camera 17 or 117 (step 420 ).
- the reflected light enters camera 17 or 117 via focusing optics 31 or 131 (step 430 ).
- Focusing optics 31 or 131 focus light 29 onto linear image sensors 26 a 26 c or 126 a - 126 b (step 440 ).
- Each pixel within each of linear image sensors 26 a 26 c or 126 a - 126 b converts the light into a charge which is converted into an electrical signal (step 450 ) that is transmitted to processor 32 .
- Processor 32 compares the signals transmitted from each of the plurality of linear image sensors 26 a 26 c or 126 a - 126 b (step 455 ). Stronger signals, for example, represent a larger quantity of light received by a pixel, which indicates better focusing at the particular pixel location.
- Processor 32 preferably selects one or more linear image sensors 26 a - 26 c or 126 a - 126 b , which are in relatively good focus compared with other ones of the plurality of linear image sensors 26 a 26 c or 126 a - 126 b , based on a comparison of the signal outputs of the pixels of the linear image sensors (step 470 ). Alternatively, only certain in-focus pixels from one or more of linear image sensors 26 a 26 c or 126 a - 126 b can be selected.
- a height detection device for example, an overhead laser triangulation device or a light curtain, can be used for providing processor 32 with object height information to be used in selecting one or more sensors 26 a 26 c or 126 a - 126 b that are in focus.
- This periodic rate can be determined by processor 32 using speed information of objects 8 and 9 (preferably determined by tachometer 19 as shown in the embodiment of FIGS. 1A and lB), and the desired sample density.
- the image data is forwarded to output means 34 for decoding, storage and/or display (step 480 ) using methods that are well known to those of skill in the art.
- the periodic rate at which information is read out from the selected sensors can alternatively be set at a fixed value without utilizing speed information of the objects and without using a tachometer.
- camera 117 of scanning system 110 is configured to scan at a fixed rate for barcodes within its depth of field (DoF) on packages passing at any suitable speed.
- DoF depth of field
- the image data is forwarded to an output device for decoding, storage and/or display (See step 480 , FIG. 5 ).
- This type of scanning system allows easy compatibility with many pre-existing package conveying systems and is suitable for lower cost applications.
- an alternative camera 717 replaces camera 17 in the system 10 for scanning an object.
- Camera 717 is configured in a manner similar to camera 17 of FIG. 2 , but utilizes actuators 736 a - 736 c , preferably servo motors, to translate linear image sensors 726 a - 726 c in a direction toward or away from lens 31 .
- actuators 736 a - 736 c preferably servo motors, to translate linear image sensors 726 a - 726 c in a direction toward or away from lens 31 .
- a custom optic is not required, but may be provided if suitable for a particular application.
- Sensors 726 a - 726 c preferably travel along a path perpendicular to lens 31 . In this manner, each sensor 726 a - 726 c can be individually focused to more precisely sense light reflected from objects on transport system 7 .
- sensors 726 a - 726 c not only allows greater precision in focusing, but also increases the effective range of depth of field of system 10 . While three sensors 726 a - 726 c with corresponding respective actuators 736 a - 736 c are shown, any suitable number of sensors can be used. It should be clear to those skilled in the art that for each additional linear image sensor and associated actuator provided, an additional adjacent object of unique height can be simultaneously scanned. For example, a system with two linear image sensors can simultaneously scan two side-by-side objects of different heights, and a system with three linear image sensors can simultaneously scan three side-by-side objects of different heights.
- a height detection device for example, an overhead laser triangulation device 740 or alternatively a light curtain can be used for providing processor 32 with object height information for selecting and focusing sensors 726 a - 726 c.
- Scanning system 10 utilizing camera 717 in place of camera 17 executes an alternative bar code symbol reading and decoding procedure 800 shown in FIG. 8 .
- a light source 12 illuminates a subject bar code symbol 18 (step 810 ).
- Light is reflected from bar code symbol 18 toward camera 717 (step 820 ).
- the reflected light enters camera 717 via focusing optics 31 (step 830 ).
- Focusing optics 31 focuses light 29 onto linear image sensors 726 a - 726 c (step 840 ).
- Each pixel on each of linear image sensors 726 a - 726 c converts the light into a charge that is converted into an electrical signal (step 850 ).
- Processor 32 compares the signals transmitted from each of linear image sensors 726 a - 726 c (step 855 ). Stronger signals, for example, represent a larger quantity of light received by a pixel, which indicates better focusing at the particular pixel row. Processor 32 preferably selects one or more linear image sensors 726 a - 726 c that are in relatively good focus compared with other ones of the plurality of linear image sensors 726 a - 726 c based on a comparison of the signal outputs of the pixels of the linear image sensors (step 870 ). Alternatively, only certain in-focus pixels from one or more of the linear image sensors 726 a - 726 c can be selected, or all of the linear image sensors may be selected, read and stored for further processing.
- Processor 32 controls one or more of actuators 736 a - 736 c to translate the respective selected one or more linear image sensors 726 a - 726 c (step 875 ) based on charge emitted from selected linear image sensor(s) 726 a - 726 c .
- actuators 736 a - 736 c receive feedback at discrete time intervals from processor 32 to enable actuators 736 a - 736 c to position selected linear image sensor(s) 726 a - 726 c in a best focus position.
- the processor can receive signals from height detection devices, for example the light curtain or overhead laser triangulation device 740 , in order to select and position appropriate linear image sensor(s) 726 a - 726 c.
- the information from selected linear image sensor(s) 726 a - 726 c , from all linear image sensors or from only selected pixels within selected sensors 726 a - 726 c is read out at a periodic rate.
- This periodic rate is determined by processor 32 preferably using speed information of objects 8 and 9 , (determined by the tachometer 19 ), and the desired sample density.
- the image data is forwarded to output means 34 for decoding, storage and/or display (step 880 ) using methods that are well known to those of skill in the art.
- a system 910 has a camera 917 preferably having a single linear image sensor 926 and first and second adjustable adjacent mirrors 927 a and 927 b , which reflect light received through optics 931 onto linear image sensor 926 .
- Optics 931 preferably includes a single lens, but alternatively, the optics can have multiple lenses in any suitable configuration.
- System 910 transports objects, for example objects 908 and 909 , on a transport system 907 in a manner similar to that provided in the first embodiment shown in FIGS. 1A and 1B .
- Camera 917 utilizes first, second, third and fourth actuators 936 a - d , which are preferably voice coil actuators, or alternatively servo motors, to adjust adjustable mirrors 927 a and 927 b .
- First and second actuators 936 a and 936 b pivot mirrors 927 a and 927 b about an axis 941 in a direction toward or away from linear image sensor 926 .
- a custom optic is not required, but may be provided if suitable for a particular application.
- mirrors 927 a and 927 b preferably travel along a path generally perpendicular to linear image sensor 926 .
- a first portion of linear image sensor 926 receives reflected light from first adjustable mirror 927 a and a second portion of linear image sensor 926 receives reflected light from second adjustable mirror 927 b .
- two separate portions of linear image sensor 926 corresponding to two separate portions of the conveyor, can be individually focused to more precisely sense light reflected from objects, for example the objects 908 and 909 on transport system 907 , as shown in FIG. 9B .
- third and fourth actuators 936 c and 936 d are provided to translate mirrors 927 a and 927 b along axis 941 .
- the mirrors can be translated axially so that the entire object image is reflected by one mirror onto linear image sensor 926 so that any coded symbology can be read without gap G disrupting it.
- a single actuator can replace third and fourth actuators 936 c and 936 d to translate both of mirrors 927 a and 927 b together.
- any suitable number of mirrors can be used. It should be clear to those skilled in the art that for each additional mirror and associated actuator provided, it would be possible to simultaneously scan an additional adjacent object, which can have a unique height. For example, a system with two mirrors can simultaneously scan two side-by-side objects of different heights, and a system with three mirrors can simultaneously scan three side-by-side objects of the same or different heights.
- a height detection device for example, overhead laser triangulation device 740 as shown in FIG. 7 , can be used to provide processor 32 with object height information for each of the objects for adjusting mirrors 927 a and 927 b .
- Scanning system 910 utilizing camera 917 executes an alternative bar code symbol reading and decoding procedure 1000 shown in FIG. 10 .
- a light source illuminates subject bar code symbols 918 a and 918 b (step 1010 ).
- Light is reflected from bar code symbols 918 a and 918 h toward camera 917 (step 1020 ).
- the reflected light enters camera 917 via focusing optics 931 (step 1030 ), which focus light 29 onto mirrors 927 a and 927 b thereby directing light 29 onto linear image sensor 926 (step 1040 ).
- Each pixel on linear image sensor 926 converts the light into a charge which is converted into an electrical signal (step 1050 ).
- Processor 32 compares the signals transmitted from different portions of linear image sensor 926 (step 1055 ). Stronger signals, for example, represent a larger quantity of light received by particular pixels 928 - 928 n indicating better focusing at a particular portion of the linear image sensor.
- Processor 32 controls one or more of actuators 936 a - d to displace one or more of mirrors 927 a and 927 b (step 1075 ) based on charge emitted from the different portions of linear image sensor 926 .
- actuators 936 a - d receive feedback at discrete time intervals from processor 32 to enable actuators 936 a - d to position mirrors 927 a and 927 b in a best focus position.
- the processor can receive signals from height detection devices, for example a light curtain or overhead laser triangulation device 740 shown in FIG. 7 , in order to select and position one or both of mirrors 927 a and 927 b.
- the information from linear image sensor 926 or from only selected pixels within linear image sensor 926 is read out at a periodic rate.
- This periodic rate can be set or optionally can be determined by processor 32 , preferably using speed information for objects 908 and 909 , and the desired sample density.
- the image data is forwarded to output means 34 for decoding, storage and/or display (step 1080 ) using methods that are well known to those of skill in the art.
- the present invention has the ability to accurately image multiple side-by-side cartons of varying height being transported on a conveyor system.
Abstract
A system for imaging an object comprising a surface for receiving the object to be imaged and an imaging device. The imaging device comprises a housing and a fixed focusing optical device mounted to the housing, whereby the fixed focusing optics comprises a lens for focusing light reflected from the object. A plurality of independent linear imaging sensors are positioned in the housing at different distances from the fixed focusing lens so that a face of each of the plurality of independent linear imaging sensors is aligned parallel to the surface. A processor is coupled to the plurality of independent linear imaging sensors for receiving an output signal from each of the plurality of independent linear imaging sensors representative of the reflected light. The lens defines a plane parallel to the surface, and the plurality of independent linear imaging sensors are adapted to receive reflected light from the object.
Description
- This application claims priority to U.S. Provisional Patent Application No. 60/737,118, filed Nov. 16, 2005, the entire disclosure of which is incorporated by reference herein.
- The present invention relates generally to optical scanning systems, More particularly, this invention relates to a scanning system having a large depth of field which is capable of focusing on and scanning a plurality of targets over a wide range of distances.
- Various optical scanning systems have been developed for reading and decoding coded symbologies. Such optical scanning systems typically incorporate a light source and a light detecting means, such as a CMOS- or CCD-based image sensor. A portion of the light which is reflected from the coded symbology is detected by the image sensor and converted into an electrical signal, which is the basis for the digital image of the coded symbology that has been scanned. The digital image is then processed and decoded according to the specific type of coded symbology.
- One disadvantage with prior known CMOS- and CCD-based scanning systems is that they require a mechanical focusing apparatus in order to focus the light reflected from the coded symbology onto the image sensor. Since a mechanical focusing apparatus typically operates slower than the image sensor and other electronics within the scanner, the focusing apparatus tends to be the limiting factor in the speed of operation of the scanning system. Additionally, focusing parts are subject to higher wear and maintenance requirements.
- Although there are scanning systems that are available which do not require a moveable focusing apparatus, these scanning systems typically have a very small depth of field.
- U.S. Pat. No. 6,783,068 discloses a large depth of field camera which was developed by Accu-Sort Systems, Inc., the assignee of the present invention. This reference discloses an area array sensor used for scanning a large depth of field. An object height sensor is used to determine an object height so that the correct row of pixels is read. While the system described in U.S. Pat. No. 6,783,068 is effective for scanning a large depth of field, suitable area array sensors are currently available only at a very high cost. Further, such sensors typically transmit large quantities of information that can burden internal or external processing systems.
- Accordingly, there exists a need for an efficient and inexpensive scanning system that can accurately read and decode coded symbologies over a large depth of field to focus light on an image sensor, and which can read symbologies on side-by-side objects of different heights.
- The present invention recognizes and addresses the foregoing disadvantages, and others, of prior art constructions and methods. In one preferred embodiment, the present invention provides a system for imaging an object comprising a surface for receiving an object to be imaged and an imaging device. The imaging device comprising a housing, a fixed focusing optical device mounted to the housing and comprising a lens for focusing light reflected from the object on the surface, a plurality of independent linear imaging sensors are positioned in the housing at different distances from the fixed focusing lens so that a face of each of the plurality of independent linear imaging sensors is aligned parallel to the surface and a processor coupled to the plurality of independent linear imaging sensors for receiving an output signal from each of the plurality of independent linear imaging sensors representative of the reflected light. The lens defines a plane parallel to the surface, and the plurality of independent linear imaging sensors are adapted to receive reflected light from the object on the surface.
- In some embodiments, the plurality of independent linear imaging sensors are located at different distances from the lens such that a plane intersecting each linear row of pixels for each of the plurality of independent linear imaging sensors intersects the lens plane at the Sheimpflug angle such that the focal point of each of the plurality of independent linear imaging sensors differ with respect to the surface. In other embodiments, a height measuring device detects a height of the object on the surface prior to reaching the imaging device, wherein the detected height is used by the processor for selectively reading the output signal from one of the plurality of independent linear imaging sensors. In yet other embodiments, the processor forms a two-dimensional image from the output signals of the plurality of independent linear imaging sensors. In still yet other embodiments, the processor is adapted to read an output signal from individual pixels located in a different one of the plurality of independent linear imaging sensors based on an intensity of the output of the individual pixels. In some or all of the above embodiments, the surface is moveable in a direction perpendicular to an axis of each of the plurality of independent linear imaging sensors.
- In another preferred embodiment, a system for imaging an object comprises a surface for receiving an object to be imaged and an imaging device comprising a housing, a plurality of independent linear imaging sensors positioned in the housing so that a face of each of the plurality of independent linear imaging sensors is parallel to the surface, where the plurality of independent linear imaging sensors is adapted to receive reflected light from the object on the surface, a fixed focusing optical device mounted to the housing and comprising a single lens for focusing the reflected light from the object on the surface onto each of the plurality of independent linear imaging sensors, and a processor coupled to the plurality of independent linear imaging sensors for reading an output signal from each of the plurality of independent linear imaging sensors. The single lens may be an irregular lens that provides a plurality of focal lengths each corresponding to a respective one of the plurality of independent linear imaging sensors. The system may also comprise a plurality of linear actuators for independently moving each of the plurality of independent linear imaging sensors with respect to the fixed focusing optical device. In addition to the linear actuators, a height measuring device may be used for detecting a height of the object on the surface prior to reaching the imaging device, wherein the detected height is used by each of the plurality of linear actuators for moving each of the plurality of independent linear imaging sensors with respect to the fixed focusing device. A feedback loop from the processor to each of the plurality of linear actuators may be used to assist in positioning each of the plurality of independent linear imaging sensors.
- In yet another preferred embodiment, a system for imaging an object comprising a surface for receiving an object to be imaged and an imaging device comprising. The imaging device comprises a housing, a fixed focusing optical device mounted to the housing and comprising a lens for focusing light reflected from the object on the surface, the fixed focusing optical device lens defining a plane parallel to the surface, a linear imaging sensor positioned in the housing radially outward from an axis normal to the lens plane, first and second mirrors rotatably coupled to a pivot axis and positioned so as to direct light entering the fixed focusing optical device onto the linear imaging sensor, and a processor coupled to the linear imaging sensor for selectively reading an output signal from individual pixels in the linear imaging sensor based on an intensity of the output signal from each of the individual pixels. Each of the first and the second mirrors are independently moveable about the pivot axis along a path generally perpendicular to the linear image sensor with respect to each other so that different portions of the surface can be focused onto different portions of the linear sensor. The system may also comprise a first and second actuator for respectively rotating the first and second mirrors about the pivot axis with respect to the linear imaging sensor. Additionally, the system may comprise third and fourth actuators for respectively translating the first and the second mirrors along the pivot axis with respect to the lineal imaging sensor. Moreover, a third mirror may be rotatably coupled to the pivot axis and positioned so as to direct light entering the fixed focusing optical device onto the linear imaging sensor. A height measuring device may be used to detect a height of the object on the surface, wherein the detected height information is used by the first and the second actuators for moving each of the first and second mirrors to focus on the object on the surface.
- A full and enabling disclosure of the present invention, including the best mode thereof directed to one of ordinary skill in the art, is set forth in the specification, which makes reference to the appended drawings, in which:
-
FIG. 1A is a top perspective view of a coded symbology scanning system in accordance with preferred embodiments of the present invention; -
FIG. 1B is bottom perspective view of the coded symbology scanning system ofFIG. 1A ; -
FIG. 2 is a camera according to a first preferred embodiment of the present invention for use in the scanning system ofFIG. 1A ; -
FIG. 3A illustrates a relationship between an object plane, an alternate camera with camera optics, and an image plane defined by a second preferred embodiment of the present invention for use in the scanning system ofFIG. 1A ; -
FIG. 3B illustrates a 3-D spatial relationship between the object plane, camera optics, and image plane ofFIG. 3A ; -
FIG. 3C shows the object plane ofFIG. 3B in greater detail; -
FIG. 3D shows the image plane ofFIG. 3B in greater detail; -
FIGS. 4A and 4B illustrate a 3-D spatial relationship between an object plane, camera optics, and an image plane defined by a third preferred embodiment of the present invention; -
FIG. 5 is a flow diagram of a preferred symbol reading and decoding method; -
FIG. 6 is another alternate camera according to a fourth preferred embodiment of the present invention for use in the scanning system ofFIG. 1A ; -
FIG. 7 illustrates a relationship between an object plane, camera optics and an image plane defined by an embodiment of the scanning system ofFIG. 1A including the alternate camera ofFIG. 6 ; -
FIG. 8 is a flow diagram of another alternative preferred symbol reading and decoding method; -
FIG. 9A is a camera according to a fifth preferred embodiment of the present invention for use in the scanning system ofFIG. 9B ; -
FIGS. 9B is a top perspective view of a coded symbology scanning system in accordance with the fifth preferred embodiment of the present invention; and -
FIG. 10 is a flow diagram of another alternative preferred symbol reading and decoding method. - Repeat use of reference characters in the present specification and drawings is intended to represent same or analogous features or elements of the invention.
- Reference will now be made in detail to presently preferred embodiments of the invention, one or more examples of which are illustrated in the accompanying drawings. Each example is provided by way of explanation of the invention, not limitation of the invention, In fact, it will be apparent to those skilled in the art that modifications and variations can be made in the present invention without departing from the scope or spirit thereof. For instance, features illustrated or described as part of one embodiment may be used on another embodiment to yield a still further embodiment. Thus, it is intended that the present invention covers such modifications and variations as come within the scope of the appended claims and their equivalents. Additional aspects and advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
- Certain terminology is used in the following description for convenience only and is not considered limiting. Words such as “front”, “back”, “top” and “bottom” designate directions in the drawings to which reference is made. This terminology includes the words specifically noted above, derivatives thereof and words of similar import. Additionally, the terms “a” and “one” are defined as including one or more of the referenced item unless specifically noted. The phrase “at least one” followed by a list of two or more items (such as A, B, or C) means any individual one of A, B or C as well as any combination thereof.
- Referring to
FIGS. 1A and 1B , ascanning system 10 made in accordance with the present invention is shown.Scanning system 10 is able to scan any type of object. For simplicity, however,scanning system 10 will be explained with reference to scanning a coded symbology, and in particular a bar code symbol.Scanning system 10 includes alight source 12 and acamera 17 located within ahousing 14. In a first preferred embodiment of the present invention,camera 17 includes a plurality of independent linear image sensors 26 a-26 c arranged side by side along a plane (FIG. 2 ). Alternatively, any suitable number of independent linear image sensors greater than 1 can be used. Those skilled in the art should know that any line scan image sensor may be utilized without departing from the spirit and scope of the present invention, for example either a CMOS or a CCD linear image sensor. -
Light source 12 preferably includes a plurality ofLEDs 13 mounted in a focusingcontainer 15 for directing light emitted byLEDs 13.Light source 12 facilitates detection of subjectbar code symbols bar code symbols objects 8 and 9. Preferably, objects 8 and 9 are supported by a movingtransport system 7. Thescanning system 10 is aligned such that light reflected frombar code symbols camera housing 14 to be detected by linear image sensors 26 a-26 c. As will be explained in detail hereinafter, linear image sensors 26 a-26 c are optically aligned parallel to the surface of thetransport system 7 and perpendicular to the motion oftransport system 7. It should be recognized by those skilled in the art thatFIGS. 1A and 1B are illustrative only and are not drawn to scale. - A
tachometer 19 monitors the speed of the surface oftransport system 7 andobjects 8 and 9 and provides this information to a processor 32 (shown inFIG. 2 ) to control the read-out rate of the plurality of linear image sensors 26 a-26 c. Those skilled in the art should know that there are several other methods of monitoring the speed oftransport system 7 andobjects 8 and 9 and providing this information toprocessor 32 to control the read-out rate of linear image sensors 26 a-26 c that may be used as well. - Referring to
FIG. 2 , in a first preferred embodiment of the present invention,camera 17 includes fixed focusingoptics 31, a plurality of independent linear image sensors 26 a-26 c, aprocessor 32 and an output means 34. Linear image sensors 26 a-26 c include a column of pixels 28 a-28 n. In this embodiment, fixed focusingoptics 31 includes at least one custom optic having more than one focal length, for example, an irregularly shaped lens. Preferably, the custom optic provides a unique focal length for each oflinear image sensors 26 a 26 c such that objects of different heights positioned on thetransport system 7, corresponding to different depths of field, are in focus (or nearly in focus) on different ones of linear image sensors 26 a-26 c. This configuration allows two or more adjacent or side-by side items having different heights, forexample objects 8 and 9 positioned ontransport system 7 inFIGS. 1A and 1B , to be viewed in focus (or nearly in focus) simultaneously by different ones oflinear image sensors 26 a 26 c. - The focusing
optics 31 focuses reflected light 29 onto the appropriate one oflinear image sensors 26 a 26 c based on a depth of field to an object being scanned.Linear image sensors 26 a 26 c detect reflected light 29 and each pixel within each oflinear image sensors 26 a 26 c converts the incident light energy into a stored electrical charge and ultimately into an electrical signal. The electrical signal comprises a value that represents the amount of light detected by a pixel. If the pixel's analog signal is digitized to have eight-bit resolution, the number of gray scale levels will be 256. - The resolution of the linear scan image is dependent on the number of
pixels 28 a 28 n in each oflinear image sensors 26 a 26 c .Linear image sensors 26 a 26 c are preferably aligned parallel to each other and transverse to the direction of travel oftransport system 7. Each of thelinear image sensors 26 a 26 c can be considered one row in a series of parallel lines when projected into the object plane. - According to a second preferred embodiment of the present invention, independent
linear image sensors 26 a 26 c are preferably positioned at different distances fromoptics 31 so that an image plane I intersects each row of pixels in each of the linear image sensors, where the image plane is disposed at the Scheimpflug angle with respect to amedian plane 33 ofoptics 31, as shown inFIG. 3A . Thus, each pixel in anyspecific column 28 a 28 n will have a corresponding unique focus point in anobject plane 0 that also intersects withmedian plane 33 and image plane I along aline 35. Thesystem 10 of the present invention utilizes the selective positioning oflinear image sensors 26 a 26 c in order to read out information pertaining to the objects within the depth of field that are in focus or nearly in focus on each linear image sensor. - A
processor 32 accesses thelinear image sensors 26 a 26 c to obtain the desired information. The processor receives the data from all independent linear imaging sensors and may (1) process and store the output of one linear image sensor that is receiving data in focus or (2) process and store all data from each of the linear image sensors. The processor uses the processed data to form one or more 2-D images from the individual line scans detected by each linear image sensor. That is, if the data from one linear image scanner is selected, then a single 2-D image is formed from the plurality of line scans captured by the selected linear imager. If, on the other hand, two or more independent linear image sensors are selected, than an independent 2-D image is formed for each selected linear image sensor using that sensors respective data. Atachometer 19 can be used to determine a scan rate forsensors 26 a 26 c . This information is then output to output means 34 for further processing. Output means 34 may be a digital display, an image processing system, a positioning system, or any other type of system that will accept an output from an imaging camera. - Referring to
FIG. 3A ,linear image sensors 26 a 26 c are positioned at different distances fromoptics 31 so that image plane I intersects each row of pixels in each of the linear image sensors, where image plane I is disposed at an angle with respect tomedian plane 33 throughoptics 31. The angle, (hereinafter, the Scheimpflug angle), is determined using the Scheimpflug condition. The Scheimpflug angle is dependent on the focal length ofoptics 31 and a desired working range defining the depth of field (DoF) along object plane O. By positioninglinear image sensors 26 a 26 c along the Scheimpflug angle, an object positioned within the depth of field (DoF) is imaged in focus, or close to in focus, onto the active portion of one of thelinear image sensors 26 a 26 c. - Focusing is improved for a wider array of objects of differing heights over a given depth of field (DoF) by increasing the number of linear image sensors and decreasing spacing between adjacent linear imaging sensors. Each horizontal line in object plane O has a corresponding horizontal line in image plane I that is in focus. For example, each of
linear image sensors 26 a 26 c in image plane I corresponds to a specific distance in object plane O; and each such distance in object plane O corresponds to a line parallel to a surface oftransport system 7 and perpendicular to the direction of motion oftransport system 7. Referring toFIG. 3B , the spatial relationship between object plane O,optics 31 and image plane I is shown.FIG. 3C shows object plane O ofFIG. 3B in greater detail, andFIG. 3D shows image plane I ofFIG. 3B in greater detail. - As shown in
FIGS. 3A-3D , an object O1 in object plane O at a distance D1 fromoptics 31 will be in focus or close to in focus in image plane I onlinear image sensor 26 a on its string of pixels I1. Likewise, an object O2 in object plane O at distance D2 will be in focus or close to in focus in image plane I onlinear image sensor 26 b at its string of pixels I2, and an object O3 in object plane O at distance D3 will be in focus or close to in focus in image plane I on thelinear image sensor 26 c at its string of pixels I3. Accordingly, for any plane parallel to the surface of thetransport system 7 in object plane O within the depth of field (DoF), a corresponding one oflinear image sensors 26 a 26 c will be in focus or close to in focus. Therefore,linear image sensors 26 a 26 c having an object plane focal point distance, for example one of distances D1, D2, D3 that best corresponds to an object height, for example a height represented by objects O1, 02, 03, will be in optimal focus. - Depending on the quality of illumination and the distance between
linear image sensors 26 a 26 c , each of thesensors 26 a 26 c will receive some light regardless of the height of object O1, O2, O3. However, the linear image sensor in optimal focus will receive the most light. Preferably,processor 32 selects the data received from one ofsensor 26 a 26 c that receives the most light, indicating optimal focus, to be processed and stored. Through comparing the intensity of the signals read by eachlinear image sensors 26 a 26 c , the appropriate in focus linear image sensor 26 can be read and the data further processed. The data from linear image sensor 26 that is selected byprocessor 32 is sent to output means 34 for further processing, decoding, storage and/or display as is well known by those skilled in the art. Data is periodically read out oflinear image sensors 26 a 26 c as will be described in greater detail hereinafter in order to read coded symbology on objects carried past the scanner. - Referring to
FIGS. 4A and 4B , a third preferred embodiment of the present invention is shown in which a system 110 includes acompact camera 117 having only twolinear image sensors optics 131.Linear image sensors optics 131 so that a common image plane I intersects each row of pixels in each of the linear image sensors, where image plane I is disposed at an angle with respect to amedian plane 133 throughoptics 131 at the Scheimpflug angle. As discussed above with reference to the first and second embodiments, an object plane O intersects with image plane I andmedian plane 133 alongline 135 so that each horizontal line in an object plane O has a corresponding horizontal line in image plane I that is in focus. Thus,sensor 126 b is generally in focus between O1 and O2 andsensor 126 a is generally in focus between O2 and O3, thereby increasing the overall depth of field ofcamera 117. Preferably, system 110 includes a conveyor for transporting objects where symbology to be read is within the depth of field (DoF) of the camera. In this embodiment,linear image sensors - The first, second and third embodiments of the invention allow two or more singulated or non-singulated objects having substantially different heights positioned on
transport system 7 inFIG. 1 , forexample objects 8 and 9, to be viewed in focus (or nearly in focus) by differentlinear image sensors 26 a 26 c or 126 a-126 b without a mechanical focusing or height detection device. Accordingly, information received simultaneously from adjacent objects can be processed simultaneously by the system recognizing and reading in focus data from more than one linear image sensor. - In operation,
scanning systems 10 and 110 according to the first, second or third preferred embodiments of the present invention execute the bar code symbol reading anddecoding procedure 400 shown inFIG. 5 . Alight source 12 illuminates a subject bar code symbol 18 (step 410). Light is reflected frombar code symbol 18 towardcamera 17 or 117 (step 420). The reflected light enterscamera optics 31 or 131 (step 430). Focusingoptics linear image sensors 26 a 26 c or 126 a-126 b (step 440). Each pixel within each oflinear image sensors 26 a 26 c or 126 a-126 b converts the light into a charge which is converted into an electrical signal (step 450) that is transmitted toprocessor 32.Processor 32 compares the signals transmitted from each of the plurality oflinear image sensors 26 a 26 c or 126 a-126 b (step 455). Stronger signals, for example, represent a larger quantity of light received by a pixel, which indicates better focusing at the particular pixel location.Processor 32 preferably selects one or more linear image sensors 26 a-26 c or 126 a-126 b, which are in relatively good focus compared with other ones of the plurality oflinear image sensors 26 a 26 c or 126 a-126 b, based on a comparison of the signal outputs of the pixels of the linear image sensors (step 470). Alternatively, only certain in-focus pixels from one or more oflinear image sensors 26 a 26 c or 126 a-126 b can be selected. Also alternatively, a height detection device, for example, an overhead laser triangulation device or a light curtain, can be used for providingprocessor 32 with object height information to be used in selecting one ormore sensors 26 a 26 c or 126 a-126 b that are in focus. - The information from the selected linear image sensor(s) 26 a 26 c or 126 a-126 b, or from only selected pixels within selected
sensors 26 a 26 c or 126 a--126 b, is read out at a periodic rate. This periodic rate can be determined byprocessor 32 using speed information ofobjects 8 and 9 (preferably determined bytachometer 19 as shown in the embodiment ofFIGS. 1A and lB), and the desired sample density. The image data is forwarded to output means 34 for decoding, storage and/or display (step 480) using methods that are well known to those of skill in the art. - The periodic rate at which information is read out from the selected sensors can alternatively be set at a fixed value without utilizing speed information of the objects and without using a tachometer. Referring to the third embodiment shown in
FIGS. 4A and 4B ,camera 117 of scanning system 110 is configured to scan at a fixed rate for barcodes within its depth of field (DoF) on packages passing at any suitable speed. The image data is forwarded to an output device for decoding, storage and/or display (Seestep 480,FIG. 5 ). This type of scanning system allows easy compatibility with many pre-existing package conveying systems and is suitable for lower cost applications. - Referring to
FIGS. 6 and 7 , in a fourth preferred embodiment of the present invention, analternative camera 717 replacescamera 17 in thesystem 10 for scanning an object.Camera 717 is configured in a manner similar tocamera 17 ofFIG. 2 , but utilizes actuators 736 a-736 c, preferably servo motors, to translate linear image sensors 726 a-726 c in a direction toward or away fromlens 31. In this embodiment, a custom optic is not required, but may be provided if suitable for a particular application. Sensors 726 a-726 c preferably travel along a path perpendicular tolens 31. In this manner, each sensor 726 a-726 c can be individually focused to more precisely sense light reflected from objects ontransport system 7. - The ability to move sensors 726 a-726 c not only allows greater precision in focusing, but also increases the effective range of depth of field of
system 10. While three sensors 726 a-726 c with corresponding respective actuators 736 a-736 c are shown, any suitable number of sensors can be used. It should be clear to those skilled in the art that for each additional linear image sensor and associated actuator provided, an additional adjacent object of unique height can be simultaneously scanned. For example, a system with two linear image sensors can simultaneously scan two side-by-side objects of different heights, and a system with three linear image sensors can simultaneously scan three side-by-side objects of different heights. A height detection device, for example, an overheadlaser triangulation device 740 or alternatively a light curtain can be used for providingprocessor 32 with object height information for selecting and focusing sensors 726 a-726 c. -
Scanning system 10 utilizingcamera 717 in place ofcamera 17 executes an alternative bar code symbol reading anddecoding procedure 800 shown inFIG. 8 . Alight source 12 illuminates a subject bar code symbol 18 (step 810). Light is reflected frombar code symbol 18 toward camera 717 (step 820). The reflected light enterscamera 717 via focusing optics 31 (step 830). Focusingoptics 31 focuses light 29 onto linear image sensors 726 a-726 c (step 840). Each pixel on each of linear image sensors 726 a-726 c converts the light into a charge that is converted into an electrical signal (step 850). -
Processor 32 compares the signals transmitted from each of linear image sensors 726 a-726 c (step 855). Stronger signals, for example, represent a larger quantity of light received by a pixel, which indicates better focusing at the particular pixel row.Processor 32 preferably selects one or more linear image sensors 726 a-726 c that are in relatively good focus compared with other ones of the plurality of linear image sensors 726 a-726 c based on a comparison of the signal outputs of the pixels of the linear image sensors (step 870). Alternatively, only certain in-focus pixels from one or more of the linear image sensors 726 a-726 c can be selected, or all of the linear image sensors may be selected, read and stored for further processing. -
Processor 32 controls one or more of actuators 736 a-736 c to translate the respective selected one or more linear image sensors 726 a-726 c (step 875) based on charge emitted from selected linear image sensor(s) 726 a-726 c. Preferably, during actuation, actuators 736 a-736 c receive feedback at discrete time intervals fromprocessor 32 to enable actuators 736 a-736 c to position selected linear image sensor(s) 726 a-726 c in a best focus position. Alternatively, in addition to or instead of receiving signals from linear image sensors 726 a-726 c, the processor can receive signals from height detection devices, for example the light curtain or overheadlaser triangulation device 740, in order to select and position appropriate linear image sensor(s) 726 a-726 c. - The information from selected linear image sensor(s) 726 a-726 c, from all linear image sensors or from only selected pixels within selected sensors 726 a-726 c is read out at a periodic rate. This periodic rate is determined by
processor 32 preferably using speed information ofobjects 8 and 9, (determined by the tachometer 19), and the desired sample density. The image data is forwarded to output means 34 for decoding, storage and/or display (step 880) using methods that are well known to those of skill in the art. - Referring to a fifth embodiment of the invention as shown in
FIGS. 9A and 9B , asystem 910 has acamera 917 preferably having a singlelinear image sensor 926 and first and second adjustableadjacent mirrors optics 931 ontolinear image sensor 926.Optics 931 preferably includes a single lens, but alternatively, the optics can have multiple lenses in any suitable configuration.System 910 transports objects, for example objects 908 and 909, on atransport system 907 in a manner similar to that provided in the first embodiment shown inFIGS. 1A and 1B . -
Camera 917 utilizes first, second, third and fourth actuators 936 a-d, which are preferably voice coil actuators, or alternatively servo motors, to adjustadjustable mirrors second actuators axis 941 in a direction toward or away fromlinear image sensor 926. In this embodiment, a custom optic is not required, but may be provided if suitable for a particular application. During pivoting, mirrors 927 a and 927 b preferably travel along a path generally perpendicular tolinear image sensor 926. In this manner, a first portion oflinear image sensor 926 receives reflected light from firstadjustable mirror 927 a and a second portion oflinear image sensor 926 receives reflected light from secondadjustable mirror 927 b. Accordingly, two separate portions oflinear image sensor 926, corresponding to two separate portions of the conveyor, can be individually focused to more precisely sense light reflected from objects, for example theobjects 908 and 909 ontransport system 907, as shown inFIG. 9B . - The ability to move
mirrors system 910. Preferably, third andfourth actuators mirrors axis 941. In this manner, if an object with a barcode or other coded symbology is positioned in a location such that there is a potential for the object to be located in a gap G betweenmirrors linear image sensor 926 so that any coded symbology can be read without gap G disrupting it. Alternatively, a single actuator can replace third andfourth actuators mirrors - While two
mirrors laser triangulation device 740 as shown inFIG. 7 , can be used to provideprocessor 32 with object height information for each of the objects for adjustingmirrors -
Scanning system 910 utilizingcamera 917 executes an alternative bar code symbol reading anddecoding procedure 1000 shown inFIG. 10 . A light source illuminates subjectbar code symbols bar code symbols 918 a and 918 h toward camera 917 (step 1020). The reflected light enterscamera 917 via focusing optics 931 (step 1030), which focus light 29 ontomirrors light 29 onto linear image sensor 926 (step 1040). Each pixel onlinear image sensor 926 converts the light into a charge which is converted into an electrical signal (step 1050).Processor 32 compares the signals transmitted from different portions of linear image sensor 926 (step 1055). Stronger signals, for example, represent a larger quantity of light received by particular pixels 928-928 n indicating better focusing at a particular portion of the linear image sensor. -
Processor 32 controls one or more of actuators 936 a-d to displace one or more ofmirrors linear image sensor 926. Preferably, during actuation, actuators 936 a-d receive feedback at discrete time intervals fromprocessor 32 to enable actuators 936 a-d to position mirrors 927 a and 927 b in a best focus position. Alternatively, in addition to or instead of receiving signals from the different portions oflinear image sensor 926, the processor can receive signals from height detection devices, for example a light curtain or overheadlaser triangulation device 740 shown inFIG. 7 , in order to select and position one or both ofmirrors - The information from
linear image sensor 926 or from only selected pixels withinlinear image sensor 926 is read out at a periodic rate. This periodic rate can be set or optionally can be determined byprocessor 32, preferably using speed information forobjects 908 and 909, and the desired sample density. The image data is forwarded to output means 34 for decoding, storage and/or display (step 1080) using methods that are well known to those of skill in the art. - As disclosed above, the present invention has the ability to accurately image multiple side-by-side cartons of varying height being transported on a conveyor system.
- While specific embodiments of the present invention have been shown and described, many modifications and variations could he made by one skilled in the art without departing from the spirit and scope of the invention. For example, although a stationary scanner has been utilized in the above embodiments, the embodiments and principles described herein may also be applied to mobile or hand-held scanners. The above description serves to illustrate and not limit the particular form in any way.
Claims (20)
1. A system for imaging an object comprising:
a. a surface for receiving an object to be imaged; and
b. an imaging device comprising:
i. a housing;
ii. a fixed focusing optical device mounted to said housing and comprising a lens for focusing light reflected from the object on said surface, said lens defining a plane parallel to said surface;
iii. a plurality of independent linear imaging sensors positioned in said housing at different distances from said fixed focusing lens so that a face of each of said plurality of independent linear imaging sensors is aligned parallel to said lens plane, wherein said plurality of independent linear imaging sensors are adapted to receive said reflected light from the object on said surface; and
c. a processor coupled to said plurality of independent linear imaging sensors receives an output signal from each of said plurality of independent linear imaging sensors representative of said reflected light.
2. The system for imaging an object of claim 1 , wherein said processor reads said output signal from said plurality of independent linear imaging sensors and selectively stores the data from one of said plurality of imaging sensors based on an intensity of light received by each of said plurality of independent linear imaging sensors, wherein said intensity of light is directly related to an amplitude of said output signal from each of said plurality of independent linear imaging sensors.
3. The system for imaging an object of claim 1 , further comprising a plurality of light sources mounted in said housing proximate said fixed focusing optical device for illuminating at least one surface of the object placed on said surface.
4. The system for imaging an object of claim 1 , wherein said plurality of independent linear imaging sensors are located at distances from said lens plane such that a plane intersecting each linear row of pixels for each of said plurality of independent linear imaging sensors intersects said lens plane at the Sheimpflug angle such that the focal point of each of said plurality of independent linear imaging sensors differ with respect to said surface.
5. The system for imaging an object of claim 1 , further comprising a height measuring device for detecting a height of the object on said surface prior to reaching said imaging device, wherein said detected height is used by said processor for selectively reading said output signal from one of said plurality of independent linear imaging sensors.
6. The system for imaging an object of claim 1 , wherein said processor forms a two-dimensional image from said output signals of said plurality of independent linear imaging sensors.
7. The system for imaging an object of claim 1 , wherein said processor is adapted to read an output signal from individual pixels located in a different one of said plurality of independent linear imaging sensors based on an intensity of said output of said individual pixels.
8. The system for imaging an object of claim 1 , wherein said surface is moveable in a direction perpendicular to an axis of each of said plurality of independent linear imaging sensors.
9. A system for imaging an object comprising:
a. a surface for receiving an object to be imaged; and
b. an imaging device comprising:
i. a housing;
ii. a plurality of independent linear imaging sensors being positioned in said housing so that a face of each of said plurality of independent linear imaging sensors is parallel to said surface, said plurality of independent linear imaging sensors being adapted to receive reflected light from the object on said surface; and
iii. a fixed focusing optical device mounted to said housing and comprising a single lens for focusing said reflected light from the object on said surface onto each of said plurality of independent linear imaging sensors;
c. a processor coupled to said plurality of independent linear imaging sensors for reading an output signal from each of said plurality of independent linear imaging sensors.
10. The system for imaging an object of claim 9 , wherein said single lens is an irregular lens that provides a plurality of focal lengths each corresponding to a respective one of said plurality of independent linear imaging sensors.
11. The system for imaging an object of claim 9 , further comprising a plurality of linear actuators for independently moving each of said plurality of independent linear imaging sensors with respect to said fixed focusing optical device.
12. The system for imaging an object of claim 11 , further comprising a height measuring device for detecting a height of the object on said surface prior to reaching said imaging device, wherein said detected height is used by each of said plurality of linear actuators for moving each of said plurality of independent linear imaging sensors with respect to said fixed focusing device.
13. The system for imaging an object of claim 12 , further comprising a feedback loop from said processor to each of said plurality of linear actuators to assist in positioning each of said plurality of independent linear imaging sensors.
14. The system for imaging an object of claim 13 , wherein said processor is adapted to read an output signal from individual pixels located in a different one of said plurality of independent linear imaging sensors based on an intensity of said output of said individual pixels.
15. The system for imaging an object of claim 10 , further comprising a height measuring device for detecting a height of the object on said surface prior to reaching said imaging device, wherein said detected height is used by said processor for selectively reading said output signal from one of said plurality of independent linear imaging sensors.
16. A system for imaging an object comprising:
a. a surface for receiving an object to be imaged; and
b. an imaging device comprising:
i. a housing;
ii. a fixed focusing optical device mounted to said housing and comprising a lens for focusing light reflected from the object on said surface, said fixed focusing optical device lens defining a plane parallel to said surface;
iii. a linear imaging sensor positioned in said housing radially outward from an axis normal to said plane;
iv. first and second mirrors rotatably coupled to a pivot axis and positioned so as to direct light entering said fixed focusing optical device onto said linear imaging sensor; and
c. a processor coupled to said linear imaging sensor for reading an output signal from individual pixels in said linear imaging sensor based on an intensity of said output signal from each of said individual pixels,
wherein each of said first and said second mirrors are independently moveable about said pivot axis along a path generally perpendicular to said linear image sensor with respect to each other so that different portions of said surface can be focused onto different portions of said linear sensor.
17. The system for imaging an object of claim 16 , further comprising a first and second actuator for respectively rotating said first and second mirrors about said pivot axis with respect to said linear imaging sensor.
18. The system for imaging an object of claim 17 , further comprising third and fourth actuators for respectively translating said first and said second mirrors along said pivot axis with respect to said lineal imaging sensor.
19. The system for imaging an object of claim 16 , further comprising a third mirror rotatably coupled to said pivot axis and positioned so as to direct light entering said fixed focusing optical device onto said linear imaging sensor.
20. The system for imaging an object of claim 16 , further comprising a height measuring device for detecting a height of the object on said surface, wherein said detected height information is used by said first and said second actuators for moving each of said first and second mirrors to focus on the object on said surface.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/559,669 US20070164202A1 (en) | 2005-11-16 | 2006-11-14 | Large depth of field line scan camera |
EP06255861A EP1791343A3 (en) | 2005-11-16 | 2006-11-16 | Large depth of field line scan camera |
US12/562,478 US8110790B2 (en) | 2005-11-16 | 2009-09-18 | Large depth of field line scan camera |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US73711805P | 2005-11-16 | 2005-11-16 | |
US11/559,669 US20070164202A1 (en) | 2005-11-16 | 2006-11-14 | Large depth of field line scan camera |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/562,478 Continuation US8110790B2 (en) | 2005-11-16 | 2009-09-18 | Large depth of field line scan camera |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070164202A1 true US20070164202A1 (en) | 2007-07-19 |
Family
ID=37907537
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/559,669 Abandoned US20070164202A1 (en) | 2005-11-16 | 2006-11-14 | Large depth of field line scan camera |
US12/562,478 Expired - Fee Related US8110790B2 (en) | 2005-11-16 | 2009-09-18 | Large depth of field line scan camera |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/562,478 Expired - Fee Related US8110790B2 (en) | 2005-11-16 | 2009-09-18 | Large depth of field line scan camera |
Country Status (2)
Country | Link |
---|---|
US (2) | US20070164202A1 (en) |
EP (1) | EP1791343A3 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100200726A1 (en) * | 2009-02-09 | 2010-08-12 | Cedrone Louis J | Apparatus and method for supporting and aligning imaging equipment on a web converting manufacturing line |
US20130242086A1 (en) * | 2012-03-13 | 2013-09-19 | Google Inc. | Method and System for Identifying Depth Data Associated with an Object |
US20140036069A1 (en) * | 2012-07-31 | 2014-02-06 | Sick Ag | Camera system and method for detection of flow of objects |
US20140141539A1 (en) * | 2012-11-22 | 2014-05-22 | Samsung Electronics Co., Ltd. | Apparatus and method of recognizing an object, and apparatus and method of mounting a semiconductor chip |
US20150334374A1 (en) * | 2014-05-14 | 2015-11-19 | 3M Innovative Properties Company | 3d image capture apparatus with depth of field extension |
US9294672B2 (en) | 2014-06-20 | 2016-03-22 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax and tilt artifacts |
US9374516B2 (en) | 2014-04-04 | 2016-06-21 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
WO2016102988A1 (en) * | 2014-12-24 | 2016-06-30 | Datalogic Automation, Inc. | Multiline scanner |
US9383550B2 (en) | 2014-04-04 | 2016-07-05 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
US9386222B2 (en) | 2014-06-20 | 2016-07-05 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax artifacts |
US9398264B2 (en) | 2012-10-19 | 2016-07-19 | Qualcomm Incorporated | Multi-camera system using folded optics |
US9438889B2 (en) | 2011-09-21 | 2016-09-06 | Qualcomm Incorporated | System and method for improving methods of manufacturing stereoscopic image sensors |
US9485495B2 (en) | 2010-08-09 | 2016-11-01 | Qualcomm Incorporated | Autofocus for stereo images |
CN106225765A (en) * | 2016-07-25 | 2016-12-14 | 浙江大学 | A kind of many line scan image sensors obtain device and the formation method of hyperfocal distance scanning imagery |
US9541740B2 (en) | 2014-06-20 | 2017-01-10 | Qualcomm Incorporated | Folded optic array camera using refractive prisms |
US9549107B2 (en) | 2014-06-20 | 2017-01-17 | Qualcomm Incorporated | Autofocus for folded optic array cameras |
US9819863B2 (en) | 2014-06-20 | 2017-11-14 | Qualcomm Incorporated | Wide field of view array camera for hemispheric and spherical imaging |
US9832381B2 (en) | 2014-10-31 | 2017-11-28 | Qualcomm Incorporated | Optical image stabilization for thin cameras |
US10013764B2 (en) | 2014-06-19 | 2018-07-03 | Qualcomm Incorporated | Local adaptive histogram equalization |
US10178373B2 (en) | 2013-08-16 | 2019-01-08 | Qualcomm Incorporated | Stereo yaw correction using autofocus feedback |
US10832023B2 (en) | 2017-12-15 | 2020-11-10 | Cognex Corporation | Dual-imaging vision system camera and method for using the same |
WO2022011036A1 (en) * | 2020-07-08 | 2022-01-13 | Cognex Corporation | System and method for extending depth of field for 2d vision system cameras in the presence of moving objects |
US11301655B2 (en) | 2017-12-15 | 2022-04-12 | Cognex Corporation | Vision imaging system having a camera and dual aimer assemblies |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2556533A1 (en) * | 2005-08-24 | 2007-02-24 | Degudent Gmbh | Method of determining the shape of a dental technology object and apparatus performing the method |
CN101867678B (en) * | 2010-04-27 | 2012-07-11 | 王震 | Oversized picture scanner and scanning method thereof |
EP3368235A4 (en) * | 2015-10-30 | 2019-07-03 | Seurat Technologies, Inc. | Chamber systems for additive manufacturing |
CN109414819A (en) * | 2016-01-08 | 2019-03-01 | 4D空间天才股份有限公司 | The robot obtained for automated image |
KR20180086741A (en) * | 2017-01-23 | 2018-08-01 | 주식회사 가치소프트 | Camera module for focusing based on rotation and method for processing product using the camera module |
CN107292380B (en) * | 2017-06-29 | 2020-11-10 | 吉林大学 | Card counter and new method for counting cards by using card counter |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4404594A (en) * | 1981-11-02 | 1983-09-13 | Itek Corporation | Imaging system with enlarged depth of field |
US4672458A (en) * | 1984-12-28 | 1987-06-09 | North American Philips Corporation | Projection television systems having lens tilt compensation member |
US4833724A (en) * | 1986-08-14 | 1989-05-23 | Amada Engineering & Service Co., Inc. | Imaging device |
US4939369A (en) * | 1988-10-04 | 1990-07-03 | Loral Fairchild Corporation | Imaging and tracking sensor designed with a sandwich structure |
US4963756A (en) * | 1989-10-13 | 1990-10-16 | Hewlett-Packard Company | Focused line identifier for a bar code reader |
US4965840A (en) * | 1987-11-27 | 1990-10-23 | State University Of New York | Method and apparatus for determining the distances between surface-patches of a three-dimensional spatial scene and a camera system |
US5025313A (en) * | 1989-10-16 | 1991-06-18 | Eastman Kodak Company | System for minimizing optical distortions and chromatic aberrations in a linear color scanner |
US5231443A (en) * | 1991-12-16 | 1993-07-27 | The Research Foundation Of State University Of New York | Automatic ranging and automatic focusing |
US5521366A (en) * | 1994-07-26 | 1996-05-28 | Metanetics Corporation | Dataform readers having controlled and overlapped exposure integration periods |
US5521700A (en) * | 1992-03-31 | 1996-05-28 | Nidek Co., Ltd. | Meter for testing progressive multifocus lenses |
US5530514A (en) * | 1994-07-19 | 1996-06-25 | Eastman Kodak Company | Direct focus feedback autofocus system |
US5689347A (en) * | 1995-03-03 | 1997-11-18 | Canon Kabushiki Kaisha | Signal processing apparatus |
US5841881A (en) * | 1994-09-22 | 1998-11-24 | Nec Corporation | Label/window position detecting device and method of detecting label/window position |
US5923017A (en) * | 1997-01-23 | 1999-07-13 | United Parcel Service Of America | Moving-light indicia reader system |
US6028299A (en) * | 1995-12-05 | 2000-02-22 | Sony Corporation | Linear image sensor device comprising first and second linear image sensor sections having first and second sensitivities |
US20020040932A1 (en) * | 2000-10-06 | 2002-04-11 | Citizen Electronics Co., Ltd. | Hand scanner |
US20030019930A1 (en) * | 2000-03-21 | 2003-01-30 | Accu-Sort Systems, Inc. | Large depth of field line scan camera |
US6621063B2 (en) * | 2001-06-21 | 2003-09-16 | Psc Scanning, Inc. | Omni-directional optical code reader using scheimpflug optics |
US7040538B2 (en) * | 2001-10-26 | 2006-05-09 | Symbol Technologies, Inc. | Bar code reader including linear sensor array and hybrid camera and bar code reader |
US20060180670A1 (en) * | 2004-12-01 | 2006-08-17 | Psc Scanning, Inc. | Triggering illumination for a data reader |
US20070115473A1 (en) * | 2005-11-23 | 2007-05-24 | Jean-Luc Legoupil | Optical method and device for detecting surface and structural defects of a travelling hot product |
Family Cites Families (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3989947A (en) | 1971-03-01 | 1976-11-02 | Hughes Aircraft Company | Telescope cluster |
FR2222666B1 (en) | 1973-03-19 | 1976-11-05 | Thomson Brandt | |
US4667255A (en) | 1979-09-24 | 1987-05-19 | Datacopy Corporation | Electronic camera employing a solid-state image sensor |
US4457580A (en) | 1980-07-11 | 1984-07-03 | Mattel, Inc. | Display for electronic games and the like including a rotating focusing device |
JPH0620223B2 (en) * | 1986-06-30 | 1994-03-16 | 富士写真フイルム株式会社 | Image information reading and recording device |
US5308966A (en) | 1986-08-08 | 1994-05-03 | Norand Corporation | Hand-held instant bar code reader having automatic focus control for operation over a range of distances |
JPS6360416A (en) | 1986-09-01 | 1988-03-16 | Konica Corp | Automatic focusing device |
GB2196504B (en) | 1986-10-18 | 1990-09-26 | Gen Electric Plc | A focussing technique for an optical sensing system |
US5811828A (en) | 1991-09-17 | 1998-09-22 | Norand Corporation | Portable reader system having an adjustable optical focusing means for reading optical information over a substantial range of distances |
US5196696A (en) * | 1992-02-10 | 1993-03-23 | Ncr Corporation | Counter rotating optical scanning unit with Fresnel reflection filtering |
US5245172A (en) | 1992-05-12 | 1993-09-14 | United Parcel Service Of America, Inc. | Voice coil focusing system having an image receptor mounted on a pivotally-rotatable frame |
US5442167A (en) | 1993-04-16 | 1995-08-15 | Intermec Corporation | Method and apparatus for automatic image focusing |
US5610730A (en) | 1994-08-16 | 1997-03-11 | Osipchuk; Yuri V. | Low-cost high-resolution digital scanning camera |
US5485263A (en) | 1994-08-18 | 1996-01-16 | United Parcel Service Of America, Inc. | Optical path equalizer |
JPH09247553A (en) | 1996-03-01 | 1997-09-19 | Minolta Co Ltd | Line sensor camera |
US6233014B1 (en) | 1996-04-17 | 2001-05-15 | Minolta Co., Ltd. | Line sensor camera without distortion in photo image |
US5793424A (en) | 1996-04-22 | 1998-08-11 | Visidyne, Inc. | Synchronously gated high speed CCD imager system |
US5717512A (en) | 1996-05-15 | 1998-02-10 | Chmielewski, Jr.; Thomas A. | Compact image steering and focusing device |
US5909302A (en) | 1996-08-02 | 1999-06-01 | Guissin; Rami | Staring scanner |
ATE244425T1 (en) | 1996-12-30 | 2003-07-15 | Datalogic Spa | METHOD AND MACHINE FOR READING AND ASSIGNING AN OPTICAL CODE |
JPH112859A (en) | 1997-06-12 | 1999-01-06 | Minolta Co Ltd | Camera |
US6002815A (en) | 1997-07-16 | 1999-12-14 | Kinetic Sciences, Inc. | Linear sensor imaging method and apparatus |
US6185044B1 (en) | 1998-01-19 | 2001-02-06 | Fuji Photo Optical Co., Ltd. | TV lens with still-taking function |
US6064423A (en) | 1998-02-12 | 2000-05-16 | Geng; Zheng Jason | Method and apparatus for high resolution three dimensional display |
JPH11295051A (en) | 1998-02-25 | 1999-10-29 | Hewlett Packard Co <Hp> | Confocal microscopy apparatus and method therefor |
US6130993A (en) | 1998-04-20 | 2000-10-10 | Asahi Kogaku Kogyo Kabushiki Kaisha | Camera having an image stabilizer |
EP0984319B1 (en) * | 1998-07-08 | 2004-01-02 | Hewlett-Packard Company, A Delaware Corporation | Camera with keystone distortion correction |
US6741284B1 (en) | 1998-09-25 | 2004-05-25 | Pentax Corporation | SLR digital still camera |
US6801260B1 (en) | 2000-02-15 | 2004-10-05 | Accu-Sort Systems, Inc. | Automatic focusing camera with moving mirror between fixed lens and fixed image sensor |
US6414779B1 (en) * | 2000-11-30 | 2002-07-02 | Opeical Biopsy Technologies, Inc. | Integrated angled-dual-axis confocal scanning endoscopes |
DE10340924A1 (en) | 2003-09-05 | 2005-03-31 | Sick Ag | identification device |
-
2006
- 2006-11-14 US US11/559,669 patent/US20070164202A1/en not_active Abandoned
- 2006-11-16 EP EP06255861A patent/EP1791343A3/en not_active Withdrawn
-
2009
- 2009-09-18 US US12/562,478 patent/US8110790B2/en not_active Expired - Fee Related
Patent Citations (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4404594A (en) * | 1981-11-02 | 1983-09-13 | Itek Corporation | Imaging system with enlarged depth of field |
US4672458A (en) * | 1984-12-28 | 1987-06-09 | North American Philips Corporation | Projection television systems having lens tilt compensation member |
US4833724A (en) * | 1986-08-14 | 1989-05-23 | Amada Engineering & Service Co., Inc. | Imaging device |
US4965840A (en) * | 1987-11-27 | 1990-10-23 | State University Of New York | Method and apparatus for determining the distances between surface-patches of a three-dimensional spatial scene and a camera system |
US4939369A (en) * | 1988-10-04 | 1990-07-03 | Loral Fairchild Corporation | Imaging and tracking sensor designed with a sandwich structure |
US4963756A (en) * | 1989-10-13 | 1990-10-16 | Hewlett-Packard Company | Focused line identifier for a bar code reader |
US5025313A (en) * | 1989-10-16 | 1991-06-18 | Eastman Kodak Company | System for minimizing optical distortions and chromatic aberrations in a linear color scanner |
US5231443A (en) * | 1991-12-16 | 1993-07-27 | The Research Foundation Of State University Of New York | Automatic ranging and automatic focusing |
US5521700A (en) * | 1992-03-31 | 1996-05-28 | Nidek Co., Ltd. | Meter for testing progressive multifocus lenses |
US5530514A (en) * | 1994-07-19 | 1996-06-25 | Eastman Kodak Company | Direct focus feedback autofocus system |
US5521366A (en) * | 1994-07-26 | 1996-05-28 | Metanetics Corporation | Dataform readers having controlled and overlapped exposure integration periods |
US5841881A (en) * | 1994-09-22 | 1998-11-24 | Nec Corporation | Label/window position detecting device and method of detecting label/window position |
US5689347A (en) * | 1995-03-03 | 1997-11-18 | Canon Kabushiki Kaisha | Signal processing apparatus |
US6028299A (en) * | 1995-12-05 | 2000-02-22 | Sony Corporation | Linear image sensor device comprising first and second linear image sensor sections having first and second sensitivities |
US5923017A (en) * | 1997-01-23 | 1999-07-13 | United Parcel Service Of America | Moving-light indicia reader system |
US20030019930A1 (en) * | 2000-03-21 | 2003-01-30 | Accu-Sort Systems, Inc. | Large depth of field line scan camera |
US6783068B2 (en) * | 2000-03-21 | 2004-08-31 | Accu-Sort Systems, Inc. | Large depth of field line scan camera |
US20020040932A1 (en) * | 2000-10-06 | 2002-04-11 | Citizen Electronics Co., Ltd. | Hand scanner |
US6621063B2 (en) * | 2001-06-21 | 2003-09-16 | Psc Scanning, Inc. | Omni-directional optical code reader using scheimpflug optics |
US20040065853A1 (en) * | 2001-06-21 | 2004-04-08 | Psc Scanning, Inc. | Omni-directional optical code reader using scheimpflug optics |
US7040538B2 (en) * | 2001-10-26 | 2006-05-09 | Symbol Technologies, Inc. | Bar code reader including linear sensor array and hybrid camera and bar code reader |
US20060180670A1 (en) * | 2004-12-01 | 2006-08-17 | Psc Scanning, Inc. | Triggering illumination for a data reader |
US20070115473A1 (en) * | 2005-11-23 | 2007-05-24 | Jean-Luc Legoupil | Optical method and device for detecting surface and structural defects of a travelling hot product |
Cited By (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100200726A1 (en) * | 2009-02-09 | 2010-08-12 | Cedrone Louis J | Apparatus and method for supporting and aligning imaging equipment on a web converting manufacturing line |
US8083201B2 (en) | 2009-02-09 | 2011-12-27 | The Procter & Gamble Company | Apparatus and method for supporting and aligning imaging equipment on a web converting manufacturing line |
US9485495B2 (en) | 2010-08-09 | 2016-11-01 | Qualcomm Incorporated | Autofocus for stereo images |
US9438889B2 (en) | 2011-09-21 | 2016-09-06 | Qualcomm Incorporated | System and method for improving methods of manufacturing stereoscopic image sensors |
US20130242086A1 (en) * | 2012-03-13 | 2013-09-19 | Google Inc. | Method and System for Identifying Depth Data Associated with an Object |
US9959634B2 (en) * | 2012-03-13 | 2018-05-01 | Google Llc | Method and system for identifying depth data associated with an object |
US20140036069A1 (en) * | 2012-07-31 | 2014-02-06 | Sick Ag | Camera system and method for detection of flow of objects |
CN103581496A (en) * | 2012-07-31 | 2014-02-12 | 西克股份公司 | Camera system and method for detection of flow of objects |
US10165183B2 (en) | 2012-10-19 | 2018-12-25 | Qualcomm Incorporated | Multi-camera system using folded optics |
US9838601B2 (en) | 2012-10-19 | 2017-12-05 | Qualcomm Incorporated | Multi-camera system using folded optics |
US9398264B2 (en) | 2012-10-19 | 2016-07-19 | Qualcomm Incorporated | Multi-camera system using folded optics |
US9218654B2 (en) * | 2012-11-22 | 2015-12-22 | Samsung Electronics Co., Ltd. | Apparatus and method of recognizing an object, and apparatus and method of mounting a semiconductor chip |
US20140141539A1 (en) * | 2012-11-22 | 2014-05-22 | Samsung Electronics Co., Ltd. | Apparatus and method of recognizing an object, and apparatus and method of mounting a semiconductor chip |
US10178373B2 (en) | 2013-08-16 | 2019-01-08 | Qualcomm Incorporated | Stereo yaw correction using autofocus feedback |
US9374516B2 (en) | 2014-04-04 | 2016-06-21 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
US9383550B2 (en) | 2014-04-04 | 2016-07-05 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
US9973680B2 (en) | 2014-04-04 | 2018-05-15 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
US9860434B2 (en) | 2014-04-04 | 2018-01-02 | Qualcomm Incorporated | Auto-focus in low-profile folded optics multi-camera system |
US20150334374A1 (en) * | 2014-05-14 | 2015-11-19 | 3M Innovative Properties Company | 3d image capture apparatus with depth of field extension |
US9591286B2 (en) * | 2014-05-14 | 2017-03-07 | 3M Innovative Properties Company | 3D image capture apparatus with depth of field extension |
US9967543B2 (en) | 2014-05-14 | 2018-05-08 | 3M Innovative Properties Company | 3D image capture apparatus with depth of field extension |
US10013764B2 (en) | 2014-06-19 | 2018-07-03 | Qualcomm Incorporated | Local adaptive histogram equalization |
US9733458B2 (en) | 2014-06-20 | 2017-08-15 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax artifacts |
US9549107B2 (en) | 2014-06-20 | 2017-01-17 | Qualcomm Incorporated | Autofocus for folded optic array cameras |
US9843723B2 (en) | 2014-06-20 | 2017-12-12 | Qualcomm Incorporated | Parallax free multi-camera system capable of capturing full spherical images |
US9854182B2 (en) | 2014-06-20 | 2017-12-26 | Qualcomm Incorporated | Folded optic array camera using refractive prisms |
US9819863B2 (en) | 2014-06-20 | 2017-11-14 | Qualcomm Incorporated | Wide field of view array camera for hemispheric and spherical imaging |
US9294672B2 (en) | 2014-06-20 | 2016-03-22 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax and tilt artifacts |
US9386222B2 (en) | 2014-06-20 | 2016-07-05 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax artifacts |
US10084958B2 (en) | 2014-06-20 | 2018-09-25 | Qualcomm Incorporated | Multi-camera system using folded optics free from parallax and tilt artifacts |
US9541740B2 (en) | 2014-06-20 | 2017-01-10 | Qualcomm Incorporated | Folded optic array camera using refractive prisms |
US9832381B2 (en) | 2014-10-31 | 2017-11-28 | Qualcomm Incorporated | Optical image stabilization for thin cameras |
WO2016102988A1 (en) * | 2014-12-24 | 2016-06-30 | Datalogic Automation, Inc. | Multiline scanner |
US10380448B2 (en) | 2014-12-24 | 2019-08-13 | Datalogic Usa, Inc. | Multiline scanner and electronic rolling shutter area imager based tunnel scanner |
CN106225765A (en) * | 2016-07-25 | 2016-12-14 | 浙江大学 | A kind of many line scan image sensors obtain device and the formation method of hyperfocal distance scanning imagery |
US10832023B2 (en) | 2017-12-15 | 2020-11-10 | Cognex Corporation | Dual-imaging vision system camera and method for using the same |
US11301655B2 (en) | 2017-12-15 | 2022-04-12 | Cognex Corporation | Vision imaging system having a camera and dual aimer assemblies |
WO2022011036A1 (en) * | 2020-07-08 | 2022-01-13 | Cognex Corporation | System and method for extending depth of field for 2d vision system cameras in the presence of moving objects |
Also Published As
Publication number | Publication date |
---|---|
EP1791343A2 (en) | 2007-05-30 |
US8110790B2 (en) | 2012-02-07 |
EP1791343A3 (en) | 2008-11-19 |
US20100006742A1 (en) | 2010-01-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8110790B2 (en) | Large depth of field line scan camera | |
US6783068B2 (en) | Large depth of field line scan camera | |
US7726573B2 (en) | Compact autofocus bar code reader with moving mirror | |
US6689998B1 (en) | Apparatus for optical distancing autofocus and imaging and method of using the same | |
US6053409A (en) | Dynamic focusing apparatus for an optical imaging system using a deformable mirror | |
US7204420B2 (en) | Scanner and method for eliminating specular reflection | |
US6621063B2 (en) | Omni-directional optical code reader using scheimpflug optics | |
US8286877B2 (en) | System and method for data reading using raster scanning | |
US6484066B1 (en) | Image life tunnel scanner inspection system using extended depth of field technology | |
US5198648A (en) | Code sensor with multi-faceted reflector for sensing plural image distances | |
US8496173B2 (en) | Camera-based code reader and method for its adjusted manufacturing | |
US20100013934A1 (en) | System for image acquisition | |
US7793840B2 (en) | Imaging zoom lens assembly for an imaging-based bar code reader | |
US9204032B2 (en) | Camera and method for taking sharp images | |
JP6734367B2 (en) | Camera and article processing apparatus using the same | |
US8028919B2 (en) | Imaging bar code reader with single prism focus adjustment | |
KR102187151B1 (en) | Optoelectronic sensor and method for adjusting focus | |
US6837433B2 (en) | Variable focal length imaging device | |
US8056806B2 (en) | Machine-readable symbol reader and method employing an ultracompact light concentrator with adaptive field of view | |
JPH0799529B2 (en) | Image recognition apparatus and image detection method | |
JP2022187471A (en) | Camera and method for detecting object moved through detection zone | |
JPH06103392A (en) | Bar-code reader | |
JPH03214279A (en) | Symbol reading device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ACCU-SORT SYSTEMS, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WURZ, DAVID A.;BROBST, THOMAS J.;FREYMAN, CHRISTOPHER ALAN;REEL/FRAME:019092/0834;SIGNING DATES FROM 20070222 TO 20070320 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |