WO2011028620A1 - Multiplexed biometric imaging and dual-imager biometric sensor - Google Patents

Multiplexed biometric imaging and dual-imager biometric sensor Download PDF

Info

Publication number
WO2011028620A1
WO2011028620A1 PCT/US2010/046852 US2010046852W WO2011028620A1 WO 2011028620 A1 WO2011028620 A1 WO 2011028620A1 US 2010046852 W US2010046852 W US 2010046852W WO 2011028620 A1 WO2011028620 A1 WO 2011028620A1
Authority
WO
WIPO (PCT)
Prior art keywords
facet
light
imaging
illumination
angle
Prior art date
Application number
PCT/US2010/046852
Other languages
French (fr)
Inventor
Ryan Martin
Robert K. Rowe
Steve P. Corcoran
Gary Rogers
Original Assignee
Lumidigm, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lumidigm, Inc. filed Critical Lumidigm, Inc.
Priority to CN2010800482377A priority Critical patent/CN102576408A/en
Priority to EP10752477A priority patent/EP2471023A1/en
Priority to DE112010003414T priority patent/DE112010003414T5/en
Priority to BR122013021647A priority patent/BR122013021647A2/en
Priority to BR112012004177A priority patent/BR112012004177A2/en
Publication of WO2011028620A1 publication Critical patent/WO2011028620A1/en
Priority to IL218238A priority patent/IL218238A0/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/12Fingerprints or palmprints
    • G06V40/13Sensors therefor
    • G06V40/1312Sensors therefor direct reading, e.g. contactless acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/143Sensing or illuminating at different wavelengths
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/12Fingerprints or palmprints
    • G06V40/13Sensors therefor
    • G06V40/1324Sensors therefor by using geometrical optics, e.g. using prisms

Definitions

  • Multispectral sensors can acquire images of an object under a plurality of distinct illumination and/or imaging conditions. Images from multispectral sensors can be used for various purposes such as for biometric imaging, bar code reading and authentication, documentation authentication, and quality assurance, to name a few.
  • TIR total internal reflectance
  • Multispectral sensors and others based on direct imaging can acquire images under a much broader range of conditions of the skin, the local environment, and the nature of the interface between the skin and sensor.
  • images are generally of lower contrast than the corresponding TIR image, if the TIR image is of high quality.
  • Multispectral sensors often have difficulty determining when an object is properly placed at the sensor for imaging, determining whether an object is a proper object for imaging, and/or distinguishing between an object for imaging and background objects.
  • Various embodiments of the invention are described herein for use with multispectral imaging. Implementation of these embodiments, for example, can be used to improve biometric imaging. According to some embodiments of the invention, systems and methods are disclosed for multiplexing multispectral imaging. This multiplexing can occur, for example, using a color filter array with pixels (or mosaics) that are substantially color matched with illumination sources that are distinct in another multispectral dimension. Such multispectral systems can be used for quick multispectral imaging.
  • a relatively steep-angled illumination source can be used for object identification, location detection, and/or background filtering.
  • object identification can be performed by identifying and/or comparing the illumination pattern of an image with the expected illumination pattern for such an object. This comparison can compare geometric, spectral and/or intensity patterns between the purported object and the pattern of a known object.
  • location detection can be made by comprising the relative position of an illumination pattern within an image with a known or expected location pattern within an image.
  • an imaged object can be distinguished from background objects by comparing the strength or weakness of color filter pixels (or mosaics) associated with the illumination wavelength with color filter pixels not associated with the illumination wavelength.
  • a multispectral sensor includes at least two illumination sources and an imager with a color filter array.
  • the two illumination sources can illuminate a platen with light having distinct illumination angles as well as distinct characteristics, such as distinct wavelengths or wavelength bands and/or distinct polarization conditions.
  • the color filter array which may be integral with the imager or separate therefrom, can filter light based on the specific distinctions between the two illumination sources.
  • the color filter array can include a pixels that filter based on color and the color of these pixels can be matched with the color of the illumination sources.
  • a single image of an object at the platen can be acquired. Individual pixels of the image will then be highly associated with one or the other illumination source. Because of the filtering by the color filter array, this image will include information about illumination angle.
  • Other color filter arrays can filter based on polarization to achieve a similar effect.
  • a single illumination source can be used to illuminate a platen with a relatively steep exit angle. Where the exit angle is the angle the light leaves the platen. Sequential imaging of the platen can occur using an imager. Each frame can be analyzed to determine whether an object is in contact with the platen, if the object is an expected object, and/or for background discrimination purposes.
  • an illumination pattern can be collected and the pattern on each frame can be analyzed. The position of the object relative to the platen can be a function of the relative location of the illumination pattern within an image, the width of the illumination pattern, the intensity profile of the illumination pattern, and/or the spectral content of the illumination pattern. Based on calibration of a sensor with expected object types, a computer can analyze each illumination pattern to determine whether the object is placed on the platen or to determine if the object is the expected object.
  • the illumination source can include a single wavelength or wavelength band and the imager can include a color filter array with some pixels or mosaics associated with the wavelength of the illumination source. Images can be analyzed to determine if an imaged feature is the object to be imaged or background noise by analyzing the differences in the number of pixels within an illumination pattern associated with the illumination source wavelength and other pixels.
  • a multispectral dual imaging system In various embodiments of the invention, a multispectral dual imaging system is disclosed. Two imagers can be used to simultaneously collect two images of a skin site under different imaging conditions. For example, one imager can directly image the skin site, while another images the skin site under TIR conditions. Various illuminators can be included to illuminate the skin site. In some embodiments, a prism can be used that allows the two imagers to image the skin site under both direct and TIR conditions. [0013] Various configuration can be used. For instance, a first illumination source (e.g. an LED) can be used to illuminating a finger through a first facet of a multifaceted prism.
  • a first illumination source e.g. an LED
  • Light from the first illumination source can undergo total internal reflectance at a second facet of the multifaceted prism prior to illuminating the object.
  • a first imager can image light scattered from the finger and passing through the first facet at an angle less than the critical angle and undergoing total internal reflectance at the second facet.
  • a second imager can image light scattered from the finger and passing though the first facet at an angle less than the critical angle.
  • the second imager can be located in a position such that it does not image light from the second illumination source that is totally internally reflected at the first facet.
  • the second imager can be located out of phase from the second illumination source. That is light from the second illumination source can only image light from the second imager after it is absorbed and/or scattered by the object. And, in this embodiment, light totally internally reflected at the first facet is not imaged by the second imager.
  • the second imager can be located at an azimuthal angle out of line with the second illumination source; such as an angle less than 170°. In some embodiments, this angle is 90°.
  • the second imager can also be located to image light at greater or less than the critical angle.
  • a first illumination source can illuminate an object located at a first facet of a multifaceted prism.
  • the system can include a second illumination source that can illuminate the object through a second facet and through the first facet at an angle greater than the critical angle of the first facet.
  • a first imager can image light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoes total internal reflectance at the second facet.
  • a second imager can be used to image light totally internally reflected from the first facet.
  • the fingerprint sensor of the present invention may be used to collect non- fingerprint images; for example, money, documents, bar codes, manufactured parts, etc.
  • optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions.
  • the dual imager of the present invention may be used to collect finger or hand print images as well as iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, barcode images, security document images, and perform a variety of other such functions.
  • the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality.
  • Figure 1 A shows a top view block diagram of a multiplexed biometric imaging system according to some embodiments of the invention.
  • Figure IB shows a top view block diagram of a multiplexed biometric imaging system according to some embodiments of the invention.
  • Figure 2 is an example of a sensor with multiplexed image data using a color filter array according to some embodiments of the invention.
  • Figure 3 is an example of a sensor with multiplexed image data using a color filter array according to some embodiments of the invention.
  • Figure 4A shows an example of a color filter array according to some embodiments of the invention.
  • Figure 4B shows the color response curve from a typical Bayer filter array.
  • Figure 5 shows a flow chart for creating a multiplexed image according to some embodiments of the invention.
  • Figure 6 illustrates an optical system that can be used to detect the presence of an object at a platen according to some embodiments of the invention.
  • Figure 7 shows an optical system and optical paths according to some embodiments of the invention.
  • Figures 8 A- 1 IB illustrate how the height of an object above the platen can be determined according to some embodiments of the invention.
  • Figure 12 shows an actual image that results when the finger or other object surface is located about 5 mm above the surface of a platen using embodiments of the invention.
  • Figure 13 shows an actual image that results when the finger or other object surface is located at the surface of a platen using embodiments of the invention
  • Figure 14 shows a flowchart for determining whether to image an object present at or near the platen according to some embodiments of the invention.
  • Figure 15 shows a block diagram of a computational system that can be used in conjunction with various other embodiments of the invention.
  • Figure 16 provides an isometric view of a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 17 provides a top view of a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 18 provides a front view of a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 19 depicts the simultaneous illumination of an image region of a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 20 depicts a total-internal-reflectance imaging path within a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 21 depicts a direct imaging ray bundle within a dual-imager biometric sensor, in accordance with various embodiments.
  • Figure 22 provides a schematic diagram of a dual-imager biometric system, in accordance with various embodiments.
  • Figure 23 provides a flow diagram of a method for creating composite biometric images from a two imager biometric system, in accordance with various embodiments.
  • Figure 24A shows an example of a TIR-imaging biometric sensor according to some embodiments of the invention.
  • Figure 24B shows an example of a direct-imaging biometric sensor according to some embodiments of the invention.
  • Figure 25 A shows an example of a TIR-illumination biometric sensor according to some embodiments of the invention.
  • Figure 25B shows an example of a direct-illumination biometric sensor according to some embodiments of the invention.
  • Embodiments of the present invention provide improvements to biometric imaging and multispectral imaging systems, process, and/or techniques. These improvements can extend to other realms of endeavor.
  • TIR imaging sensors are often able to perform such a function by analyzing an image sequence and initiating an acquisition when a significant change in the image is detected. Such a method works because TIR imagers are substantially unable to view objects that are not in direct contact with the optical sensor and have certain other required characteristics.
  • direct imaging sensors such as multispectral fingerprint sensors are able to view the external environment and any nearby objects directly through the sensor surface. As such, direct imaging sensors see a finger that is in contact with the sensor as well as a finger that is nearby but not in contact with the sensor. The resulting image may be in focus or out of focus depending on the distance and the depth of field of the imaging system as well as the motion of the finger.
  • Embodiments of the invention to solve this problem by providing a reliable means of making such determination using direct imaging are desirable.
  • Conventional optical fingerprint sensors typically incorporate a single TIR imager, which can require that a number of conditions be met to provide a good fingerprint image. These conditions can include the requirement that the fingerprint ridges are intact, clean, in optical contact with the sensor surface, and that the skin is of proper moisture content, i.e., neither too dry nor too wet. If any of these conditions aren't met, the resulting fingerprint image can be degraded or even missing entirely. However, when these conditions are met, the resulting image is generally of high quality and useful for both automated analysis and analysis by a human observer.
  • Multispectral fingerprint sensors and others based on direct imaging are able to acquire images under a much broader range of conditions of the skin, the local environment, and the nature of the interface between the skin and sensor.
  • images are generally of lower contrast than the corresponding TIR image, if the TIR image is of high quality.
  • finger For purposes of this disclosure, the terms “finger,” “fingerprint,” and “fingerprint image” are meant to include sites and images collected from a single finger, multiple fingers, intermediate finger joints, the palm, the entire palmar surface of the hand, and/or any other skin site on the body, as well as other animate or inanimate objects such as documents, barcodes, credentials, and the like.
  • multispectral imaging refers to methods and systems for acquiring multiple images of a finger during a single measurement session, wherein at least two of the multiple images are collected under different optical conditions.
  • Different optical conditions may include, but not limited to, different illumination wavelengths, different illumination angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different illumination polarization conditions, different imaging angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different imaging focal planes, different imaging spatial resolutions, different imaging temporal resolutions, different imaging polarization conditions, and other such conditions that substantially alter the resulting images.
  • TIR imaging refers to a method of imaging known in the art wherein the optical axis of the imaging system lies at an angle relative to the normal of the sensor imaging surface and that is greater than the optical critical angle of that surface.
  • Figure 24A A block diagram showing TIR imaging is shown in Figure 24A. In this diagram, imager 150 images light from platen 105 at angle 9i masing that is greater than the critical angle, Q cr tecai, as measured from normal 116 of facet 105. This is TIR imaging.
  • Illumination sources 122 can be positioned at various elevation and azimuth angles and.
  • Figure 24B shows imager 150 that images light from platen 105 at angle 0 ima ging that is less than the critical angle, 9 critica i.
  • Figure 25 A illustrates TIR illumination.
  • illumination source 120 illuminates platen 105 at an angle, (3 ⁇ 47/ ⁇ «/ « ⁇ 3 ⁇ 4 0 «, that is greater than the critical angle
  • 6 C riticai- Figure 25B shows a non-TIR illumination system, with the illumination source 120 illuminating platen 105 at an angle, #// « W m ⁇ 1 ⁇ 2 ow > that is less than the critical angle, e criiica i.
  • Embodiments of the invention can incorporate any combination of TIR imaging, direct imaging, TIR illumination, and direct illumination. Moreover, multiple illumination sources and/or imagers can be at multiple angles of elevation and/or aziumuth.
  • the critical angle is a function of the index of refraction of the two media on either side of an interface and is approximately 42 degrees for a glass-air interface. Because the optical axis of the TIR imaging system lies beyond the critical angle of the sensor surface, the surface acts as a mirror (as seen by the imager) when untouched, and can cease to act as a mirror in those locations in which a material with suitable optical characteristics comes into direct contact with the sensor surface.
  • critical angle will refer to the angle established by the sensor (i.e., the platen surface) and the surrounding environment, which is assumed to be air for most purposes.
  • light will change angles at boundaries between media due to phenomena such as refraction, reflection, diffraction and other such effects.
  • a ray angle is referred to in the present application as being greater than or less than the critical angle, for example, the statement refers to the angle of the ray at the operative boundary such as the sensor imaging surface rather than the angle of the same ray at any other boundary or media, unless explicitly stated as such.
  • direct imaging refers to a method of imaging wherein the optical axis of the imaging system lies at an angle relative to the sensor imaging surface that is less than the optical critical angle of that surface.
  • the system shown in Figure 24B is one sample of a direct imaging system.
  • Figures 1A and IB are block diagrams of an imaging system according to some embodiments of the invention.
  • Figure 1A shows a side view of such a system with a direct imaging imager 150. That is, it images the platen at an angle less than the critical angle, # m w ca /, as measured from the normal.
  • Multiple illumination sources 120, 121, and 122 are shown illuminating the platen at various angles ⁇ , ⁇ 2 ⁇ , and #/22 ⁇ Note that angle ⁇ 12 2 is greater than the critical angle e cr iticai-
  • Figure IB shows a top view of the imaging system shown in FIG. 1 A; the finger is not shown for clarity purposes.
  • Each illumination source 120, 121, and 122 illuminates the platen from different azimuth angles ⁇ , ⁇ , and ⁇ ⁇ , relative to a portion of the platen.
  • Multiplexed biometric imaging systems and methods are provided according to some embodiments of the invention.
  • Multiplexed imaging systems can acquire images under different illumination conditions more efficiently than simply acquiring a sequence of image frames under each of the desired conditions.
  • image data may be collected in a multiplexed manner.
  • the wavelength or spectral characteristics of an imaged object can be used to multiplex information from different illumination angles and/or optical polarizations together into a single image.
  • Figure 1 A shows a side view, block diagram of system that can be used for multiplexed biometric imaging according to some embodiments of the invention.
  • Figure IB shows a side view.
  • the system can include platen 103 that can be any number of shapes.
  • the platen can include imaging surface
  • finger 110 can be placed for imaging. While a finger is shown, any type of object can be imaged. These objects can include objects for biometric purposes such as hands, multiple fingers, faces, eyes, etc. And the objects can include money, documents, objects with codes or barcodes, etc.
  • Finger 110 can be illuminated using a plurality of illumination sources.
  • Three illumination sources 120, 121, and 122 are shown in this embodiment. In some embodiments, only two illumination sources may be used. In others as many as four or five can be used. There is no limit on the number of sources that can be used.
  • Each illumination source can illuminate imaging surface 105 at a different illumination angle relative to the normal of imaging surface 105.
  • illumination source 120 illuminates imaging surface 105 at angle ⁇ 1 ⁇ 2 1 ⁇ 4 illumination source 121 illuminates imaging surface 105 at angle ⁇ 1 ⁇ 2>, and illumination source 122 illuminates imaging surface 105 at angle 0i 22 - Where Quo ⁇ 9m ⁇ 2 ⁇
  • illumination sources 120, 121, and 122 are also placed at different azimuth angels ⁇ uo, ⁇ ni, and ⁇ around a central portion of the platen.
  • An azimuth angle can be measured from any arbitrary point.
  • a first illumination source can have an elevation angle of about 43° and an azimuth angle of about 12°.
  • a second illumination source can have an elevation angle of about 46° and an azimuth angle of about 78°.
  • a third illumination source can have an elevation angle of about 43° and an azimuth angle of about 168°.
  • a fourth illumination source can have an elevation angle of about 46° and an azimuth angle of about 102°.
  • the azimuth angles relative to various illumination sources and/or imagers can be more important than the angle between the illumination sources and a reference line.
  • the wavelength of the light provided y the illumination sources can vary tremendously.
  • illumination sources 120, 121, and 122 can each provide unique monochromatic light.
  • illumination source 120 can illuminate imaging surface 105 with blue light (e.g., 440 - 490 nm)
  • illumination source can illuminate imaging surface 105 with green light (e.g., 520 - 570 nm)
  • illumination source 122 can illuminate imaging surface 105 with red light (e.g., 630 - 740 nm).
  • Monochromatic light sources can provide light of single wavelength (e.g., a laser) or light within a narrow light band surrounding a single wavelength (e.g. LED or filtered broad-band source).
  • the light sources can provide light outside the visible spectrum in spectral regions that may include ultraviolet and/or infrared wavelengths.
  • the main wavelength or wavelength band of light from each illumination source in some embodiments, is different from the other sources.
  • each light source provides red, blue, or green light.
  • Illumination sources 120, 121, and 122 can include a quasimonochromatic light sources having a narrow band of primary-color wavelengths.
  • Illumination sources 120, 121, and 122 can also include LEDs, laser diodes, or lasers.
  • Imager 150 is used to image the light scattered and/or reflected from any object placed at imaging surface 105, such as finger 110.
  • Imager 150 can be a color or black and white imager.
  • Imager 150 can image an object through various optical elements 140 and 145, as well as through color filter array 130.
  • Color filter array 130 can be integrated within imager 150 or be a stand alone element.
  • Color filter array 130 is a mosaic of tiny color filters that separates the color response based on the arrangement of the mosaic.
  • Various types of color filter arrays can be used such as, for example, a Bayer filter, a RGBE filter, a CYYM filter, a CYGM filter, or any type of RGB W filter.
  • FIG. 4A An example of a Bayer color filter array is shown in Figure 4A. This arrangement uses red 404, green 412, and blue 408 color filter elements. As seen from the color response curve for a typical Bayer filter array shown in Figure 4B, there is generally some overlap in the spectral ranges of the red 424, green 432, and blue 428 transmission characteristics of the filter elements. As evident particularly in the curves for the green 432 and blue 428 transmission characteristics, the filter array may allow the transmission of infrared light. This can be avoided with the inclusion of an infrared filter as part of the imager. In other embodiments, an infrared filter may be omitted and one or more illumination sources that emit infrared light may be incorporated. In this way, all color filter elements 404, 408, and 412 may allow the light to substantially pass through, resulting in an infrared image at imager 150.
  • the imager or other optics may incorporate chromatic and/or polarizing beam splitters to separate the multiplexed light from the finger.
  • the separated light may then be directed to individual monochromatic imagers where the number of imagers corresponds to the number of illumination conditions so separated.
  • the wavelength of light provided by illumination sources 120, 121, and 122 should be aligned or closely aligned with the elements of the filter array.
  • the number of illumination sources can be, but not necessarily, aligned with the number of different filter elements in the array. For example, for a Bayer array, there are three filter colors. Hence, three illumination sources can be used. Each illumination source can then provide light in a red, blue, or green wavelength band to be consistent with the array.
  • imager 150 can create a multiplexed image of an object illuminated by multiple illumination sources arranged to illuminate light of different wavelengths at different illumination angles. This can be done during a single illumination session and, in some embodiments, a single image can be created.
  • This single image can be considered a multiplexed image that can later be demultiplexed by color to provide information regarding the object that depends on the illumination angle of the illumination source.
  • the multiplexed image can be considered a multispectral image because it contains information not only about the illumination wavelength but also the illumination angle.
  • the illumination sources can include different polarization effects instead of or in addition to the differences in illumination angle.
  • the resulting multiplexed image can then be a multispectral image that includes polarization information as well.
  • An example of a sensor with multiplexed image data using a color filter array is shown in Figure 2. Platen 103 is provided and through which imaging optics 205 can image a finger (or an object) onto the imager 207.
  • imager 207 can be a silicon CMOS imager with an integrated Bayer filter.
  • Other color filter arrays and/or other mean of separating the light such as polarizing beam splitters or polarizing filter arrays), and/or other image array materials may be used alternatively or in conjunction with this arrangement.
  • imagers can be used as well such as silicon CCD imagers or imagers made from other photoactive material such as InGaAs, InSb, MCT and others known in the art.
  • Figure 3 shows another view of the sensor shown in Figure 2 with two of the multiplexed illumination sources illustrated.
  • blue light source 203 e.g., a blue LED, which may be combined with lenses, mirrors, optical diffusers and other components of the sort
  • green light source 205 e.g., a green LED, which may be combined with lenses, mirrors, optical diffusers and other components of the sort
  • the light traverses a path that illuminates platen 103 at two different angles.
  • the light from light sources 203 and 205 are shown as being reflected from internal mirrored surfaces 211 and 213, respectively.
  • the reflecting surfaces 211, 213 are used to increase the illumination angle (with respect to the platen surface normal) while maintaining a small sensor footprint. Once the light from light sources 203 and 205 illuminate and pass through platen 103, the light may illuminate a finger or another object from two angles simultaneously. An imaging system with a color filter array, like the one shown in Figure 1 A can then used to acquire a multiplexed image.
  • illuminators 207, 209 could be used to provide for similar multiplexed illumination during earlier or later image frames.
  • illuminators of colors other than the blue and green can be incorporated into other embodiments of the present invention.
  • a red illuminator in a different spatial and angular position from illuminators 203, 205 could be used.
  • FIG. 5 shows a flow chart of process 500 for creating a multiplexed image according to some embodiments of the invention.
  • Process 500 starts at block 500.
  • a skin site is illuminated with at least two illumination sources having different wavelengths and illumination angles at block 510.
  • the two illumination sources can illuminate the skin site at the same time or sequentially. In other embodiments, more than two illumination sources with different illumination wavelengths and different illumination angles can be used.
  • light from the skin site can be filtered using a color filter array of any type and then imaged at block 520. In some embodiments, a single multiplexed image can be acquired.
  • Process 520 can end at block 525.
  • Various post processing techniques can be used to pull out multispectral information from the multiplexed image.
  • the pixels corresponding to specific color filter mosaics may be separately extracted from the multiplexed image.
  • Each of these images will then represent images collected under different illumination conditions (as well as different spectral conditions).
  • These images may then be interpolated using linear or nonlinear techniques to produce a uniform sampling of the object over the same grid.
  • Process 500 for example, can be executed by the computational device shown in Figure 15.
  • Illumination system 600 can include an illumination source (e.g., LED, laser, etc.) and illumination optics (e.g., cylindrical lens, spherical lens, mirrors, and/or other such optical components).
  • the illumination source can be a
  • illumination assembly 603 can direct light into a beam steering component 605, which directs the light toward mirrored surface 607 and then on and through platen 603.
  • Imager 615 can be used to take video or sequential images of an area on or above platen 603.
  • Imager 615 can include a color filter array.
  • a pixel (or mosaic) of the color filter array can be substantially the same as or consistent with the wavelength of one or more of illumination source (e.g., illumination system 600).
  • the path of the illumination light 703 is illustrated in Figure 7.
  • illumination light 703 illuminates a distinct region of platen 603.
  • illumination light 703 illuminates platen 103 with a steep illumination exit angle relative to the normal of the platen surface.
  • the distinct pattern can appear at different locations in the image based on the height of the finger or other object above the platen.
  • Mirrored surface 607 can be used to provide this steep illumination exit angle relative to the air above the platen without sacrificing a compact design.
  • the exit angle of illumination (above the platen as the light exits the platen) can be any exit angle that is not perpendicular to the surface of the platen.
  • the exit angle ( ⁇ ) can be about 51° ( ⁇ 10°) from the normal of the platen.
  • the exit angle can be any angle from 40° to as close to parallel to the platen as practical (i.e. 90°).
  • the angle of illumination can be 40°, 45°, 50°, 55°, 60°, 65°, 70°, 75°, 80°, 85°.
  • Figures 8A-1 IB illustrate how the height of an object above the platen can be determined according to some embodiments of the invention. For simplicity refraction effects as light enters and exits the platen are ignored. If the refraction effects were shown, the light exiting the platen would have a larger exit angle, ⁇ , than shown.
  • This exit angle, ⁇ can be any angle described in the preceding paragraph.
  • finger 805- is positioned well above platen 803.
  • Imager 815 is used to image the finger.
  • Imager 815 can include a color filter array.
  • Illumination source 800 illuminates the platen surface and the area near the platen surface. Light rays from illumination source 800 can be confined within strip 820. The illumination exit angle of illumination source 800 can be steep relative to the normal of the platen surface. As finger 805 approaches platen 803, a portion of the finger is illuminated as shown in Figure 8A.
  • Figure 8B shows a block illustration of image 830 produced by imager 815.
  • Illumination pattern 835 is formed from the light scattered from finger 805.
  • the position of illumination pattern 805 can indicate the position of the finger relative to the platen.
  • illumination pattern is located at the far right of the image indicating a position of finger 805 well above platen 803.
  • Figure 9A shows finger 805 positioned closer to platen 803 a moment after the finger's position in Figure 8A.
  • Figure 9B shows image 830 of finger 805 with
  • FIG 10A shows finger 805 positioned even closer to platen 803 a moment after the finger's position in Figure 9 A.
  • Figure 10B shows image 830 of finger 805 with illumination pattern 835 moved toward the left to coincide with the location of the finger.
  • Figure 11 A shows finger 805 in contact with platen 803 a moment after the finger's position in Figure 10A.
  • Figure 1 IB shows image 830 of finger 805 with illumination pattern 835 moved toward the left.
  • the steepness of the illumination exit angle in the air above the platen determines the relationship between the height of the finger 805 relative to the platen 803 and the position of the illumination pattern 835 in the image 830. If the illumination exit angle is steeper (closer to 90 degrees) the position of the illumination pattern 835 will change faster as a function of the finger 805 height above the platen 803.
  • Figure 12 shows an actual image that results when the finger or other object surface is located about 5 mm above the surface of a platen. As can be seen from the image, the illuminated pattern 1235 is located toward the left side of the image.
  • Figure 13 shows another actual image that results when a finger or other object surface is located on the platen. With the finger on the platen, illumination pattern 1335 is centrally located on the image, while illumination pattern 1235 is located on the side of the image. The horizontal shift between illumination patterns is due to the steep exit angle of the illumination beam.
  • the intensity profile of the illumination pattern changes as the finger approaches the platen. Because of the exit angle of illumination and the rounded cross- sectional shape of a finger, while the finger is above the platen the intensity profile will be sharper on one side of the finger and more gradual on the other. As the finger moves closer the intensity profile becomes more and more symmetric. In some embodiments, this intensity profile can be used as further evidence of the position of the finger. Other information such as the spectral content, the shape, and the position can be used to determine the location and the object type. For example, the width of the illumination pattern may widen as the finger approaches the platen.
  • a video sequence of images can be acquired and analyzed to determine the location of the illumination pattern. If the illumination pattern is not located in the proper location then another image can be collected and/or analyzed until the illumination pattern is in the proper location. Once in the proper location, biometric imaging or other processes may begin that require the object to be in a certain location.
  • the shape of illumination pattern can indicate the type of object being imaged. For example, in Figures 12 and 13, a finger was imaged resulting in an illumination pattern resembling a slightly bowed rectangle. And the center of the platen image is brightly illuminated (shown by the red color) while the light intensity falls off on either side (blue). Thus, both the geometrical shape and the intensity of the illumination can be indicative of the object under illumination.
  • This illumination pattern may depend on the illumination angle and/or the type of object being imaged. By fixing the illumination angle and calibrating the system with the illumination angle fixed, the illumination pattern may then depend on the object type. Images can be acquired in sequential order and the illumination pattern can be analyzed.
  • biometric imaging of fingerprints for example, if the illumination pattern is inconsistent with the known illumination pattern of a finger, then it can be determined that a foreign object is being placed on the platen and biometric imaging should not start. If the illumination pattern is consistent with the known illumination pattern of a finger, then it can be determined that a finger is being placed on the platen and biometric imaging can begin. This technique can be used in conjunction with the presence detection techniques to determine when the finger is in contact with the platen prior to triggering the start of biometric imaging.
  • illumination patterns can be used.
  • the particular illumination pattern is not critical, as long as it is easily recognized and distinct from objects in the background.
  • the shape of the illumination pattern may be altered using appropriate optics in the illumination assembly as well as elsewhere along the optical path.
  • the spectral properties of the illumination pattern may be distinguished from background materials and/or ambient lighting conditions ("background noise").
  • the illumination source can be a blue LED that emits light strongly detected by the blue pixels of a color filter array and weakly detected by the other pixels of the color filter array used in the imager.
  • the illumination pattern can then be analyzed. And if it is determined that there illumination pattern provides a strong signal in the blue pixels and a weak signal in the other pixels, then the illumination pattern can safely be determined to not be background noise. If, however, the blue pixels do not register a strong signal relative to the other pixels, then the illumination pattern is likely a background and can be ignored. While this example uses a blue LED relative to blue pixels, other wavelengths can be used.
  • FIG. 14 shows a flowchart outlining process 1400 for determining whether to image an object present at or near the platen according to some embodiments of the invention.
  • Process 1400 starts at block 1405.
  • the imaging surface and/or the area nearby the imaging surface is illuminated.
  • this illumination can occur at a steep angle relative to the normal.
  • this illumination can occur using a monochromatic light source.
  • the wavelength of this monochromatic light source for example, can be matched with one of the pixels and/or mosaics of a color filter array of the imager.
  • the area near the imaging surface is imaged.
  • the image is analyzed to determine if the shape of the illumination pattern is consistent with the expected object.
  • This illumination pattern may be a geometric illumination pattern and/or an intensity pattern. If the illumination patter is inconsistent with the expected pattern, then process 600 returns to block 1415 and another image is acquired. In some embodiments, process 600 pauses a short time period prior to returning to block 1410. If the pattern is consistent with the expected object, then process 1400 proceeds to block 1425.
  • process 1400 determines whether the illumination pattern is formed in a location consistent with the expect object being placed on the platen. If the location is inconsistent, then process 1400 returns to block 1415 and another image is acquired. If it is consistent, then process 1400 proceeds to block 1430. At block 1430 process 1400 determines whether the illumination pattern is consistent with background noise. This can be done by analyzing the pixels of the image in conjunction with the color filter array. If the pixels within the illumination pattern are strongly associated with light illumination wavelength and/or weakly associated with light associated with other pixels, then the object is not a background object and process 1400 proceeds to 1435. Otherwise process 1400 returns back to block 1415. At block 1435 imaging or other tests or measurements can occur once it is determined that the proper object is placed in the proper location and is not background noise. Process 1400 ends at block 1440.
  • Process 1400 for example, can be executed by the computational device shown in Figure 15.
  • Computational Device for example, can be executed by the computational device shown in Figure 15.
  • Figure 15 shows a block diagram of computational unit 1500 that can be used inc conjunction or as part of a biometric sensor system.
  • the figure broadly illustrates how individual system elements may be implemented in a separated or more integrated manner.
  • the drawing also illustrates how each of the four imagers 1510 may include a dedicated processor 1515 and/or dedicated memory 1520.
  • Each dedicated memory 1520 may include operational programs, data processing programs, instructions for carrying out methods described herein, and/or image processing programs operable on the dedicated processors 1515.
  • the dedicated memory 1 20 may include programs that can be executed by CPU 1502 and/or provide image processing.
  • the computational device is shown comprised of hardware elements that are electrically coupled via bus 1530.
  • the bus 1530 may also be coupled with the one or more LED(s) 1505, a proximity sensor (or presence sensor) 1512 and four imaging subsystems 1504 according to various embodiments.
  • imager memory 1520 may be shared amongst imagers 1515 and/or with the computational device 1502.
  • an imaging subsystem may include an imager 1510, a processor 1515, and memory 1520.
  • an imaging subsystem 1504 may also include light sources and/or optical elements. Imaging subsystems 1504 may be modular and additional imaging subsystems may be easily added to the system Thus, biometric sensor subsystems may include any number of imaging subsystems 1504.
  • the various imaging subsystems may be spatially modular in that each imaging subsystem is used to image a different spatial location.
  • the various imaging subsystems in another embodiment, may be multispectrally modular in that each imaging subsystem is used to image a different multispectral condition.
  • an imaging subsystem 1504 may also include various optical elements such as, for example, color filter arrays, color filters, polarizers, etc and/or the imager 1510 may be placed at various angles relative to the imaging location.
  • the various imaging subsystems in another embodiment, may provide focus modularity in that each imaging subsystem is used to image a different focal point or focal plane.
  • the hardware elements may include a central processing unit (CPU) 1550, an input/output device(s) 1535, a storage device 1555, a computer-readable storage 1540, a network interface card (NIC) 1545, a processing acceleration unit 1548 such as a DSP or special-purpose processor, and a memory 1560.
  • the computer-readable storage 1540 may include a computer-readable storage medium and a computer readable medium reader, the combination comprehensively representing remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing computer- readable information.
  • the NIC 1545 may comprise a wired, wireless, modem, and/or other type of interfacing connection and permits data to be exchanged with external devices.
  • computational unit can include illumination source interface 1557 and/or imager interface 1556.
  • These interfaces can be standard input/output interfaces such USB, firewire, UART, RS232 or a proprietary interface.
  • imager interface 1556 can be used to instruct the imager (e.g., any of the imagers or sensors described herein) to acquire an image and/or export an image.
  • Illumination source interface 1557 can be use to activate and/or deactivate any of a plurality of illumination sources singularly, as a group, or sequentially.
  • the computational unit 1500 may also comprises software elements, shown as being currently located within working memory 1560, including an operating system 1565 and other programs and/or code 1570, such as a program or programs designed to implement methods described herein. It will be apparent to those skilled in the art that substantial variations may be used in accordance with specific requirements. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets), or both. Further, connection to other computing devices such as network input/output devices may be employed.
  • Computational unit 1500 can be used to carry out processes shown in any of the figures and described in the specification. Specific instructions and/or program code can also be stored in memory 1518 or 1512 and executed by CPU 1502.
  • Embodiments of the invention may include systems, methods, and devices that may collect images of the finger using two different types of imagers that acquire images of the finger during the same measurement session.
  • the imagers may acquire images simultaneously.
  • the imagers may acquire two video sequences wherein the frames of the video sequences are synchronized such that the two sequences are substantially temporally aligned to aid processing in the presence of motion of the finger or other object being imaged.
  • the two imagers may comprise a TIR imaging system and/or a direct imaging system. In some embodiments either or both the imaging systems may be multispectral imaging systems.
  • a prism with multiple facets can be used with facets arranged for simultaneous direct and TIR imaging of a skin site.
  • the prism can be used in conjunction with one or more illumination sources and/or one or more imagers.
  • a first facet of the prism may be adapted for placement of a purported skin site by an individual.
  • the first facet may include an imaging surface for placement of the purported skin site.
  • the first facet or just the imaging surface of the first facet may alternatively be referred to as a "platen".
  • a second facet may be oriented with respect to the first facet such that a portion of light totally internally reflected at the first facet substantially passes through the second facet.
  • the sensor may include one or more illumination sources disposed to illuminate the imaging surface of the prism and/or the purported skin site through the first facet of the prism.
  • the sensor may include a direct imaging assembly disposed to receive light scattered from the purported skin site and reflected from the second facet of the prism.
  • the sensor may include a total-internal- reflectance imaging assembly disposed to receive light scattered beneath a surface of the purported skin site at the first facet and substantially passes through a third facet of the prism.
  • Figure 16 shows an isometric view of dual-imager biometric sensor 1601 according to some embodiments of the invention.
  • Sensor 1601 includes prism 1610 and further designates a region of the platen that is the active image surface 1615 used to image the finger or other obj ect; active image surface 1615 may be referred to as a platen or as a facet of prism 1610.
  • prism 1610 may include multiple facets.
  • a dual-imager biometric sensor may include one or more illumination sources.
  • dual imager sensor 1601 includes two illumination sources 1631 and 1632.
  • a dual imager sensor may include other components, such as reflective surface like mirrors 1621, 1622, 1623, 1624, and/or 1625, which may be used to facilitate directing light to different imaging components.
  • a dual imaging sensor may also include one or more total-internal-reflectance imaging assemblies.
  • Figure 21 for example shows TIR imaging assembly 1641.
  • a dual imaging biometric sensor may also include a direct imaging assembly 1642.
  • Figure 17 shows the top view of dual-imager biometric sensor 1601 and Figure 18 shows a front view of sensor 1601.
  • Two illumination assemblies 1631 and 1632 are shown in Figure 17.
  • Figure 18 shows a third illumination assembly 1633 along with illumination assembly 1631.
  • Each illumination assembly may comprise an illumination source as well as lenses, mirrors and/or other optics as necessary to shape and direct the illumination light. In some instances, more or less illuminators may be present.
  • the illuminators 1631, 1632, and/or 1633 may come on simultaneously during an acquisition sequence, or sequentially, or pulsed or encoded in some manner.
  • the illuminators 1631, 1632, and/or 1633 may be of
  • Some embodiments may utilize illumination assemblies that may include illumination sources without other optics to shape and direct the illumination source. Some embodiments of a dual imaging sensor may also include other illumination assemblies or illumination sources as will be discussed below. For example, some embodiments may include an illuminator such as a flat light guide illuminated with LEDs, or a cold cathode fluorescent lamp (CCFL), or other illuminators known in the art that may be used as part of a TTR imaging.
  • an illuminator such as a flat light guide illuminated with LEDs, or a cold cathode fluorescent lamp (CCFL), or other illuminators known in the art that may be used as part of a TTR imaging.
  • CCFL cold cathode fluorescent lamp
  • Illumination light from any or all sources may illuminate the image region 1615.
  • Light from a finger touching the sensor may then be imaged by the TIR image assembly 1641 after being reflected from mirrored surfaces 1621 and 1624.
  • a TIR image assembly, such as 1641 may not need reflecting surfaces such as 1621 and 1624 to make an image.
  • Other embodiments may include more or less imaging surfaces that may be in different configurations.
  • the finger touching the active image region 1615 may also be imaged by direct image assembly 1642 after reflections from surfaces 1622, 1623, and/or 1625.
  • a direct image assembly such as 1642
  • Other embodiments may use more or less reflecting surfaces, which also may be position and/or oriented in different configurations.
  • direct image assembly 1642 may contain a color imager, as well as lenses, mirrors, optical filters, grating and other such optical elements as may be necessary to form an image as known in the art.
  • TIR image assembly 1641 may contain a color imager or a monochromatic imager, as well as lenses, mirrors, optical filters, grating and other such optical elements as may be necessary to form an image as known in the art.
  • FIG. 19 shows a three-dimensional view of sensor 1601. This illustration shows the simultaneous illumination of the image region 1615 of prism 1610 by illuminators 1631, 1632, and/or 1633.
  • illuminators 1631, 1632, and/or 1633 may be different colors (such as green, red and blue, respectively, merely by way of example; ultraviolet and/or infrared light may be used instead of or in addition to various visible wavelengths), in accordance with another aspect of various embodiments.
  • the central wavelengths of the illuminators such as 1631, 1632, and/or 1633 as well as the spectral width and selection of which illuminator emits which color can all vary and such choices are within the scope of various embodiments.
  • an spectral width and selection of which illuminator emits which color can all vary and such choices are within the scope of various embodiments.
  • illuminator such as 1631 and/or 1632
  • illuminator may be positioned such that light transmitted from the illuminator is totally internally reflected at a facet, such as facet 1611, of prism 1610 and then illuminates image region 1615.
  • Some illuminators, such as 1633, may be positioned and/or oriented such that light passes through a facet, such as facet 1611 and then illuminates image region 1615.
  • FIG. 20 depicts a TIR imaging path 1651.
  • the TIR imaging assembly 1641 which may be referred to as a TIR imager in some embodiments, can view the active region 1615 of prisml610 through prism facet 1612 and reflective surfaces 1621 and 1624.
  • the angle between the TIR image bundle 1651 and the platen facet 1612 is such that the bundle 1651 can pass through the facet 1612.
  • the image bundle 1651 is substantially perpendicular to facet 1612.
  • facet 1612 may be oriented approximately perpendicular to an angle equal to or greater than the a critical angle with respect to facet 1615
  • image bundle 1651 is oriented at an angle greater than the critical angle defined by the surface 1615 and the surrounding media, which is usually air. Because of this, it may be desirable to provide a black background for the TIR imager 1641 to view in cases where the TIR imager 1641 is operating in a dark-field illumination mode.
  • facet 1611 of prism 1610 may be oriented such that rays in the TIR bundle 1615 reflected through TIR at surface 1615 may be substantially able to pass out of prism 1610 through facet 1611.
  • a light trap (e.g., optical absorber) 1660, which may comprise a material with a matte black surface finish, may be positioned such that substantially all of the TIR bundle 1651 lands on the trap 1660.
  • the light trap 1660 may be replaced by an illuminator, such as a flat light guide illuminated with LEDs, or a cold cathode fluorescent lamp (CCFL), or other illuminators known in the art.
  • the TIR imaging system operates in a bright-field illumination mode.
  • Figure 21 depicts direct imaging ray bundle 1652 seen by direct imaging assembly 1642 in accordance with various embodiments.
  • Direct image bundle 1652 may pass through facet 1613.
  • direct image bundle 1652 is approximately normal to facet 1613.
  • direct image bundle 1652 is oriented such that it is reflected due to TIR at facet 1611 and passes through image region 1615.
  • facet 1611 may serve a dual role of being highly reflective for the direct imaging bundle 1652 and also being transmissive for TIR imaging bundle 1651 as illustrated in Figure 22. In some embodiments, facet 1611 may serve a dual role of being highly reflective for light emanating from illuminators 1631 and/or 1632 and also being transmissive for light emanating from illuminator 1633. In some embodiments, facet 1611 may be variously reflective and transmissive for some combination of illumination and imaging bundles.
  • FIG. 22 provides a diagram of a dual imager sensor or system 1607 in accordance with various embodiments.
  • System 1607 includes prism 1610 with multiple facets 1611, 1612, 1613, and 1615.
  • System 1607 includes a direct imager 1644 and a TIR imager 1643.
  • Direct imaging assembly 1642 of Figure 21, for example, may be an example of direct imager 1644.
  • Some embodiments may include direct imager 1644 that may include other optical components as described herein, such as mirrors, 1622, 1623, and 1625.
  • TIR imaging assembly 1641 of Figure 20, for example, may be an example of TIR imager 1643.
  • Some embodiments may include TIR imager 1643 that may include other optical components as described herein, such as mirrors 1621 and/or 1624.
  • Direct imager 1644 and TIR imager 1643 may be oriented and/or positioned in different configurations in different embodiments.
  • facets of prism 1610 may provide different functions.
  • facet 1611 may serve a dual role of being highly reflective for direct imaging and also being transmissive for TIR imaging.
  • a TIR bundle 1653 directed towards facet 1611 will transmit through facet 1611.
  • a direct imaging bundle 1654 may reflect at facet 1611.
  • direct imaging bundle 1653 totally internally reflects at facet 1611.
  • system 1607 shows an example of a facet 1613 that direct imaging bundle 1654 may pass through and then reach a direct imager 1644.
  • a TIR bundle 1655 may pass through facet 1612 and then reach a TIR imager 1643.
  • Facets as seen in Figure 22 may have different relationships with each other in order to affect the functions described for different embodiments.
  • facet 1615 and 1611 may be configured such that facet 1611 is oriented approximately perpendicular to an angle equal to or greater than the critical angle of facet 1615 such that light that undergoes TIR at facet 1615 may pass through facet 1611 before or after such TIR interaction.
  • facet 1612 may be oriented approximately perpendicular to an angle equal to or greater than the critical angle of facet 1615 such that light totally internally reflect at facet 1615 may pass through facet 1612.
  • Prism 1610 may include facets such as 1613 that are oriented in different ways.
  • facet 1613 may be oriented relatively perpendicular to light that passed through facet 1615 and reflected from facet 1611 such that the light substantially passes through facet 1613 .
  • System 1607 may also include different illumination sources.
  • Figures 1 to 6, 16 to 21, and 24 to 25 show different examples with illumination sources. These figures also show different examples of orientations with respect to imagers.
  • Illumination sources may be position and/or oriented in different ways to illuminate a finger, for example, place on prism 1610 along facet 1615, for example.
  • Some embodiments may also include a controller, which may control imagers such as 1643 and 1643 along with illumination sources.
  • sensors and systems may each collect images simultaneously using direct imagers such as 1642 or 1644 and TIR imager 1641 or 1643. If either of the images is multiplexed in the manner of the different embodiments, the image may be decomposed into the images corresponding to the different illumination conditions followed by interpolation and/or other processing of the resulting images. The resulting collection of images (both TIR and direct) may then be processed further in such a manner that a single, composite image of the fingerprint is generated. This composite image may then be reported to a host computer or other connected device.
  • direct imagers such as 1642 or 1644 and TIR imager 1641 or 1643.
  • the TIR imager and direct imagers of sensors and systems may acquire synchronized video streams of images. These video streams may be acquired while the finger is rolled in the manner known in the art. The rolled image sequences may then be combined to produce a composite rolled print image.
  • the fingerprint sensor of different embodiments may be used to collect non-fingerprint images.
  • the direct imager may be used to acquire such images.
  • Other image data so collected may include barcodes and documents.
  • Optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions.
  • the direct imager of the different embodiments may be used to collect iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, and perform a variety of other such functions.
  • the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality.
  • FIG. 23 shows a flow diagram for process 2300 for creating composite biometric images from a two imager biometric system.
  • Process 2300 may be implemented, for example, using sensors and systems as seen in Figures 15 to 22.
  • Process 2300 can start at block 2305.
  • a purported skin site is illuminated through a first facet of a platen using one or more illumination sources.
  • light from the purported skin site is received.
  • a direct image of the purported skin site is generated from a first portion of the received light that is reflected at a second facet of the platen.
  • a total-internal- reflectance image of the purported skin site is generated from a second portion of the received light that passes through a third facet of the platen oriented.
  • Process 2300 can end at block 2330. Any of these blocks can occur simultaneously. For instance, blocks 2310 and 2315 may occur at the same time as well as blocks 2320 and 2325.
  • a third portion of light is totally internally reflected at the first facet of the platen and then passes through the second facet of the platen.
  • Some embodiments may further comprise absorbing the third portion of the light at a light trap located proximate to the second facet such that totally internally reflected light at the first facet that passes through the second facet is substantially absorbed by the light trap.
  • the light received by the direct imaging assembly that is reflected from the second facet of the platen is totally internally reflected at the second facet.
  • the second and third facets are oriented at substantially critical angles with respect to the first facet.
  • Some embodiments of process 2300 may include generating a composite image of the purported skin site using the total-internal-reflectance image and the direct image of the purported skin site.
  • Some embodiments may include receiving light from the purported skin site that includes receiving the first portion of the received light at a total-internal-reflectance imaging assembly and receiving the second portion of the received light at a direct imaging assembly.
  • Some embodiments may further comprise decomposing at least one of the images into multiple images corresponding to different illumination conditions.
  • Some embodiments of process 2300 may further comprise illuminating a non-skin site through the platen using one or more illumination sources. Light from the non-skin site is received and then a direct image of the non-skin site from a second portion of the received light may be generated.
  • the non-skin site may include barcodes, documents, or optical security markings, for example.
  • the non-skin site may includes an iris, a portion of a face, a surveillance location, a moving object, or a lighting condition, for example.
  • Process 900 for example, can be executed by the computational device shown in Figure 15.
  • the TIR imager and direct imager may each collect images simultaneously. If either of the images is multiplexed in the manner described in some of the embodiments described herein, the image may be decomposed into the images corresponding to the different illumination conditions followed by interpolation and/or other processing of the resulting images. The resulting collection of images (both TIR and direct) may then be processed to create a single, composite image of the fingerprint. This composite image may then be reported to a host computer or other connected device.
  • the TIR imager and direct imager may acquire synchronized video streams of images. These video streams may be acquired while the finger is rolled in the manner known in the art. The rolled image sequences may then be combined to produce a composite rolled print image.
  • a first illumination source e.g. an LED
  • a first facet e.g., facet 1615
  • a multifaceted prism e.g., prism 1610
  • Light from the first illumination source can undergo total internal reflectance at a second facet (e.g., 1611) of the multifaceted prism prior to illuminating the object.
  • a first imager e.g., 1644
  • a second imager can image light scattered from the finger and passing through the first facet at an angle less than the critical angle.
  • the second imager can be located in a position such that it does not image light from the second illumination source that is totally internally reflected at the first facet.
  • the second imager can be located out of phase from the second illumination source. That is light from the second illumination source can only image light from the second imager after it is absorbed and/or scattered by the object. And, in this embodiment, light totally internally reflected at the first facet is not imaged by the second imager.
  • the second imager can be located at an azimuthal angle out of line with the second illumination source; such as an angle less than 170°. In some embodiments, this angle is 90°.
  • the second imager can also be located to image light at greater or less than the critical angle.
  • a first illumination source can illuminate an object located at a first facet (e.g., 1615) of a multifaceted prism (e.g., 1610).
  • the system can include a second illumination source that can illuminate the object through a second facet (e.g., 1611) and through the first facet at an angle greater than the critical angle of the first facet.
  • a first imager e.g., 1644 can image light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoes total internal reflectance at the second facet.
  • a second imager (1643) can be used to image light totally internally reflected from the first facet.
  • the fingerprint sensor of the present invention may be used to collect non-fingerprint images; for example, money, documents, bar codes, manufactured parts, etc.
  • optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions.
  • the dual imager of the present invention may be used to collect finger or hand print images as well as iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, barcode images, security document images, and perform a variety of other such functions.
  • the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality.

Abstract

Some embodiments disclose systems and methods for a multiplexed multispectral imaging, object discrimination, background discrimination, and/or object identification. In some embodiments, a multispectral sensor is provided that includes at least two illumination sources and an imager with a color filter array. The two illumination sources can illuminate a platen with light having distinct illumination angles and distinct characteristics. The color filter array, which may be integral with the imager or separate therefrom, can filter light based on specific distinctions between the two illumination sources. A single image of an object at the platen can be acquired. Individual pixels of the image will then be highly associated with one or the other illumination source. Because of the filtering by the color filter array, this image will include information about the illumination angle. Some embodiments disclose a dual-imager biometric sensor. The two imagers may include a direct imager and a TIR imager.

Description

MULTIPLEXED BIOMETRIC IMAGING AND DUAL-IMAGER
BIOMETRIC SENSOR
CROSS-REFERENCES TO RELATED APPLICATIONS
[0001] This application is a non-provisional, and claims the benefit, of commonly assigned U.S. Provisional Application No. 61/237,189, filed August 26, 2009, entitled "Multiplexed Biometric Imaging and Dual-Imager Biometric Sensor," the entirety of which is herein incorporated by reference for all purposes.
BACKGROUND
[0002] Multispectral sensors can acquire images of an object under a plurality of distinct illumination and/or imaging conditions. Images from multispectral sensors can be used for various purposes such as for biometric imaging, bar code reading and authentication, documentation authentication, and quality assurance, to name a few.
[0003] Conventional optical fingerprint sensors typically incorporate a single total internal reflectance (TIR) imager, which can require that a number of conditions be met to provide a good fingerprint image. These conditions can include the requirement that the fingerprint ridges are intact, clean, in optical contact with the sensor surface, and that the skin is of proper moisture content, i.e., neither too dry nor too wet. If any of these conditions aren't met, the resulting fingerprint image can be degraded or even missing entirely. However, when these conditions are met, the resulting image is generally of high quality and useful for both automated analysis and analysis by a human observer.
[0004] Multispectral sensors and others based on direct imaging can acquire images under a much broader range of conditions of the skin, the local environment, and the nature of the interface between the skin and sensor. However, such images are generally of lower contrast than the corresponding TIR image, if the TIR image is of high quality.
[0005] It can be a time consuming process for a sensor to illuminate and acquire images from the plurality of distinct optical and/or imaging conditions. Multispectral sensors often have difficulty determining when an object is properly placed at the sensor for imaging, determining whether an object is a proper object for imaging, and/or distinguishing between an object for imaging and background objects.
SUMMARY
[0006] Various embodiments of the invention are described herein for use with multispectral imaging. Implementation of these embodiments, for example, can be used to improve biometric imaging. According to some embodiments of the invention, systems and methods are disclosed for multiplexing multispectral imaging. This multiplexing can occur, for example, using a color filter array with pixels (or mosaics) that are substantially color matched with illumination sources that are distinct in another multispectral dimension. Such multispectral systems can be used for quick multispectral imaging.
[0007] According to various embodiments of the invention, a relatively steep-angled illumination source can be used for object identification, location detection, and/or background filtering. According to some embodiments, object identification can be performed by identifying and/or comparing the illumination pattern of an image with the expected illumination pattern for such an object. This comparison can compare geometric, spectral and/or intensity patterns between the purported object and the pattern of a known object. In some embodiments, location detection can be made by comprising the relative position of an illumination pattern within an image with a known or expected location pattern within an image. In some embodiments, an imaged object can be distinguished from background objects by comparing the strength or weakness of color filter pixels (or mosaics) associated with the illumination wavelength with color filter pixels not associated with the illumination wavelength.
[0008] For example, according to some embodiments of the invention a multispectral sensor is provided that includes at least two illumination sources and an imager with a color filter array. The two illumination sources can illuminate a platen with light having distinct illumination angles as well as distinct characteristics, such as distinct wavelengths or wavelength bands and/or distinct polarization conditions. The color filter array, which may be integral with the imager or separate therefrom, can filter light based on the specific distinctions between the two illumination sources. For example, the color filter array can include a pixels that filter based on color and the color of these pixels can be matched with the color of the illumination sources. A single image of an object at the platen can be acquired. Individual pixels of the image will then be highly associated with one or the other illumination source. Because of the filtering by the color filter array, this image will include information about illumination angle. Other color filter arrays can filter based on polarization to achieve a similar effect.
[0009] As another example, a single illumination source can be used to illuminate a platen with a relatively steep exit angle. Where the exit angle is the angle the light leaves the platen. Sequential imaging of the platen can occur using an imager. Each frame can be analyzed to determine whether an object is in contact with the platen, if the object is an expected object, and/or for background discrimination purposes. In some embodiments, an illumination pattern can be collected and the pattern on each frame can be analyzed. The position of the object relative to the platen can be a function of the relative location of the illumination pattern within an image, the width of the illumination pattern, the intensity profile of the illumination pattern, and/or the spectral content of the illumination pattern. Based on calibration of a sensor with expected object types, a computer can analyze each illumination pattern to determine whether the object is placed on the platen or to determine if the object is the expected object.
[0010] In other embodiments, the illumination source can include a single wavelength or wavelength band and the imager can include a color filter array with some pixels or mosaics associated with the wavelength of the illumination source. Images can be analyzed to determine if an imaged feature is the object to be imaged or background noise by analyzing the differences in the number of pixels within an illumination pattern associated with the illumination source wavelength and other pixels.
[0011] Various embodiments of the invention are also described herein involving
multispectral dual imaging. [0012] In various embodiments of the invention, a multispectral dual imaging system is disclosed. Two imagers can be used to simultaneously collect two images of a skin site under different imaging conditions. For example, one imager can directly image the skin site, while another images the skin site under TIR conditions. Various illuminators can be included to illuminate the skin site. In some embodiments, a prism can be used that allows the two imagers to image the skin site under both direct and TIR conditions. [0013] Various configuration can be used. For instance, a first illumination source (e.g. an LED) can be used to illuminating a finger through a first facet of a multifaceted prism. Light from the first illumination source can undergo total internal reflectance at a second facet of the multifaceted prism prior to illuminating the object. A first imager can image light scattered from the finger and passing through the first facet at an angle less than the critical angle and undergoing total internal reflectance at the second facet. A second imager can image light scattered from the finger and passing though the first facet at an angle less than the critical angle. In some embodiments, the second imager can be located in a position such that it does not image light from the second illumination source that is totally internally reflected at the first facet.
[0014] In some embodiments, the second imager can be located out of phase from the second illumination source. That is light from the second illumination source can only image light from the second imager after it is absorbed and/or scattered by the object. And, in this embodiment, light totally internally reflected at the first facet is not imaged by the second imager. For example, the second imager can be located at an azimuthal angle out of line with the second illumination source; such as an angle less than 170°. In some embodiments, this angle is 90°. The second imager can also be located to image light at greater or less than the critical angle.
[0015] In another embodiment, a first illumination source can illuminate an object located at a first facet of a multifaceted prism. The system can include a second illumination source that can illuminate the object through a second facet and through the first facet at an angle greater than the critical angle of the first facet. A first imager can image light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoes total internal reflectance at the second facet. A second imager can be used to image light totally internally reflected from the first facet.
[0016] In some embodiments, the fingerprint sensor of the present invention may be used to collect non- fingerprint images; for example, money, documents, bar codes, manufactured parts, etc. In some of these images optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions. [0017] In some embodiments, the dual imager of the present invention may be used to collect finger or hand print images as well as iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, barcode images, security document images, and perform a variety of other such functions. In some embodiments the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality.
BRIEF DESCRIPTION OF THE DRAWINGS
[0018] The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
[0019] The present disclosure is described in conjunction with the appended figures.
[0020] Figure 1 A shows a top view block diagram of a multiplexed biometric imaging system according to some embodiments of the invention.
[0021] Figure IB shows a top view block diagram of a multiplexed biometric imaging system according to some embodiments of the invention.
[0022] Figure 2 is an example of a sensor with multiplexed image data using a color filter array according to some embodiments of the invention.
[0023] Figure 3 is an example of a sensor with multiplexed image data using a color filter array according to some embodiments of the invention. [0024] Figure 4A shows an example of a color filter array according to some embodiments of the invention.
[0025] Figure 4B shows the color response curve from a typical Bayer filter array.
[0026] Figure 5 shows a flow chart for creating a multiplexed image according to some embodiments of the invention. [0027] Figure 6 illustrates an optical system that can be used to detect the presence of an object at a platen according to some embodiments of the invention. [0028] Figure 7 shows an optical system and optical paths according to some embodiments of the invention.
[0029] Figures 8 A- 1 IB illustrate how the height of an object above the platen can be determined according to some embodiments of the invention.
[0030] Figure 12 shows an actual image that results when the finger or other object surface is located about 5 mm above the surface of a platen using embodiments of the invention.
[0031] Figure 13 shows an actual image that results when the finger or other object surface is located at the surface of a platen using embodiments of the invention
[0032] Figure 14 shows a flowchart for determining whether to image an object present at or near the platen according to some embodiments of the invention.
[0033] Figure 15 shows a block diagram of a computational system that can be used in conjunction with various other embodiments of the invention.
[0034] Figure 16 provides an isometric view of a dual-imager biometric sensor, in accordance with various embodiments.
[0035] Figure 17 provides a top view of a dual-imager biometric sensor, in accordance with various embodiments.
[0036] Figure 18 provides a front view of a dual-imager biometric sensor, in accordance with various embodiments.
[0037] Figure 19 depicts the simultaneous illumination of an image region of a dual-imager biometric sensor, in accordance with various embodiments.
[0038] Figure 20 depicts a total-internal-reflectance imaging path within a dual-imager biometric sensor, in accordance with various embodiments.
[0039] Figure 21 depicts a direct imaging ray bundle within a dual-imager biometric sensor, in accordance with various embodiments.
[0040] Figure 22 provides a schematic diagram of a dual-imager biometric system, in accordance with various embodiments. [0041] Figure 23 provides a flow diagram of a method for creating composite biometric images from a two imager biometric system, in accordance with various embodiments.
[0042] Figure 24A shows an example of a TIR-imaging biometric sensor according to some embodiments of the invention. [0043] Figure 24B shows an example of a direct-imaging biometric sensor according to some embodiments of the invention.
[0044] Figure 25 A shows an example of a TIR-illumination biometric sensor according to some embodiments of the invention.
[0045] Figure 25B shows an example of a direct-illumination biometric sensor according to some embodiments of the invention.
DETAILED DESCRIPTION
[0046] The ensuing description provides preferred exemplary embodiment(s) only, and is not intended to limit the scope, applicability or configuration of the disclosure. Rather, the ensuing description of the preferred exemplary embodiment(s) will provide those skilled in the art with an enabling description for implementing a preferred exemplary embodiment. It being understood that various changes may be made in the function and arrangement of elements without departing from the spirit and scope as set forth in the appended claims.
[0047] Embodiments of the present invention provide improvements to biometric imaging and multispectral imaging systems, process, and/or techniques. These improvements can extend to other realms of endeavor.
[0048] Conventional optical fingerprint sensors acquire a single image of the fingerprint (or a set of images taken under substantially similar optical conditions) during a single measurement session. On the other hand, multispectral imagers acquire multiple images during a single measurement session under different optical conditions. Collecting such a set of images can require more acquisition time than collecting a single image. Embodiments of the invention provide methods and systems capable of solving this problem by acquiring multispectral images in a more efficient manner than with traditional serial acquisition; for example, by using multiplexed techniques. [0049] In some cases, it is also desirable to initiate a fingerprint acquisition sequence automatically. Conventional optical sensors based on TIR imaging are often able to perform such a function by analyzing an image sequence and initiating an acquisition when a significant change in the image is detected. Such a method works because TIR imagers are substantially unable to view objects that are not in direct contact with the optical sensor and have certain other required characteristics. In contrast, direct imaging sensors such as multispectral fingerprint sensors are able to view the external environment and any nearby objects directly through the sensor surface. As such, direct imaging sensors see a finger that is in contact with the sensor as well as a finger that is nearby but not in contact with the sensor. The resulting image may be in focus or out of focus depending on the distance and the depth of field of the imaging system as well as the motion of the finger. In an effort to achieve maximum image quality, it is therefore often preferable to initiate a fingerprint image acquisition only after the finger is in firm contact with the sensor. Embodiments of the invention to solve this problem by providing a reliable means of making such determination using direct imaging are desirable. [0050] Conventional optical fingerprint sensors typically incorporate a single TIR imager, which can require that a number of conditions be met to provide a good fingerprint image. These conditions can include the requirement that the fingerprint ridges are intact, clean, in optical contact with the sensor surface, and that the skin is of proper moisture content, i.e., neither too dry nor too wet. If any of these conditions aren't met, the resulting fingerprint image can be degraded or even missing entirely. However, when these conditions are met, the resulting image is generally of high quality and useful for both automated analysis and analysis by a human observer.
[0051] Multispectral fingerprint sensors and others based on direct imaging are able to acquire images under a much broader range of conditions of the skin, the local environment, and the nature of the interface between the skin and sensor. However, such images are generally of lower contrast than the corresponding TIR image, if the TIR image is of high quality. Some embodiments of the invention solve this problem by providing a fingerprint sensor that is able to collect both TIR and direct images during the same measurement session to ensure that biometric images are always collected regardless of the conditions of the finger, sensor surface and/or the local environment. [0052] For purposes of this disclosure, the terms "finger," "fingerprint," and "fingerprint image" are meant to include sites and images collected from a single finger, multiple fingers, intermediate finger joints, the palm, the entire palmar surface of the hand, and/or any other skin site on the body, as well as other animate or inanimate objects such as documents, barcodes, credentials, and the like.
[0053] The terms "multispectral imaging," "MSI," and "multi-imaging" refer to methods and systems for acquiring multiple images of a finger during a single measurement session, wherein at least two of the multiple images are collected under different optical conditions. Different optical conditions may include, but not limited to, different illumination wavelengths, different illumination angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different illumination polarization conditions, different imaging angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different imaging focal planes, different imaging spatial resolutions, different imaging temporal resolutions, different imaging polarization conditions, and other such conditions that substantially alter the resulting images. Also, unless otherwise specified, the angle of incidence, angle of illumination, angle of imaging, etc. is measured relative to the normal of the incident surface. [0054] The terms "total internal reflectance imaging" and "TIR imaging" refer to a method of imaging known in the art wherein the optical axis of the imaging system lies at an angle relative to the normal of the sensor imaging surface and that is greater than the optical critical angle of that surface. A block diagram showing TIR imaging is shown in Figure 24A. In this diagram, imager 150 images light from platen 105 at angle 9imasing that is greater than the critical angle, Qcrtecai, as measured from normal 116 of facet 105. This is TIR imaging.
Illumination sources 122 can be positioned at various elevation and azimuth angles and. Figure 24B shows imager 150 that images light from platen 105 at angle 0imaging that is less than the critical angle, 9criticai.
[0055] Figure 25 A illustrates TIR illumination. In this example, illumination source 120 illuminates platen 105 at an angle, (¾7/Μ«/«α¾0«, that is greater than the critical angle, 6Criticai- Figure 25B shows a non-TIR illumination system, with the illumination source 120 illuminating platen 105 at an angle, #//«Wm<½ow> that is less than the critical angle, ecriiicai.
[0056] Various configurations of illumination and imaging are shown in Figures 24A-25B. Embodiments of the invention can incorporate any combination of TIR imaging, direct imaging, TIR illumination, and direct illumination. Moreover, multiple illumination sources and/or imagers can be at multiple angles of elevation and/or aziumuth.
[0057] The critical angle is a function of the index of refraction of the two media on either side of an interface and is approximately 42 degrees for a glass-air interface. Because the optical axis of the TIR imaging system lies beyond the critical angle of the sensor surface, the surface acts as a mirror (as seen by the imager) when untouched, and can cease to act as a mirror in those locations in which a material with suitable optical characteristics comes into direct contact with the sensor surface.
[0058] In locations where a finger or other material contacts a sensor surface, a new critical angle is established. However, for purposes of the present disclosure, the term "critical angle" will refer to the angle established by the sensor (i.e., the platen surface) and the surrounding environment, which is assumed to be air for most purposes. Also, as known in the art, light will change angles at boundaries between media due to phenomena such as refraction, reflection, diffraction and other such effects. When a ray angle is referred to in the present application as being greater than or less than the critical angle, for example, the statement refers to the angle of the ray at the operative boundary such as the sensor imaging surface rather than the angle of the same ray at any other boundary or media, unless explicitly stated as such.
[0059] The term "direct imaging" refers to a method of imaging wherein the optical axis of the imaging system lies at an angle relative to the sensor imaging surface that is less than the optical critical angle of that surface. For example, the system shown in Figure 24B is one sample of a direct imaging system.
[0060] Figures 1A and IB are block diagrams of an imaging system according to some embodiments of the invention. Figure 1A shows a side view of such a system with a direct imaging imager 150. That is, it images the platen at an angle less than the critical angle, #mwca/, as measured from the normal. Multiple illumination sources 120, 121, and 122 are shown illuminating the platen at various angles θηο, Θΐ2ΐ, and #/22· Note that angle θ122 is greater than the critical angle ecriticai- Figure IB shows a top view of the imaging system shown in FIG. 1 A; the finger is not shown for clarity purposes. Each illumination source 120, 121, and 122, illuminates the platen from different azimuth angles φηο, φηι, and φ ι, relative to a portion of the platen.
[0061] Multiplexed Biometric Imaging
[0062] Multiplexed biometric imaging systems and methods are provided according to some embodiments of the invention. Multiplexed imaging systems can acquire images under different illumination conditions more efficiently than simply acquiring a sequence of image frames under each of the desired conditions. Such image data may be collected in a multiplexed manner. In some embodiments, the wavelength or spectral characteristics of an imaged object can be used to multiplex information from different illumination angles and/or optical polarizations together into a single image.
[0063] Figure 1 A shows a side view, block diagram of system that can be used for multiplexed biometric imaging according to some embodiments of the invention. Figure IB shows a side view. The system can include platen 103 that can be any number of shapes.
Various polygonal-shaped platens are described later. The platen can include imaging surface
105, for example, where a purported skin site can be placed for imaging during a biometric acquisition. At imaging surface 105 finger 110 can be placed for imaging. While a finger is shown, any type of object can be imaged. These objects can include objects for biometric purposes such as hands, multiple fingers, faces, eyes, etc. And the objects can include money, documents, objects with codes or barcodes, etc.
[0064] Finger 110 can be illuminated using a plurality of illumination sources. Three illumination sources 120, 121, and 122 are shown in this embodiment. In some embodiments, only two illumination sources may be used. In others as many as four or five can be used. There is no limit on the number of sources that can be used. Each illumination source can illuminate imaging surface 105 at a different illumination angle relative to the normal of imaging surface 105. For instance, illumination source 120 illuminates imaging surface 105 at angle έ½¼ illumination source 121 illuminates imaging surface 105 at angle έ½>, and illumination source 122 illuminates imaging surface 105 at angle 0i22- Where Quo≠9m≠θη2· These angles can be greater than or less than the critical angle at the interface near the platen as measured relative to the normal to the platen. And these angles can be called "elevation angles". [0065] Furthermore, each of illumination sources 120, 121, and 122 are also placed at different azimuth angels φ uo, φ ni, and ηι around a central portion of the platen. An azimuth angle can be measured from any arbitrary point. As shown in Figure IB, the azimuth angles are measured relative to a reference line perpendicular to the standard approach of a finger. In a system with four illumination sources, for example, a first illumination source can have an elevation angle of about 43° and an azimuth angle of about 12°. A second illumination source can have an elevation angle of about 46° and an azimuth angle of about 78°. A third illumination source can have an elevation angle of about 43° and an azimuth angle of about 168°. A fourth illumination source can have an elevation angle of about 46° and an azimuth angle of about 102°. In some embodiments, the azimuth angles relative to various illumination sources and/or imagers can be more important than the angle between the illumination sources and a reference line. The wavelength of the light provided y the illumination sources can vary tremendously.
[0066] Furthermore, illumination sources 120, 121, and 122 can each provide unique monochromatic light. For example, illumination source 120 can illuminate imaging surface 105 with blue light (e.g., 440 - 490 nm), illumination source can illuminate imaging surface 105 with green light (e.g., 520 - 570 nm), and illumination source 122 can illuminate imaging surface 105 with red light (e.g., 630 - 740 nm). Monochromatic light sources can provide light of single wavelength (e.g., a laser) or light within a narrow light band surrounding a single wavelength (e.g. LED or filtered broad-band source). In some embodiments, the light sources can provide light outside the visible spectrum in spectral regions that may include ultraviolet and/or infrared wavelengths. The main wavelength or wavelength band of light from each illumination source, in some embodiments, is different from the other sources. In the preceding example, each light source provides red, blue, or green light. [0067] Illumination sources 120, 121, and 122 can include a quasimonochromatic light sources having a narrow band of primary-color wavelengths. Illumination sources 120, 121, and 122 can also include LEDs, laser diodes, or lasers.
[0068] Imager 150 is used to image the light scattered and/or reflected from any object placed at imaging surface 105, such as finger 110. Imager 150 can be a color or black and white imager. Imager 150 can image an object through various optical elements 140 and 145, as well as through color filter array 130. Color filter array 130 can be integrated within imager 150 or be a stand alone element. Color filter array 130 is a mosaic of tiny color filters that separates the color response based on the arrangement of the mosaic. Various types of color filter arrays can be used such as, for example, a Bayer filter, a RGBE filter, a CYYM filter, a CYGM filter, or any type of RGB W filter.
[0069] An example of a Bayer color filter array is shown in Figure 4A. This arrangement uses red 404, green 412, and blue 408 color filter elements. As seen from the color response curve for a typical Bayer filter array shown in Figure 4B, there is generally some overlap in the spectral ranges of the red 424, green 432, and blue 428 transmission characteristics of the filter elements. As evident particularly in the curves for the green 432 and blue 428 transmission characteristics, the filter array may allow the transmission of infrared light. This can be avoided with the inclusion of an infrared filter as part of the imager. In other embodiments, an infrared filter may be omitted and one or more illumination sources that emit infrared light may be incorporated. In this way, all color filter elements 404, 408, and 412 may allow the light to substantially pass through, resulting in an infrared image at imager 150.
[0070] As an alternative to a color filter the imager or other optics may incorporate chromatic and/or polarizing beam splitters to separate the multiplexed light from the finger. The separated light may then be directed to individual monochromatic imagers where the number of imagers corresponds to the number of illumination conditions so separated.
[0071] Regardless of the type of color filter array that is used, in some embodiments, the wavelength of light provided by illumination sources 120, 121, and 122 should be aligned or closely aligned with the elements of the filter array. And, in some embodiments, the number of illumination sources can be, but not necessarily, aligned with the number of different filter elements in the array. For example, for a Bayer array, there are three filter colors. Hence, three illumination sources can be used. Each illumination source can then provide light in a red, blue, or green wavelength band to be consistent with the array.
[0072] Thus, imager 150 can create a multiplexed image of an object illuminated by multiple illumination sources arranged to illuminate light of different wavelengths at different illumination angles. This can be done during a single illumination session and, in some embodiments, a single image can be created. This single image can be considered a multiplexed image that can later be demultiplexed by color to provide information regarding the object that depends on the illumination angle of the illumination source. Furthermore, the multiplexed image can be considered a multispectral image because it contains information not only about the illumination wavelength but also the illumination angle.
[0073] In some embodiments, the illumination sources can include different polarization effects instead of or in addition to the differences in illumination angle. The resulting multiplexed image can then be a multispectral image that includes polarization information as well. [0074] An example of a sensor with multiplexed image data using a color filter array is shown in Figure 2. Platen 103 is provided and through which imaging optics 205 can image a finger (or an object) onto the imager 207. In some embodiments, imager 207 can be a silicon CMOS imager with an integrated Bayer filter. Other color filter arrays and/or other mean of separating the light (such as polarizing beam splitters or polarizing filter arrays), and/or other image array materials may be used alternatively or in conjunction with this arrangement.
Moreover, various imagers can be used as well such as silicon CCD imagers or imagers made from other photoactive material such as InGaAs, InSb, MCT and others known in the art.
[0075] Figure 3 shows another view of the sensor shown in Figure 2 with two of the multiplexed illumination sources illustrated. In this embodiment, blue light source 203 (e.g., a blue LED, which may be combined with lenses, mirrors, optical diffusers and other components of the sort), and green light source 205 (e.g., a green LED, which may be combined with lenses, mirrors, optical diffusers and other components of the sort), are turned on simultaneously. The light traverses a path that illuminates platen 103 at two different angles. The light from light sources 203 and 205 are shown as being reflected from internal mirrored surfaces 211 and 213, respectively. The reflecting surfaces 211, 213 are used to increase the illumination angle (with respect to the platen surface normal) while maintaining a small sensor footprint. Once the light from light sources 203 and 205 illuminate and pass through platen 103, the light may illuminate a finger or another object from two angles simultaneously. An imaging system with a color filter array, like the one shown in Figure 1 A can then used to acquire a multiplexed image.
[0076] In addition to the two illuminators 203, 205, other illuminators 207, 209 could be used to provide for similar multiplexed illumination during earlier or later image frames. In addition, illuminators of colors other than the blue and green can be incorporated into other embodiments of the present invention. For example, a red illuminator in a different spatial and angular position from illuminators 203, 205 could be used.
[0077] Figure 5 shows a flow chart of process 500 for creating a multiplexed image according to some embodiments of the invention. Process 500 starts at block 500. At block 510 a skin site is illuminated with at least two illumination sources having different wavelengths and illumination angles at block 510. The two illumination sources can illuminate the skin site at the same time or sequentially. In other embodiments, more than two illumination sources with different illumination wavelengths and different illumination angles can be used. At block 515, light from the skin site can be filtered using a color filter array of any type and then imaged at block 520. In some embodiments, a single multiplexed image can be acquired. Process 520 can end at block 525. Various post processing techniques can be used to pull out multispectral information from the multiplexed image. For example, the pixels corresponding to specific color filter mosaics may be separately extracted from the multiplexed image. Each of these images will then represent images collected under different illumination conditions (as well as different spectral conditions). These images may then be interpolated using linear or nonlinear techniques to produce a uniform sampling of the object over the same grid.
[0078] Process 500, for example, can be executed by the computational device shown in Figure 15.
[0079] Presence and Object Detection [0080] Some embodiments of the invention can be used to detect the presence and/or location of an object prior to imaging. Figure 6 illustrates an optical system that can be used to detect the presence of an object at a platen. Illumination system 600 can include an illumination source (e.g., LED, laser, etc.) and illumination optics (e.g., cylindrical lens, spherical lens, mirrors, and/or other such optical components). The illumination source can be a
monochromatic source. For example, a red, blue or green monochromatic source can be used. In some embodiments, illumination assembly 603 can direct light into a beam steering component 605, which directs the light toward mirrored surface 607 and then on and through platen 603. Imager 615 can be used to take video or sequential images of an area on or above platen 603. Imager 615 can include a color filter array. And, in some embodiments, a pixel (or mosaic) of the color filter array can be substantially the same as or consistent with the wavelength of one or more of illumination source (e.g., illumination system 600). The path of the illumination light 703 is illustrated in Figure 7.
[0081] In Figure 7, illumination light 703 illuminates a distinct region of platen 603. In some embodiments and as shown in the figure, illumination light 703 illuminates platen 103 with a steep illumination exit angle relative to the normal of the platen surface. In so doing, the distinct pattern can appear at different locations in the image based on the height of the finger or other object above the platen. Mirrored surface 607 can be used to provide this steep illumination exit angle relative to the air above the platen without sacrificing a compact design. For example, the exit angle of illumination (above the platen as the light exits the platen) can be any exit angle that is not perpendicular to the surface of the platen. For example, the exit angle (Θ) can be about 51° (±10°) from the normal of the platen. In other embodiments, the exit angle can be any angle from 40° to as close to parallel to the platen as practical (i.e. 90°).0 For example, the angle of illumination can be 40°, 45°, 50°, 55°, 60°, 65°, 70°, 75°, 80°, 85°. [0082] Figures 8A-1 IB illustrate how the height of an object above the platen can be determined according to some embodiments of the invention. For simplicity refraction effects as light enters and exits the platen are ignored. If the refraction effects were shown, the light exiting the platen would have a larger exit angle, Θ, than shown. This exit angle, Θ, can be any angle described in the preceding paragraph. In figure 8A finger 805-is positioned well above platen 803. Imager 815 is used to image the finger. Imager 815 can include a color filter array. Illumination source 800 illuminates the platen surface and the area near the platen surface. Light rays from illumination source 800 can be confined within strip 820. The illumination exit angle of illumination source 800 can be steep relative to the normal of the platen surface. As finger 805 approaches platen 803, a portion of the finger is illuminated as shown in Figure 8A. Figure 8B shows a block illustration of image 830 produced by imager 815. Illumination pattern 835 is formed from the light scattered from finger 805. The position of illumination pattern 805 can indicate the position of the finger relative to the platen. In Figure 8B illumination pattern is located at the far right of the image indicating a position of finger 805 well above platen 803. [0083] Figure 9A shows finger 805 positioned closer to platen 803 a moment after the finger's position in Figure 8A. And Figure 9B shows image 830 of finger 805 with
illumination pattern 835 moved toward the left to coincide with the movement of the finger. Figure 10A shows finger 805 positioned even closer to platen 803 a moment after the finger's position in Figure 9 A. And Figure 10B shows image 830 of finger 805 with illumination pattern 835 moved toward the left to coincide with the location of the finger. Finally, Figure 11 A shows finger 805 in contact with platen 803 a moment after the finger's position in Figure 10A. And Figure 1 IB shows image 830 of finger 805 with illumination pattern 835 moved toward the left. Thus, by monitoring the position of illumination pattern 835 in sequential images the position of finger 805 can be determined relative to platen 803. As can be seen from these figures, the steepness of the illumination exit angle in the air above the platen determines the relationship between the height of the finger 805 relative to the platen 803 and the position of the illumination pattern 835 in the image 830. If the illumination exit angle is steeper (closer to 90 degrees) the position of the illumination pattern 835 will change faster as a function of the finger 805 height above the platen 803. [0084] Figure 12 shows an actual image that results when the finger or other object surface is located about 5 mm above the surface of a platen. As can be seen from the image, the illuminated pattern 1235 is located toward the left side of the image. Figure 13 shows another actual image that results when a finger or other object surface is located on the platen. With the finger on the platen, illumination pattern 1335 is centrally located on the image, while illumination pattern 1235 is located on the side of the image. The horizontal shift between illumination patterns is due to the steep exit angle of the illumination beam.
[0085] In some embodiments, the intensity profile of the illumination pattern changes as the finger approaches the platen. Because of the exit angle of illumination and the rounded cross- sectional shape of a finger, while the finger is above the platen the intensity profile will be sharper on one side of the finger and more gradual on the other. As the finger moves closer the intensity profile becomes more and more symmetric. In some embodiments, this intensity profile can be used as further evidence of the position of the finger. Other information such as the spectral content, the shape, and the position can be used to determine the location and the object type. For example, the width of the illumination pattern may widen as the finger approaches the platen.
[0086] A video sequence of images can be acquired and analyzed to determine the location of the illumination pattern. If the illumination pattern is not located in the proper location then another image can be collected and/or analyzed until the illumination pattern is in the proper location. Once in the proper location, biometric imaging or other processes may begin that require the object to be in a certain location.
[0087] In some embodiments, the shape of illumination pattern can indicate the type of object being imaged. For example, in Figures 12 and 13, a finger was imaged resulting in an illumination pattern resembling a slightly bowed rectangle. And the center of the platen image is brightly illuminated (shown by the red color) while the light intensity falls off on either side (blue). Thus, both the geometrical shape and the intensity of the illumination can be indicative of the object under illumination. This illumination pattern may depend on the illumination angle and/or the type of object being imaged. By fixing the illumination angle and calibrating the system with the illumination angle fixed, the illumination pattern may then depend on the object type. Images can be acquired in sequential order and the illumination pattern can be analyzed.
[0088] In biometric imaging of fingerprints, for example, if the illumination pattern is inconsistent with the known illumination pattern of a finger, then it can be determined that a foreign object is being placed on the platen and biometric imaging should not start. If the illumination pattern is consistent with the known illumination pattern of a finger, then it can be determined that a finger is being placed on the platen and biometric imaging can begin. This technique can be used in conjunction with the presence detection techniques to determine when the finger is in contact with the platen prior to triggering the start of biometric imaging.
[0089] Various other objects can have correspondingly unique illumination patterns that can be used. The particular illumination pattern is not critical, as long as it is easily recognized and distinct from objects in the background. In order to achieve this, the shape of the illumination pattern may be altered using appropriate optics in the illumination assembly as well as elsewhere along the optical path.
[0090] In addition, the spectral properties of the illumination pattern may be distinguished from background materials and/or ambient lighting conditions ("background noise"). For example, the illumination source can be a blue LED that emits light strongly detected by the blue pixels of a color filter array and weakly detected by the other pixels of the color filter array used in the imager. The illumination pattern can then be analyzed. And if it is determined that there illumination pattern provides a strong signal in the blue pixels and a weak signal in the other pixels, then the illumination pattern can safely be determined to not be background noise. If, however, the blue pixels do not register a strong signal relative to the other pixels, then the illumination pattern is likely a background and can be ignored. While this example uses a blue LED relative to blue pixels, other wavelengths can be used.
[0091] Using various embodiments described herein sequential images of the skin site can be acquired while being illuminated with a single illumination source having a steep illumination angle. Prior to operating a function that requires the presence of a specific type of object at a platen for imaging, such as in biometric imaging or multispectral imaging, the following can occur in any order and with or without any one step:
1. Analyzing the illumination pattern to ensure that the shape of the illumination pattern is consistent with the shape of the expected object.
2. Analyzing the location of the illumination pattern to determine if it's location is consistent with the expected object being in contact with the platen. The illumination pattern can include the geometric pattern and/or the intensity profile. 3. Analyzing the illumination pattern to determine if it has a strong signal for the pixels that are consistent with the wavelength of the illumination source and/or determine if the illumination pattern has a weak signal for pixels not consistent with the wavelength of the illumination source. [0092] Figure 14 shows a flowchart outlining process 1400 for determining whether to image an object present at or near the platen according to some embodiments of the invention.
Process 1400 starts at block 1405. At block 1410 the imaging surface and/or the area nearby the imaging surface is illuminated. In some embodiments, this illumination can occur at a steep angle relative to the normal. In some embodiments, this illumination can occur using a monochromatic light source. The wavelength of this monochromatic light source, for example, can be matched with one of the pixels and/or mosaics of a color filter array of the imager.
[0093] At block 1415 the area near the imaging surface is imaged. At block 1420 the image is analyzed to determine if the shape of the illumination pattern is consistent with the expected object. This illumination pattern may be a geometric illumination pattern and/or an intensity pattern. If the illumination patter is inconsistent with the expected pattern, then process 600 returns to block 1415 and another image is acquired. In some embodiments, process 600 pauses a short time period prior to returning to block 1410. If the pattern is consistent with the expected object, then process 1400 proceeds to block 1425.
[0094] At block 1425 process 1400 determines whether the illumination pattern is formed in a location consistent with the expect object being placed on the platen. If the location is inconsistent, then process 1400 returns to block 1415 and another image is acquired. If it is consistent, then process 1400 proceeds to block 1430. At block 1430 process 1400 determines whether the illumination pattern is consistent with background noise. This can be done by analyzing the pixels of the image in conjunction with the color filter array. If the pixels within the illumination pattern are strongly associated with light illumination wavelength and/or weakly associated with light associated with other pixels, then the object is not a background object and process 1400 proceeds to 1435. Otherwise process 1400 returns back to block 1415. At block 1435 imaging or other tests or measurements can occur once it is determined that the proper object is placed in the proper location and is not background noise. Process 1400 ends at block 1440.
[0095] Process 1400, for example, can be executed by the computational device shown in Figure 15. [0096] Computational Device
[0097] Figure 15 shows a block diagram of computational unit 1500 that can be used inc conjunction or as part of a biometric sensor system. The figure broadly illustrates how individual system elements may be implemented in a separated or more integrated manner. Moreover, the drawing also illustrates how each of the four imagers 1510 may include a dedicated processor 1515 and/or dedicated memory 1520. Each dedicated memory 1520 may include operational programs, data processing programs, instructions for carrying out methods described herein, and/or image processing programs operable on the dedicated processors 1515. For example, the dedicated memory 1 20 may include programs that can be executed by CPU 1502 and/or provide image processing. The computational device is shown comprised of hardware elements that are electrically coupled via bus 1530. The bus 1530, depending on the configuration, may also be coupled with the one or more LED(s) 1505, a proximity sensor (or presence sensor) 1512 and four imaging subsystems 1504 according to various embodiments. In another embodiment, imager memory 1520 may be shared amongst imagers 1515 and/or with the computational device 1502. [0098] In such embodiments, an imaging subsystem may include an imager 1510, a processor 1515, and memory 1520. In other embodiments, an imaging subsystem 1504 may also include light sources and/or optical elements. Imaging subsystems 1504 may be modular and additional imaging subsystems may be easily added to the system Thus, biometric sensor subsystems may include any number of imaging subsystems 1504. The various imaging subsystems, in one embodiment, may be spatially modular in that each imaging subsystem is used to image a different spatial location. The various imaging subsystems, in another embodiment, may be multispectrally modular in that each imaging subsystem is used to image a different multispectral condition. Accordingly, in such an embodiment, an imaging subsystem 1504 may also include various optical elements such as, for example, color filter arrays, color filters, polarizers, etc and/or the imager 1510 may be placed at various angles relative to the imaging location. The various imaging subsystems, in another embodiment, may provide focus modularity in that each imaging subsystem is used to image a different focal point or focal plane. [0099] The hardware elements may include a central processing unit (CPU) 1550, an input/output device(s) 1535, a storage device 1555, a computer-readable storage 1540, a network interface card (NIC) 1545, a processing acceleration unit 1548 such as a DSP or special-purpose processor, and a memory 1560. The computer-readable storage 1540 may include a computer-readable storage medium and a computer readable medium reader, the combination comprehensively representing remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing computer- readable information. The NIC 1545 may comprise a wired, wireless, modem, and/or other type of interfacing connection and permits data to be exchanged with external devices.
[0100] Moreover, computational unit can include illumination source interface 1557 and/or imager interface 1556. These interfaces can be standard input/output interfaces such USB, firewire, UART, RS232 or a proprietary interface. Regardless of the communication protocols, imager interface 1556 can be used to instruct the imager (e.g., any of the imagers or sensors described herein) to acquire an image and/or export an image. Illumination source interface 1557 can be use to activate and/or deactivate any of a plurality of illumination sources singularly, as a group, or sequentially. [0101] The computational unit 1500 may also comprises software elements, shown as being currently located within working memory 1560, including an operating system 1565 and other programs and/or code 1570, such as a program or programs designed to implement methods described herein. It will be apparent to those skilled in the art that substantial variations may be used in accordance with specific requirements. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets), or both. Further, connection to other computing devices such as network input/output devices may be employed.
[0102] Computational unit 1500 can be used to carry out processes shown in any of the figures and described in the specification. Specific instructions and/or program code can also be stored in memory 1518 or 1512 and executed by CPU 1502.
[0103] Dual-Imager Biometric Sensor
[0104] Embodiments of the invention may include systems, methods, and devices that may collect images of the finger using two different types of imagers that acquire images of the finger during the same measurement session. In some cases, the imagers may acquire images simultaneously. In some cases the imagers may acquire two video sequences wherein the frames of the video sequences are synchronized such that the two sequences are substantially temporally aligned to aid processing in the presence of motion of the finger or other object being imaged. In some embodiments, the two imagers may comprise a TIR imaging system and/or a direct imaging system. In some embodiments either or both the imaging systems may be multispectral imaging systems.
[0105] In some embodiments of the invention a prism with multiple facets can be used with facets arranged for simultaneous direct and TIR imaging of a skin site. The prism can be used in conjunction with one or more illumination sources and/or one or more imagers. A first facet of the prism may be adapted for placement of a purported skin site by an individual. The first facet may include an imaging surface for placement of the purported skin site. The first facet or just the imaging surface of the first facet may alternatively be referred to as a "platen". A second facet may be oriented with respect to the first facet such that a portion of light totally internally reflected at the first facet substantially passes through the second facet. The sensor may include one or more illumination sources disposed to illuminate the imaging surface of the prism and/or the purported skin site through the first facet of the prism. The sensor may include a direct imaging assembly disposed to receive light scattered from the purported skin site and reflected from the second facet of the prism. The sensor may include a total-internal- reflectance imaging assembly disposed to receive light scattered beneath a surface of the purported skin site at the first facet and substantially passes through a third facet of the prism.
[0106] Figure 16 shows an isometric view of dual-imager biometric sensor 1601 according to some embodiments of the invention. Sensor 1601 includes prism 1610 and further designates a region of the platen that is the active image surface 1615 used to image the finger or other obj ect; active image surface 1615 may be referred to as a platen or as a facet of prism 1610. In some embodiments, prism 1610 may include multiple facets. A dual-imager biometric sensor may include one or more illumination sources. For example, dual imager sensor 1601 includes two illumination sources 1631 and 1632. A dual imager sensor may include other components, such as reflective surface like mirrors 1621, 1622, 1623, 1624, and/or 1625, which may be used to facilitate directing light to different imaging components. A dual imaging sensor may also include one or more total-internal-reflectance imaging assemblies. Figure 21 for example shows TIR imaging assembly 1641. A dual imaging biometric sensor may also include a direct imaging assembly 1642.
[0107] Figure 17 shows the top view of dual-imager biometric sensor 1601 and Figure 18 shows a front view of sensor 1601. Two illumination assemblies 1631 and 1632 are shown in Figure 17. And Figure 18 shows a third illumination assembly 1633 along with illumination assembly 1631. Each illumination assembly may comprise an illumination source as well as lenses, mirrors and/or other optics as necessary to shape and direct the illumination light. In some instances, more or less illuminators may be present. The illuminators 1631, 1632, and/or 1633 may come on simultaneously during an acquisition sequence, or sequentially, or pulsed or encoded in some manner.
[0108] In some embodiments the illuminators 1631, 1632, and/or 1633 may be of
substantially different wavelengths in accordance with other aspects of various embodiments. Some embodiments may utilize illumination assemblies that may include illumination sources without other optics to shape and direct the illumination source. Some embodiments of a dual imaging sensor may also include other illumination assemblies or illumination sources as will be discussed below. For example, some embodiments may include an illuminator such as a flat light guide illuminated with LEDs, or a cold cathode fluorescent lamp (CCFL), or other illuminators known in the art that may be used as part of a TTR imaging.
[0109] Illumination light from any or all sources, such as 1631, 1632, and/or 1633, may illuminate the image region 1615. Light from a finger touching the sensor may then be imaged by the TIR image assembly 1641 after being reflected from mirrored surfaces 1621 and 1624. In some embodiments, a TIR image assembly, such as 1641, may not need reflecting surfaces such as 1621 and 1624 to make an image. Other embodiments may include more or less imaging surfaces that may be in different configurations. The finger touching the active image region 1615 may also be imaged by direct image assembly 1642 after reflections from surfaces 1622, 1623, and/or 1625. In some embodiments, a direct image assembly, such as 1642, may not need reflection surfaces such as 1622, 1623, and/or 1625. Other embodiments may use more or less reflecting surfaces, which also may be position and/or oriented in different configurations. [0110] In one embodiment, direct image assembly 1642 may contain a color imager, as well as lenses, mirrors, optical filters, grating and other such optical elements as may be necessary to form an image as known in the art. TIR image assembly 1641 may contain a color imager or a monochromatic imager, as well as lenses, mirrors, optical filters, grating and other such optical elements as may be necessary to form an image as known in the art. In some embodiments that incorporate a monochromatic TIR imager, there may be a optical band pass filter in TIR image assembly 1641 that substantially passes the light from some number of the illuminators, such as 1631, 1632, and/or 1633, while blocking the light from others. In some embodiments, an optical filter that blocks out-of-band light ambient light may be included in either or both imaging paths. [0111] Figure 19 shows a three-dimensional view of sensor 1601. This illustration shows the simultaneous illumination of the image region 1615 of prism 1610 by illuminators 1631, 1632, and/or 1633. In some embodiments, illuminators 1631, 1632, and/or 1633 may be different colors (such as green, red and blue, respectively, merely by way of example; ultraviolet and/or infrared light may be used instead of or in addition to various visible wavelengths), in accordance with another aspect of various embodiments. The central wavelengths of the illuminators such as 1631, 1632, and/or 1633 as well as the spectral width and selection of which illuminator emits which color can all vary and such choices are within the scope of various embodiments. In some embodiments, as shown in Figure 4 for example, an
illuminator, such as 1631 and/or 1632, may be positioned such that light transmitted from the illuminator is totally internally reflected at a facet, such as facet 1611, of prism 1610 and then illuminates image region 1615. Some illuminators, such as 1633, may be positioned and/or oriented such that light passes through a facet, such as facet 1611 and then illuminates image region 1615.
[0112] Figure 20 depicts a TIR imaging path 1651. The TIR imaging assembly 1641 , which may be referred to as a TIR imager in some embodiments, can view the active region 1615 of prisml610 through prism facet 1612 and reflective surfaces 1621 and 1624. The angle between the TIR image bundle 1651 and the platen facet 1612 is such that the bundle 1651 can pass through the facet 1612. In some embodiments, the image bundle 1651 is substantially perpendicular to facet 1612. In some embodiments, facet 1612 may be oriented approximately perpendicular to an angle equal to or greater than the a critical angle with respect to facet 1615
[0113] In some embodiments, image bundle 1651 is oriented at an angle greater than the critical angle defined by the surface 1615 and the surrounding media, which is usually air. Because of this, it may be desirable to provide a black background for the TIR imager 1641 to view in cases where the TIR imager 1641 is operating in a dark-field illumination mode. In order to accomplish this, facet 1611 of prism 1610 may be oriented such that rays in the TIR bundle 1615 reflected through TIR at surface 1615 may be substantially able to pass out of prism 1610 through facet 1611. In some embodiments, a light trap (e.g., optical absorber) 1660, which may comprise a material with a matte black surface finish, may be positioned such that substantially all of the TIR bundle 1651 lands on the trap 1660. [0114] In an alternative embodiment, the light trap 1660 may be replaced by an illuminator, such as a flat light guide illuminated with LEDs, or a cold cathode fluorescent lamp (CCFL), or other illuminators known in the art. In this alternate embodiment, the TIR imaging system operates in a bright-field illumination mode.
[0115] Figure 21 depicts direct imaging ray bundle 1652 seen by direct imaging assembly 1642 in accordance with various embodiments. Direct image bundle 1652 may pass through facet 1613. In some embodiment, direct image bundle 1652 is approximately normal to facet 1613. In some embodiments, direct image bundle 1652 is oriented such that it is reflected due to TIR at facet 1611 and passes through image region 1615.
[0116] In some embodiments, facet 1611 may serve a dual role of being highly reflective for the direct imaging bundle 1652 and also being transmissive for TIR imaging bundle 1651 as illustrated in Figure 22. In some embodiments, facet 1611 may serve a dual role of being highly reflective for light emanating from illuminators 1631 and/or 1632 and also being transmissive for light emanating from illuminator 1633. In some embodiments, facet 1611 may be variously reflective and transmissive for some combination of illumination and imaging bundles.
[0117] Figure 22 provides a diagram of a dual imager sensor or system 1607 in accordance with various embodiments. System 1607 includes prism 1610 with multiple facets 1611, 1612, 1613, and 1615. System 1607 includes a direct imager 1644 and a TIR imager 1643. Direct imaging assembly 1642 of Figure 21, for example, may be an example of direct imager 1644. Some embodiments may include direct imager 1644 that may include other optical components as described herein, such as mirrors, 1622, 1623, and 1625. TIR imaging assembly 1641 of Figure 20, for example, may be an example of TIR imager 1643. Some embodiments may include TIR imager 1643 that may include other optical components as described herein, such as mirrors 1621 and/or 1624. Direct imager 1644 and TIR imager 1643 may be oriented and/or positioned in different configurations in different embodiments.
[0118] As described above, different facets of prism 1610 may provide different functions. For example, facet 1611 may serve a dual role of being highly reflective for direct imaging and also being transmissive for TIR imaging. As can be seen, a TIR bundle 1653 directed towards facet 1611 will transmit through facet 1611. A direct imaging bundle 1654 may reflect at facet 1611. In some embodiments, direct imaging bundle 1653 totally internally reflects at facet 1611. In addition, system 1607 shows an example of a facet 1613 that direct imaging bundle 1654 may pass through and then reach a direct imager 1644. Furthermore, a TIR bundle 1655 may pass through facet 1612 and then reach a TIR imager 1643.
[0119] Facets as seen in Figure 22 may have different relationships with each other in order to affect the functions described for different embodiments. For example, facet 1615 and 1611 may be configured such that facet 1611 is oriented approximately perpendicular to an angle equal to or greater than the critical angle of facet 1615 such that light that undergoes TIR at facet 1615 may pass through facet 1611 before or after such TIR interaction. Similarly facet 1612 may be oriented approximately perpendicular to an angle equal to or greater than the critical angle of facet 1615 such that light totally internally reflect at facet 1615 may pass through facet 1612.
[0120] Prism 1610 may include facets such as 1613 that are oriented in different ways. For example, facet 1613 may be oriented relatively perpendicular to light that passed through facet 1615 and reflected from facet 1611 such that the light substantially passes through facet 1613 . [0121] System 1607 may also include different illumination sources. Figures 1 to 6, 16 to 21, and 24 to 25 show different examples with illumination sources. These figures also show different examples of orientations with respect to imagers. Illumination sources may be position and/or oriented in different ways to illuminate a finger, for example, place on prism 1610 along facet 1615, for example. Some embodiments may also include a controller, which may control imagers such as 1643 and 1643 along with illumination sources.
[0122] In some embodiments, sensors and systems may each collect images simultaneously using direct imagers such as 1642 or 1644 and TIR imager 1641 or 1643. If either of the images is multiplexed in the manner of the different embodiments, the image may be decomposed into the images corresponding to the different illumination conditions followed by interpolation and/or other processing of the resulting images. The resulting collection of images (both TIR and direct) may then be processed further in such a manner that a single, composite image of the fingerprint is generated. This composite image may then be reported to a host computer or other connected device.
[0123] In some embodiments, the TIR imager and direct imagers of sensors and systems may acquire synchronized video streams of images. These video streams may be acquired while the finger is rolled in the manner known in the art. The rolled image sequences may then be combined to produce a composite rolled print image.
[0124] In some embodiments, the fingerprint sensor of different embodiments may be used to collect non-fingerprint images. In particular, the direct imager may be used to acquire such images. Other image data so collected may include barcodes and documents. Optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions. [0125] In some embodiments, the direct imager of the different embodiments may be used to collect iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, and perform a variety of other such functions. In some embodiments the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality. [0126] Figure 23 shows a flow diagram for process 2300 for creating composite biometric images from a two imager biometric system. Process 2300 may be implemented, for example, using sensors and systems as seen in Figures 15 to 22. Process 2300 can start at block 2305. At block 2310, a purported skin site is illuminated through a first facet of a platen using one or more illumination sources. At block 2315, light from the purported skin site is received. At block 2320, a direct image of the purported skin site is generated from a first portion of the received light that is reflected at a second facet of the platen. At block 2320, a total-internal- reflectance image of the purported skin site is generated from a second portion of the received light that passes through a third facet of the platen oriented. Process 2300 can end at block 2330. Any of these blocks can occur simultaneously. For instance, blocks 2310 and 2315 may occur at the same time as well as blocks 2320 and 2325.
[0127] In some embodiments of process 2300, a third portion of light is totally internally reflected at the first facet of the platen and then passes through the second facet of the platen. Some embodiments may further comprise absorbing the third portion of the light at a light trap located proximate to the second facet such that totally internally reflected light at the first facet that passes through the second facet is substantially absorbed by the light trap.
[0128] In some embodiments of process 2300, the light received by the direct imaging assembly that is reflected from the second facet of the platen is totally internally reflected at the second facet. In some embodiments, the second and third facets are oriented at substantially critical angles with respect to the first facet. [0129] Some embodiments of process 2300 may include generating a composite image of the purported skin site using the total-internal-reflectance image and the direct image of the purported skin site. Some embodiments may include receiving light from the purported skin site that includes receiving the first portion of the received light at a total-internal-reflectance imaging assembly and receiving the second portion of the received light at a direct imaging assembly. Some embodiments may further comprise decomposing at least one of the images into multiple images corresponding to different illumination conditions.
[0130] Some embodiments of process 2300 may further comprise illuminating a non-skin site through the platen using one or more illumination sources. Light from the non-skin site is received and then a direct image of the non-skin site from a second portion of the received light may be generated. The non-skin site may include barcodes, documents, or optical security markings, for example. In some embodiments, the non-skin site may includes an iris, a portion of a face, a surveillance location, a moving object, or a lighting condition, for example.
[0131] Process 900, for example, can be executed by the computational device shown in Figure 15.
[0132] In some embodiments of the invention, the TIR imager and direct imager may each collect images simultaneously. If either of the images is multiplexed in the manner described in some of the embodiments described herein, the image may be decomposed into the images corresponding to the different illumination conditions followed by interpolation and/or other processing of the resulting images. The resulting collection of images (both TIR and direct) may then be processed to create a single, composite image of the fingerprint. This composite image may then be reported to a host computer or other connected device.
[0133] In some embodiments, the TIR imager and direct imager may acquire synchronized video streams of images. These video streams may be acquired while the finger is rolled in the manner known in the art. The rolled image sequences may then be combined to produce a composite rolled print image.
[0134] Various other configuration can be used. For instance, a first illumination source (e.g. an LED) can be used to illuminating a finger through a first facet (e.g., facet 1615) of a multifaceted prism (e.g., prism 1610). Light from the first illumination source can undergo total internal reflectance at a second facet (e.g., 1611) of the multifaceted prism prior to illuminating the object. A first imager (e.g., 1644) can image light scattered from the finger and passing though the first facet at an angle less than the critical angle and undergoing total internal reflectance at the second facet. A second imager can image light scattered from the finger and passing through the first facet at an angle less than the critical angle. In some embodiments, the second imager can be located in a position such that it does not image light from the second illumination source that is totally internally reflected at the first facet.
[0135] In some embodiments, the second imager can be located out of phase from the second illumination source. That is light from the second illumination source can only image light from the second imager after it is absorbed and/or scattered by the object. And, in this embodiment, light totally internally reflected at the first facet is not imaged by the second imager. For example, the second imager can be located at an azimuthal angle out of line with the second illumination source; such as an angle less than 170°. In some embodiments, this angle is 90°. The second imager can also be located to image light at greater or less than the critical angle.
[0136] In another embodiment, a first illumination source can illuminate an object located at a first facet (e.g., 1615) of a multifaceted prism (e.g., 1610). The system can include a second illumination source that can illuminate the object through a second facet (e.g., 1611) and through the first facet at an angle greater than the critical angle of the first facet. A first imager (e.g., 1644) can image light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoes total internal reflectance at the second facet. A second imager (1643) can be used to image light totally internally reflected from the first facet.
[0137] In some embodiments, the fingerprint sensor of the present invention may be used to collect non-fingerprint images; for example, money, documents, bar codes, manufactured parts, etc. In some of these images optical security markings such as holograms, color-changing ink and other such markings may be present and used to confirm that the documents or barcodes by assessing the images that correspond to different illumination conditions.
[0138] In some embodiments, the dual imager of the present invention may be used to collect finger or hand print images as well as iris images, facial images, surveillance images, detect motion, detect ambient lighting conditions, barcode images, security document images, and perform a variety of other such functions. In some embodiments the direct imager may include an automatic, variable focus ("autofocus") mechanism to facilitate additional imaging functionality.

Claims

WHAT IS CLAIMED IS:
1. A biometric method comprising:
illuminating a purported skin site under two unique illumination conditions during the same illumination session, the first illumination condition comprising light with a first wavelength illuminating the purported skin site at a first angle, and the second illumination condition comprising light with a second wavelength illuminating the purported skin site at a second angle, wherein the first wavelength and the second wavelength are distinct and the first angle and second angle are distinct;
filtering light from the purported skin site using a color filter array that distinguishes between light of the first wavelength and light of the second wavelength; and imaging the filtered light with a single imager.
2. The method according to claim 1, wherein the imaging produces a single multiplexed image of the skin site.
3. The method according to claim 1, wherein the imaging creates a single image of the purported skin site.
4. The method according to claim 1 further comprising extracting a first image of the purported skin site imaged with light of the first wavelength, and extracting a second image of the purported skin site imaged with light of the second wavelength.
5. The method according to claim 1 wherein the first light source is monochromatic, and the second light source is monochromatic.
6. The method according to claim 1, wherein the filtering filters the light using a color filter array.
7. The method according to claim 6, wherein the color filter array comprises a Bayer filter.
8. The method according to claim 1, wherein the first wavelength is between 440 nm and 490 nm.
9. The method according to claim 1, wherein the first wavelength is between 520 nm and 570 nm.
10. The method according to claim 1 further comprising illuminating the purported skin site with light of a third wavelength illuminating the purported skin site at a third angle, wherein the third wavelength is distinct from both the first and the second wavelengths and the third angle is distinct from the second and third angles, wherein the filter distinguishes between light of the third, second and first wavelengths.
11. A system comprising:
a platen having an imaging site;
a first illumination source configured to illuminate the imaging site with light having a first wavelength and at a first angle;
a second illumination source configured to illuminate the imaging site with light having a second wavelength and at a second angle, wherein the first wavelength and the second wavelength are distinct, and the first angle and the second angle are distinct;
a color filter array configured to filter light having the first wavelength from light having the second wavelength; and
an imager configured to acquire an image the skin site through the filter array.
12. The system according to claim 11, wherein either or both the first angle and the second angle are greater than the critical angle defined by the interface at the imaging site and measured relative to the normal.
13. The system according to claim 11 , wherein the color filter array is part of the imager.
14. The system according to claim 11, wherein the image acquired through the color filter array is a multiplexed image.
15. system according to claim 11, wherein the color filter array comprises a Bayer filter.
16. An object location method comprising: illuminating an area near a platen at an exit angle relative to the normal of the platen;
acquiring an image of an object at or near the platen;
identifying a location of an illumination pattern within the image, wherein the illumination pattern is an imaged response of the object being illuminated; and
determining the location of the object relative to the platen from the location of the illumination pattern within the image.
17. The method according to claim 16, wherein the determining determines whether the object is in contact with the platen.
18. The method according to claim 16 further comprising determining a shape of the illumination pattern and determining whether the shape of the illumination pattern is consistent with an expected object.
19. The method according to claim 18, wherein the shape of the illumination pattern includes a geometric shape and/or an illumination pattern.
20. The method according to claim 18, wherein the acquiring, identifying, and determining steps are repeated until the illumination pattern is consistent with the expected object.
21. An object discrimination method comprising:
illuminating a purported skin site of a platen at an exit angle relative to the normal of the platen;
acquiring an image of an object at or near the platen;
identifying the shape of an illumination pattern within the image that is associated with an object; and
determining whether the shape of the illumination pattern is consistent with the shape of an illumination pattern of an expected object.
22. The method according to claim 21 , wherein the illumination pattern includes geometric pattern.
23. The method according to claim 21 , wherein the illumination pattern includes an intensity pattern.
24. The method according to claim 21 , wherein the determining step comprises determining whether the illumination pattern is consistent with an imaged finger.
25. A background discrimination method comprising:
illuminating a purported skin site of a platen with an illumination source having a first wavelength;
acquiring an image of an object at or near the platen, wherein the image is acquired with an imaging system that includes a color filter array that includes a plurality of first pixels associated with the first wavelength and a plurality of second pixels associated with a second wavelength;
identifying an illumination pattern of the obj ect; and
determining whether a higher concentration of first pixels within the illumination pattern are activated than the second pixels within the illumination pattern.
26. The method according to claim 25, wherein the first wavelength is between 440 nm and 490 nm.
27. A system comprising:
a platen having an imaging site;
an illumination source configured to illuminate the imaging site at an exit angle relative to the normal of the imaging site and with light having a first wavelength;
an imager configured to acquire an image of the skin site using a color filter array, wherein the color filter array filters colors according to a plurality of wavelengths and wherein one of the plurality of wavelengths includes the first wavelength; and
a computational unit communicatively coupled with the imager..
28. The system according to claim 27, wherein the computational unit is configured to determine whether an object is in contact with the imaging site from an image of the obj ect provided by the imager.
29. The system according to claim 28, wherein the computational unit is configured to identify an illumination pattern of the object within an image and determine whether the illumination pattern is consistent with an object in contact with the skin site.
30. The system according to claim 27, wherein the computational unit is configured to determine whether an object is consistent with an expected object.
31. The system according to claim 30, wherein the computational unit is configured to identify an illumination pattern of the object within an image and determine whether the illumination pattern is consistent with an expected object.
32. The system according to claim 31 , wherein the illumination pattern includes either or both a geometric pattern and an intensity pattern.
33. The system according to claim 27, wherein
the color filter array includes first pixels associated with the first wavelength; and
the computational unit is configured to whether an object within an image is consistent with a background object.
34. The system according to claim 33, wherein the computational unit is configured to identify an illumination pattern within the image and determine whether the concentration of first pixels activated within the illumination pattern is greater than the concentration of first pixels without the illumination pattern.
35. A biometric system comprising:
first illumination means for illuminating a finger through a first facet of a multifaceted prism, wherein light from the first illumination means undergoes total internal reflectance at a second surface of the multifaceted prism prior to illuminating the object; and first imaging means for imaging light scattered from the finger and passing though the first facet at an angle less than the critical angle and undergoing total internal reflectance at the second facet, wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
36. The system of claim 35, further comprising second imaging means for imaging light scattered from the finger and passing though the first facet at an angle greater than the critical angle.
37. The system claim 35, further comprising a light trap located proximate to the second facet such that light undergoing total internal reflectance at the first facet that passes through the second facet is substantially absorbed by the light trap.
38. The system of claim 36, further comprising a controller interfaced with the first illumination means, the first imaging means, and the second imaging means, the controller including:
instructions to illuminate the finger with the first illumination means during a single illumination session;
instructions to generate a total-internal reflectance image of the finger from light received by the second imaging means; and
instructions to generate a direct image of the finger from light received by the second imaging means.
39. The system of claim 38, wherein the controller interface further includes instructions to generate composite image of the finger using the first imaging means image and the second imaging means image of the finger.
40. The system of claim 36, wherein the first imaging means and the second imaging means are configured to receive light simultaneously.
41. The system of claim 36, wherein the first imaging means and the second imaging means are configured to acquire synchronized video streams.
42. The system of claim 36, wherein at least one of first imaging means or the second imaging means comprises a multispectral imaging means
43. The system of claim 35, further comprising a second illumination means for illuminating the finger through the first facet of a multifaceted prism, wherein light from the second illumination means passes through the second surface of the multifaceted prism prior to illuminating the finger.
44. The system of claim 43, wherein the first illumination means and the second illumination means comprise different wavelengths.
45. A method comprising:
illuminating an object positioned at a first facet with light that undergoes total internal reflectance at a second facet;
imaging light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoes total internal reflectance at the second facet; and
imaging light scattered by the object that passes through the first facet at an angle greater than the critical angle,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
46. The method of claim 45, further comprising absorbing light at a light trap located proximate to the second facet, wherein the light undergoes total-internal- reflectance at the first facet and then passes through the second facet.
47. The method of claim 45, further comprising generating a composite image of the object using the image from light scattered by the object that passes through the first facet at the angle less than the critical angle and undergoes total internal reflectance at the second facet and the image from light scattered by the object that passes through the first facet at the angle greater than the critical angle.
48. The method of claim 45, further comprising decomposing at least one of the images into a plurality of images corresponding to different illumination conditions.
49. The method of claim 45, wherein the object comprise a barcode, a document, or money.
50. A multifaceted prism comprising:
a first facet; and
a second facet, wherein the first facet and the second facet are oriented relative to each other such that:
an object positioned at the first facet can be illuminated with light that undergoes total internal reflectance at the second facet,
a portion of light scattered by the object that passes through the first facet at an angle less than the critical angle can be imaged after undergoing total internal reflectance at the second facet, and
a portion of light scattered by the object positioned at the first facet that passes through the first facet at an angle greater than the critical angle can be imaged without interacting with the second facet,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
51. The multifaceted prism of claim 50, further comprising: a third facet oriented with respect to the first and second facets such the portion of light scattered by the object that passes through the first facet at an angle less than the critical angle and undergoing total internal reflectance at the second facet then passes through the third facet.
52. The multifaceted prism of claim 51 ,wherein the third facet is oriented substantially perpendicular to the first facet.
53. The multifaceted prism of claim 50, further comprising: a forth facet oriented with respect to the first and second facets such that the portion of light scattered by the object positioned at the first facet that passes through the first facet at an angle greater than the critical angle passes through the fourth facet.
54. The multifaceted prism of claim 53,wherein the fourth facet is oriented at an angle greater than or equal the critical angle with respect to the first facet.
55. A system comprising:
first illumination means for illuminating an object located at a first facet of a multifaceted prism;
second illumination means for illuminating the object through a second facet and through the first facet at an angle greater than the critical angle of the first facet;
first imaging means for imaging light scattered by the object, passes through the first facet at an angle less than the critical angle, and undergoes total internal reflectance at the second facet; and
second imaging means for imaging light totally internally reflected from the first facet,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
56. The system according to claim 55, wherein the first illumination means and the second illumination means have distinct wavelengths.
57. The system according to claim 55, wherein the second imaging means is oriented to image light totally internally reflected from the interface of the object and the first facet.
58. The system according to claim 57, wherein the first illumination means illuminates the object at an angle less than the critical angle of the first facet.
59. The system according to claim 55, wherein the first imaging means and the second imaging means are configured to acquire synchronized video streams.
60. A system comprising:
first illumination means for illuminating an object located at a first facet of a multifaceted prism; second illumination means for illuminating the object through a second facet and through the first facet at an angle greater than the critical angle of the first facet;
first imaging means for imaging light scattered by the object, passes through the first facet at an angle less than the critical angle, and undergoes total internal reflectance at the second facet; and
second imaging means for imaging light scattered from the object at points where the object is in contact with the platen,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
61. The system according to claim 60, wherein the second imaging means images the platen at an angle less than the critical angle.
62. The system according to claim 60, wherein the second imaging means images the platen at an elevation angle greater than the critical angle.
63. The system according to claim 60, wherein the second imaging means images the platen at an azimuth angle less than 170° relative to the second illumination means.
64. The system according to claim 60, wherein the first illumination means and the second illumination means have distinct wavelengths.
65. A method comprising:
illuminating an object located at a first facet of a prism at an angle less than the critical angle;
illuminating the object at the first facet at an angle greater than the critical angle, wherein the light enters the prism through a second facet of the prism;
imaging light that is scattered by the object at an angle less than the critical angle relative to the first facet and totally internally reflected at the second facet using a first imaging system; and
imaging light from the first facet using a second imaging system, wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
66. The method of claim 65, wherein imaging light from the first facet using a second imaging system further comprises imaging light totally internally reflected from the first facet using the second imaging system.
67. The method of claim 65, wherein imaging light from the first facet using a second imaging system further comprises imaging light scattered by the object at object- facet interface using the second imaging system.
68. The method of claim 65, wherein illuminating the object located at the first facet of the prism at an angle less than the critical angle and illuminating the object at the first facet at the angle greater than the critical angle with light having distinct wavelengths.
69. The method of claim 65 further comprising generating a composite image of the object using the images created by a first imaging system and the second imaging system.
70. A system comprising:
a multifaceted prism having a first and second facet;
a first illumination source configured to illuminate the first facet at an angle less than the critical angle;
a second illumination source configured to illuminate the first facet at an angle greater than the critical angle, wherein light from the second illumination source enters the prism through the second facet;
a first imager configured to image light from the first facet that is totally internally reflected by the second facet; and
a second imager configured to image light from the first illumination source that is totally internally reflected from the first facet,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
71. A system comprising:
a multifaceted prism having a first and second facet; a first illumination source configured to illuminate the first facet at an angle less than the critical angle;
a second illumination source configured to illuminate the first facet at an angle greater than the critical angle, wherein light from the second illumination source enters the prism through the second facet;
a first imager configured to image light from the first facet that is totally internally reflected by the second facet; and
a second imager configured to image light scattered from the interface of the object and the first facet,
wherein the critical angle is defined by an interface of the first facet with an external environment and is measured relative to the normal of the first facet.
72. The system according to claim 71, wherein the second imager images light at an elevation angle less than the critical angle.
73. The system according to claim 71, wherein the second imager images light at an elevation angle greater than the critical angle.
74. The system according to claim 71, wherein the second imager images light at an azimuth angle out of phase with the second illumination source.
PCT/US2010/046852 2009-08-26 2010-08-26 Multiplexed biometric imaging and dual-imager biometric sensor WO2011028620A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
CN2010800482377A CN102576408A (en) 2009-08-26 2010-08-26 Multiplexed biometric imaging and dual-imager biometric sensor
EP10752477A EP2471023A1 (en) 2009-08-26 2010-08-26 Multiplexed biometric imaging and dual-imager biometric sensor
DE112010003414T DE112010003414T5 (en) 2009-08-26 2010-08-26 Biometric multiplex imaging and biometric dual imager sensor
BR122013021647A BR122013021647A2 (en) 2009-08-26 2010-08-26 object locating, object discrimination and background methods, and
BR112012004177A BR112012004177A2 (en) 2009-08-26 2010-08-26 biometric method and system, system, method, object location, object discrimination, and background methods, and multifaceted prism
IL218238A IL218238A0 (en) 2009-08-26 2012-02-21 Multiplexed biometric imaging and dual-imager biometric sensor

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US23718909P 2009-08-26 2009-08-26
US61/237,189 2009-08-26

Publications (1)

Publication Number Publication Date
WO2011028620A1 true WO2011028620A1 (en) 2011-03-10

Family

ID=42813510

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/046852 WO2011028620A1 (en) 2009-08-26 2010-08-26 Multiplexed biometric imaging and dual-imager biometric sensor

Country Status (7)

Country Link
US (2) US8731250B2 (en)
EP (1) EP2471023A1 (en)
CN (1) CN102576408A (en)
BR (2) BR122013021647A2 (en)
DE (1) DE112010003414T5 (en)
IL (1) IL218238A0 (en)
WO (1) WO2011028620A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3019264A1 (en) * 2014-03-31 2015-10-02 Morpho LIGHTING OPTICS
EP2639675B1 (en) * 2012-03-15 2017-12-13 Geoffrey Lee Wen-Chieh High resolution and high sensitivity optically activated motion detection device using multiple color light sources
EP3690830A1 (en) * 2013-04-09 2020-08-05 Bundesdruckerei GmbH Imaging system for capturing a feature of an identification document
US10845893B2 (en) 2013-06-04 2020-11-24 Wen-Chieh Geoffrey Lee High resolution and high sensitivity three-dimensional (3D) cursor maneuvering device
US11216150B2 (en) 2019-06-28 2022-01-04 Wen-Chieh Geoffrey Lee Pervasive 3D graphical user interface with vector field functionality
US11307730B2 (en) 2018-10-19 2022-04-19 Wen-Chieh Geoffrey Lee Pervasive 3D graphical user interface configured for machine learning
GB2602793A (en) * 2021-01-06 2022-07-20 Foster & Freeman Ltd Improved coaxial examination of surfaces

Families Citing this family (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7460696B2 (en) 2004-06-01 2008-12-02 Lumidigm, Inc. Multispectral imaging biometrics
US7751594B2 (en) 2003-04-04 2010-07-06 Lumidigm, Inc. White-light spectral biometric sensors
US8229185B2 (en) 2004-06-01 2012-07-24 Lumidigm, Inc. Hygienic biometric sensors
US8787630B2 (en) 2004-08-11 2014-07-22 Lumidigm, Inc. Multispectral barcode imaging
US7995808B2 (en) 2006-07-19 2011-08-09 Lumidigm, Inc. Contactless multispectral biometric capture
US8175346B2 (en) 2006-07-19 2012-05-08 Lumidigm, Inc. Whole-hand multispectral biometric imaging
US8355545B2 (en) 2007-04-10 2013-01-15 Lumidigm, Inc. Biometric detection using spatial, temporal, and/or spectral techniques
WO2008134135A2 (en) 2007-03-21 2008-11-06 Lumidigm, Inc. Biometrics based on locally consistent features
US20100246902A1 (en) * 2009-02-26 2010-09-30 Lumidigm, Inc. Method and apparatus to combine biometric sensing and other functionality
EP2471023A1 (en) 2009-08-26 2012-07-04 Lumidigm, Inc. Multiplexed biometric imaging and dual-imager biometric sensor
US8570149B2 (en) 2010-03-16 2013-10-29 Lumidigm, Inc. Biometric imaging using an optical adaptive interface
CN103339642A (en) 2010-09-20 2013-10-02 鲁米迪格姆有限公司 Machine-readable symbols
US9483677B2 (en) 2010-09-20 2016-11-01 Hid Global Corporation Machine-readable symbols
EP2639760A4 (en) * 2010-11-08 2017-12-13 Nec Corporation Biometric authentication apparatus and biometric authentication method
US9372118B1 (en) * 2011-03-07 2016-06-21 Fluxdata, Inc. Apparatus and method for multispectral based detection
WO2012133110A1 (en) 2011-03-25 2012-10-04 日本電気株式会社 Authentication device and authentication method
US9213438B2 (en) * 2011-06-02 2015-12-15 Omnivision Technologies, Inc. Optical touchpad for touch and gesture recognition
US9292916B2 (en) * 2011-08-09 2016-03-22 Hid Global Corporation Methods and systems for estimating genetic characteristics from biometric measurements
WO2013040448A1 (en) 2011-09-16 2013-03-21 Life Technologies Corporation Simultaneous acquisition of biometric data and nucleic acid
WO2013044154A1 (en) 2011-09-23 2013-03-28 Life Technologies Corporation Simultaneous aquisition of biometric data and nucleic acid
US8553235B1 (en) * 2012-01-18 2013-10-08 Wen-Chieh Geoffrey Lee High resolution and high sensitivity optically activated touch sensing device using multiple color light sources
US9719130B2 (en) 2012-02-22 2017-08-01 Life Technologies Corporation Sample collection devices, kits and methods of use
DE102012102983A1 (en) * 2012-04-05 2013-10-10 Carl Zeiss Microscopy Gmbh Method and apparatus for determining a critical angle of an excitation light beam
US20160034772A1 (en) * 2013-03-15 2016-02-04 Ellis I. Betensky Method and apparatus for acquiring biometric image
US10599932B2 (en) 2014-06-09 2020-03-24 Lawrence Livermore National Security, Llc Personal electronic device for performing multimodal imaging for non-contact identification of multiple biometric traits
US9659205B2 (en) * 2014-06-09 2017-05-23 Lawrence Livermore National Security, Llc Multimodal imaging system and method for non-contact identification of multiple biometric traits
US10527497B2 (en) * 2015-05-22 2020-01-07 Irnova Ab Infrared imaging detector
US10303916B2 (en) * 2015-07-31 2019-05-28 Fotonation Limited Image processing apparatus
IL240872A (en) * 2015-08-27 2016-11-30 Elbit Systems Land & C4I Ltd System and method for object authenticity detection
KR20180048823A (en) 2015-08-28 2018-05-10 위르겐 막스 Method and device for sensing the surface structure and composition of a sample
US10984363B2 (en) 2015-09-04 2021-04-20 International Business Machines Corporation Summarization of a recording for quality control
CN108475324B (en) 2015-09-09 2022-03-25 格马尔托股份有限公司 Non-contact friction ridge capturing device
WO2017044343A1 (en) * 2015-09-09 2017-03-16 3M Innovative Properties Company Non-contact friction ridge capture device
US10235556B2 (en) * 2015-12-13 2019-03-19 Joshua Noel Hogan Frustrated total internal reflection fingerprint detector
US10616218B2 (en) * 2016-10-03 2020-04-07 Gentex Corporation Driver identification and authentication systems and methods
FR3072199B1 (en) * 2017-10-11 2020-12-25 Idemia Identity & Security France METHOD OF DETECTION OF THE PRESENCE OF A BODY PART WITH A FINGERPRINT ON A FINGERPRINT SENSOR
EP3578126B1 (en) * 2018-06-08 2023-02-22 Stryker European Operations Holdings LLC Surgical navigation system
KR20230053417A (en) * 2021-10-14 2023-04-21 삼성전자주식회사 Apparatus and method for acquiring image and electronic device including the same

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050271258A1 (en) 2004-06-01 2005-12-08 Lumidigm, Inc. Multispectral imaging biometrics
WO2006093508A2 (en) * 2004-06-01 2006-09-08 Lumidigm, Inc. Multispectral imaging biometrics
US20070014437A1 (en) * 2004-09-02 2007-01-18 Sony Corporation Information processing device

Family Cites Families (329)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3508830A (en) 1967-11-13 1970-04-28 Shell Oil Co Apparatus for light scattering measurements
US3619060A (en) 1968-11-19 1971-11-09 Joseph E Johnson Identification device
US4035083A (en) 1972-05-30 1977-07-12 Woodriff Ray A Background correction in spectro-chemical analysis
US3854319A (en) 1972-10-16 1974-12-17 Borg Warner Alcoholic breath simulator
US3872443A (en) * 1973-01-26 1975-03-18 Novar Electronics Corp Individual identification apparatus and method using frequency response
USRE29008E (en) 1973-01-26 1976-10-19 Novar Electronics Corporation Individual identification apparatus and method using frequency response
US3910701A (en) 1973-07-30 1975-10-07 George R Henderson Method and apparatus for measuring light reflectance absorption and or transmission
DE2606991A1 (en) 1976-02-20 1977-08-25 Nils Dr Med Kaiser DEVICE FOR DETERMINING THE CONTENT OF METABOLIC PRODUCTS IN THE BLOOD
US4142797A (en) * 1977-02-11 1979-03-06 Barnes Engineering Company Common path interferometer
US4322163A (en) * 1977-10-25 1982-03-30 Fingermatrix Inc. Finger identification
US4170987A (en) 1977-11-28 1979-10-16 California Institute Of Technology Medical diagnosis system and method with multispectral imaging
US4260220A (en) * 1979-06-15 1981-04-07 Canadian Patents And Development Limited Prism light guide having surfaces which are in octature
DE2934190A1 (en) * 1979-08-23 1981-03-19 Müller, Gerhard, Prof. Dr.-Ing., 7080 Aalen METHOD AND DEVICE FOR MOLECULAR SPECTROSCOPY, ESPECIALLY FOR DETERMINING METABOLISM PRODUCTS
DE3215879A1 (en) 1982-04-29 1983-11-03 Fa. Carl Zeiss, 7920 Heidenheim DEVICE FOR SPECTRUM MEASUREMENT IN THE BLOOD RAIL
DE3378551D1 (en) 1982-06-25 1988-12-29 Oskar Oehler Light collector device and utilization thereof for spectroscopy
US4654530A (en) 1983-10-31 1987-03-31 Dybwad Jens P Refractively scanned interferometer
US4537484A (en) 1984-01-30 1985-08-27 Identix Incorporated Fingerprint imaging apparatus
US4699149A (en) 1984-03-20 1987-10-13 Joseph Rice Apparatus for the identification of individuals
DE3587083T2 (en) 1984-04-18 1993-06-03 Nec Corp IDENTIFICATION SYSTEM BY TESTING FINGERPRINTS.
GB2163548B (en) 1984-08-09 1987-11-25 Perkin Elmer Ltd Interferometric apparatus particularly for use in ft spectrophotometer
JPS61182174A (en) 1985-02-07 1986-08-14 Nippon Fuaamu Uea Kk Personal identifying device
US4661706A (en) * 1985-02-25 1987-04-28 Spectra-Tech Inc. Blocker device for eliminating specular reflectance from a diffuse reflection spectrum
US4653880A (en) 1985-03-01 1987-03-31 Spectra-Tech Inc. Reflective beam splitting objective
US4655225A (en) * 1985-04-18 1987-04-07 Kurabo Industries Ltd. Spectrophotometric method and apparatus for the non-invasive
US4747147A (en) 1985-09-03 1988-05-24 Sparrow Malcolm K Fingerprint recognition and retrieval system
US4656562A (en) * 1985-09-13 1987-04-07 Santa Barbara Research Center Optical integrator means for intensity modification of Gaussian beam
US4730882A (en) 1986-02-10 1988-03-15 Spectra-Tech, Inc. Multiple internal reflectance spectroscopy system
US4712912A (en) 1986-03-10 1987-12-15 Spectra-Tech, Inc. Spectrophotometric image scrambler for full aperture microspectroscopy
US4866644A (en) 1986-08-29 1989-09-12 Shenk John S Optical instrument calibration system
US4810875A (en) 1987-02-02 1989-03-07 Wyatt Technology Corporation Method and apparatus for examining the interior of semi-opaque objects
JPS63252239A (en) 1987-04-09 1988-10-19 Sumitomo Electric Ind Ltd Reflection type oxymeter
EP0290013B1 (en) 1987-05-06 1996-03-13 Fuji Photo Film Co., Ltd. Densitometer and its use
US4787708A (en) 1987-05-08 1988-11-29 Tir Systems Ltd. Apparatus for continuously controlled emission of light from prism light guide
US4853542A (en) 1987-06-08 1989-08-01 Nicolas J. Harrick Collecting hemispherical attachment for spectrophotometry
US4857735A (en) 1987-10-23 1989-08-15 Noller Hans G Light emitting diode spectrophotometer
JPH0827235B2 (en) 1987-11-17 1996-03-21 倉敷紡績株式会社 Spectroscopic method for measuring sugar concentration
DK163194C (en) 1988-12-22 1992-06-22 Radiometer As METHOD OF PHOTOMETRIC IN VITRO DETERMINING A BLOOD GAS PARAMETER IN A BLOOD TEST
US4787013A (en) 1987-11-30 1988-11-22 Santa Barbara Research Center Intermediate range intensity modification of gaussian beam using optical integration means
US4882492A (en) 1988-01-19 1989-11-21 Biotronics Associates, Inc. Non-invasive near infrared measurement of blood analyte concentrations
US4830496A (en) 1988-01-22 1989-05-16 General Scanning, Inc. Interferometer
US4859064A (en) 1988-05-09 1989-08-22 Spectra-Tech, Inc. Diffuse reflectance spectroscopy system and method
US5361758A (en) 1988-06-09 1994-11-08 Cme Telemetrix Inc. Method and device for measuring concentration levels of blood constituents non-invasively
US5055658A (en) 1988-07-25 1991-10-08 Cockburn John B Security system employing digitized personal physical characteristics
DE68919394T2 (en) 1988-09-16 1995-03-30 Fujitsu Ltd System for detecting a biological object and fingerprint comparison system using this system.
US5402778A (en) 1993-01-19 1995-04-04 Nim Incorporated Spectrophotometric examination of tissue of small dimension
CA2003131C (en) 1988-11-25 1998-06-23 Seigo Igaki Biological object detection apparatus
CA2004457A1 (en) 1988-12-06 1990-06-06 Seigo Igaki Minutia data extraction in fingerprint identification
US5353799A (en) 1991-01-22 1994-10-11 Non Invasive Technology, Inc. Examination of subjects using photon migration with high directionality techniques
US5782755A (en) 1993-11-15 1998-07-21 Non-Invasive Technology, Inc. Monitoring one or more solutes in a biological system using optical techniques
US5237178A (en) 1990-06-27 1993-08-17 Rosenthal Robert D Non-invasive near-infrared quantitative measurement instrument
US5068536A (en) 1989-01-19 1991-11-26 Futrex, Inc. Method for providing custom calibration for near infrared instruments for measurement of blood glucose
US5028787A (en) 1989-01-19 1991-07-02 Futrex, Inc. Non-invasive measurement of blood glucose
US6066847A (en) 1989-01-19 2000-05-23 Futrex Inc. Procedure for verifying the accuracy of non-invasive blood glucose measurement instruments
US5204532A (en) 1989-01-19 1993-04-20 Futrex, Inc. Method for providing general calibration for near infrared instruments for measurement of blood glucose
US4936680A (en) 1989-04-03 1990-06-26 General Electric Company Method of, and apparatus for, edge enhancement of fingerprint minutia
US5178142A (en) * 1989-05-23 1993-01-12 Vivascan Corporation Electromagnetic method and apparatus to measure constituents of human or animal tissue
US5258922A (en) 1989-06-12 1993-11-02 Wieslaw Bicz Process and device for determining of surface structures
US4975581A (en) 1989-06-21 1990-12-04 University Of New Mexico Method of and apparatus for determining the similarity of a biological analyte from a model constructed from known biological fluids
CA2025330C (en) 1989-09-18 2002-01-22 David W. Osten Characterizing biological matter in a dynamic condition using near infrared spectroscopy
CA2028261C (en) 1989-10-28 1995-01-17 Won Suck Yang Non-invasive method and apparatus for measuring blood glucose concentration
US5070874A (en) 1990-01-30 1991-12-10 Biocontrol Technology, Inc. Non-invasive determination of glucose concentration in body of patients
US5222496A (en) 1990-02-02 1993-06-29 Angiomedics Ii, Inc. Infrared glucose sensor
US5222495A (en) 1990-02-02 1993-06-29 Angiomedics Ii, Inc. Non-invasive blood analysis by near infrared absorption measurements using two closely spaced wavelengths
US5146102A (en) 1990-02-22 1992-09-08 Kabushiki Kaisha Toshiba Fingerprint image input apparatus including a cylindrical lens
US5051602A (en) 1990-03-02 1991-09-24 Spectra-Tech, Inc. Optical system and method for sample analyzation
US5015100A (en) 1990-03-02 1991-05-14 Axiom Analytical, Inc. Apparatus and method for normal incidence reflectance spectroscopy
US5019715A (en) 1990-03-02 1991-05-28 Spectra-Tech, Inc. Optical system and method for sample analyzation
US5177802A (en) 1990-03-07 1993-01-05 Sharp Kabushiki Kaisha Fingerprint input apparatus
US5115133A (en) * 1990-04-19 1992-05-19 Inomet, Inc. Testing of body fluid constituents through measuring light reflected from tympanic membrane
GB2243211A (en) 1990-04-20 1991-10-23 Philips Electronic Associated Analytical instrument and method of calibrating an analytical instrument
US5419321A (en) 1990-05-17 1995-05-30 Johnson & Johnson Professional Products Limited Non-invasive medical sensor
IE77034B1 (en) 1990-06-27 1997-11-19 Futrex Inc Non-invasive masurement of blood glucose
US5158082A (en) 1990-08-23 1992-10-27 Spacelabs, Inc. Apparatus for heating tissue with a photoplethysmograph sensor
US5351686A (en) 1990-10-06 1994-10-04 In-Line Diagnostics Corporation Disposable extracorporeal conduit for blood constituent monitoring
US5459677A (en) 1990-10-09 1995-10-17 Board Of Regents Of The University Of Washington Calibration transfer for analytical instruments
GB9027480D0 (en) 1990-12-19 1991-02-06 Philips Electronic Associated Interferometer
US5243546A (en) 1991-01-10 1993-09-07 Ashland Oil, Inc. Spectroscopic instrument calibration
US5230702A (en) 1991-01-16 1993-07-27 Paradigm Biotechnologies Partnership Hemodialysis method
US6198532B1 (en) 1991-02-22 2001-03-06 Applied Spectral Imaging Ltd. Spectral bio-imaging of the eye
US5303026A (en) 1991-02-26 1994-04-12 The Regents Of The University Of California Los Alamos National Laboratory Apparatus and method for spectroscopic analysis of scattering media
US5163094A (en) 1991-03-20 1992-11-10 Francine J. Prokoski Method for identifying individuals from analysis of elemental shapes derived from biosensor data
US5638818A (en) 1991-03-21 1997-06-17 Masimo Corporation Low noise optical probe
GB9106672D0 (en) 1991-03-28 1991-05-15 Abbey Biosystems Ltd Method and apparatus for glucose concentration monitoring
US5441053A (en) 1991-05-03 1995-08-15 University Of Kentucky Research Foundation Apparatus and method for multiple wavelength of tissue
EP0522913B1 (en) 1991-06-28 1996-12-27 U.S. Divers Company, Inc. Regulator with improved high pressure seat
DE69227545T2 (en) * 1991-07-12 1999-04-29 Mark R Robinson Oximeter for the reliable clinical determination of blood oxygen saturation in a fetus
US5291560A (en) * 1991-07-15 1994-03-01 Iri Scan Incorporated Biometric personal identification system based on iris analysis
US5268749A (en) 1991-07-26 1993-12-07 Kollmorgen Corporation Apparatus and method for providing uniform illumination of a sample plane
EP0527703B1 (en) 1991-08-12 1995-06-28 AVL Medical Instruments AG Device for measuring at least one gaseous concentration level in particular the oxygen concentration level in blood
US5223715A (en) 1991-09-20 1993-06-29 Amoco Corporation Process for spectrophotometric analysis
JPH07508426A (en) 1991-10-17 1995-09-21 サイエンティフィック ジェネリクス リミテッド Blood sample measuring device and method
US5311021A (en) 1991-11-13 1994-05-10 Connecticut Instrument Corp. Spectroscopic sampling accessory having dual measuring and viewing systems
US5225678A (en) 1991-11-13 1993-07-06 Connecticut Instrument Corporation Spectoscopic sampling accessory having dual measuring and viewing systems
US5681273A (en) 1991-12-23 1997-10-28 Baxter International Inc. Systems and methods for predicting blood processing parameters
US5413098A (en) 1991-12-24 1995-05-09 Sextant Medical Corporation Path constrained spectrophotometer and method for determination of spatial distribution of light or other radiation scattering and absorbing substances in a radiation scattering medium
WO1993012712A1 (en) 1991-12-31 1993-07-08 Vivascan Corporation Blood constituent determination based on differential spectral analysis
US5335288A (en) 1992-02-10 1994-08-02 Faulkner Keith W Apparatus and method for biometric identification
US5331958A (en) 1992-03-31 1994-07-26 University Of Manitoba Spectrophotometric blood analysis
US5853370A (en) 1996-09-13 1998-12-29 Non-Invasive Technology, Inc. Optical system and method for non-invasive imaging of biological tissue
US5257086A (en) 1992-06-09 1993-10-26 D.O.M. Associates Int'l Optical spectrophotometer having a multi-element light source
US5355880A (en) 1992-07-06 1994-10-18 Sandia Corporation Reliable noninvasive measurement of blood gases
US5792050A (en) 1992-07-06 1998-08-11 Alam; Mary K. Near-infrared noninvasive spectroscopic determination of pH
US5321265A (en) 1992-07-15 1994-06-14 Block Myron J Non-invasive testing
US5672875A (en) 1992-07-15 1997-09-30 Optix Lp Methods of minimizing scattering and improving tissue sampling in non-invasive testing and imaging
US5818048A (en) 1992-07-15 1998-10-06 Optix Lp Rapid non-invasive optical analysis using broad bandpass spectral processing
US5452723A (en) 1992-07-24 1995-09-26 Massachusetts Institute Of Technology Calibrated spectrographic imaging
US5348003A (en) 1992-09-03 1994-09-20 Sirraya, Inc. Method and apparatus for chemical analysis
US6172743B1 (en) 1992-10-07 2001-01-09 Chemtrix, Inc. Technique for measuring a blood analyte by non-invasive spectrometry in living tissue
MX9306336A (en) 1992-10-13 1995-01-31 Baxter Int METHOD AND APPARATUS FOR MONITORING OR MONITORING A HEMODIALYSIS TREATMENT, WHICH IS CARRIED OUT IN A HEMODIALISIS MACHINE.
US6315772B1 (en) 1993-09-24 2001-11-13 Transmedica International, Inc. Laser assisted pharmaceutical delivery and fluid removal
US5379764A (en) 1992-12-09 1995-01-10 Diasense, Inc. Non-invasive determination of analyte concentration in body of mammals
US5360004A (en) 1992-12-09 1994-11-01 Diasense, Inc. Non-invasive determination of analyte concentration using non-continuous radiation
US5559504A (en) 1993-01-08 1996-09-24 Kabushiki Kaisha Toshiba Surface shape sensor, identification device using this sensor, and protected system using this device
US5313941A (en) 1993-01-28 1994-05-24 Braig James R Noninvasive pulsed infrared spectrophotometer
US5515847A (en) 1993-01-28 1996-05-14 Optiscan, Inc. Self-emission noninvasive infrared spectrophotometer
US5987346A (en) 1993-02-26 1999-11-16 Benaron; David A. Device and method for classification of tissue
US5301375A (en) * 1993-03-09 1994-04-12 Osmond John S Primer tank-checkvalve syphon toilet flushing apparatus
US5483335A (en) * 1993-03-18 1996-01-09 Tobias; Reginald Multiplex spectroscopy
US5460177A (en) 1993-05-07 1995-10-24 Diasense, Inc. Method for non-invasive measurement of concentration of analytes in blood using continuous spectrum radiation
US5596992A (en) 1993-06-30 1997-01-28 Sandia Corporation Multivariate classification of infrared spectra of cell and tissue samples
US5308315A (en) * 1993-07-27 1994-05-03 Raja N. Khuri Method for determining the adequacy of dialysis
US5435309A (en) * 1993-08-10 1995-07-25 Thomas; Edward V. Systematic wavelength selection for improved multivariate spectral analysis
EP0670143B1 (en) 1993-08-12 2003-05-28 Kurashiki Boseki Kabushiki Kaisha Non invasive method and instrument for measuring blood sugar level
EP0683641A4 (en) 1993-08-24 1998-07-15 Mark R Robinson A robust accurate non-invasive analyte monitor.
US5459317A (en) 1994-02-14 1995-10-17 Ohio University Method and apparatus for non-invasive detection of physiological chemicals, particularly glucose
US5505726A (en) * 1994-03-21 1996-04-09 Dusa Pharmaceuticals, Inc. Article of manufacture for the photodynamic therapy of dermal lesion
US5568251A (en) 1994-03-23 1996-10-22 National Research Council Of Canada Authenticating system
JPH09510636A (en) 1994-03-24 1997-10-28 ミネソタ マイニング アンド マニュファクチャリング カンパニー Biometric personal identification system
GB9409064D0 (en) 1994-05-06 1994-06-22 Perkin Elmer Ltd Improvements in or relating to optical interferometers
US5507723A (en) 1994-05-24 1996-04-16 Baxter International, Inc. Method and system for optimizing dialysis clearance
US5523054A (en) 1995-01-31 1996-06-04 Johnson & Johnson Clinical Diagnostics, Inc. Test element for quantitative NIR spectroscopic analysis
JP3261264B2 (en) 1994-07-13 2002-02-25 株式会社堀場製作所 Multicomponent aqueous solution analysis method and analyzer
US5539207A (en) 1994-07-19 1996-07-23 National Research Council Of Canada Method of identifying tissue
US5572596A (en) 1994-09-02 1996-11-05 David Sarnoff Research Center, Inc. Automated, non-invasive iris recognition system and method
CA2200455A1 (en) 1994-09-20 1996-03-28 Louis R. Piloco Apparatus for illumination stabilization and homogenization
US5613014A (en) * 1994-10-12 1997-03-18 Martin Marietta Corp. Fingerprint matching system
NZ300915A (en) 1995-02-09 1998-12-23 Foss Electric As Method for standardizing a spectrometer generating an optical spectrum from a sample
FR2734360B1 (en) * 1995-05-19 1997-07-04 Elf Antar France METHOD OF CORRECTING A SIGNAL DELIVERED BY A MEASURING INSTRUMENT
US5743262A (en) 1995-06-07 1998-04-28 Masimo Corporation Blood glucose monitoring system
US5761330A (en) 1995-06-07 1998-06-02 Mytec Technologies, Inc. Hybrid optical-digital method and apparatus for fingerprint verification
US5724268A (en) * 1995-06-29 1998-03-03 Chiron Diagnostics Corporation Apparatus and methods for the analytical determination of sample component concentrations that account for experimental error
SG38866A1 (en) 1995-07-31 1997-04-17 Instrumentation Metrics Inc Liquid correlation spectrometry
US5606164A (en) 1996-01-16 1997-02-25 Boehringer Mannheim Corporation Method and apparatus for biological fluid analyte concentration measurement using generalized distance outlier detection
CA2179338C (en) 1995-08-07 2000-04-25 Gordon Albert Thomas Apparatus and method for spectroscopic product recognition and identification
JP3579686B2 (en) 1995-08-07 2004-10-20 アークレイ株式会社 Measuring position reproducing method, measuring position reproducing device, and optical measuring device using the same
US5729619A (en) * 1995-08-08 1998-03-17 Northrop Grumman Corporation Operator identity, intoxication and drowsiness monitoring system and method
US5655530A (en) 1995-08-09 1997-08-12 Rio Grande Medical Technologies, Inc. Method for non-invasive blood analyte measurement with improved optical interface
US5636633A (en) 1995-08-09 1997-06-10 Rio Grande Medical Technologies, Inc. Diffuse reflectance monitoring apparatus
US6240306B1 (en) 1995-08-09 2001-05-29 Rio Grande Medical Technologies, Inc. Method and apparatus for non-invasive blood analyte measurement with fluid compartment equilibration
US6152876A (en) 1997-04-18 2000-11-28 Rio Grande Medical Technologies, Inc. Method for non-invasive blood analyte measurement with improved optical interface
US6212424B1 (en) 1998-10-29 2001-04-03 Rio Grande Medical Technologies, Inc. Apparatus and method for determination of the adequacy of dialysis by non-invasive near-infrared spectroscopy
US5793881A (en) 1995-08-31 1998-08-11 Stiver; John A. Identification system
JPH0991434A (en) 1995-09-28 1997-04-04 Hamamatsu Photonics Kk Human body collation device
US5751835A (en) 1995-10-04 1998-05-12 Topping; Allen Method and apparatus for the automated identification of individuals by the nail beds of their fingernails
US6240309B1 (en) 1995-10-06 2001-05-29 Hitachi, Ltd. Optical measurement instrument for living body
US6025597A (en) * 1995-10-17 2000-02-15 Optiscan Biomedical Corporation Non-invasive infrared absorption spectrometer for measuring glucose or other constituents in a human or other body
US6041247A (en) 1995-11-29 2000-03-21 Instrumentarium Corp Non-invasive optical measuring sensor and measuring method
US5719399A (en) 1995-12-18 1998-02-17 The Research Foundation Of City College Of New York Imaging and characterization of tissue based upon the preservation of polarized light transmitted therethrough
US5929443A (en) 1995-12-18 1999-07-27 The Research Foundation City College Of New York Imaging of objects based upon the polarization or depolarization of light
US6045502A (en) 1996-01-17 2000-04-04 Spectrx, Inc. Analyzing system with disposable calibration device
US6226541B1 (en) 1996-01-17 2001-05-01 Spectrx, Inc. Apparatus and method for calibrating measurement systems
US5860421A (en) * 1996-01-17 1999-01-19 Spectrx, Inc. Apparatus and method for calibrating measurement systems
US6040578A (en) * 1996-02-02 2000-03-21 Instrumentation Metrics, Inc. Method and apparatus for multi-spectral analysis of organic blood analytes in noninvasive infrared spectroscopy
US5747806A (en) 1996-02-02 1998-05-05 Instrumentation Metrics, Inc Method and apparatus for multi-spectral analysis in noninvasive nir spectroscopy
US5859420A (en) 1996-02-12 1999-01-12 Dew Engineering And Development Limited Optical imaging device
EP0955867A1 (en) 1996-02-23 1999-11-17 Diasense, Inc. Method and apparatus for non-invasive blood glucose sensing
US5672864A (en) 1996-02-26 1997-09-30 Eastman Kodak Company Light integrator
US6229908B1 (en) 1996-04-26 2001-05-08 Edmonds, Iii Dean Stockett Driver alcohol ignition interlock
WO1997041527A1 (en) 1996-05-01 1997-11-06 Xros, Inc. Compact, simple, 2d raster, image-building fingerprint scanner
US5796858A (en) 1996-05-10 1998-08-18 Digital Persona, Inc. Fingerprint sensing system using a sheet prism
JP3604231B2 (en) 1996-05-16 2004-12-22 富士写真フイルム株式会社 Method and apparatus for measuring glucose concentration
US5828066A (en) * 1996-07-02 1998-10-27 Messerschmidt; Robert G. Multisource infrared spectrometer
AU3596597A (en) * 1996-07-08 1998-02-02 Animas Corporation Implantable sensor and system for in vivo measurement and control of fluid constituent levels
US5978495A (en) 1996-07-17 1999-11-02 Intelnet Inc. Method and apparatus for accurate determination of the identity of human beings
US5963657A (en) 1996-09-09 1999-10-05 Arete Associates Economical skin-pattern-acquisition and analysis apparatus for access control; systems controlled thereby
US6148094A (en) 1996-09-30 2000-11-14 David J. Kinsella Pointing device with biometric sensor
DE69631530T2 (en) 1996-10-09 2004-07-08 Perkin-Elmer Ltd., Beaconsfield Interferogram digitization for Fourier transformation spectroscopy
US5737439A (en) * 1996-10-29 1998-04-07 Smarttouch, Llc. Anti-fraud biometric scanner that accurately detects blood flow
JP4212007B2 (en) * 1996-11-26 2009-01-21 パナソニック電工株式会社 Blood component concentration analyzer
US6122042A (en) 1997-02-07 2000-09-19 Wunderman; Irwin Devices and methods for optically identifying characteristics of material objects
US5902033A (en) 1997-02-18 1999-05-11 Torch Technologies Llc Projector system with hollow light pipe optics
WO1998037805A1 (en) 1997-02-26 1998-09-03 Diasense, Inc. Individual calibration of blood glucose for supporting noninvasive self-monitoring blood glucose
US6081612A (en) 1997-02-28 2000-06-27 Electro Optical Sciences Inc. Systems and methods for the multispectral imaging and characterization of skin tissue
US6159147A (en) 1997-02-28 2000-12-12 Qrs Diagnostics, Llc Personal computer card for collection of real-time biological data
AU740638B2 (en) * 1997-02-28 2001-11-08 Electro-Optical Sciences, Inc. Systems and methods for the multispectral imaging and characterization of skin tissue
US5850623A (en) 1997-03-14 1998-12-15 Eastman Chemical Company Method for standardizing raman spectrometers to obtain stable and transferable calibrations
JP2001515596A (en) 1997-03-14 2001-09-18 ローズマウント アナリティカル インコーポレイテッド Improved low noise Raman analysis system
US5792053A (en) 1997-03-17 1998-08-11 Polartechnics, Limited Hybrid probe for tissue type recognition
TW352335B (en) 1997-03-25 1999-02-11 Matsushita Electric Works Ltd Method of determining a glucose concentration in a target by using near-infrared spectroscopy
KR100259475B1 (en) 1997-04-14 2000-06-15 최환수 Method for the identification of individuals using the pattern of blood vessels
US6008889A (en) 1997-04-16 1999-12-28 Zeng; Haishan Spectrometer system for diagnosis of skin disease
US6193153B1 (en) * 1997-04-16 2001-02-27 Francis Lambert Method and apparatus for non-intrusive biometric capture
US6125192A (en) 1997-04-21 2000-09-26 Digital Persona, Inc. Fingerprint recognition system
US6031609A (en) * 1997-05-29 2000-02-29 The Regents Of The University Of California Fourier transform spectrometer using a multielement liquid crystal display
US6560352B2 (en) 1999-10-08 2003-05-06 Lumidigm, Inc. Apparatus and method of biometric identification or verification of individuals using optical spectroscopy
US7890158B2 (en) * 2001-06-05 2011-02-15 Lumidigm, Inc. Apparatus and method of biometric determination using specialized optical spectroscopy systems
US6628809B1 (en) 1999-10-08 2003-09-30 Lumidigm, Inc. Apparatus and method for identification of individuals by near-infrared spectrum
FR2761180B1 (en) 1997-08-07 1999-05-07 Sagem SECURE FINGERPRINT READER
US6246751B1 (en) 1997-08-11 2001-06-12 International Business Machines Corporation Apparatus and methods for user identification to deny access or service to unauthorized users
US6115673A (en) 1997-08-14 2000-09-05 Instrumentation Metrics, Inc. Method and apparatus for generating basis sets for use in spectroscopic analysis
FI973454A (en) 1997-08-22 1999-02-23 Instrumentarium Oy A resilient device in a measuring sensor for observing the properties of living tissue
GB2329015B (en) 1997-09-05 2002-02-13 Samsung Electronics Co Ltd Method and device for noninvasive measurement of concentrations of blood components
US6043492A (en) * 1997-10-27 2000-03-28 Industrial Technology Research Institute Non-invasive blood glucose meter
US6937885B1 (en) 1997-10-30 2005-08-30 Hypermed, Inc. Multispectral/hyperspectral medical instrument
US6141101A (en) 1997-11-12 2000-10-31 Plx, Inc. Monolithic optical assembly
US6044285A (en) 1997-11-12 2000-03-28 Lightouch Medical, Inc. Method for non-invasive measurement of an analyte
US5949543A (en) 1997-11-12 1999-09-07 Plx, Inc. Monolithic optical assembly and associated retroreflector with beamsplitter assembly
US6028773A (en) 1997-11-14 2000-02-22 Stmicroelectronics, Inc. Packaging for silicon sensors
US6122737A (en) 1997-11-14 2000-09-19 Digital Persona, Inc. Method for using fingerprints to distribute information over a network
WO1999027489A1 (en) 1997-11-20 1999-06-03 Quo Technologies, L.L.C. Method and system for biometric recognition using unique internal distinguishing characteristics
US6070093A (en) 1997-12-02 2000-05-30 Abbott Laboratories Multiplex sensor and method of use
US6100811A (en) 1997-12-22 2000-08-08 Trw Inc. Fingerprint actuation of customized vehicle features
US6041410A (en) 1997-12-22 2000-03-21 Trw Inc. Personal identification fob
US6640124B2 (en) 1998-01-30 2003-10-28 The Schepens Eye Research Institute Imaging apparatus and methods for near simultaneous observation of directly scattered light and multiply scattered light
US6006119A (en) 1998-02-04 1999-12-21 Polestar Technologies, Inc. Non-invasive optical measurement of blood hematocrit
JP2002501803A (en) 1998-02-05 2002-01-22 イン−ラインダイアグノスティックスコーポレイション Non-invasive blood component monitoring method and apparatus
US6181414B1 (en) 1998-02-06 2001-01-30 Morphometrix Technologies Inc Infrared spectroscopy for medical imaging
US6201608B1 (en) 1998-03-13 2001-03-13 Optical Biopsy Technologies, Inc. Method and apparatus for measuring optical reflectivity and imaging through a scattering medium
US6539101B1 (en) 1998-04-07 2003-03-25 Gerald R. Black Method for identity verification
DE19818229A1 (en) 1998-04-24 1999-10-28 Hauke Rudolf Contactless method for hand- and fingerprint recognition
US6241663B1 (en) 1998-05-18 2001-06-05 Abbott Laboratories Method for improving non-invasive determination of the concentration of analytes in a biological sample
DE19824354A1 (en) 1998-05-30 1999-12-02 Philips Patentverwaltung Device for verifying signals
US6324310B1 (en) 1998-06-02 2001-11-27 Digital Persona, Inc. Method and apparatus for scanning a fingerprint using a linear sensor
DE19830830C2 (en) 1998-07-09 2000-11-23 Siemens Ag Process for the live detection of human skin
US6188781B1 (en) * 1998-07-28 2001-02-13 Digital Persona, Inc. Method and apparatus for illuminating a fingerprint through side illumination of a platen
US6057925A (en) 1998-08-28 2000-05-02 Optical Coating Laboratory, Inc. Compact spectrometer device
US6005722A (en) 1998-09-04 1999-12-21 Hewlett-Packard Company Optical display system including a light valve
WO2000019383A2 (en) 1998-09-11 2000-04-06 Loquitor Technologies Llc Generation and detection of induced current using acoustic energy
US6157041A (en) 1998-10-13 2000-12-05 Rio Grande Medical Technologies, Inc. Methods and apparatus for tailoring spectroscopic calibration models
US6723267B2 (en) * 1998-10-28 2004-04-20 Dsm N.V. Process of making highly oriented polyolefin fiber
US6438396B1 (en) 1998-11-05 2002-08-20 Cytometrics, Inc. Method and apparatus for providing high contrast imaging
US6353226B1 (en) 1998-11-23 2002-03-05 Abbott Laboratories Non-invasive sensor capable of determining optical parameters in a sample having multiple layers
US6631199B1 (en) 1998-12-08 2003-10-07 Allen W. L. Topping Automated identification through analysis of optical birefringence within nail beds
US6154658A (en) 1998-12-14 2000-11-28 Lockheed Martin Corporation Vehicle information and safety control system
US6175407B1 (en) 1998-12-17 2001-01-16 Identix Incorporated Apparatus and method for optically imaging features on the surface of a hand
US6280381B1 (en) 1999-07-22 2001-08-28 Instrumentation Metrics, Inc. Intelligent system for noninvasive blood analyte prediction
US6493566B1 (en) 1999-01-22 2002-12-10 Instrumentation Metrics, Inc. Classification system for sex determination and tissue characterization
US6501982B1 (en) 1999-01-22 2002-12-31 Sensys Medical, Inc. System for the noninvasive estimation of relative age
AU2987700A (en) 1999-02-08 2000-08-25 Amy S. Zelson Fingerprint analysis method
US6097035A (en) 1999-02-22 2000-08-01 Digital Persona, Inc. Fingerprint detection apparatus with partial fingerprint images
US6301815B1 (en) 1999-03-04 2001-10-16 Colt's Manufacturing Company, Inc. Firearms and docking station system for limiting use of firearm
US6046808A (en) 1999-04-09 2000-04-04 Three Lc, Inc. Radiation filter, spectrometer and imager using a micro-mirror array
JP2001033381A (en) 1999-07-26 2001-02-09 Shimadzu Corp Light application method and apparatus for measuring organism
JP2003508734A (en) 1999-08-31 2003-03-04 シーエムイー テレメトリックス インク. Apparatus for verifying the accuracy of a spectrum analyzer
WO2001018332A1 (en) 1999-09-06 2001-03-15 Siemens Aktiengesellschaft Activation of secured objects
US7047419B2 (en) 1999-09-17 2006-05-16 Pen-One Inc. Data security system
EP1233697A4 (en) * 1999-10-07 2005-06-22 Alexander K Mills Device and method for noninvasive continuous determination of physiologic characteristics
US6816605B2 (en) 1999-10-08 2004-11-09 Lumidigm, Inc. Methods and systems for biometric identification of individuals using linear optical spectroscopy
US6504614B1 (en) * 1999-10-08 2003-01-07 Rio Grande Medical Technologies, Inc. Interferometer spectrometer with reduced alignment sensitivity
US7054470B2 (en) 1999-12-02 2006-05-30 International Business Machines Corporation System and method for distortion characterization in fingerprint and palm-print image sequences and using this distortion as a behavioral biometrics
JP3966397B2 (en) 1999-12-27 2007-08-29 シャープ株式会社 Fingerprint detector
AU2781401A (en) 2000-01-10 2001-07-24 Tarian, Llc Device using histological and physiological biometric marker for authentication and activation
DE10001929A1 (en) 2000-01-19 2001-08-09 Skidata Ag Authorization control facility
US6292576B1 (en) * 2000-02-29 2001-09-18 Digital Persona, Inc. Method and apparatus for distinguishing a human finger from a reproduction of a fingerprint
WO2001069520A2 (en) 2000-03-10 2001-09-20 Ethentica, Inc. Biometric sensor
WO2001072216A2 (en) 2000-03-28 2001-10-04 Board Of Regents, The University Of Texas System Methods and apparatus for diagnositic multispectral digital imaging
US6799275B1 (en) 2000-03-30 2004-09-28 Digital Persona, Inc. Method and apparatus for securing a secure processor
KR100333138B1 (en) 2000-04-29 2002-04-19 안준영 Method of discriminating latent fingerprint in optical fingerprint input apparatus
US7536557B2 (en) 2001-03-22 2009-05-19 Ensign Holdings Method for biometric authentication through layering biometric traits
US6483929B1 (en) 2000-06-08 2002-11-19 Tarian Llc Method and apparatus for histological and physiological biometric operation and authentication
JP3807222B2 (en) 2000-10-30 2006-08-09 カシオ計算機株式会社 Image data reading device
IL140267A0 (en) 2000-12-13 2003-09-17 Milsys Ltd Dual processor trusted computing environment
EP1217573A1 (en) 2000-12-22 2002-06-26 Fingerpin AG Device for capturing finger papillary ridges
WO2002054337A1 (en) 2001-01-04 2002-07-11 Advanced Biometrics, Inc. Method and apparatus for subcutaneous identification
US7921297B2 (en) 2001-01-10 2011-04-05 Luis Melisendro Ortiz Random biometric authentication utilizing unique biometric signatures
US6606509B2 (en) 2001-03-16 2003-08-12 Nellcor Puritan Bennett Incorporated Method and apparatus for improving the accuracy of noninvasive hematocrit measurements
US8095193B2 (en) 2001-04-11 2012-01-10 Trutouch Technologies, Inc. Apparatus and method for controlling operation of vehicles or machinery by intoxicated or impaired individuals
US7126682B2 (en) 2001-04-11 2006-10-24 Rio Grande Medical Technologies, Inc. Encoded variable filter spectrometer
US6574490B2 (en) 2001-04-11 2003-06-03 Rio Grande Medical Technologies, Inc. System for non-invasive measurement of glucose in humans
DE10123330A1 (en) 2001-05-14 2002-11-28 Infineon Technologies Ag Detection of falsified fingerprints, e.g. a silicon casting of a fingerprint, using a dynamic, software-based method for detection of falsified fingerprints that is quick and efficient
JP2003006627A (en) 2001-06-18 2003-01-10 Nec Corp Fingerprint input device
US7303120B2 (en) 2001-07-10 2007-12-04 American Express Travel Related Services Company, Inc. System for biometric security using a FOB
US7272248B2 (en) * 2001-07-16 2007-09-18 Activcard Ireland Limited Biometric imaging device compensating for non-biometric parameters
AU2002355272A1 (en) 2001-07-25 2003-02-17 Argose, Inc. Adjunct quantitative system and method for non-invasive measurement of in vivo analytes
JP2003050993A (en) * 2001-08-06 2003-02-21 Omron Corp Method and device for reading fingerprint
JP2003075135A (en) 2001-08-31 2003-03-12 Nec Corp Fingerprint image input device and organism discrimination method by fingerprint image
KR20040038906A (en) 2001-10-02 2004-05-08 마츠시타 덴끼 산교 가부시키가이샤 Image sensing apparatus
DE10153808B4 (en) 2001-11-05 2010-04-15 Tst Biometrics Holding Ag Method for non-contact, optical generation of unrolled fingerprints and apparatus for carrying out the method
EP1353292B1 (en) 2002-04-12 2011-10-26 STMicroelectronics (Research & Development) Limited Biometric sensor apparatus and methods
CN1662931A (en) 2002-05-09 2005-08-31 索尼株式会社 Bio-pattern detecting means, bio-pattern detecting device, biometrics method and biometrics device
US6825930B2 (en) 2002-06-04 2004-11-30 Cambridge Research And Instrumentation, Inc. Multispectral imaging system
WO2004001680A1 (en) * 2002-06-20 2003-12-31 Wayfare Identifiers Inc. Biometric document authentication system
US20040008875A1 (en) * 2002-07-09 2004-01-15 Miguel Linares 3-D fingerprint identification system
JP4387643B2 (en) * 2002-07-31 2009-12-16 富士通株式会社 Processing device with personal recognition function
US6853444B2 (en) 2002-08-30 2005-02-08 Waleed S. Haddad Non-contact optical imaging system for biometric identification
FR2844898B1 (en) * 2002-09-20 2005-03-11 Thales Sa COLOR IMAGE PROCESSING METHOD FOR REMOVING SHADES AND REFLECTIONS.
US7496214B2 (en) 2002-09-25 2009-02-24 The Hong Kong Polytechnic University Method of palm print identification
DE10246411A1 (en) 2002-10-05 2004-04-15 ASTRA Gesellschaft für Asset Management mbH & Co. KG Device for the contactless optical detection of biometric properties of at least one part of the body
JP2004206412A (en) 2002-12-25 2004-07-22 Casio Comput Co Ltd Card type device and authentication system
FR2850191B1 (en) 2003-01-21 2006-04-28 Atmel Grenoble Sa METHOD AND DEVICE FOR SECURING PERSON RECOGNITION
FR2850190B1 (en) 2003-01-21 2006-04-28 Atmel Grenoble Sa METHOD AND DEVICE FOR RECOGNIZING PERSON
US7539330B2 (en) 2004-06-01 2009-05-26 Lumidigm, Inc. Multispectral liveness determination
US7347365B2 (en) * 2003-04-04 2008-03-25 Lumidigm, Inc. Combined total-internal-reflectance and tissue imaging systems and methods
US7751594B2 (en) 2003-04-04 2010-07-06 Lumidigm, Inc. White-light spectral biometric sensors
US7668350B2 (en) * 2003-04-04 2010-02-23 Lumidigm, Inc. Comparative texture analysis of tissue for biometric spoof detection
US7545963B2 (en) 2003-04-04 2009-06-09 Lumidigm, Inc. Texture-biometrics sensor
US7627151B2 (en) 2003-04-04 2009-12-01 Lumidigm, Inc. Systems and methods for improved biometric feature definition
US7394919B2 (en) 2004-06-01 2008-07-01 Lumidigm, Inc. Multispectral biometric imaging
KR20060002923A (en) 2003-04-04 2006-01-09 루미다임 인크. Multispectral biometric sensor
DE10315923A1 (en) 2003-04-08 2004-10-28 Tbs Holding Ag Procedure to detect data of uneven surfaces for biometric data, using non-contact optical sensing of surface
JP2003308520A (en) 2003-04-09 2003-10-31 Casio Comput Co Ltd Image data reading device
JP3733357B2 (en) 2003-05-28 2006-01-11 キヤノン株式会社 Fingerprint input device and personal authentication system using the same
US20050007582A1 (en) * 2003-07-07 2005-01-13 Lumidigm, Inc. Methods and apparatus for collection of optical reference measurements for monolithic sensors
US6958194B1 (en) 2003-10-21 2005-10-25 Foveon, Inc. Imager with improved sensitivity
US7398925B2 (en) 2003-12-09 2008-07-15 First Data Corporation Systems and methods for assessing the risk of a financial transaction using biometric information
US7263213B2 (en) 2003-12-11 2007-08-28 Lumidigm, Inc. Methods and systems for estimation of personal characteristics from biometric measurements
WO2005114557A2 (en) 2004-05-13 2005-12-01 Proximex Multimodal high-dimensional data fusion for classification and identification
US7508965B2 (en) 2004-06-01 2009-03-24 Lumidigm, Inc. System and method for robust fingerprint acquisition
US8229185B2 (en) 2004-06-01 2012-07-24 Lumidigm, Inc. Hygienic biometric sensors
KR100607577B1 (en) 2004-11-05 2006-08-02 (주)니트 젠 Method and apparatus for distinguishing forged fingerprint for optical fingerprint acquisition apparatus
GB2438103B8 (en) 2005-01-14 2010-09-15 Ultra Scan Corp Multimodal fusion decision logic system.
EP1844368A2 (en) 2005-01-21 2007-10-17 Ver-Tec Security Systems Limited Hologram imaging techniques and holograms
US20060171571A1 (en) 2005-02-01 2006-08-03 Chan Michael T Systems and methods for quality-based fusion of multiple biometrics for authentication
US7801338B2 (en) 2005-04-27 2010-09-21 Lumidigm, Inc. Multispectral biometric sensors
US20070052827A1 (en) * 2005-09-07 2007-03-08 Omnivision Technologies, Inc. Coated wafer level camera modules and associated methods
US7515252B2 (en) 2005-12-30 2009-04-07 Cardinal Health 303, Inc. Optical fingerprint imaging system and method with protective film
US9327538B2 (en) 2006-01-05 2016-05-03 Ppg Industries Ohio, Inc. Bragg diffracting security markers
US8355545B2 (en) * 2007-04-10 2013-01-15 Lumidigm, Inc. Biometric detection using spatial, temporal, and/or spectral techniques
CN103336941A (en) 2006-07-19 2013-10-02 光谱辨识公司 Multibiometric multispectral imager
US8175346B2 (en) 2006-07-19 2012-05-08 Lumidigm, Inc. Whole-hand multispectral biometric imaging
US7995808B2 (en) 2006-07-19 2011-08-09 Lumidigm, Inc. Contactless multispectral biometric capture
US7804984B2 (en) * 2006-07-31 2010-09-28 Lumidigm, Inc. Spatial-spectral fingerprint spoof detection
US7801339B2 (en) * 2006-07-31 2010-09-21 Lumidigm, Inc. Biometrics with spatiospectral spoof detection
US8090163B2 (en) * 2006-10-10 2012-01-03 West Virginia University Research Corp. Multi-resolutional texture analysis fingerprint liveness systems and methods
FR2911204B1 (en) * 2007-01-09 2009-02-27 Sagem Defense Securite METHOD FOR PROCESSING AN IMAGE OF AN IMPRINT
WO2008134135A2 (en) 2007-03-21 2008-11-06 Lumidigm, Inc. Biometrics based on locally consistent features
US8050468B2 (en) * 2008-03-05 2011-11-01 Honeywell International Inc. Fingerprint acquisition system
US20100246902A1 (en) 2009-02-26 2010-09-30 Lumidigm, Inc. Method and apparatus to combine biometric sensing and other functionality
EP2471023A1 (en) 2009-08-26 2012-07-04 Lumidigm, Inc. Multiplexed biometric imaging and dual-imager biometric sensor

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050271258A1 (en) 2004-06-01 2005-12-08 Lumidigm, Inc. Multispectral imaging biometrics
WO2006093508A2 (en) * 2004-06-01 2006-09-08 Lumidigm, Inc. Multispectral imaging biometrics
US20070014437A1 (en) * 2004-09-02 2007-01-18 Sony Corporation Information processing device

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2639675B1 (en) * 2012-03-15 2017-12-13 Geoffrey Lee Wen-Chieh High resolution and high sensitivity optically activated motion detection device using multiple color light sources
US11429202B2 (en) 2012-03-15 2022-08-30 Lee Wen Chieh Geoffrey High resolution and high sensitivity optically activated motion detection device using multiple color light sources
US10120460B2 (en) 2012-03-15 2018-11-06 Wen-Chieh Geoffrey Lee High resolution and high sensitivity optically activated motion detection device using multiple color light sources
EP3693934A1 (en) * 2013-04-09 2020-08-12 Bundesdruckerei GmbH Imaging system for capturing a feature of an identification document
EP3690830A1 (en) * 2013-04-09 2020-08-05 Bundesdruckerei GmbH Imaging system for capturing a feature of an identification document
US10845893B2 (en) 2013-06-04 2020-11-24 Wen-Chieh Geoffrey Lee High resolution and high sensitivity three-dimensional (3D) cursor maneuvering device
US9719661B2 (en) 2014-03-31 2017-08-01 Morpho Lighting lens for biometric measurement device
FR3019264A1 (en) * 2014-03-31 2015-10-02 Morpho LIGHTING OPTICS
EP2927577A1 (en) * 2014-03-31 2015-10-07 Morpho Lighting optic for biometric measurement device
US11307730B2 (en) 2018-10-19 2022-04-19 Wen-Chieh Geoffrey Lee Pervasive 3D graphical user interface configured for machine learning
US11216150B2 (en) 2019-06-28 2022-01-04 Wen-Chieh Geoffrey Lee Pervasive 3D graphical user interface with vector field functionality
GB2602793A (en) * 2021-01-06 2022-07-20 Foster & Freeman Ltd Improved coaxial examination of surfaces
GB2602793B (en) * 2021-01-06 2023-02-08 Foster & Freeman Ltd Improved coaxial examination of surfaces

Also Published As

Publication number Publication date
US20110211055A1 (en) 2011-09-01
BR122013021647A2 (en) 2016-04-05
CN102576408A (en) 2012-07-11
US20110085708A1 (en) 2011-04-14
US8731250B2 (en) 2014-05-20
US8872908B2 (en) 2014-10-28
IL218238A0 (en) 2012-04-30
EP2471023A1 (en) 2012-07-04
DE112010003414T5 (en) 2012-12-06
BR112012004177A2 (en) 2016-03-29

Similar Documents

Publication Publication Date Title
US8731250B2 (en) Multiplexed biometric imaging
EP1839233B1 (en) Biometric recognition/verification using multispectral imaging
US8831297B2 (en) Contactless multispectral biometric capture
US8175346B2 (en) Whole-hand multispectral biometric imaging
US9767340B2 (en) Latent fingerprint detectors and fingerprint scanners therefrom
US8355545B2 (en) Biometric detection using spatial, temporal, and/or spectral techniques
CN100376922C (en) Personnel identifier and fingerprint image camera method
CN101506827B (en) System and method for robust fingerprint acquisition
KR20090032128A (en) Multibiometric multispectral imager
WO2008154589A2 (en) Contactless multispectral biometric capture
IL184442A (en) Biometric recognition/verification using multispectral imaging
FR2757975A1 (en) Optical finger print detector e.g. for access control

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080048237.7

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10752477

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 218238

Country of ref document: IL

WWE Wipo information: entry into national phase

Ref document number: 112010003414

Country of ref document: DE

Ref document number: 1120100034141

Country of ref document: DE

WWE Wipo information: entry into national phase

Ref document number: 2010752477

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 539/MUMNP/2012

Country of ref document: IN

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112012004177

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 112012004177

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20120227