WO1986005018A1 - Iris recognition system - Google Patents
Iris recognition system Download PDFInfo
- Publication number
- WO1986005018A1 WO1986005018A1 PCT/US1986/000227 US8600227W WO8605018A1 WO 1986005018 A1 WO1986005018 A1 WO 1986005018A1 US 8600227 W US8600227 W US 8600227W WO 8605018 A1 WO8605018 A1 WO 8605018A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- pupil
- iris
- descriptors
- eye
- image
- Prior art date
Links
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/12—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for looking at the eye fundus, e.g. ophthalmoscopes
- A61B3/1216—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for looking at the eye fundus, e.g. ophthalmoscopes for diagnostics of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/141—Control of illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C9/00—Individual registration on entry or exit
- G07C9/30—Individual registration on entry or exit not involving the use of a pass
- G07C9/32—Individual registration on entry or exit not involving the use of a pass in combination with an identity check
- G07C9/37—Individual registration on entry or exit not involving the use of a pass in combination with an identity check using biometric data, e.g. fingerprints, iris scans or voice recognition
Definitions
- the invention relates to methods and apparatus for identification of physical characteristics of a human being or other animal. More specifically, the invention relates to the recognition of the physical characteristics of the eyes.
- Fingerprints are one of the best known because of their extensive use by law enforcement agencies. Identification by fingerprint is especially helpful where an individual has left his fingerprint during the commission of a crime, but in general, the technology of fingerprint identification relies heavily on human judgment. In addition, the collection and detection of fingerprints is typically difficult.
- U.S. Patent 4,109,237 has disclosed an identification technique making use of the retinal vasculature patterns of the human eye.
- this technique can be automated, and is simpler than detecting, collecting, and iden tifying fingerprints.
- some individuals may attempt to alter their fingerprints in order to avoid identification, altering the retinal vasculature pattern would be difficult and dangerous, and might impair or destroy the vision in that eye.
- the present invention is based on the discovery that the iris and pupil of the eye, especially the human eye, may be used for identification. Therefore, the present invention provides an identification technique based upon the recognition of the unique features of the iris and pupil, referred to herein as "iris identification".
- Iris identification shares the advantages of retinal identification, but can be performed more easily because the iris and pupil are more readily visible.
- some of the features of the iris and pupil are highly regular, providing a geometric pattern which facilitates automatic identification.
- An important feature of the iris is that its visible features change substantially as the pupil of the eye expands or contracts. Therefore, an important aspect of effective iris identification is to provide a technique which takes into account these changing features.
- the method of identification according to the invention includes illuminating an eye, obtaining an image of the iris and pupil of the eye, and comparing the obtained image with stored image information to identify the eye.
- Illuminating the eye may include driving the pupil of the eye to a predetermined size, which may be done by directing light onto the retina of the eye and adjusting the intensity of the light until the pupil reaches the predetermined size.
- the pupil may be driven to a plurality of predetermined sizes, and a respective image of the iris and the pupil may be obtained at each of the predetermined sizes.
- the image obtained at each predetermined pupil size may be compared with stored image information from an eye with the same pupil size.
- illuminating the eye includes obliquely illuminating the iris to provide shadows of elevation-dependent features.
- the iris may be illuminated by a plurality of spatially separated, relatively monochromatic light sources, so that each shadow of one of the elevation-dependent features will lack a color corresponding to one of the light sources.
- the comparing of the obtained image with stored image information may be a point-by-point comparison with a stored reference image.
- a set of descriptors may be. derived from the just-obtained image and these descriptors may be compared with reference descriptors derived from a previous image.
- the system according to the invention includes means for illuminating an eye, means for obtaining an image of the iris and pupil of the eye, and means for comparing the obtained image with stored image information to make an identification.
- the illuminating means may include an adjustable light source for illuminating the eye at a selected value.
- the illuminating means may include control means which receive the obtained image of the eye and detect the size of the pupil, the control means being operable for adjusting the adjustable light source to a selected light intensity for controlling the size of the pupil.
- the control means may further be operable to control the size of the pupil to obtain a plurality of predetermined pupil sizes.
- the illuminating means may also include a plurality of relatively monochromatic, spatially separated light sources for providing shadows of elevation-dependent features. Each resulting shadow of an elevationdependent feature will lack a color corresponding to the color of one of the light sources.
- the image-obtaining means may include a camera, and the comparing means may include a processor responsive to the output from the camera.
- the processor may include storage means for storing reference descriptors derived from a previous image, so that the processor may be operable to derive a set of descriptors from the justobtained image and to compare these descriptors with the reference descriptors stored in the storage means.
- Fig. 1 is a schematic view of the iris and pupil of an eye, showing the major visible features.
- Fig. 2 is a general flow chart showing the major functions performed in the method of the invention.
- Fig. 3 is a simplified block diagram showing the flow of information through a system according to the invention.
- Fig. 4 is a flow chart showing in greater detail an embodiment of the method of Fig. 2.
- Fig. 5 is a flow chart showing a method of measuring rate of pupil size change according to the invention.
- Fig. 6 is a simplified flow chart showing a method of side illumination according to the invention.
- Fig. 7 is a general flow chart showing a method of comparing according to the invention.
- Fig. 8 is a schematic diagram showing the electrical and optical characteristics of a system according to the invention.
- Fig. 9 is a plan view of a viewing screen for use in the system of Fig. 8.
- Fig. 10 is a functional block diagram of one embodiment of the system of Fig. 3.
- Fig. 11 is a diagram illustrating the illumination control circuit of Fig. 10.
- Fig. 12 is a flow chart of the operation of the processor of Fig. 10 according to the invention.
- Fig. 1 is a simplified drawing of an eye 10, showing in detail the features of the iris 20 and pupil 30.
- Eye 10 as shown, is a human eye, but some other animals have a similar eye structure, permitting identification by the method of the invention.
- pupil 30 is a black, light receptive orifice, while iris 20 may have any of a broad range of hues and may be irregularly pigmented.
- iris 20 and pupil 30 are closely related in function. More precisely, one function of iris 20 is to control the size of pupil 30. Therefore, in addition to the visible features of iris 20 relating to pigmentation, a number of visible features relate to the movements made by iris 20 in order to contract and dilate pupil 30.
- iris 20 may be divided into the ciliary area 22, which is an annular region at its periphery, and the pupillary area 24, which is an annular area bordering pupil 30.
- the tissue of iris 20 is soft and loosely woven, and the illumination which enters the pupil 30 and falls on the retina of eye 10 controls muscles in iris 20 causing the size of pupil 30 to grow larger or smaller to regulate the amount of light entering pupil 30.
- the change in the size of pupil 30 results from involuntary reflexes and is not under conscious control.
- pigment spots 26 are random concentrations of pigment-bearing cells in the visible surface of the iris 20, and generally appear in the ciliary area 22.
- the color of the pigment spots 26 may be nearly black -- darker than the darkest brown iris 20.
- pigment spots 26 are quite stable in number, position and color.
- Crypts 28, unlike pigment spots 26, relate both to pigmentation and to the surface structure of the iris 20.
- Iris 20 includes an anterior layer which is directly visible and a posterior layer behind it, which is very darkly pigmented. The anterior layer ranges in pigment from a very light, almost white color to a dark brown color.
- the anterior layer is very light or white
- the iris 20 appears blue due to the pigment in the posterior layer, and if iris 20 is relatively light in color, the areas in which it is relatively thin will appear somewhat darker, due to the dark color of the posterior layer.
- These darkened areas may occur, for example, in the features referred to as crypts 28, which resemble sharply demarcated excavations and are typically located adjacent to the collarette 32, the boundary between ciliary area 22 and pupillary area 24.
- Collarette 32 is the boundary between the ciliary area 22 and the pupillary area 24, and is typically a sinuous line which forms an elevated ridge running roughly parallel with the margin of the pupil 30.
- the collarette 32 is the thickest part of iris 20.
- radial furrows 34 Extending radially in relation to the center of pupil 30 are radial furrows 34.
- a typical radial furrow 34 may begin near pupil 30 and extend through collarette 32.
- Radial furrows 34 are creases in the anterior layer of iris 20, from which loose tissue may bulge outward, and it is this loose tissue which permits the iris 20 to expand or contract, changing the size of pupil 30.
- the positions of radial furrows 34 in relation to each other are stable.
- concentric furrows 36 are creases from which loose tissue may bulge outward, but their shape is generally circular, concentric with the pupil, so that they permit expansion and contraction of the iris in a different direction than radial furrows 34.
- Concentric furrows 36 typically appear in the ciliary area 22, near the periphery of iris 20.
- Eye 10 may also have any of several rare anomalous visible features (not shown) unrelated to pigment or to controlling the size of pupil 30. Due to aging or trauma, atrophic areas may appear on iris 20, resulting in a "moth-eaten” texture. Tumors may grow on the iris. Congenital filaments may occur connecting the iris to the lens of the eye. Finally, the most striking visible feature of eye 10 is typically pupil 30. As shown in Fig. 1, pupil 30 may not be exactly circular in shape, and its deviation from a circle is a visible characteristic. At the margin of pupil 30 is a protruding portion of the posterior layer of iris 20, the pigment frill 38. Pigment frill 38 typically has a very dark brown color.
- the present invention is based on the discovery that the iris of the eye may be used for identification. This discovery involves both the uniqueness of the iris and the stability of the iris over time. It has been discovered that every iris is unique, particularly in the detailed structure of the front or anterior layer. Not only are the irises of the eyes of identical twins different, but the iris of each eye of any person is different from that of his other eye. Furthermore, although specific details of the appearance of an iris vary dramatically, depending on level and direction of illumination, the basic, significant features of the iris remain extremely stable and do not change over a period of many years.
- the method of the invention may be generally understood from Fig. 2, a flow chart showing the basic functions performed in identifying an eye from the visible features of the iris and pupil.
- the first function is to illuminate the eye. Because the visible features of iris 20 and pupil 30 vary depending on illumination, the illumination of the eye will be performed to obtain a desired set of features.
- an image of iris 20 and pupil 30 is obtained, including the desired features.
- this image is compared with stored image information in order to identify the eye.
- Fig. 3 is a block diagram showing the major functional components of the system according to the invention.
- Illuminating means 70 provides light having desired characteristics to the eye 10 under observation. Eye 10 responds to the illumination provided, so that iris 20 and pupil 30 take a characteristic shape.
- iris 20 reflects light to imaging means 80, which obtains an image of iris 20 and pupil 30. This image is then compared with stored image information by comparing means 90 in order to identify the eye.
- FIG. 4-7 A number of methods for implementing the invention are shown in Figs. 4-7. Each of these figures is a flow chart showing a sequence of steps performed in a process of identifying an eye from the iris and pupil.
- Fig. 4 shows a basic method of performing the invention in which the pupil is driven to at least one predetermined size at which an image is taken.
- the pupil is illuminated at one extreme of the range of illumination, either an extreme of darkness or of brightness.
- the extreme may be determined by the capabilities of the system or may be arrived at by a straightforward series of steps.
- the illumination may gradually be increased from a moderate level of illumination to an illumination beyond which further illumination does not decrease the size of the pupil.
- a more efficient approach is to drive the pupil until its size is larger or smaller than the largest or smallest of the predetermined sizes at which an image will be obtained. From this extreme illumination, it will then be possible to drive the pupil to the predetermined size for imaging.
- the pupil is driven to the first predetermined size in step 44. Assuming that the extreme of illumination was darkness, the illumination will be increased in step 44 until the pupil contracts to a predetermined size. When the pupil reaches this size, an image is obtained and stored in step 52 for subsequent comparison. Then, in step 54, a test is run to determine whether the number n, a count of the images obtained, is greater than or equal to the number N of predetermined sizes at which images are to be obtained. If not, the count of the images obtained, n, is incremented in step 46, and steps 44, 52 and 54 are repeated for the next predetermined size.
- step 54 will determine that n is equal to N , and image comparison will be done in step 62.
- step 62 the image at each of the N predetermined sizes is compared with stored image information obtained from an eye illuminated to have the same predetermined pupil size.
- Fig. 5 illustrates another variation which can be made in the basic method of the invention. This variation includes measuring the rate at which the pupil size changes between two predetermined sizes.
- step 142 light from a light source is directed onto the retina of the eye in order to control the size of the pupil through reflex action.
- step 144 the position or intensity of the light source is adjusted in order to obtain the first predetermined pupil size.
- the light source may be moved from a central position in which it directs light onto the central area of the retina toward an oblique position in which it directs light at an angle through the pupil onto the more peripheral parts of the retina, or vice versa.
- a test is performed in step 146 to determine whether the first predetermined size has been obtained. If not, the adjustment is again performed in step 144, and this loop is repeated until the first predetermined size is obtained.
- the illumination of the retina is changed at a predetermined rate in step 148 .
- the changing of the illumination is continued until the test in step 150 determines that a second predetermined size has been attained.
- the time necessary to go from the first predetermined size to the second predetermined size at the predetermined rate will provide a measure of the rate of change, in step 152.
- This measure could also be related to the change in size between the first and second predetermined size and to the specific predetermined rate at which the illumination is changed in step 148.
- the resulting value may then be used to identify the eye. It should be borne in mind, however, that the rate of change of pupil size may vary for individuals whose reflexes are affected by emotion, drugs, and so forth, so that the rate of change of pupil diameter should generally provide only a secondary indication of identity.
- step 44 could first be performed to obtain the first predetermined size, corresponding to steps 144 and 146 in Fig. 5. After the performance of steps 52, 54 and 46, step 44 could again be performed by changing the illumination at the predetermined rate to obtain the second predetermined size, corresponding to steps 148 and 150 in Fig. 5. Then, in addition to obtaining the image in step 52, the rate of change of pupil size could also be determined in step 152.
- Fig. 6 illustrates a routine which may similarly be included in step 52 of Fig. 4, to obtain an image of the iris of an eye in which the elevation-dependent features are shown with greater contrast. This is done by illuminating the iris from a plurality of oblique positions. The illumination from each position is provided by a relatively monochromatic light source to produce shadows of elevation-dependent features lacking the color of that light source. In step 250, illumination from position m is performed using the color m of the corresponding relatively monochromatic light source. This illumination will produce shadows lacking the color m, and an image of these shadows is obtained in step 252. Then, in step 254, a test is performed to determine whether m has reached M, the total number of positions from which the iris is to be illuminated.
- n is incremented in step 256, and the loop including steps 250, 252 and 254 is repeated until m equals M.
- theshadow images may be processed in step 258 to prepare them for subsequent comparison with reference images.
- the method of Fig. 6 is especially beneficial for obtaining images of elevation-dependent features such as radial furrows 34.
- radial furrows 34 may extend radially in any direction, so that illumination from any one direction will provide an excellent shadow image of radial furrows 34 which extend perpendicular to that direction of illumination. Radial furrows which extend parallel to the direction of illumination, however, will produce poor shadow images. Therefore , to produce the best images of radial furrows 34, it is desirable to obtain a number of shadow images and to combine them in processing step 258.
- the comparison of images having the same pupil sizes in step 62 of Fig. 4 may be performed in many ways.
- One straightforward method would be to store the reference image as a pixel-by-pixel image resulting from a scanning camera such as a conventional video camera.
- the image obtained would then be obtained with a similar video camera, as discussed below in relation to Fig. 8, to obtain a pixel-by-pixel image of the eye.
- the two images could then be compared, pixel-by-pixel, to determine whether they were the .same image.
- the comparison could be made in a way which would take into account differences in the overall intensity of the two images. This method of comparison, although simple to implement, is extremely sensitive to registration.
- test would only result in a conclusion that the two images were of the same eye if the eyes shown in the two images were of the same size and were positioned identically. Therefore, it will be preferable to perform a comparison algorithm which is not so sensitive to registration.
- Fig. 7 illustrates another method of comparison which could be used in the present invention.
- descriptors are extracted from the image obtained, which may be the image obtained in step 52 of Fig. 4.
- One example of descriptors which could be extracted would be the angular positions and lengths of the radial furrows 34.
- Another set of descriptors would be the shape of the pupil 30 or of the pigment frill 38.
- Other descriptor sets could be obtained relating to the positions and sizes of pigment spots 26, crypts 28 and anomalous features such as atrophic areas, tumors and congenital filaments.
- generally circular features such as collarettes 32 and concentric furrows 36 could be described by respective descriptors.
- illuminating means 70 As noted above, the necessary equipment for performing iris recognition according to the invention could take many forms. The major functional components of any such system, however, will be those shown in Fig. 3, including illuminating means 70, imaging means 80 and comparing means 90.
- Fig. 8 shows a system for illuminating and obtaining an image of the eye.
- the system includes a simplified electrical circuit for controlling illumination and an optical arrangement in which the image of fixation target 71 seen by eye 10 and the reflected image of iris 20 and pupil 30 to camera 84 are approximately coaxial. It is presently preferred to provide a target image, upon which eye 10 may be fixated, concentric with visual axis 16 for reference purposes. It is also preferred for camera 84 to capture the reflected image concentric or nearly concentric with visual axis 16, because if the image were obtained far from the axis 16, it would be distorted, which would then have to be compensated by appropriate processing. Additional illumination of iris 20 is provided obliquely to create good shadows of elevation-dependent features.
- the source of light is a filament 72a which receives power from power source 72b connected in series through variable resistor 72c, which may be an electrically controlled potentiometer.
- the light emitted by filament 72a is collimated by lens 73 and illuminates target 71, which may have any appropriate shape.
- the image of target 71 is then focused by movable lens 74 and transmitted through monitor beam-splitter 76 onto axial beam-splitting plane 82.
- the image is then reflected from the central mirrored area 82a through pupil 30 onto retina 12.
- the optical system shown in Fig. 8 includes means for placing target 71 at any desired optical distance from eye 10.
- Target 71 can thus be made visible to subjects with a wide variety of refractive states.
- the refraction or bending of light rays at the air interface of cornea 14, for example, is substantial and varies for different subjects.
- the necessary adjustment of optical distance is obtained by moving lens 74 toward or away from target 71 until target 71 is optically conjugate with retina 12.
- lens 74 may also be a variable sphero-cylindrical lens system of any of the well-known types.
- Target 71 will be optically conjugate with retina 12 when, as shown in Fig. 8, the light from each point on target 71 comes to a focus at a corresponding point on retina 12. The convergence or divergence of light from target 71 due to the adjustment of lens 74 will thus compensate for the optics of eye 10.
- Oblique light sources 78a and 78b provide illumination to the iris 20 and may be controlled similarly to filament 72a by using a variable resistor (not shown). The resulting reflection will depend on the positioning of oblique light sources 78a and 78b, because the iris is a diffuse reflector, reflecting light according to a cosine law in which the amount of reflected light in the direction normal to the iris varies as the cosine of the incident angle of the light.
- Axial beam-splitting plane 82 serves the two functions of directing the image of target 71 from lens 74 toward eye 10 and also transmitting the reflected image of iris 20 to camera 84.
- beam splitting plane 82 is a thin transparent sheet such as glass on which mirror 82a is. formed for reflecting the image.
- Mirror 82a may be a microscopically thin, reflective film or coating, such as a metal, on the central part of plane 82 and on the side toward eye 10, to provide a first surface mirror. It may, for example, have an elliptical shape.
- the reflected image of target 71 is directed through pupil 30 onto the retina 12 of eye 10 to control the size of pupil 30.
- Mirror 82a will create a non-transparent area in the center of plane 82, preventing an intense reflection of the target image from cornea 14 from reaching camera 84. Since the reflected target image may be the brightest object in the field of view of camera 84, this may significantly reduce the noise in the image of iris 20 and pupil 30.
- Monitor beam-splitter 76 will provide an image of the intense reflection from cornea 14, as reflected by mirror 82a, and this reflection image may be used to position the eye according to known eye-positioning techniques. Proper positioning of the eye will help to provide a standardized image of iris 20 and pupil 30 and will ensure that the intense reflection of the target image does not reach camera 84.
- Fig. 9 shows viewing screen 86 of camera 84, upon which are marked peripheral points 88a and central point 88b.
- Peripheral points 88a may be used to measure the size of pupil 30, as shown.
- Central point 88b may be useful in aligning pupil 30 so that size may be properly tested.
- a grid may also be provided on viewing screen 86 to assist in the determination of the size of pupil 30 when it does not meet peripheral points 88a.
- concentric markings could be provided on screen 86, as disclosed in U.S. Patent 3,966,310, issued to Larson.
- Figs. 8 and 9 the distance between camera 84 and eye 10 must be fixed in order to provide a standardized image. This can be done by providing a headrest (not shown) against which the forehead above eye 10 may be positioned.
- the headrest may also be equipped with appropriate additional surfaces for darkening a space around eye 10. This permits the system to provide a condition of nearly zero illumination to retina 12.
- variable resistor 72c is manually adjusted to its maximum resistance, effectively turning off filament 72a.
- Oblique light sources 78a and 78b are activated to provide a desired level of illumination upon iris 20.
- variable resistor 72c is gradually turned to a lower resistance until the image of target 71 begins to illuminate retina 12.
- the operator looking at viewing screen 86, will see the pupil beginning to decrease in size as the illumination from filament 72a increases.
- the operator may activate camera 84 to obtain an image of iris 20 and pupil 30. If further, predetermined sizes are marked by points in viewing screen 86, the operator may continue to adjust variable resistor 72c until those additional predetermined sizes of pupil 30 are shown in viewing screen 86 and are imaged when camera 84 is activated.
- Fig. 8 cold be modified in many ways within the scope of the invention.
- the beam-splitting plane could be defined by two prisms positioned against each other or by a half-silvered mirror or other partially reflecting surface. Such arrangements, however, will result in a greater loss of intensity of the collimated light and of the reflected image than the arrangement of
- Fig. 10 shows a block diagram of electrical components performing the functions set forth in Fig. 3.
- Iris recognition system 100 in Fig. 10 includes illumination control circuit 170, camera 180 and processor 190.
- Illumination control circuit 170 and camera 180 operate under the control of processor 190.
- illumination control circuit 170 illuminates eye 10 both to control the size of pupil 30 and to provide a reflected image of iris 20 and pupil 30.
- This reflected image is received by camera 180, which may be a video camera which converts the reflected image into an image signal for processor 190.
- Processor 190 then compares the image signal with a reference image in order to identify the human being.
- Fig. 11 shows in greater detail the illumination system of iris recognition system 100.
- illumination control circuit 170 controls light sources 172, 174, 176 and 178. In general, these light sources should be as small as possible to minimize noise in the reflected image from eye 10.
- Light source 172 which illuminates retina 12, as discussed in relation to Fig. 8, may do so through a fixation target 71 and a beam-splitting plane 82. Light source 172 will thus provide a target image upon which eye 10 will fix and focus, thereby becoming aligned for imaging.
- the target image should preferably be a distant object to discourage accommodation which would reduce pupil size.
- Rays from the target image may be relatively tightly collimated and subtend a small angle in order to provide a more critical, centralized alignment, or may be less tightly collimated if additional iris illumination is desired.
- the target image could also be compound, such as cross hairs in one plane which must be aligned on a designated part of a background at another plane to obtain alignment of the pupil, although this would depend on the acuity of the eye 10.
- a polarizing ring-sight device could be used to create concentric circles on the retina independent of the refraction of eye 10.
- the target image should not be extremely bright, because that would also reduce the size of pupil 30.
- the target image may be used to cause the subject to align pupil 30 for optimal imaging.
- the head rest (not shown) can be moved in order to orient the head.
- the illumination system shown in Fig. 11 includes oblique light sources 174, 176 and 178.
- each of these oblique light sources 174, 176 and 178 may provide a relatively monochromatic beam of light, so that the shadows resulting from the oblique positioning of these light sources will lack the corresponding color, facilitating the subsequent proces sing of the shadow images.
- Each source 174, 176 and 178 could be a relatively monochromatic source such as an LED or a solid state laser.
- Fig. 11 shows an arrangement in which three light sources are provided at approximately 120° angles from each other, the number of light sources and their relative angular positions could be varied in many ways. In addition, other arrangements could be used to obtain additional information about elevation-dependent features of iris 20. For example, more than one camera 180 could be provided, or a single camera 180 could be provided which could be moved between a plurality of positions in order to obtain three-dimensional information. Any other means for obtaining an image could be used, including laser-based techniques such as holography and including flying spot scanning by mechanical or electrooptical techniques.
- Fig. 12 shows a flow chart of the operation of processor 190 in Fig. 10.
- Processor 190 will at first be initialized in step 342, in order to provide appropriate memory contents and in order to perform the necessary loading of programs.
- processor 190 may provide control signals causing camera 180 to begin to operate and providing power to illumination control circuit 170.
- processor 190 provides control signals to illumination control circuit 170 causing pupil-driving light source 172 to illuminate the retina 12 at the extreme illumination of the range available.
- oblique light sources 174, 176 and 178 may be activated to provide sufficient illumination so that camera 180 is able to provide an image of pupil 30 to processor 190.
- the pupil is then driven across the range of its size in step 346. This is performed by gradually increasing the illumination from pupil-driving light source 172. After each increase in intensity, the pupil size is measured in step 348 and tested in step 352 to determine whether it equals the desired predetermined size x n . For the first iteration, the first predetermined size x 1 will be obtained after the necessary reitteration of the loop including steps 346, 348 and 352.
- Oblique light sources 374, 376 and 378 will also be activated, either at the same time or in sequence according to a method like that shown in Fig. 6, in order to obtain an image of the iris 20 and pupil 30 of eye 10. Oblique light sources 374, 376 and 378 should be activated in a way which minimizes the constriction of pupil 30. Therefore, it may be necessary to activate them before or while the illumination from light source 172 is increased.
- processor 190 will then extract the nth descriptor set in step 362.
- the first descriptor set will be obtained.
- step 364 n is tested to determined if it is equal to N, the total number of descriptor sets to be obtained. If not, n is incremented in step 366, and the loop including steps 346, 348, 352, 354, 362 and 364 is repeated for the next value of n. This loop will be repeated until step 364. determines that n equals N.
- n N
- the resulting descriptor sets are compared with one or more reference descriptor sets in step 368. If the objective is to determine whether the eye 10 is that of one of a population of people, each descriptor set resulting from step 362 will be compared with the corresponding descriptor sets of the members of that population. On the other hand, if the objective is simply to confirm that an eye 10 is that of a specific individual, it is only necessary to compare the descrip tor sets resulting from step 362 with the corresponding descriptor set of that individual, which may, for example, be stored on a credit card, identification card, or other identifying document, or in the memory of a computer system to which the processor 190 has access.
- step 368 an appropriate indication may be provided to the operator of iris recognition system 100 that the eye 10 has been identified or has not been identified, completing the operation shown in Fig. 12. At this point, the operator may choose to initiate further iris recognition by repeating the process shown in Fig. 12.
- Processor 190 in extracting descriptors in step 362, may perform a wide variety of algorithms.
- a number of such algorithms are set forth in Ballard, D.H. and Brown, CM., Computer Vision. Prentice-Hall, Englewood Cliffs, New Jersey, 1982; Duda; R.O. and Hart, P.E., Pattern Recognition and Scene Analysis, Wiley, New York, 1973; and Rosenfeld, A., Digital Picture Processing, Academic Press, New York, 1982.
- Many of these algorithms use classical mathematical techniques such as difference operators, taking a differential equation of selected points and testing a rate of change, and mask matching or surface fitting, fitting to a known shape at various orientations. Although these and other such algorithms are known, they have not previously been applied to the specific features of the iris and pupil.
- the location, size and deviation from circularity of the pupil could provide a number of descriptors.
- the pupil could first be detected using a boundary detection algorithm or an edge detection algorithm, which may, for example, detect an abrupt gray level change.
- Systems performing pupil size measuring algorithms are disclosed, for example, in U.S. Patent 3,598,107, issued to Ishikawa et al, and U.S. Patent 3,533,683, issued to Stark et al.
- a Hough transform algorithm for detecting circles, which maps curves into the transform space according to characteristics such as curvature, could be used on a boundary or edge which is detected.
- the received image could be subjected to a thresholding algorithm after which a region-growing or aggregation algorithm, such as the blob coloring algorithm discussed by Ballard and Brown, could be performed to find the largest connected region of pixels with intensity values below threshold.
- a region-growing or aggregation algorithm such as the blob coloring algorithm discussed by Ballard and Brown
- the region could be grown outward from a central dark pixel in the image, progressively aggregating the adjacent dark pixels until the pupil boundary is reached, beyond which adjacent pixels will not be dark. This will provide a measure both of pupil size and of location, as the center of the pupil may be determined from its boundary. This may be done, for example, by taking slices through the dark region in a number of radial directions to select a largest diameter.
- the least sum of squares from its center to the boundary could be taken and compared with that of the second largest diameter.
- the center and radius giving the least sum of squares would define the circle most closely approximating the pupil boundary, and the residual sum of squares would be a quantitative measure of deviation of the pupil from circularity, a "figure of merit" for the pupil. This figure of merit would be a useful descriptor of the pupil.
- the characteristics of the radial furrows could also be used to obtain descriptors. From the center of the pupil, the precision of radiality of the radial furrows can be obtained. Line detection algorithms and streak or thick line detection algorithms could be used to locate the radial furrows. The radial furrows may then be compared with true straight lines radiating from the center by obtaining a sum of squares of differences, indicating radiality. Some furrows, rather than being entirely straight, include, for example, two offset parts, each of which is generally radial, the two being connected by a non-radial jog. A measure of thickness could be obtained for such a furrow by determining how wide a streak is necessary to cover the furrow. Also, the Hough transform algorithm for detecting straight lines could be applied to obtain such descriptors as the length, angular location and direction of radial furrows.
- the concentric furrows may be detected and described using edge detection algorithms and the Hough transform algorithms for detecting curved lines.
- the collarette and the contours of the crypts along it could be described using edge detection algorithms and chain coding algorithms.
- points of high curvature along these contours could be located and represented for matching purposes, which may be more efficient.
- Isolated crypts could be described by a thresholding algorithm, followed by a region-growing algorithm and then algorithms obtaining the center of gravity and the lower order central moments., as described by Duda and Hart.
- Spot detection algorithms detecting high contrast spikes in areas of relatively constant gray level, could detect small pigment spots. More generally, algorithms like those for measuring the pupil could be used to obtain location and lower-order central moments of pigement spots, and other such features. Also, high-curvature points along the contour of a feature could be detected and described. Finally, the color of the iris generally could be found by an algorithm obtaining a histogram in a three-dimensional RGB (Red-Green-Blue) color space, as described by Ballard and Brown at page 153. The peak in the histogram will provide a descriptor of color.
- RGB Red-Green-Blue
- Processor 190 may control the operation of illumination control circuit 170 and camera 180 in any appropriate manner to obtain images for processing.
- the image from camera 180 may be constantly received and processed by processor 190 to determine the appropriate level of illumination by illumination control circuit 170.
- processor 190 may control illumination means 170 to maintain a temporarily constant level of illumination while camera 180 obtains a complete image for processing.
Abstract
Methods and apparatus for identifying an eye, especially a human eye (10), on the basis of the visible features of the iris (20) and pupil (30). The eye is first illuminated until the pupil reaches a predetermined size, at which an image of the iris and pupil is obtained. This image is then compared with stored image information for identification. The stored image information is previously obtained from an eye, the pupil of which was similarly brought to the same predetermined size. The illumination of the iris may include oblique illumination from several positions around the circumference of the iris. The illumination from each position may be relatively monochromatic, so that the resulting shadow will lack the color of the light source (172) at that position providing better contrast for elevation-dependent features. A system for performing iris recognition (100) may include a processor (190) which controls an illumination control circuit (170) and a camera (180) to obtain images at several predetermined sizes of the pupil.
Description
IRIS RECOGNITION SYSTEM
BACKGROUND OF THE INVENTION
1. Field of the Invention
The invention relates to methods and apparatus for identification of physical characteristics of a human being or other animal. More specifically, the invention relates to the recognition of the physical characteristics of the eyes.
2. Description of the Prior Art
Several known techniques are traditionally used to identify a human being from physical characteristics. Fingerprints are one of the best known because of their extensive use by law enforcement agencies. Identification by fingerprint is especially helpful where an individual has left his fingerprint during the commission of a crime, but in general, the technology of fingerprint identification relies heavily on human judgment. In addition, the collection and detection of fingerprints is typically difficult.
More recently, U.S. Patent 4,109,237 has disclosed an identification technique making use of the retinal vasculature patterns of the human eye. Among the advantages of this technique are that it can be automated, and is simpler than detecting, collecting, and iden
tifying fingerprints. Furthermore, although some individuals may attempt to alter their fingerprints in order to avoid identification, altering the retinal vasculature pattern would be difficult and dangerous, and might impair or destroy the vision in that eye.
It would be advantageous to have an identification technique making use of the unique characteristics of the eye but not requiring pupil dilation and the complex alignment procedure of retinal identification, including focusing on the retina and maintaining alignment with the direction of gaze of the subject. It would also be advantageous to have such an identification technique which could be used with minimal cooperation of the subject.
SUMMARY OF THE INVENTION
The present invention is based on the discovery that the iris and pupil of the eye, especially the human eye, may be used for identification. Therefore, the present invention provides an identification technique based upon the recognition of the unique features of the iris and pupil, referred to herein as "iris identification".
Iris identification shares the advantages of retinal identification, but can be performed more easily because the iris and pupil are more readily visible. In addition, some of the features of the iris and pupil are highly regular, providing a geometric pattern which facilitates automatic identification. An important feature of the iris, however, is that its visible features change substantially as the pupil of the eye expands or contracts. Therefore, an important aspect of effective iris identification is to provide a technique which takes into account these changing features.
The method of identification according to the invention includes illuminating an eye, obtaining an image of the iris and pupil of the eye, and comparing the obtained image with stored image information to identify the eye. Illuminating the eye may include driving the pupil of the eye to a predetermined size, which may be done by directing light onto the retina of the eye and adjusting the intensity of the light until the pupil reaches the predetermined size. The pupil may be driven to a plurality of predetermined sizes, and a respective image of the iris and the pupil may be obtained at each of the predetermined sizes. The image obtained at each predetermined pupil size may be compared with stored image information from an eye with the same pupil size.
In another embodiment, illuminating the eye includes obliquely illuminating the iris to provide shadows of elevation-dependent features. The iris may be illuminated by a plurality of spatially separated, relatively monochromatic light sources, so that each shadow of one of the elevation-dependent features will lack a color corresponding to one of the light sources.
The comparing of the obtained image with stored image information may be a point-by-point comparison with a stored reference image. In an alternative form of comparison, a set of descriptors may be. derived from the just-obtained image and these descriptors may be compared with reference descriptors derived from a previous image.
The system according to the invention includes means for illuminating an eye, means for obtaining an image of the iris and pupil of the eye, and means for comparing the obtained image with stored image information to make an identification. The illuminating means may include an adjustable light source for illuminating the eye at a selected value. In addition, the illuminating means
may include control means which receive the obtained image of the eye and detect the size of the pupil, the control means being operable for adjusting the adjustable light source to a selected light intensity for controlling the size of the pupil. The control means may further be operable to control the size of the pupil to obtain a plurality of predetermined pupil sizes.
The illuminating means may also include a plurality of relatively monochromatic, spatially separated light sources for providing shadows of elevation-dependent features. Each resulting shadow of an elevationdependent feature will lack a color corresponding to the color of one of the light sources.
The image-obtaining means may include a camera, and the comparing means may include a processor responsive to the output from the camera. The processor may include storage means for storing reference descriptors derived from a previous image, so that the processor may be operable to derive a set of descriptors from the justobtained image and to compare these descriptors with the reference descriptors stored in the storage means.
Other objects, features and advantages of the invention will be apparent from the following description, together with the accompanying drawings and the appended claims.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 is a schematic view of the iris and pupil of an eye, showing the major visible features.
Fig. 2 is a general flow chart showing the major functions performed in the method of the invention.
Fig. 3 is a simplified block diagram showing the flow of information through a system according to the invention.
Fig. 4 is a flow chart showing in greater detail an embodiment of the method of Fig. 2.
Fig. 5 is a flow chart showing a method of measuring rate of pupil size change according to the invention.
Fig. 6 is a simplified flow chart showing a method of side illumination according to the invention.
Fig. 7 is a general flow chart showing a method of comparing according to the invention.
Fig. 8 is a schematic diagram showing the electrical and optical characteristics of a system according to the invention.
Fig. 9 is a plan view of a viewing screen for use in the system of Fig. 8.
Fig. 10 is a functional block diagram of one embodiment of the system of Fig. 3.
Fig. 11 is a diagram illustrating the illumination control circuit of Fig. 10.
Fig. 12 is a flow chart of the operation of the processor of Fig. 10 according to the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
I . General Description
Fig. 1 is a simplified drawing of an eye 10, showing in detail the features of the iris 20 and pupil 30. Eye 10, as shown, is a human eye, but some other animals have a similar eye structure, permitting identification by the method of the invention. As shown, pupil 30 is a black, light receptive orifice, while iris 20 may have any of a broad range of hues and may be irregularly pigmented.
The iris 20 and pupil 30 are closely related in function. More precisely, one function of iris 20 is to control the size of pupil 30. Therefore, in addition to
the visible features of iris 20 relating to pigmentation, a number of visible features relate to the movements made by iris 20 in order to contract and dilate pupil 30. In general, iris 20 may be divided into the ciliary area 22, which is an annular region at its periphery, and the pupillary area 24, which is an annular area bordering pupil 30. In general, the tissue of iris 20 is soft and loosely woven, and the illumination which enters the pupil 30 and falls on the retina of eye 10 controls muscles in iris 20 causing the size of pupil 30 to grow larger or smaller to regulate the amount of light entering pupil 30. The change in the size of pupil 30 results from involuntary reflexes and is not under conscious control.
Among the pigment-related features of iris 20 are pigment spots 26, some types of which are known as moles, freckles, or nevi, and crypts 28. Pigment spots 26 are random concentrations of pigment-bearing cells in the visible surface of the iris 20, and generally appear in the ciliary area 22. The color of the pigment spots 26 may be nearly black -- darker than the darkest brown iris 20. In general, pigment spots 26 are quite stable in number, position and color. Crypts 28, unlike pigment spots 26, relate both to pigmentation and to the surface structure of the iris 20. Iris 20 includes an anterior layer which is directly visible and a posterior layer behind it, which is very darkly pigmented. The anterior layer ranges in pigment from a very light, almost white color to a dark brown color. If the anterior layer is very light or white, the iris 20 appears blue due to the pigment in the posterior layer, and if iris 20 is relatively light in color, the areas in which it is relatively thin will appear somewhat darker, due to the dark color of the posterior layer. These darkened areas may occur, for example, in the features referred to as crypts
28, which resemble sharply demarcated excavations and are typically located adjacent to the collarette 32, the boundary between ciliary area 22 and pupillary area 24.
In addition to the pigment-related features of the iris 20, several other visible features relate to its function of controlling the size of pupil 30. Collarette 32, discussed briefly above, is the boundary between the ciliary area 22 and the pupillary area 24, and is typically a sinuous line which forms an elevated ridge running roughly parallel with the margin of the pupil 30. The collarette 32 is the thickest part of iris 20.
Extending radially in relation to the center of pupil 30 are radial furrows 34. A typical radial furrow 34 may begin near pupil 30 and extend through collarette 32. Radial furrows 34 are creases in the anterior layer of iris 20, from which loose tissue may bulge outward, and it is this loose tissue which permits the iris 20 to expand or contract, changing the size of pupil 30. The positions of radial furrows 34 in relation to each other are stable.
Similarly, concentric furrows 36 are creases from which loose tissue may bulge outward, but their shape is generally circular, concentric with the pupil, so that they permit expansion and contraction of the iris in a different direction than radial furrows 34. Concentric furrows 36 typically appear in the ciliary area 22, near the periphery of iris 20.
Eye 10 may also have any of several rare anomalous visible features (not shown) unrelated to pigment or to controlling the size of pupil 30. Due to aging or trauma, atrophic areas may appear on iris 20, resulting in a "moth-eaten" texture. Tumors may grow on the iris. Congenital filaments may occur connecting the iris to the lens of the eye.
Finally, the most striking visible feature of eye 10 is typically pupil 30. As shown in Fig. 1, pupil 30 may not be exactly circular in shape, and its deviation from a circle is a visible characteristic. At the margin of pupil 30 is a protruding portion of the posterior layer of iris 20, the pigment frill 38. Pigment frill 38 typically has a very dark brown color.
The present invention, as discussed above, is based on the discovery that the iris of the eye may be used for identification. This discovery involves both the uniqueness of the iris and the stability of the iris over time. It has been discovered that every iris is unique, particularly in the detailed structure of the front or anterior layer. Not only are the irises of the eyes of identical twins different, but the iris of each eye of any person is different from that of his other eye. Furthermore, although specific details of the appearance of an iris vary dramatically, depending on level and direction of illumination, the basic, significant features of the iris remain extremely stable and do not change over a period of many years. Even features which do develop over time, such as the atrophic areas discussed above, usually develop rather slowly, so that an updated iris image will permit identification for a substantial length of time. A sudden or rapid change in such a feature may result in a failure to identify an individual, but this may alert the individual to the possibility of pathology of the eye. In general, the visible features of the iris and pupil may be used to identify an eye.
The method of the invention may be generally understood from Fig. 2, a flow chart showing the basic functions performed in identifying an eye from the visible features of the iris and pupil. As shown in block 40, the first function is to illuminate the eye. Because
the visible features of iris 20 and pupil 30 vary depending on illumination, the illumination of the eye will be performed to obtain a desired set of features. Then, in block 50, an image of iris 20 and pupil 30 is obtained, including the desired features. In block 60, this image is compared with stored image information in order to identify the eye.
Fig. 3 is a block diagram showing the major functional components of the system according to the invention. Illuminating means 70 provides light having desired characteristics to the eye 10 under observation. Eye 10 responds to the illumination provided, so that iris 20 and pupil 30 take a characteristic shape. In addition, iris 20 reflects light to imaging means 80, which obtains an image of iris 20 and pupil 30. This image is then compared with stored image information by comparing means 90 in order to identify the eye.
From these basic principles, a variety of methods and systems may be provided for identifying an eye from the iris and pupil.
II. Methods of Identification
A number of methods for implementing the invention are shown in Figs. 4-7. Each of these figures is a flow chart showing a sequence of steps performed in a process of identifying an eye from the iris and pupil.
Fig. 4 shows a basic method of performing the invention in which the pupil is driven to at least one predetermined size at which an image is taken. In step 42, the pupil is illuminated at one extreme of the range of illumination, either an extreme of darkness or of brightness. The extreme may be determined by the capabilities of the system or may be arrived at by a straightforward series of steps. For example, the illumination may gradually be increased from a moderate level
of illumination to an illumination beyond which further illumination does not decrease the size of the pupil. A more efficient approach, however, is to drive the pupil until its size is larger or smaller than the largest or smallest of the predetermined sizes at which an image will be obtained. From this extreme illumination, it will then be possible to drive the pupil to the predetermined size for imaging.
From the size which it has after step 42, the pupil is driven to the first predetermined size in step 44. Assuming that the extreme of illumination was darkness, the illumination will be increased in step 44 until the pupil contracts to a predetermined size. When the pupil reaches this size, an image is obtained and stored in step 52 for subsequent comparison. Then, in step 54, a test is run to determine whether the number n, a count of the images obtained, is greater than or equal to the number N of predetermined sizes at which images are to be obtained. If not, the count of the images obtained, n, is incremented in step 46, and steps 44, 52 and 54 are repeated for the next predetermined size. When all of the predetermined sizes have been imaged, step 54 will determine that n is equal to N , and image comparison will be done in step 62. In step 62, the image at each of the N predetermined sizes is compared with stored image information obtained from an eye illuminated to have the same predetermined pupil size.
Many variations may be made in the basic method of Fig. 4. For example, the largest size which the pupil attains as illumination is decreased could be treated as the first predetermined size, so that an image would be obtained after step 42. If the maximum size of the pupil remains relatively constant over time, this size will be valuable as a descriptor of the eye.
Fig. 5 illustrates another variation which can be made in the basic method of the invention. This variation includes measuring the rate at which the pupil size changes between two predetermined sizes. In step 142, light from a light source is directed onto the retina of the eye in order to control the size of the pupil through reflex action. Then, in step 144, the position or intensity of the light source is adjusted in order to obtain the first predetermined pupil size. In other words, as an alternative to varying the light intensity, the light source may be moved from a central position in which it directs light onto the central area of the retina toward an oblique position in which it directs light at an angle through the pupil onto the more peripheral parts of the retina, or vice versa. After each adjustment, in step 144, a test is performed in step 146 to determine whether the first predetermined size has been obtained. If not, the adjustment is again performed in step 144, and this loop is repeated until the first predetermined size is obtained.
After the first predetermined size is obtained, the illumination of the retina is changed at a predetermined rate in step 148 . The changing of the illumination is continued until the test in step 150 determines that a second predetermined size has been attained. At that point, the time necessary to go from the first predetermined size to the second predetermined size at the predetermined rate will provide a measure of the rate of change, in step 152. This measure could also be related to the change in size between the first and second predetermined size and to the specific predetermined rate at which the illumination is changed in step 148. The resulting value may then be used to identify the eye. It should be borne in mind, however, that the rate of change of pupil size may vary for individuals whose reflexes are
affected by emotion, drugs, and so forth, so that the rate of change of pupil diameter should generally provide only a secondary indication of identity.
The methods shown in Figs. 4 and 5 could be combined in a simple manner by making minor modifications to the steps performed in Fig. 4. Step 44 could first be performed to obtain the first predetermined size, corresponding to steps 144 and 146 in Fig. 5. After the performance of steps 52, 54 and 46, step 44 could again be performed by changing the illumination at the predetermined rate to obtain the second predetermined size, corresponding to steps 148 and 150 in Fig. 5. Then, in addition to obtaining the image in step 52, the rate of change of pupil size could also be determined in step 152.
Fig. 6 illustrates a routine which may similarly be included in step 52 of Fig. 4, to obtain an image of the iris of an eye in which the elevation-dependent features are shown with greater contrast. This is done by illuminating the iris from a plurality of oblique positions. The illumination from each position is provided by a relatively monochromatic light source to produce shadows of elevation-dependent features lacking the color of that light source. In step 250, illumination from position m is performed using the color m of the corresponding relatively monochromatic light source. This illumination will produce shadows lacking the color m, and an image of these shadows is obtained in step 252. Then, in step 254, a test is performed to determine whether m has reached M, the total number of positions from which the iris is to be illuminated. If not, m is incremented in step 256, and the loop including steps 250, 252 and 254 is repeated until m equals M. When m reaches M, theshadow images may be processed in step 258 to prepare them for subsequent comparison with reference images.
The method of Fig. 6 is especially beneficial for obtaining images of elevation-dependent features such as radial furrows 34. As can be seen from Fig. 1, however, radial furrows 34 may extend radially in any direction, so that illumination from any one direction will provide an excellent shadow image of radial furrows 34 which extend perpendicular to that direction of illumination. Radial furrows which extend parallel to the direction of illumination, however, will produce poor shadow images. Therefore , to produce the best images of radial furrows 34, it is desirable to obtain a number of shadow images and to combine them in processing step 258.
The comparison of images having the same pupil sizes in step 62 of Fig. 4 may be performed in many ways. One straightforward method would be to store the reference image as a pixel-by-pixel image resulting from a scanning camera such as a conventional video camera. The image obtained would then be obtained with a similar video camera, as discussed below in relation to Fig. 8, to obtain a pixel-by-pixel image of the eye. The two images could then be compared, pixel-by-pixel, to determine whether they were the .same image. The comparison could be made in a way which would take into account differences in the overall intensity of the two images. This method of comparison, although simple to implement, is extremely sensitive to registration. In other words, the test would only result in a conclusion that the two images were of the same eye if the eyes shown in the two images were of the same size and were positioned identically. Therefore, it will be preferable to perform a comparison algorithm which is not so sensitive to registration.
One comparison algorithm which would not be so sensitive to registration is the correlation algorithm, which has many variations. The effect of an offset in
the positioning of the two images could be eliminated based on an algorithm for determining the center of gravity of the pupil in each image. These algorithms could thus be combined to reduce registration sensitivity.
Another comparison method which would not be highly sensitive to registration would be a method using optical signal processing. Hecht, J. , "Light Modulators Help Crunch Image Data", High Technology, Vol. 5, No. 1 (January, 1985), p. 69-70, 72, discusses techniques for optical comparison which would be readily applicable to the present invention.
Fig. 7 illustrates another method of comparison which could be used in the present invention. In step 64, descriptors are extracted from the image obtained, which may be the image obtained in step 52 of Fig. 4. One example of descriptors which could be extracted would be the angular positions and lengths of the radial furrows 34. Another set of descriptors would be the shape of the pupil 30 or of the pigment frill 38. Other descriptor sets could be obtained relating to the positions and sizes of pigment spots 26, crypts 28 and anomalous features such as atrophic areas, tumors and congenital filaments. Similarly, generally circular features such as collarettes 32 and concentric furrows 36 could be described by respective descriptors. All of these descriptors would then be compared with the descriptors of a reference image in step 66. Because several features. of the iris 20 and pupil 30 lend themselves readily to mathematical description, this method of comparison should be effective and should not be sensitive to registration. A number of algorithms for extracting descriptors are discussed below.
Many types of equipment could be used to implement the invention and to perform the methods set forth above. One example of circuitry for performing these functions is described below.
III. Identification Systems
As noted above, the necessary equipment for performing iris recognition according to the invention could take many forms. The major functional components of any such system, however, will be those shown in Fig. 3, including illuminating means 70, imaging means 80 and comparing means 90.
Fig. 8 shows a system for illuminating and obtaining an image of the eye. The system includes a simplified electrical circuit for controlling illumination and an optical arrangement in which the image of fixation target 71 seen by eye 10 and the reflected image of iris 20 and pupil 30 to camera 84 are approximately coaxial. It is presently preferred to provide a target image, upon which eye 10 may be fixated, concentric with visual axis 16 for reference purposes. It is also preferred for camera 84 to capture the reflected image concentric or nearly concentric with visual axis 16, because if the image were obtained far from the axis 16, it would be distorted, which would then have to be compensated by appropriate processing. Additional illumination of iris 20 is provided obliquely to create good shadows of elevation-dependent features.
In the system of Fig. 8, the source of light is a filament 72a which receives power from power source 72b connected in series through variable resistor 72c, which may be an electrically controlled potentiometer. The light emitted by filament 72a is collimated by lens 73 and illuminates target 71, which may have any appropriate shape.
The image of target 71 is then focused by movable lens 74 and transmitted through monitor beam-splitter 76 onto axial beam-splitting plane 82. The image is then reflected from the central mirrored area 82a through pupil 30 onto retina 12.
The optical system shown in Fig. 8 includes means for placing target 71 at any desired optical distance from eye 10. Target 71 can thus be made visible to subjects with a wide variety of refractive states. The refraction or bending of light rays at the air interface of cornea 14, for example, is substantial and varies for different subjects. The necessary adjustment of optical distance is obtained by moving lens 74 toward or away from target 71 until target 71 is optically conjugate with retina 12. If necessary to compensate astigmatism, lens 74 may also be a variable sphero-cylindrical lens system of any of the well-known types. Target 71 will be optically conjugate with retina 12 when, as shown in Fig. 8, the light from each point on target 71 comes to a focus at a corresponding point on retina 12. The convergence or divergence of light from target 71 due to the adjustment of lens 74 will thus compensate for the optics of eye 10.
Oblique light sources 78a and 78b provide illumination to the iris 20 and may be controlled similarly to filament 72a by using a variable resistor (not shown). The resulting reflection will depend on the positioning of oblique light sources 78a and 78b, because the iris is a diffuse reflector, reflecting light according to a cosine law in which the amount of reflected light in the direction normal to the iris varies as the cosine of the incident angle of the light.
Axial beam-splitting plane 82 serves the two functions of directing the image of target 71 from lens 74 toward eye 10 and also transmitting the reflected image of iris 20 to camera 84. In the arrangement shown in Fig. 8, which is not to scale, beam splitting plane 82 is a thin transparent sheet such as glass on which mirror 82a is. formed for reflecting the image. Mirror 82a may be a microscopically thin, reflective film or coating,
such as a metal, on the central part of plane 82 and on the side toward eye 10, to provide a first surface mirror. It may, for example, have an elliptical shape. The reflected image of target 71 is directed through pupil 30 onto the retina 12 of eye 10 to control the size of pupil 30. The reflected light from iris 20, which will emerge from eye 10 quite divergently, may pass through the clear, transparent area of beam splitting plane 82 which surrounds mirror 82a. Mirror 82a, however, will create a non-transparent area in the center of plane 82, preventing an intense reflection of the target image from cornea 14 from reaching camera 84. Since the reflected target image may be the brightest object in the field of view of camera 84, this may significantly reduce the noise in the image of iris 20 and pupil 30. Monitor beam-splitter 76 will provide an image of the intense reflection from cornea 14, as reflected by mirror 82a, and this reflection image may be used to position the eye according to known eye-positioning techniques. Proper positioning of the eye will help to provide a standardized image of iris 20 and pupil 30 and will ensure that the intense reflection of the target image does not reach camera 84.
Fig. 9 shows viewing screen 86 of camera 84, upon which are marked peripheral points 88a and central point 88b. Peripheral points 88a may be used to measure the size of pupil 30, as shown. When the perimeter of pupil 30 touches each of the peripheral points 88a, pupil 30 has been driven to a predetermined size at which an image should be obtained. Central point 88b may be useful in aligning pupil 30 so that size may be properly tested. As shown in Fig. 9, a grid may also be provided on viewing screen 86 to assist in the determination of the size of pupil 30 when it does not meet peripheral
points 88a. Alternatively, concentric markings could be provided on screen 86, as disclosed in U.S. Patent 3,966,310, issued to Larson.
It will be apparent from Figs. 8 and 9 that the distance between camera 84 and eye 10 must be fixed in order to provide a standardized image. This can be done by providing a headrest (not shown) against which the forehead above eye 10 may be positioned. The headrest may also be equipped with appropriate additional surfaces for darkening a space around eye 10. This permits the system to provide a condition of nearly zero illumination to retina 12.
The system shown in Figs. 8 and 9 may be manually operated to obtain the predetermined pupil size touching peripheral points 88a. First, variable resistor 72c is manually adjusted to its maximum resistance, effectively turning off filament 72a. Oblique light sources 78a and 78b are activated to provide a desired level of illumination upon iris 20. Then, variable resistor 72c is gradually turned to a lower resistance until the image of target 71 begins to illuminate retina 12. The operator, looking at viewing screen 86, will see the pupil beginning to decrease in size as the illumination from filament 72a increases. When the perimeter of pupil 30, as seen through viewing screen 86, just touches peripheral points 88a, the operator may activate camera 84 to obtain an image of iris 20 and pupil 30. If further, predetermined sizes are marked by points in viewing screen 86, the operator may continue to adjust variable resistor 72c until those additional predetermined sizes of pupil 30 are shown in viewing screen 86 and are imaged when camera 84 is activated.
The optics of Fig. 8 cold be modified in many ways within the scope of the invention. For example, rather than providing a transparent beam-splitting plane
82 with a centered mirror 82a, the beam-splitting plane could be defined by two prisms positioned against each other or by a half-silvered mirror or other partially reflecting surface. Such arrangements, however, will result in a greater loss of intensity of the collimated light and of the reflected image than the arrangement of
Fig. 8. If such an arrangement is desirable, therefore, it may be appropriate to provide additional illumination of the iris 20, as discussed below in relation to Fig. 11.
Fig. 10 shows a block diagram of electrical components performing the functions set forth in Fig. 3. Iris recognition system 100 in Fig. 10 includes illumination control circuit 170, camera 180 and processor 190. Illumination control circuit 170 and camera 180 operate under the control of processor 190. In response to control signals, illumination control circuit 170 illuminates eye 10 both to control the size of pupil 30 and to provide a reflected image of iris 20 and pupil 30. This reflected image is received by camera 180, which may be a video camera which converts the reflected image into an image signal for processor 190. Processor 190 then compares the image signal with a reference image in order to identify the human being.
Fig. 11 shows in greater detail the illumination system of iris recognition system 100. As shown in Fig. 11, illumination control circuit 170 controls light sources 172, 174, 176 and 178. In general, these light sources should be as small as possible to minimize noise in the reflected image from eye 10. Light source 172, which illuminates retina 12, as discussed in relation to Fig. 8, may do so through a fixation target 71 and a beam-splitting plane 82. Light source 172 will thus provide a target image upon which eye 10 will fix and focus, thereby becoming aligned for imaging. The target image
should preferably be a distant object to discourage accommodation which would reduce pupil size. Rays from the target image may be relatively tightly collimated and subtend a small angle in order to provide a more critical, centralized alignment, or may be less tightly collimated if additional iris illumination is desired. The target image could also be compound, such as cross hairs in one plane which must be aligned on a designated part of a background at another plane to obtain alignment of the pupil, although this would depend on the acuity of the eye 10. Alternatively, a polarizing ring-sight device could be used to create concentric circles on the retina independent of the refraction of eye 10. The target image should not be extremely bright, because that would also reduce the size of pupil 30.
As mentioned above, the target image may be used to cause the subject to align pupil 30 for optimal imaging. In addition, however, it would be possible to provide an automatic alignment system using a feedback circuit and a servo-control mechanism to adjust the relative positions of the iris 20 and the camera 180. This adjustment of position could be accomplished with rotatable mirrors, piano-parallel plates, or prismatic devices such as Risley prisms . Alternatively , the head rest (not shown) can be moved in order to orient the head.
In addition to pupil-driving light source 172, the illumination system shown in Fig. 11 includes oblique light sources 174, 176 and 178. As discussed above in relation to Fig. 6, each of these oblique light sources 174, 176 and 178 may provide a relatively monochromatic beam of light, so that the shadows resulting from the oblique positioning of these light sources will lack the corresponding color, facilitating the subsequent proces
sing of the shadow images. Each source 174, 176 and 178 could be a relatively monochromatic source such as an LED or a solid state laser.
Although Fig. 11 shows an arrangement in which three light sources are provided at approximately 120° angles from each other, the number of light sources and their relative angular positions could be varied in many ways. In addition, other arrangements could be used to obtain additional information about elevation-dependent features of iris 20. For example, more than one camera 180 could be provided, or a single camera 180 could be provided which could be moved between a plurality of positions in order to obtain three-dimensional information. Any other means for obtaining an image could be used, including laser-based techniques such as holography and including flying spot scanning by mechanical or electrooptical techniques.
Fig. 12 shows a flow chart of the operation of processor 190 in Fig. 10. Processor 190 will at first be initialized in step 342, in order to provide appropriate memory contents and in order to perform the necessary loading of programs. In addition, as a part of initialization, processor 190 may provide control signals causing camera 180 to begin to operate and providing power to illumination control circuit 170. In step 344, processor 190 provides control signals to illumination control circuit 170 causing pupil-driving light source 172 to illuminate the retina 12 at the extreme illumination of the range available. In addition, oblique light sources 174, 176 and 178 may be activated to provide sufficient illumination so that camera 180 is able to provide an image of pupil 30 to processor 190.
When the extreme size of the pupil has been obtained in step 344, the pupil is then driven across the range of its size in step 346. This is performed by
gradually increasing the illumination from pupil-driving light source 172. After each increase in intensity, the pupil size is measured in step 348 and tested in step 352 to determine whether it equals the desired predetermined size xn. For the first iteration, the first predetermined size x1 will be obtained after the necessary reitteration of the loop including steps 346, 348 and 352. Oblique light sources 374, 376 and 378 will also be activated, either at the same time or in sequence according to a method like that shown in Fig. 6, in order to obtain an image of the iris 20 and pupil 30 of eye 10. Oblique light sources 374, 376 and 378 should be activated in a way which minimizes the constriction of pupil 30. Therefore, it may be necessary to activate them before or while the illumination from light source 172 is increased.
From this image, processor 190 will then extract the nth descriptor set in step 362. For the image obtained when the pupil size is x1, the first descriptor set will be obtained. Then, in step 364, n is tested to determined if it is equal to N, the total number of descriptor sets to be obtained. If not, n is incremented in step 366, and the loop including steps 346, 348, 352, 354, 362 and 364 is repeated for the next value of n. This loop will be repeated until step 364. determines that n equals N.
When n equals N, the resulting descriptor sets are compared with one or more reference descriptor sets in step 368. If the objective is to determine whether the eye 10 is that of one of a population of people, each descriptor set resulting from step 362 will be compared with the corresponding descriptor sets of the members of that population. On the other hand, if the objective is simply to confirm that an eye 10 is that of a specific individual, it is only necessary to compare the descrip
tor sets resulting from step 362 with the corresponding descriptor set of that individual, which may, for example, be stored on a credit card, identification card, or other identifying document, or in the memory of a computer system to which the processor 190 has access.
Upon the completion of step 368 , an appropriate indication may be provided to the operator of iris recognition system 100 that the eye 10 has been identified or has not been identified, completing the operation shown in Fig. 12. At this point, the operator may choose to initiate further iris recognition by repeating the process shown in Fig. 12.
Processor 190, in extracting descriptors in step 362, may perform a wide variety of algorithms. A number of such algorithms are set forth in Ballard, D.H. and Brown, CM., Computer Vision. Prentice-Hall, Englewood Cliffs, New Jersey, 1982; Duda; R.O. and Hart, P.E., Pattern Recognition and Scene Analysis, Wiley, New York, 1973; and Rosenfeld, A., Digital Picture Processing, Academic Press, New York, 1982. Many of these algorithms use classical mathematical techniques such as difference operators, taking a differential equation of selected points and testing a rate of change, and mask matching or surface fitting, fitting to a known shape at various orientations. Although these and other such algorithms are known, they have not previously been applied to the specific features of the iris and pupil.
The location, size and deviation from circularity of the pupil could provide a number of descriptors. The pupil could first be detected using a boundary detection algorithm or an edge detection algorithm, which may, for example, detect an abrupt gray level change. Systems performing pupil size measuring algorithms are disclosed, for example, in U.S. Patent 3,598,107, issued to Ishikawa et al, and U.S. Patent 3,533,683, issued to Stark et al.
A Hough transform algorithm for detecting circles, which maps curves into the transform space according to characteristics such as curvature, could be used on a boundary or edge which is detected. Alternatively, the received image could be subjected to a thresholding algorithm after which a region-growing or aggregation algorithm, such as the blob coloring algorithm discussed by Ballard and Brown, could be performed to find the largest connected region of pixels with intensity values below threshold. Since the pupil is central, the region could be grown outward from a central dark pixel in the image, progressively aggregating the adjacent dark pixels until the pupil boundary is reached, beyond which adjacent pixels will not be dark. This will provide a measure both of pupil size and of location, as the center of the pupil may be determined from its boundary. This may be done, for example, by taking slices through the dark region in a number of radial directions to select a largest diameter. To eliminate the chance that the largest diameter is not the true diameter, as, for example, if it extended to a notch or other irregularity in the boundary, the least sum of squares from its center to the boundary could be taken and compared with that of the second largest diameter. The center and radius giving the least sum of squares would define the circle most closely approximating the pupil boundary, and the residual sum of squares would be a quantitative measure of deviation of the pupil from circularity, a "figure of merit" for the pupil. This figure of merit would be a useful descriptor of the pupil.
The characteristics of the radial furrows could also be used to obtain descriptors. From the center of the pupil, the precision of radiality of the radial furrows can be obtained. Line detection algorithms and streak or thick line detection algorithms could be used
to locate the radial furrows. The radial furrows may then be compared with true straight lines radiating from the center by obtaining a sum of squares of differences, indicating radiality. Some furrows, rather than being entirely straight, include, for example, two offset parts, each of which is generally radial, the two being connected by a non-radial jog. A measure of thickness could be obtained for such a furrow by determining how wide a streak is necessary to cover the furrow. Also, the Hough transform algorithm for detecting straight lines could be applied to obtain such descriptors as the length, angular location and direction of radial furrows.
The concentric furrows may be detected and described using edge detection algorithms and the Hough transform algorithms for detecting curved lines. Similarly, the collarette and the contours of the crypts along it could be described using edge detection algorithms and chain coding algorithms. In addition, points of high curvature along these contours could be located and represented for matching purposes, which may be more efficient. Isolated crypts could be described by a thresholding algorithm, followed by a region-growing algorithm and then algorithms obtaining the center of gravity and the lower order central moments., as described by Duda and Hart.
Spot detection algorithms, detecting high contrast spikes in areas of relatively constant gray level, could detect small pigment spots. More generally, algorithms like those for measuring the pupil could be used to obtain location and lower-order central moments of pigement spots, and other such features. Also, high-curvature points along the contour of a feature could be detected and described.
Finally, the color of the iris generally could be found by an algorithm obtaining a histogram in a three-dimensional RGB (Red-Green-Blue) color space, as described by Ballard and Brown at page 153. The peak in the histogram will provide a descriptor of color.
As noted above, the embodiment of Figs. 8-12 could be modified in many ways within the scope of the invention. Processor 190 may control the operation of illumination control circuit 170 and camera 180 in any appropriate manner to obtain images for processing. For example, the image from camera 180 may be constantly received and processed by processor 190 to determine the appropriate level of illumination by illumination control circuit 170. At the same time, when it is desired to obtain an image at one of the predetermined sizes of pupil 30, processor 190 may control illumination means 170 to maintain a temporarily constant level of illumination while camera 180 obtains a complete image for processing.
Although the present invention has been described in connection with a plurality of preferred embodiments thereof, many other variations and modifications will now become apparent to those skilled in the art. It is preferred, therefore, that the present invention be limited not by the specific disclosure herein, but only by the appended claims.
Claims
1. A method of identification, comprising: illuminating an eye having an iris and a pupil; obtaining at least one image of at least part of the iris and pupil of the eye; and comparing the obtained image with stored image information to identify the eye.
2. The method of claim 1 in which illuminating comprises driving the pupil of the eye to at least one predetermined size, comparing the obtained image comprising comparing the obtained image with stored image information obtained from an eye with its pupil having the same predetermined size.
3. The method of claim 2 in which driving the pupil comprises directing light onto the retina of the eye and adjusting the intensity of the light until the pupil reaches the predetermined size.
4. The method of claim 2 in which driving the pupil comprises driving the pupil to a plurality of the predetermined sizes; the step of obtaining at least one image comprising obtaining a respective image at each of the predetermined sizes of the pupil.
5. The method of claim 4 in which illuminating comprises changing the intensity of illumination at a predetermined rate between at least two of the predetermined pupil sizes and measuring a rate of change of pupil size as the illumination intensity changes.
6. The method of claim 1 in which illuminating comprises illuminating the iris obliquely for providing shadows of elevation-dependent features.
7. The method of claim 6 in which illuminating the iris obliquely comprises illuminating the iris from a plurality of spatially separated light sources for providing shadows of elevation dependent features extending in a plurality of directions.
8. The method of claim 7 in which illuminating the iris from a plurality of spatially separated light sources comprises illuminating the iris from a plurality of relatively monochromatic light sources, each light source providing shadows lacking a corresponding color of light.
9. The method of claim 1 in which comparing the obtained image with stored image information comprises comparing the obtained image with a stored reference image for identifying the eye.
10. The method of claim 1 in which comparing the obtained image with stored image information comprises deriving a set of descriptors of the obtained image and comparing the derived descriptors with stored reference descriptors derived from a previous image for identifying the eye.
11. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of the pupil.
12. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of a pigment frill around the pupil.
13. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of a collarette in the iris.
14. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one concentric furrow in the iris.
15. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one radial furrow in the iris.
16. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one crypt in the iris.
17. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one pigment spot in the iris.
18. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one atrophic area in the iris.
19. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one tumor in the iris.
20. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one congenital filament in the iris.
21. A system for identification, comprising: means for illuminating an eye having an iris and a pupil; means for obtaining an image of at least part of the iris and pupil of the eye; and means for comparing the obtained image with stored image information to identify the eye.
22. The system of claim 21 in which the means for illuminating comprises an adjustable light source for illuminating the eye at a selected value in a range of light intensity.
23. The system of claim 22 in which the means for illuminating further comprises control means responsive to the obtained image of the eye for detecting the size of the pupil and operable for adjusting the adjustable light source to a selected light intensity for controlling the size of the pupil.
24. The system of claim 23 in which the control means is further operable for controlling the size of the pupil to obtain a plurality of predetermined pupil sizes; the control means being further operable for adjusting the light source at a predetermined rate across the range of light intensity between at least two of the predetermined pupil sizes for measuring the rate of change of pupil sizes as the light intensity is adjusted.
25. The system of claim 21 in which the illuminating means further comprises a plurality of spatially separated oblique light sources for illuminating the iris obliquely.
26. The system of claim 25 in which each of the oblique light sources comprises a relatively monochromatic light source for providing shadows lacking a corresponding color of light.
27. The system of claim 26 in which each relatively monochromatic light source comprises a laser.
28. The system of claim 21 in which the image obtaining means comprises a camera, the comparing means comprising a processor responsive to an output from the camera for comparing the obtained image with stored image information.
29. The system of claim 28 in which the processor comprises storage means for storing reference descriptors derived from a previous image, the processor being operable for deriving a set of descriptors of the obtained image and for comparing the derived descriptors with the reference descriptors for identifying the eye.
30. The system of claim 21 in which the image obtaining means comprises a laser.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR8605561A BR8605561A (en) | 1985-02-20 | 1986-02-04 | IRIS IDENTIFICATION PROCESS AND SYSTEM |
DE8686901250T DE3680618D1 (en) | 1985-02-20 | 1986-02-04 | IRIS DETECTION SYSTEM. |
AT86901250T ATE65851T1 (en) | 1985-02-20 | 1986-02-04 | IRIS RECOGNITION SYSTEM. |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US703,312 | 1985-02-20 | ||
US06/703,312 US4641349A (en) | 1985-02-20 | 1985-02-20 | Iris recognition system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1986005018A1 true WO1986005018A1 (en) | 1986-08-28 |
Family
ID=24824891
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US1986/000227 WO1986005018A1 (en) | 1985-02-20 | 1986-02-04 | Iris recognition system |
Country Status (10)
Country | Link |
---|---|
US (1) | US4641349A (en) |
EP (1) | EP0215818B1 (en) |
JP (1) | JPS62501889A (en) |
AT (1) | ATE65851T1 (en) |
BR (1) | BR8605561A (en) |
CA (1) | CA1244552A (en) |
DE (1) | DE3680618D1 (en) |
IL (1) | IL77920A (en) |
MX (1) | MX163339A (en) |
WO (1) | WO1986005018A1 (en) |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2622786A1 (en) * | 1987-11-10 | 1989-05-12 | Giet Francis | APPARATUS FOR EYE EXAMINATION BY IRISCOPY AND IRIGRAPHY |
EP0569291A1 (en) * | 1992-05-06 | 1993-11-10 | Parfums Christian Dior | Procedure for the making-up of a face, in particular the eyes and a device for carrying out the same |
WO1994010658A1 (en) * | 1992-11-05 | 1994-05-11 | Coms21 Limited | Secure access control system |
DE4414216C1 (en) * | 1994-04-23 | 1995-04-06 | Daimler Benz Ag | Device for protecting a motor vehicle against use by third parties, with personalisation of the driving authorisation |
WO1997046979A1 (en) * | 1996-06-06 | 1997-12-11 | British Telecommunications Public Limited Company | Personal identification |
EP0821912A2 (en) * | 1996-07-25 | 1998-02-04 | Oki Electric Industry Co., Ltd. | Animal body identifying device and body identifying system |
EP0872814A1 (en) * | 1997-04-15 | 1998-10-21 | BRITISH TELECOMMUNICATIONS public limited company | Optical apparatus |
WO1999027844A1 (en) * | 1997-12-01 | 1999-06-10 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
US5954583A (en) * | 1992-11-05 | 1999-09-21 | Com21 Limited | Secure access control system |
WO2002087442A1 (en) * | 2001-04-27 | 2002-11-07 | Bausch & Lomb Incorporated | Iris pattern recognition and alignment |
DE10194154B4 (en) * | 2000-10-07 | 2005-12-22 | Qritek Co., Ltd. | An iris identification system and method and computer readable storage medium having computer executable instructions stored therein for implementing the iris identification method |
US7623686B2 (en) | 2004-05-10 | 2009-11-24 | Panasonic Corporation | Techniques and apparatus for increasing accuracy of iris authentication by utilizing a plurality of iris images |
US7796784B2 (en) | 2002-11-07 | 2010-09-14 | Panasonic Corporation | Personal authentication method for certificating individual iris |
US9002053B2 (en) | 2011-10-07 | 2015-04-07 | Irisguard Inc. | Iris recognition systems |
US9008375B2 (en) | 2011-10-07 | 2015-04-14 | Irisguard Inc. | Security improvements for iris recognition systems |
EP2265231B1 (en) * | 2008-04-02 | 2015-08-12 | Refocus Group, Inc. | System and method for identifying a position to insert a scleral prosthesis into an eye |
EP2916262A1 (en) * | 2014-03-06 | 2015-09-09 | Sony Corporation | Information processing apparatus, information processing method, eyewear terminal, and authentication system |
WO2017062235A3 (en) * | 2015-10-08 | 2017-05-26 | Microsoft Technology Licensing, Llc | Iris recognition |
US9977977B2 (en) | 2014-09-30 | 2018-05-22 | Qualcomm Incorporated | Apparatus and method for low-power object-detection in images using computer vision feature computation hardware |
US9986211B2 (en) | 2014-09-30 | 2018-05-29 | Qualcomm Incorporated | Low-power always-on face detection, tracking, recognition and/or analysis using events-based vision sensor |
WO2018111528A1 (en) * | 2016-12-16 | 2018-06-21 | Qualcomm Incorporated | Light source modulation for iris size adjustment |
EP3223232A4 (en) * | 2014-11-20 | 2018-07-18 | Sony Corporation | Control system, information processing device, control method, and program |
US10515284B2 (en) | 2014-09-30 | 2019-12-24 | Qualcomm Incorporated | Single-processor computer vision hardware control and application execution |
US10984235B2 (en) | 2016-12-16 | 2021-04-20 | Qualcomm Incorporated | Low power data generation for iris-related detection and authentication |
US11068712B2 (en) | 2014-09-30 | 2021-07-20 | Qualcomm Incorporated | Low-power iris scan initialization |
Families Citing this family (343)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4854329A (en) * | 1987-07-21 | 1989-08-08 | Walruff James C | Apparatus and method for noninvasive testing of voluntary and involuntary motor response patterns |
US5106183A (en) * | 1987-11-25 | 1992-04-21 | Taunton Technologies, Inc. | Topography measuring apparatus |
US4902123A (en) * | 1987-11-25 | 1990-02-20 | Taunton Technologies, Inc. | Topography measuring apparatus |
US4998819A (en) * | 1987-11-25 | 1991-03-12 | Taunton Technologies, Inc. | Topography measuring apparatus |
US5016282A (en) * | 1988-07-14 | 1991-05-14 | Atr Communication Systems Research Laboratories | Eye tracking image pickup apparatus for separating noise from feature portions |
US4931865A (en) * | 1988-08-24 | 1990-06-05 | Sebastiano Scarampi | Apparatus and methods for monitoring television viewers |
US5016173A (en) * | 1989-04-13 | 1991-05-14 | Vanguard Imaging Ltd. | Apparatus and method for monitoring visually accessible surfaces of the body |
DE4102895C1 (en) * | 1991-01-31 | 1992-01-30 | Siemens Ag, 8000 Muenchen, De | |
US5231580A (en) * | 1991-04-01 | 1993-07-27 | The United States Of America As Represented By The Secretary Of The Department Of Health And Human Services | Automated method and apparatus for determining characteristics of nerve fibers |
US5291560A (en) * | 1991-07-15 | 1994-03-01 | Iri Scan Incorporated | Biometric personal identification system based on iris analysis |
US5437278A (en) * | 1992-01-10 | 1995-08-01 | Wilk; Peter J. | Medical diagnosis system and method |
WO1993014454A1 (en) * | 1992-01-10 | 1993-07-22 | Foster-Miller, Inc. | A sensory integrated data interface |
US5360005A (en) * | 1992-01-10 | 1994-11-01 | Wilk Peter J | Medical diagnosis device for sensing cardiac activity and blood flow |
US5359669A (en) * | 1992-04-13 | 1994-10-25 | Motorola, Inc. | Remote retinal scan identifier |
US5331544A (en) * | 1992-04-23 | 1994-07-19 | A. C. Nielsen Company | Market research method and system for collecting retail store and shopper market research data |
JPH0694979A (en) * | 1992-09-14 | 1994-04-08 | Nikon Corp | Camera provided with device for detecting |
KR100307792B1 (en) * | 1992-10-10 | 2001-11-30 | 존 이. 시드라즈 | Biometric measurement based on iris analysis |
DE4421242B4 (en) * | 1993-06-17 | 2005-12-08 | Pentax Corp. | Image capture device for an identification system |
US5471542A (en) * | 1993-09-27 | 1995-11-28 | Ragland; Richard R. | Point-of-gaze tracker |
DE69501327T3 (en) * | 1994-03-24 | 2005-12-22 | Minnesota Mining And Mfg. Co., Saint Paul | Biometric personal authentication system |
EP1548637A3 (en) * | 1994-08-25 | 2008-01-02 | Neotec Holdings Limited | Method and apparatus for providing identification |
AU719428B2 (en) * | 1994-09-02 | 2000-05-11 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US6714665B1 (en) * | 1994-09-02 | 2004-03-30 | Sarnoff Corporation | Fully automated iris recognition system utilizing wide and narrow fields of view |
US5572596A (en) * | 1994-09-02 | 1996-11-05 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US5842194A (en) * | 1995-07-28 | 1998-11-24 | Mitsubishi Denki Kabushiki Kaisha | Method of recognizing images of faces or general images using fuzzy combination of multiple resolutions |
US5729619A (en) * | 1995-08-08 | 1998-03-17 | Northrop Grumman Corporation | Operator identity, intoxication and drowsiness monitoring system and method |
JPH09134430A (en) * | 1995-11-08 | 1997-05-20 | Oki Electric Ind Co Ltd | Iris data collating system |
JP3427873B2 (en) * | 1995-11-17 | 2003-07-22 | 沖電気工業株式会社 | Automatic transaction device and automatic transaction system |
JP3625941B2 (en) * | 1996-01-30 | 2005-03-02 | 沖電気工業株式会社 | Iris recognition system |
JPH09212644A (en) * | 1996-02-07 | 1997-08-15 | Oki Electric Ind Co Ltd | Iris recognition device and iris recognition method |
US6259827B1 (en) | 1996-03-21 | 2001-07-10 | Cognex Corporation | Machine vision methods for enhancing the contrast between an object and its background using multiple on-axis images |
JP3580634B2 (en) * | 1996-04-19 | 2004-10-27 | 沖電気工業株式会社 | Automatic transaction equipment |
US6320610B1 (en) | 1998-12-31 | 2001-11-20 | Sensar, Inc. | Compact imaging device incorporating rotatably mounted cameras |
JP3575915B2 (en) * | 1996-05-15 | 2004-10-13 | 沖電気工業株式会社 | Automatic transaction equipment |
US5717512A (en) * | 1996-05-15 | 1998-02-10 | Chmielewski, Jr.; Thomas A. | Compact image steering and focusing device |
JP3337913B2 (en) * | 1996-06-19 | 2002-10-28 | 沖電気工業株式会社 | Iris imaging method and imaging device thereof |
JP3751368B2 (en) * | 1996-06-28 | 2006-03-01 | 沖電気工業株式会社 | Iris recognition system and iris recognition device |
US5876926A (en) | 1996-07-23 | 1999-03-02 | Beecham; James E. | Method, apparatus and system for verification of human medical data |
JP2002514098A (en) | 1996-08-25 | 2002-05-14 | センサー インコーポレイテッド | Device for iris acquisition image |
JP3587635B2 (en) * | 1996-11-15 | 2004-11-10 | 沖電気工業株式会社 | Personal recognition device using iris and automatic transaction system using this personal recognition device |
US6075881A (en) * | 1997-03-18 | 2000-06-13 | Cognex Corporation | Machine vision methods for identifying collinear sets of points from an image |
US5974169A (en) * | 1997-03-20 | 1999-10-26 | Cognex Corporation | Machine vision methods for determining characteristics of an object using boundary points and bounding regions |
US6229905B1 (en) * | 1997-03-26 | 2001-05-08 | Oki Electric Industry Co., Ltd. | Animal identification based on irial granule analysis |
US6215891B1 (en) * | 1997-03-26 | 2001-04-10 | Oki Electric Industry Co., Ltd. | Eye image recognition method eye image selection method and system therefor |
US6144754A (en) * | 1997-03-28 | 2000-11-07 | Oki Electric Industry Co., Ltd. | Method and apparatus for identifying individuals |
US6285780B1 (en) * | 1997-03-28 | 2001-09-04 | Oki Electric Industry Co., Ltd. | Apparatus for identifying individual animals and image processing method |
WO1998050875A2 (en) * | 1997-05-09 | 1998-11-12 | Gte Government Systems Corporation | Biometric certificates |
US6202151B1 (en) | 1997-05-09 | 2001-03-13 | Gte Service Corporation | System and method for authenticating electronic transactions using biometric certificates |
US5867587A (en) * | 1997-05-19 | 1999-02-02 | Northrop Grumman Corporation | Impaired operator detection and warning system employing eyeblink analysis |
US6373968B2 (en) | 1997-06-06 | 2002-04-16 | Oki Electric Industry Co., Ltd. | System for identifying individuals |
US6608647B1 (en) | 1997-06-24 | 2003-08-19 | Cognex Corporation | Methods and apparatus for charge coupled device image acquisition with independent integration and readout |
US6119096A (en) * | 1997-07-31 | 2000-09-12 | Eyeticket Corporation | System and method for aircraft passenger check-in and boarding using iris recognition |
JP2985839B2 (en) * | 1997-08-04 | 1999-12-06 | 日本電気株式会社 | Biometric matching method and apparatus, information storage medium |
JPH1169219A (en) * | 1997-08-21 | 1999-03-09 | Victor Co Of Japan Ltd | Video camera for iris photographing |
US6185318B1 (en) * | 1997-08-22 | 2001-02-06 | International Business Machines Corporation | System and method for matching (fingerprint) images an aligned string-based representation |
US5892837A (en) * | 1997-08-29 | 1999-04-06 | Eastman Kodak Company | Computer program product for locating objects in an image |
GB9721506D0 (en) * | 1997-10-10 | 1997-12-10 | Virulite Limited | Treatment of diseases |
US6072494A (en) * | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
US6532022B1 (en) * | 1997-10-15 | 2003-03-11 | Electric Planet, Inc. | Method and apparatus for model-based compositing |
US6031934A (en) * | 1997-10-15 | 2000-02-29 | Electric Planet, Inc. | Computer vision system for subject characterization |
US6007202A (en) * | 1997-10-23 | 1999-12-28 | Lasersight Technologies, Inc. | Eye illumination system and method |
EP0910986A1 (en) | 1997-10-24 | 1999-04-28 | BRITISH TELECOMMUNICATIONS public limited company | Imaging apparatus |
US6069967A (en) * | 1997-11-04 | 2000-05-30 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses |
US6064752A (en) * | 1997-11-04 | 2000-05-16 | Sensar, Inc. | Method and apparatus for positioning subjects before a single camera |
US6021210A (en) | 1997-12-01 | 2000-02-01 | Sensar, Inc. | Image subtraction to remove ambient illumination |
US6028949A (en) * | 1997-12-02 | 2000-02-22 | Mckendall; Raymond A. | Method of verifying the presence of an eye in a close-up image |
US5953440A (en) | 1997-12-02 | 1999-09-14 | Sensar, Inc. | Method of measuring the focus of close-up images of eyes |
US6850631B1 (en) | 1998-02-20 | 2005-02-01 | Oki Electric Industry Co., Ltd. | Photographing device, iris input device and iris image input method |
US6381375B1 (en) | 1998-02-20 | 2002-04-30 | Cognex Corporation | Methods and apparatus for generating a projection of an image |
US5978494A (en) * | 1998-03-04 | 1999-11-02 | Sensar, Inc. | Method of selecting the best enroll image for personal identification |
JP3271750B2 (en) | 1998-03-05 | 2002-04-08 | 沖電気工業株式会社 | Iris identification code extraction method and device, iris recognition method and device, data encryption device |
JP3868621B2 (en) | 1998-03-17 | 2007-01-17 | 株式会社東芝 | Image acquisition apparatus, image acquisition method, and recording medium |
US6199985B1 (en) | 1998-05-15 | 2001-03-13 | Christopher Scott Anderson | Pupilometer methods and apparatus |
US5956122A (en) * | 1998-06-26 | 1999-09-21 | Litton Systems, Inc | Iris recognition apparatus and method |
US20040208343A1 (en) * | 1998-07-09 | 2004-10-21 | Colorado State University Research Foundation | Apparatus and method for creating a record using biometric information |
JP3315648B2 (en) | 1998-07-17 | 2002-08-19 | 沖電気工業株式会社 | Iris code generation device and iris recognition system |
JP3610234B2 (en) | 1998-07-17 | 2005-01-12 | 株式会社メディア・テクノロジー | Iris information acquisition device and iris identification device |
JP3813023B2 (en) * | 1998-08-17 | 2006-08-23 | 沖電気工業株式会社 | Iris recognition device |
WO2000013588A1 (en) | 1998-09-04 | 2000-03-16 | Beecham James E | Verification of human medical data |
US6522772B1 (en) * | 1998-09-30 | 2003-02-18 | Ncr Corporation | Self-service checkout terminal having a biometric sensing device for verifying identity of a user and associated method |
US6424727B1 (en) * | 1998-11-25 | 2002-07-23 | Iridian Technologies, Inc. | System and method of animal identification and animal transaction authorization using iris patterns |
US6289113B1 (en) | 1998-11-25 | 2001-09-11 | Iridian Technologies, Inc. | Handheld iris imaging apparatus and method |
US6532298B1 (en) | 1998-11-25 | 2003-03-11 | Iridian Technologies, Inc. | Portable authentication device and method using iris patterns |
US6377699B1 (en) | 1998-11-25 | 2002-04-23 | Iridian Technologies, Inc. | Iris imaging telephone security module and method |
US6753919B1 (en) | 1998-11-25 | 2004-06-22 | Iridian Technologies, Inc. | Fast focus assessment system and method for imaging |
US6687402B1 (en) | 1998-12-18 | 2004-02-03 | Cognex Corporation | Machine vision methods and systems for boundary feature comparison of patterns and images |
US6381366B1 (en) | 1998-12-18 | 2002-04-30 | Cognex Corporation | Machine vision methods and system for boundary point-based comparison of patterns and images |
NL1011509C2 (en) * | 1999-01-05 | 2000-07-06 | Nedap Nv | Method for biometric identification of animals. |
US6944318B1 (en) | 1999-01-15 | 2005-09-13 | Citicorp Development Center, Inc. | Fast matching systems and methods for personal identification |
US20020056043A1 (en) * | 1999-01-18 | 2002-05-09 | Sensar, Inc. | Method and apparatus for securely transmitting and authenticating biometric data over a network |
US6332193B1 (en) | 1999-01-18 | 2001-12-18 | Sensar, Inc. | Method and apparatus for securely transmitting and authenticating biometric data over a network |
US6183086B1 (en) * | 1999-03-12 | 2001-02-06 | Bausch & Lomb Surgical, Inc. | Variable multiple color LED illumination system |
KR100320188B1 (en) * | 1999-03-23 | 2002-01-10 | 구자홍 | Forgery judgment method for iris recognition system |
GB9907515D0 (en) | 1999-04-01 | 1999-05-26 | Ncr Int Inc | Self service terminal |
JP4619384B2 (en) * | 1999-04-09 | 2011-01-26 | アイリテック株式会社 | Iris identification system |
JP4006192B2 (en) * | 1999-04-09 | 2007-11-14 | アイリテック株式会社 | Iris recognition system and method |
KR100356600B1 (en) * | 1999-04-09 | 2002-10-19 | 아이리텍 잉크 | A Method For Identifying The Iris Of Persons Based On The Shape Of Lacuna And/Or Autonomous Nervous Wreath |
US6247813B1 (en) * | 1999-04-09 | 2001-06-19 | Iritech, Inc. | Iris identification system and method of identifying a person through iris recognition |
US6820979B1 (en) * | 1999-04-23 | 2004-11-23 | Neuroptics, Inc. | Pupilometer with pupil irregularity detection, pupil tracking, and pupil response detection capability, glaucoma screening capability, intracranial pressure detection capability, and ocular aberration measurement capability |
US6700998B1 (en) * | 1999-04-23 | 2004-03-02 | Oki Electric Industry Co, Ltd. | Iris registration unit |
JP4290278B2 (en) | 1999-06-03 | 2009-07-01 | 株式会社ニデック | Ophthalmic equipment |
JP2001034754A (en) * | 1999-07-19 | 2001-02-09 | Sony Corp | Iris authentication device |
US6553494B1 (en) | 1999-07-21 | 2003-04-22 | Sensar, Inc. | Method and apparatus for applying and verifying a biometric-based digital signature to an electronic document |
CA2712321C (en) * | 1999-10-21 | 2013-07-30 | Technolas Gmbh Ophthalmologische Systeme | Iris recognition and tracking for optical treatment |
US7237898B1 (en) | 1999-10-21 | 2007-07-03 | Bausch & Lomb Incorporated | Customized corneal profiling |
WO2001035349A1 (en) | 1999-11-09 | 2001-05-17 | Iridian Technologies, Inc. | System and method of biometric authentication of electronic signatures using iris patterns |
WO2001035348A1 (en) | 1999-11-09 | 2001-05-17 | Iridian Technologies, Inc. | System and method for authentication of shipping transactions using printable and readable biometric data |
US6505193B1 (en) | 1999-12-01 | 2003-01-07 | Iridian Technologies, Inc. | System and method of fast biometric database searching using digital certificates |
US6684402B1 (en) | 1999-12-01 | 2004-01-27 | Cognex Technology And Investment Corporation | Control methods and apparatus for coupling multiple image acquisition devices to a digital data processor |
AU2001291157B2 (en) * | 2000-02-03 | 2003-12-18 | Gregg S. Homer | Method for alteration of iris pigment |
AU2003241638B2 (en) * | 2000-02-03 | 2004-01-22 | Gregg S. Homer | Method for alteration of iris pigment |
US20030011758A1 (en) * | 2000-02-09 | 2003-01-16 | Nobuyoshi Ochiai | Personal identification system |
JP2001331799A (en) * | 2000-03-16 | 2001-11-30 | Toshiba Corp | Image processor and image processing method |
EP1136937B1 (en) * | 2000-03-22 | 2006-05-10 | Kabushiki Kaisha Toshiba | Facial image forming recognition apparatus and a pass control apparatus |
JP3825222B2 (en) * | 2000-03-24 | 2006-09-27 | 松下電器産業株式会社 | Personal authentication device, personal authentication system, and electronic payment system |
US6748104B1 (en) | 2000-03-24 | 2004-06-08 | Cognex Corporation | Methods and apparatus for machine vision inspection using single and multiple templates or patterns |
US6299306B1 (en) | 2000-03-31 | 2001-10-09 | Sensar, Inc. | Method and apparatus for positioning subjects using a holographic optical element |
DK1285409T3 (en) | 2000-05-16 | 2005-08-22 | Swisscom Mobile Ag | Process of biometric identification and authentication |
JP4228520B2 (en) * | 2000-06-12 | 2009-02-25 | 沖電気工業株式会社 | Iris photography device |
JP2002101322A (en) * | 2000-07-10 | 2002-04-05 | Matsushita Electric Ind Co Ltd | Iris camera module |
JP3401502B2 (en) * | 2000-07-13 | 2003-04-28 | 松下電器産業株式会社 | Eye imaging device |
KR100400975B1 (en) * | 2000-09-01 | 2003-10-10 | 주승기 | Digital Processing System of Microscopic Image for Ophthalmology |
JP4529263B2 (en) * | 2000-09-18 | 2010-08-25 | 沖電気工業株式会社 | Iris recognition device |
KR100357441B1 (en) * | 2000-09-22 | 2002-10-19 | 에버미디어 주식회사 | Human iris recognition method using harr wavelet transform and lvq |
US7277561B2 (en) * | 2000-10-07 | 2007-10-02 | Qritek Co., Ltd. | Iris identification |
DE10052201B8 (en) * | 2000-10-20 | 2005-06-30 | Carl Zeiss Meditec Ag | Method and device for identifying a patient and an operating area |
US7224822B2 (en) * | 2000-11-02 | 2007-05-29 | Retinal Technologies, L.L.C. | System for capturing an image of the retina for identification |
US6453057B1 (en) * | 2000-11-02 | 2002-09-17 | Retinal Technologies, L.L.C. | Method for generating a unique consistent signal pattern for identification of an individual |
US6920237B2 (en) * | 2000-12-19 | 2005-07-19 | Eastman Kodak Company | Digital image processing method and computer program product for detecting human irises in an image |
US7006669B1 (en) | 2000-12-31 | 2006-02-28 | Cognex Corporation | Machine vision method and apparatus for thresholding images of non-uniform materials |
MXPA03006923A (en) * | 2001-02-02 | 2003-11-18 | Homer Gregg S | Method for alteration of iris pigment. |
US20040111625A1 (en) * | 2001-02-14 | 2004-06-10 | Duffy Dominic Gavan | Data processing apparatus and method |
GB0113255D0 (en) * | 2001-05-31 | 2001-07-25 | Scient Generics Ltd | Number generator |
KR100374708B1 (en) * | 2001-03-06 | 2003-03-04 | 에버미디어 주식회사 | Non-contact type human iris recognition method by correction of rotated iris image |
JP2002341406A (en) * | 2001-05-11 | 2002-11-27 | Matsushita Electric Ind Co Ltd | Method and device for imaging object to be authenticated |
US20040193893A1 (en) * | 2001-05-18 | 2004-09-30 | Michael Braithwaite | Application-specific biometric templates |
JP2003016434A (en) * | 2001-06-27 | 2003-01-17 | Matsushita Electric Ind Co Ltd | Individual authenticating device |
JP2003030659A (en) * | 2001-07-16 | 2003-01-31 | Matsushita Electric Ind Co Ltd | Iris authentication device and iris image pickup device |
KR20030034258A (en) * | 2001-08-04 | 2003-05-09 | 주식회사 큐리텍 | Identification system and method using iris and retina, and media that can record computer program sources thereof |
GB0119859D0 (en) * | 2001-08-15 | 2001-10-10 | Qinetiq Ltd | Eye tracking system |
JP2005503619A (en) * | 2001-09-14 | 2005-02-03 | エステーミクロエレクトロニクス ソシエテ アノニム | How to identify the ring position |
US7203343B2 (en) * | 2001-09-21 | 2007-04-10 | Hewlett-Packard Development Company, L.P. | System and method for determining likely identity in a biometric database |
NO316489B1 (en) | 2001-10-01 | 2004-01-26 | Genkey As | System, portable device and method for digital authentication, encryption and signing by generating volatile but consistent and repeatable crypton keys |
JP4172930B2 (en) * | 2001-10-31 | 2008-10-29 | 松下電器産業株式会社 | Eye imaging device and entrance / exit management system |
US7775665B2 (en) * | 2001-11-13 | 2010-08-17 | Dellavecchia Michael A | Method for optically scanning objects |
US7226166B2 (en) | 2001-11-13 | 2007-06-05 | Philadelphia Retina Endowment Fund | Optimizing the properties of electromagnetic energy in a medium using stochastic parallel perturbation gradient descent optimization adaptive optics |
US20040165147A1 (en) * | 2001-11-13 | 2004-08-26 | Della Vecchia Michael A. | Determining iris biometric and spatial orientation of an iris in accordance with same |
US7377647B2 (en) * | 2001-11-13 | 2008-05-27 | Philadelphia Retina Endowment Fund | Clarifying an image of an object to perform a procedure on the object |
AUPR912101A0 (en) * | 2001-11-26 | 2001-12-20 | Argus Solutions Pty Ltd | Computerised identity matching management |
US6665426B1 (en) * | 2002-01-29 | 2003-12-16 | West Virginia University Research Corporation | Method of biometric identification of an individual and associated apparatus |
KR100954640B1 (en) | 2002-02-05 | 2010-04-27 | 파나소닉 주식회사 | Personal authentication method and device |
US20030165259A1 (en) * | 2002-02-15 | 2003-09-04 | Balent James S. | Signal analysis using image processing techniques |
US20040052418A1 (en) * | 2002-04-05 | 2004-03-18 | Bruno Delean | Method and apparatus for probabilistic image analysis |
US7369685B2 (en) * | 2002-04-05 | 2008-05-06 | Identix Corporation | Vision-based operating method and system |
GB0228434D0 (en) * | 2002-12-05 | 2003-01-08 | Scient Generics Ltd | Error correction |
AUPS254302A0 (en) * | 2002-05-24 | 2002-06-13 | Resmed Limited | A sleepiness test |
EP1516156B1 (en) | 2002-05-30 | 2019-10-23 | AMO Manufacturing USA, LLC | Tracking torsional eye orientation and position |
DE60309176T2 (en) * | 2002-05-31 | 2007-09-06 | Scientific Generics Ltd., Harston | BIOMETRIC AUTHENTICATION SYSTEM |
JP3639291B2 (en) * | 2002-11-07 | 2005-04-20 | 松下電器産業株式会社 | Personal authentication method, iris registration device, iris authentication device, and personal authentication program |
AU2003900586A0 (en) * | 2003-02-11 | 2003-02-27 | Argus Solutions Pty Ltd | Delivery of metered doses |
US7436986B2 (en) * | 2003-03-25 | 2008-10-14 | Bausch & Lomb Incorporated | Positive patient identification |
US7599524B2 (en) * | 2003-04-04 | 2009-10-06 | Sarnoff Corporation | Method and apparatus for providing a robust object finder |
CN100571624C (en) * | 2003-04-11 | 2009-12-23 | 博士伦公司 | Be used to obtain the system and method for data and aligning and tracking eyes |
WO2004097743A1 (en) * | 2003-04-28 | 2004-11-11 | Matsushita Electric Industrial Co. Ltd. | Artificial eye distinguishing method and device, artificial eye distinguishing program, iris recognition method, false printed matter distinguishing method, and image distinguishing method |
KR200321670Y1 (en) * | 2003-05-07 | 2003-07-31 | 주식회사 큐리텍 | Iris identification camera |
US20040233192A1 (en) * | 2003-05-22 | 2004-11-25 | Hopper Stephen A. | Focally-controlled imaging system and method |
US7458683B2 (en) * | 2003-06-16 | 2008-12-02 | Amo Manufacturing Usa, Llc | Methods and devices for registering optical measurement datasets of an optical system |
AU2003903392A0 (en) | 2003-07-03 | 2003-07-17 | Argus Solutions Pty Ltd | Management control of assets |
JP3848953B2 (en) * | 2003-07-04 | 2006-11-22 | 松下電器産業株式会社 | Living body eye determination method and living body eye determination device |
WO2005002441A1 (en) * | 2003-07-04 | 2005-01-13 | Matsushita Electric Industrial Co., Ltd. | Organism eye judgment method and organism eye judgment device |
GB0316631D0 (en) * | 2003-07-16 | 2003-08-20 | Omniperception Ltd | Facial liveness assessment system |
JP3879719B2 (en) * | 2003-08-22 | 2007-02-14 | 松下電器産業株式会社 | Image input device and authentication device using the same |
US8085993B2 (en) | 2006-03-03 | 2011-12-27 | Honeywell International Inc. | Modular biometrics collection system architecture |
US8045764B2 (en) * | 2005-01-26 | 2011-10-25 | Honeywell International Inc. | Expedient encoding system |
US8090157B2 (en) * | 2005-01-26 | 2012-01-03 | Honeywell International Inc. | Approaches and apparatus for eye detection in a digital image |
US8098901B2 (en) * | 2005-01-26 | 2012-01-17 | Honeywell International Inc. | Standoff iris recognition system |
US8049812B2 (en) | 2006-03-03 | 2011-11-01 | Honeywell International Inc. | Camera with auto focus capability |
US8064647B2 (en) | 2006-03-03 | 2011-11-22 | Honeywell International Inc. | System for iris detection tracking and recognition at a distance |
US7756301B2 (en) * | 2005-01-26 | 2010-07-13 | Honeywell International Inc. | Iris recognition system and method |
US7933507B2 (en) * | 2006-03-03 | 2011-04-26 | Honeywell International Inc. | Single lens splitter camera |
US7593550B2 (en) * | 2005-01-26 | 2009-09-22 | Honeywell International Inc. | Distance iris recognition |
US8705808B2 (en) | 2003-09-05 | 2014-04-22 | Honeywell International Inc. | Combined face and iris recognition system |
US8442276B2 (en) | 2006-03-03 | 2013-05-14 | Honeywell International Inc. | Invariant radial iris segmentation |
KR20050026347A (en) * | 2003-09-09 | 2005-03-15 | 유웅덕 | Apparatus for photographing iris pattern |
US20050111704A1 (en) * | 2003-11-24 | 2005-05-26 | Verghis James S. | Iris mapping and compatibility and personality evaluation |
WO2005055125A1 (en) * | 2003-12-01 | 2005-06-16 | Sung-Bok Shin | Identification system and method having iris, and media that can record computer program sources thereof |
FR2864290B1 (en) * | 2003-12-18 | 2006-05-26 | Sagem | METHOD AND DEVICE FOR RECOGNIZING IRIS |
US7481536B2 (en) * | 2004-02-19 | 2009-01-27 | Amo Manufacturing Usa, Llc | Methods and systems for differentiating left and right eye images |
JP2005304809A (en) * | 2004-04-22 | 2005-11-04 | Matsushita Electric Ind Co Ltd | Eye image pickup device with lighting system |
GB0412175D0 (en) * | 2004-06-01 | 2004-06-30 | Smart Sensors Ltd | Identification of image characteristics |
GB0413034D0 (en) * | 2004-06-10 | 2004-07-14 | Scient Generics Ltd | Secure workflow engine |
US20050281440A1 (en) * | 2004-06-18 | 2005-12-22 | Pemer Frederick A | Iris feature detection and sensor-based edge detection |
GB0414113D0 (en) * | 2004-06-24 | 2004-07-28 | Virulite Distrib Ltd | Cosmetic uses of electromagnetic radiation |
JP3966309B2 (en) * | 2004-07-14 | 2007-08-29 | 松下電器産業株式会社 | Pupil detection device and iris authentication device |
JP4599110B2 (en) * | 2004-07-30 | 2010-12-15 | キヤノン株式会社 | Image processing apparatus and method, imaging apparatus, and program |
JP4455216B2 (en) * | 2004-08-06 | 2010-04-21 | キヤノン株式会社 | Detection device |
US7248720B2 (en) * | 2004-10-21 | 2007-07-24 | Retica Systems, Inc. | Method and system for generating a combined retina/iris pattern biometric |
US7167736B2 (en) * | 2004-11-04 | 2007-01-23 | Q Step Technologies, Inc. | Non-invasive measurement system and method for measuring the concentration of an optically-active substance |
US20060115130A1 (en) * | 2004-11-29 | 2006-06-01 | Douglas Kozlay | Eyewear with biometrics to protect displayed data |
IL165586A0 (en) * | 2004-12-06 | 2006-01-15 | Daphna Palti Wasserman | Multivariate dynamic biometrics system |
US20060147095A1 (en) * | 2005-01-03 | 2006-07-06 | Usher David B | Method and system for automatically capturing an image of a retina |
EP1693801A3 (en) * | 2005-02-16 | 2006-11-29 | David Schaufele | Biometric-based systems and methods for identity verification |
US8317327B2 (en) * | 2005-03-16 | 2012-11-27 | Lc Technologies, Inc. | System and method for eyeball surface topography as a biometric discriminator |
GB0512038D0 (en) * | 2005-06-14 | 2005-07-20 | Dougal Gordon | Therapeutic and cosmetic uses of electromagnetic radiation |
JP4498224B2 (en) * | 2005-06-14 | 2010-07-07 | キヤノン株式会社 | Image processing apparatus and method |
US7639861B2 (en) | 2005-09-14 | 2009-12-29 | Cognex Technology And Investment Corporation | Method and apparatus for backlighting a wafer during alignment |
US8111904B2 (en) * | 2005-10-07 | 2012-02-07 | Cognex Technology And Investment Corp. | Methods and apparatus for practical 3D vision system |
US8260008B2 (en) * | 2005-11-11 | 2012-09-04 | Eyelock, Inc. | Methods for performing biometric recognition of a human eye and corroboration of same |
US7801335B2 (en) * | 2005-11-11 | 2010-09-21 | Global Rainmakers Inc. | Apparatus and methods for detecting the presence of a human eye |
JP2009520178A (en) * | 2005-12-19 | 2009-05-21 | ハネウェル・インターナショナル・インコーポレーテッド | Multi-fluid refrigerant system |
US7744216B1 (en) | 2006-01-06 | 2010-06-29 | Lockheed Martin Corporation | Display system intensity adjustment based on pupil dilation |
GB0603411D0 (en) * | 2006-02-21 | 2006-03-29 | Xvista Ltd | Method of processing an image of an eye |
US8364646B2 (en) | 2006-03-03 | 2013-01-29 | Eyelock, Inc. | Scalable searching of biometric databases using dynamic selection of data subsets |
WO2007103834A1 (en) * | 2006-03-03 | 2007-09-13 | Honeywell International, Inc. | Indexing and database search system |
KR101308368B1 (en) * | 2006-03-03 | 2013-09-16 | 허니웰 인터내셔널 인코포레이티드 | An iris recognition system having image quality metrics |
DE102007016397B4 (en) * | 2006-04-03 | 2011-02-03 | Sabine Raveau | Apparatus and method for image acquisition of eyes |
US8014571B2 (en) * | 2006-05-15 | 2011-09-06 | Identix Incorporated | Multimodal ocular biometric system |
US8604901B2 (en) | 2006-06-27 | 2013-12-10 | Eyelock, Inc. | Ensuring the provenance of passengers at a transportation facility |
US8162584B2 (en) * | 2006-08-23 | 2012-04-24 | Cognex Corporation | Method and apparatus for semiconductor wafer alignment |
PL380581A1 (en) * | 2006-09-07 | 2008-03-17 | Naukowa I Akademicka Sieć Komputerowa | Testing method of eye vitality and device for thesting of eye vitality |
EP2062197A4 (en) * | 2006-09-15 | 2010-10-06 | Retica Systems Inc | Long distance multimodal biometric system and method |
WO2008091401A2 (en) * | 2006-09-15 | 2008-07-31 | Retica Systems, Inc | Multimodal ocular biometric system and methods |
US8121356B2 (en) | 2006-09-15 | 2012-02-21 | Identix Incorporated | Long distance multimodal biometric system and method |
US8965063B2 (en) | 2006-09-22 | 2015-02-24 | Eyelock, Inc. | Compact biometric acquisition system and method |
US7970179B2 (en) | 2006-09-25 | 2011-06-28 | Identix Incorporated | Iris data extraction |
TW200816060A (en) * | 2006-09-29 | 2008-04-01 | Univ Nat Chiao Tung | Iris recognition method |
EP2100253A4 (en) | 2006-10-02 | 2011-01-12 | Global Rainmakers Inc | Fraud resistant biometric financial transaction system and method |
US7809747B2 (en) * | 2006-10-23 | 2010-10-05 | Donald Martin Monro | Fuzzy database matching |
US9846739B2 (en) | 2006-10-23 | 2017-12-19 | Fotonation Limited | Fast database matching |
ES2276637B1 (en) * | 2006-11-03 | 2008-11-16 | Jose Antonio Gil Soldevilla | COLOR OBTAINING PROCEDURE FROM IRIS. |
RU2318438C1 (en) | 2006-11-28 | 2008-03-10 | Дмитрий Евгеньевич Антонов | Device for getting image of iris |
WO2008131201A1 (en) | 2007-04-19 | 2008-10-30 | Global Rainmakers, Inc. | Method and system for biometric recognition |
US8953849B2 (en) | 2007-04-19 | 2015-02-10 | Eyelock, Inc. | Method and system for biometric recognition |
US8063889B2 (en) * | 2007-04-25 | 2011-11-22 | Honeywell International Inc. | Biometric data collection system |
US20120239458A9 (en) * | 2007-05-18 | 2012-09-20 | Global Rainmakers, Inc. | Measuring Effectiveness of Advertisements and Linking Certain Consumer Activities Including Purchases to Other Activities of the Consumer |
US7652716B2 (en) * | 2007-05-31 | 2010-01-26 | Microsoft Corporation | Computer-controlled lighting for video communication |
IL184399A0 (en) * | 2007-07-03 | 2007-10-31 | Yossi Tsuria | Content delivery system |
US20090060348A1 (en) * | 2007-08-28 | 2009-03-05 | Donald Martin Monro | Determination of Image Similarity |
US9036871B2 (en) | 2007-09-01 | 2015-05-19 | Eyelock, Inc. | Mobility identity platform |
US9002073B2 (en) | 2007-09-01 | 2015-04-07 | Eyelock, Inc. | Mobile identity platform |
WO2009029757A1 (en) | 2007-09-01 | 2009-03-05 | Global Rainmakers, Inc. | System and method for iris data acquisition for biometric identification |
US8212870B2 (en) | 2007-09-01 | 2012-07-03 | Hanna Keith J | Mirror system and method for acquiring biometric data |
US9117119B2 (en) | 2007-09-01 | 2015-08-25 | Eyelock, Inc. | Mobile identity platform |
US8393734B2 (en) | 2007-09-14 | 2013-03-12 | Neuroptics, Inc. | Pupilary screening method and system |
US7824034B2 (en) * | 2007-09-19 | 2010-11-02 | Utc Fire & Security Americas Corporation, Inc. | Iris imaging system and method for the same |
US10398599B2 (en) | 2007-10-05 | 2019-09-03 | Topcon Medical Laser Systems Inc. | Semi-automated ophthalmic photocoagulation method and apparatus |
CA2744757C (en) * | 2007-11-29 | 2017-06-13 | Wavefront Biometric Technologies Pty Limited | Biometric authentication using the eye |
US20090252382A1 (en) * | 2007-12-06 | 2009-10-08 | University Of Notre Dame Du Lac | Segmentation of iris images using active contour processing |
WO2009102940A1 (en) * | 2008-02-14 | 2009-08-20 | The International Performance Registry, Llc | System and method for animal identification using iris images |
US8345936B2 (en) * | 2008-05-09 | 2013-01-01 | Noblis, Inc. | Multispectral iris fusion for enhancement and interoperability |
US8436907B2 (en) * | 2008-05-09 | 2013-05-07 | Honeywell International Inc. | Heterogeneous video capturing system |
DE602008003019D1 (en) | 2008-06-25 | 2010-11-25 | Deutsche Telekom Ag | System for extraction, identification and verification of iris features based on directionlets |
WO2009158662A2 (en) | 2008-06-26 | 2009-12-30 | Global Rainmakers, Inc. | Method of reducing visibility of illimination while acquiring high quality imagery |
GB0812753D0 (en) * | 2008-07-14 | 2008-08-20 | Dougal Gordon R P | Electromagnetic radiation and its therapeutic effect |
WO2010011932A1 (en) * | 2008-07-24 | 2010-01-28 | Kynen Llc | Mobile biometric identification system and method |
US8213782B2 (en) | 2008-08-07 | 2012-07-03 | Honeywell International Inc. | Predictive autofocusing system |
US8090246B2 (en) * | 2008-08-08 | 2012-01-03 | Honeywell International Inc. | Image acquisition system |
US8290219B2 (en) * | 2008-08-22 | 2012-10-16 | International Business Machines Corporation | Registration-free transforms for cancelable iris biometrics |
US9633261B2 (en) * | 2008-08-22 | 2017-04-25 | International Business Machines Corporation | Salting system and method for cancelable iris biometric |
US8306279B2 (en) * | 2008-09-15 | 2012-11-06 | Eyelock, Inc. | Operator interface for face and iris recognition devices |
CN101727574A (en) * | 2008-10-17 | 2010-06-09 | 深圳富泰宏精密工业有限公司 | Iris recognition system and method |
US20100278394A1 (en) * | 2008-10-29 | 2010-11-04 | Raguin Daniel H | Apparatus for Iris Capture |
US8317325B2 (en) | 2008-10-31 | 2012-11-27 | Cross Match Technologies, Inc. | Apparatus and method for two eye imaging for iris identification |
WO2010062400A1 (en) * | 2008-11-28 | 2010-06-03 | Neuroptics, Inc. | Methods, systems, and devices for monitoring anisocoria and asymmetry of pupillary reaction to stimulus |
US8280119B2 (en) | 2008-12-05 | 2012-10-02 | Honeywell International Inc. | Iris recognition system using quality metrics |
US8374404B2 (en) | 2009-02-13 | 2013-02-12 | Raytheon Company | Iris recognition using hyper-spectral signatures |
WO2010099475A1 (en) * | 2009-02-26 | 2010-09-02 | Kynen Llc | User authentication system and method |
US8195044B2 (en) | 2009-03-30 | 2012-06-05 | Eyelock Inc. | Biometric camera mount system |
US8472681B2 (en) | 2009-06-15 | 2013-06-25 | Honeywell International Inc. | Iris and ocular recognition system using trace transforms |
US8630464B2 (en) | 2009-06-15 | 2014-01-14 | Honeywell International Inc. | Adaptive iris matching using database indexing |
US8750575B2 (en) * | 2009-08-04 | 2014-06-10 | International Business Machines Corporation | Reflexive iris template |
US20110119141A1 (en) * | 2009-11-16 | 2011-05-19 | Hoyos Corporation | Siccolla Identity Verification Architecture and Tool |
US8577094B2 (en) | 2010-04-09 | 2013-11-05 | Donald Martin Monro | Image template masking |
EP2568938A1 (en) | 2010-05-10 | 2013-03-20 | Ramot at Tel-Aviv University Ltd | System for treating glaucoma by directing electromagnetic energy to the limbal area of an eye |
US8742887B2 (en) | 2010-09-03 | 2014-06-03 | Honeywell International Inc. | Biometric visitor check system |
CA2811803A1 (en) * | 2010-09-20 | 2012-03-29 | Amo Development Llc | System and methods for mitigating changes in pupil size during laser refractive surgery to maintain ablation centration |
US9753025B2 (en) | 2010-10-26 | 2017-09-05 | Bi2 Technologies, LLC | Mobile wireless hand-held identification system and breathalyzer |
US8719584B2 (en) | 2010-10-26 | 2014-05-06 | Bi2 Technologies, LLC | Mobile, wireless hand-held biometric capture, processing and communication system and method for biometric identification |
US10068080B2 (en) | 2010-10-26 | 2018-09-04 | Bi2 Technologies, LLC | Mobile wireless hand-held biometric identification system |
US9507926B2 (en) | 2010-10-26 | 2016-11-29 | Bi2 Technologies, LLC | Mobile wireless hand-held identification system and method for identification |
US8254768B2 (en) * | 2010-12-22 | 2012-08-28 | Michael Braithwaite | System and method for illuminating and imaging the iris of a person |
US8831416B2 (en) * | 2010-12-22 | 2014-09-09 | Michael Braithwaite | System and method for illuminating and identifying a person |
US10043229B2 (en) | 2011-01-26 | 2018-08-07 | Eyelock Llc | Method for confirming the identity of an individual while shielding that individual's personal data |
WO2012112788A2 (en) | 2011-02-17 | 2012-08-23 | Eyelock Inc. | Efficient method and system for the acquisition of scene imagery and iris imagery using a single sensor |
US20120268241A1 (en) | 2011-04-19 | 2012-10-25 | Eyelock Inc. | Biometric chain of provenance |
US9124798B2 (en) * | 2011-05-17 | 2015-09-01 | Eyelock Inc. | Systems and methods for illuminating an iris with visible light for biometric acquisition |
US8885882B1 (en) | 2011-07-14 | 2014-11-11 | The Research Foundation For The State University Of New York | Real time eye tracking for human computer interaction |
US8965064B2 (en) | 2011-08-22 | 2015-02-24 | Eyelock, Inc. | Systems and methods for capturing artifact free images |
GB2495328B (en) * | 2011-10-07 | 2018-05-30 | Irisguard Inc | Improvements relating to Iris cameras |
US9412022B2 (en) | 2012-09-06 | 2016-08-09 | Leonard Flom | Iris identification system and method |
US9101297B2 (en) | 2012-12-11 | 2015-08-11 | Elwha Llc | Time-based unobtrusive active eye interrogation |
US9039180B2 (en) | 2012-12-11 | 2015-05-26 | Elwah LLC | Self-aligning unobtrusive active eye interrogation |
US9039179B2 (en) | 2012-12-11 | 2015-05-26 | Elwha Llc | Unobtrusive active eye interrogation |
EP2961364B1 (en) | 2013-02-26 | 2020-05-06 | Belkin Laser Ltd. | System for glaucoma treatment |
US9495526B2 (en) | 2013-03-15 | 2016-11-15 | Eyelock Llc | Efficient prevention of fraud |
JP2014206932A (en) * | 2013-04-15 | 2014-10-30 | オムロン株式会社 | Authentication device, authentication method, control program, and recording medium |
US8958608B2 (en) | 2013-06-04 | 2015-02-17 | Ut-Battelle, Llc | Frontal view reconstruction for iris recognition |
CN105934764A (en) | 2013-12-23 | 2016-09-07 | 眼锁有限责任公司 | Methods and apparatus for power-efficient iris recognition |
BR112016015664A8 (en) | 2014-01-06 | 2020-06-09 | Eyelock Llc | device for repeatedly generating images of an iris and iris image recognition device for repetitive use |
WO2015108911A1 (en) * | 2014-01-16 | 2015-07-23 | Delta ID Inc. | Method and apparatus for controlling intensity of illumination in eye based biometric systems |
US9818114B2 (en) | 2014-08-11 | 2017-11-14 | Mastercard International Incorporated | Systems and methods for performing payment card transactions using a wearable computing device |
US9870457B2 (en) | 2014-08-15 | 2018-01-16 | California Institute Of Technology | HERMA—heartbeat microwave authentication |
BR112017004593A2 (en) | 2014-09-12 | 2017-12-05 | Eyelock Llc | methods and apparatus for directing a user's gaze on an iris recognition system |
US10425814B2 (en) | 2014-09-24 | 2019-09-24 | Princeton Identity, Inc. | Control of wireless communication device capability in a mobile device with a biometric key |
US10430557B2 (en) | 2014-11-17 | 2019-10-01 | Elwha Llc | Monitoring treatment compliance using patient activity patterns |
US9585616B2 (en) | 2014-11-17 | 2017-03-07 | Elwha Llc | Determining treatment compliance using speech patterns passively captured from a patient environment |
US9589107B2 (en) | 2014-11-17 | 2017-03-07 | Elwha Llc | Monitoring treatment compliance using speech patterns passively captured from a patient environment |
BR112017009982A2 (en) | 2014-11-19 | 2018-01-02 | Eyelock Llc | model-based prediction of an optimal convenience metric for authorization transactions |
WO2016089592A1 (en) | 2014-12-03 | 2016-06-09 | Sri Internaitonal | System and method for mobile device biometric add-on |
US10074011B2 (en) | 2015-01-20 | 2018-09-11 | Eyelock Llc | Lens system for high quality visible image acquisition and infra-red iris image acquisition |
US9509690B2 (en) | 2015-03-12 | 2016-11-29 | Eyelock Llc | Methods and systems for managing network activity using biometrics |
US10589105B2 (en) | 2015-03-27 | 2020-03-17 | The Invention Science Fund Ii, Llc | Method and system for controlling ear stimulation |
US9987489B2 (en) | 2015-03-27 | 2018-06-05 | Elwha Llc | Controlling ear stimulation in response to electrical contact sensing |
US11364380B2 (en) | 2015-03-27 | 2022-06-21 | Elwha Llc | Nerve stimulation system, subsystem, headset, and earpiece |
US10039928B2 (en) | 2015-03-27 | 2018-08-07 | Equility Llc | Ear stimulation with neural feedback sensing |
US10512783B2 (en) | 2015-03-27 | 2019-12-24 | Equility Llc | User interface method and system for ear stimulation |
US10327984B2 (en) | 2015-03-27 | 2019-06-25 | Equility Llc | Controlling ear stimulation in response to image analysis |
US10406376B2 (en) | 2015-03-27 | 2019-09-10 | Equility Llc | Multi-factor control of ear stimulation |
US10398902B2 (en) | 2015-03-27 | 2019-09-03 | Equility Llc | Neural stimulation method and system with audio output |
US9811729B2 (en) | 2015-05-12 | 2017-11-07 | Ut-Battelle, Llc | Iris recognition via plenoptic imaging |
CN104899583A (en) * | 2015-06-30 | 2015-09-09 | 成都点石创想科技有限公司 | Iris identifying method for door access monitoring system |
US11100673B2 (en) | 2015-09-24 | 2021-08-24 | Apple Inc. | Systems and methods for localization using surface imaging |
WO2017053415A1 (en) | 2015-09-24 | 2017-03-30 | Quovard Management Llc | Systems and methods for surface monitoring |
US10311299B2 (en) | 2015-12-21 | 2019-06-04 | Eyelock Llc | Reflected optic camera module for iris recognition in a computing device |
JP2019506694A (en) | 2016-01-12 | 2019-03-07 | プリンストン・アイデンティティー・インコーポレーテッド | Biometric analysis system and method |
US10366296B2 (en) | 2016-03-31 | 2019-07-30 | Princeton Identity, Inc. | Biometric enrollment systems and methods |
WO2017172695A1 (en) | 2016-03-31 | 2017-10-05 | Princeton Identity, Inc. | Systems and methods of biometric anaysis with adaptive trigger |
JP6409088B2 (en) * | 2016-04-28 | 2018-10-17 | シャープ株式会社 | Image processing method and image processing apparatus |
WO2017201147A2 (en) | 2016-05-18 | 2017-11-23 | Eyelock, Llc | Iris recognition methods and systems based on an iris stochastic texture model |
EP3465536A1 (en) | 2016-05-27 | 2019-04-10 | Jeff B. Pelz | System and method for eye tracking |
US10442439B1 (en) | 2016-08-18 | 2019-10-15 | Apple Inc. | System and method for road friction coefficient estimation |
JP6798233B2 (en) * | 2016-10-06 | 2020-12-09 | 富士ゼロックス株式会社 | Eye light measuring device and eye light measuring method |
WO2018156726A1 (en) | 2017-02-24 | 2018-08-30 | Eyelock, Llc | Systems and methods for providing illumination for iris biometric acquisition |
US10607096B2 (en) | 2017-04-04 | 2020-03-31 | Princeton Identity, Inc. | Z-dimension user feedback biometric system |
US10657401B2 (en) * | 2017-06-06 | 2020-05-19 | Microsoft Technology Licensing, Llc | Biometric object spoof detection based on image intensity variations |
WO2019023032A1 (en) | 2017-07-26 | 2019-01-31 | Princeton Identity, Inc. | Biometric security systems and methods |
CA3015802C (en) | 2017-08-31 | 2021-06-22 | Eyelock, Llc | Systems and methods of biometric acquistion using positive optical distortion |
US11382794B2 (en) | 2018-07-02 | 2022-07-12 | Belkin Laser Ltd. | Direct selective laser trabeculoplasty |
CN109657531A (en) * | 2018-09-18 | 2019-04-19 | 深圳先牛信息技术有限公司 | A kind of human face in-vivo detection method and detection device based on hot spot on eyeball |
EP3764272A1 (en) | 2019-07-08 | 2021-01-13 | Fraunhofer Gesellschaft zur Förderung der Angewand | System and method for identifying a test object |
DE102019213931B4 (en) | 2019-09-12 | 2021-10-28 | Zf Friedrichshafen Ag | Method and computer program product for identifying a vehicle user and control device for automated driving functions |
US20240013606A1 (en) * | 2022-07-08 | 2024-01-11 | EZ Vending Machines, Inc. | Verification vending machine |
US11762969B1 (en) | 2023-01-12 | 2023-09-19 | King Saud University | Systems and methods for facilitating biometric recognition |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3533683A (en) * | 1967-06-13 | 1970-10-13 | Whittaker Corp | Dynamic pupillometers using television camera system |
US3533684A (en) * | 1967-06-26 | 1970-10-13 | Lawrence Stark | Display of measurement adequacy marker system for pupillometers |
US3598107A (en) * | 1968-07-25 | 1971-08-10 | Hamamatsu T V Co Ltd | Pupillary motion observing apparatus |
US3966310A (en) * | 1974-02-15 | 1976-06-29 | Larson Merlin D | Pupillometer and method of use thereof |
US4008606A (en) * | 1975-10-20 | 1977-02-22 | The United States Of America As Represented By The Secretary Of The Navy | Ship's bottom inspection apparatus |
US4109237A (en) * | 1977-01-17 | 1978-08-22 | Hill Robert B | Apparatus and method for identifying individuals through their retinal vasculature patterns |
US4393366A (en) * | 1981-02-17 | 1983-07-12 | Eye-D Development Ii Ltd. | Rotating beam ocular identification apparatus and method |
US4523820A (en) * | 1981-05-19 | 1985-06-18 | Kaakinen Kari A | Procedure and means for establishing and recording errors of the eye |
US4533222A (en) * | 1979-10-25 | 1985-08-06 | Canon Kabushiki Kaisha | Eye examining instrument |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1192512A (en) * | 1912-02-10 | 1916-07-25 | George W Fitz | Shadow-pupillometer. |
US1887115A (en) * | 1931-04-23 | 1932-11-08 | Zeiss Carl Fa | Instrument for the examination of the reaction of the eye pupils when different parts of the retina are illuminated |
US2434890A (en) * | 1938-03-28 | 1948-01-20 | James B Saunders | Pupillometric exposure meter |
US3136839A (en) * | 1958-09-16 | 1964-06-09 | Safir Aran | Apparatus for objectively testing an optical system |
US3366786A (en) * | 1965-04-15 | 1968-01-30 | Richard P. Delano | Apparatus for producing color effects |
US3473868A (en) * | 1967-04-27 | 1969-10-21 | Space Sciences Inc | Eye position and movement monitor |
US3482904A (en) * | 1967-12-08 | 1969-12-09 | David Volk | Method and apparatus for determining the eccentricity of a conicoid surface |
US3600098A (en) * | 1969-12-29 | 1971-08-17 | Bausch & Lomb | Optical alignment method and apparatus |
US3778135A (en) * | 1972-11-29 | 1973-12-11 | American Optical Corp | Ophthalmometer having alternative viewing and measuring systems and including an improved illumination system |
JPS49136227U (en) * | 1973-03-22 | 1974-11-22 | ||
DE2414322C3 (en) * | 1974-03-25 | 1980-01-17 | Philips Patentverwaltung Gmbh, 2000 Hamburg | Method and arrangement for decoding overlay images of three-dimensional objects |
CH581986A5 (en) * | 1974-08-23 | 1976-11-30 | Bracher Damiel | |
US3915564A (en) * | 1974-09-12 | 1975-10-28 | Zeiss Stiftung | Retinal image-display system |
US4172632A (en) * | 1976-01-21 | 1979-10-30 | Holmes Lawrence Jr | Method and apparatus producing three-dimensional shadow images |
US4189215A (en) * | 1976-03-22 | 1980-02-19 | Humphrey Instruments Inc. | Method and apparatus for the corneal positioning of a patient's eye |
JPS52144174A (en) * | 1976-05-25 | 1977-12-01 | Mitsubishi Electric Corp | Composite illumination |
JPS53125379A (en) * | 1976-10-25 | 1978-11-01 | Mitsubishi Electric Corp | Mixed lights illumination method |
JPS53126791A (en) * | 1977-04-12 | 1978-11-06 | Canon Kk | Ophthalmolgic decice |
US4253743A (en) * | 1977-05-17 | 1981-03-03 | Canon Kabushiki Kaisha | Eye testing instrument |
JPS6054053B2 (en) * | 1977-11-15 | 1985-11-28 | ミノルタ株式会社 | Fundus camera for easy pupil alignment |
US4309085A (en) * | 1979-07-12 | 1982-01-05 | Morrison Robert J | Method for measuring eye features with a contact lens |
US4375320A (en) * | 1980-09-05 | 1983-03-01 | Smirmaul Heinz J | Dual image corneal radius measurement |
-
1985
- 1985-02-20 US US06/703,312 patent/US4641349A/en not_active Expired - Lifetime
-
1986
- 1986-02-04 BR BR8605561A patent/BR8605561A/en not_active IP Right Cessation
- 1986-02-04 AT AT86901250T patent/ATE65851T1/en not_active IP Right Cessation
- 1986-02-04 EP EP86901250A patent/EP0215818B1/en not_active Expired - Lifetime
- 1986-02-04 JP JP61501122A patent/JPS62501889A/en active Granted
- 1986-02-04 DE DE8686901250T patent/DE3680618D1/en not_active Expired - Lifetime
- 1986-02-04 WO PCT/US1986/000227 patent/WO1986005018A1/en active IP Right Grant
- 1986-02-19 IL IL77920A patent/IL77920A/en not_active IP Right Cessation
- 1986-02-19 MX MX1587A patent/MX163339A/en unknown
- 1986-02-19 CA CA000502207A patent/CA1244552A/en not_active Expired
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3533683A (en) * | 1967-06-13 | 1970-10-13 | Whittaker Corp | Dynamic pupillometers using television camera system |
US3533684A (en) * | 1967-06-26 | 1970-10-13 | Lawrence Stark | Display of measurement adequacy marker system for pupillometers |
US3598107A (en) * | 1968-07-25 | 1971-08-10 | Hamamatsu T V Co Ltd | Pupillary motion observing apparatus |
US3966310A (en) * | 1974-02-15 | 1976-06-29 | Larson Merlin D | Pupillometer and method of use thereof |
US4008606A (en) * | 1975-10-20 | 1977-02-22 | The United States Of America As Represented By The Secretary Of The Navy | Ship's bottom inspection apparatus |
US4109237A (en) * | 1977-01-17 | 1978-08-22 | Hill Robert B | Apparatus and method for identifying individuals through their retinal vasculature patterns |
US4533222A (en) * | 1979-10-25 | 1985-08-06 | Canon Kabushiki Kaisha | Eye examining instrument |
US4393366A (en) * | 1981-02-17 | 1983-07-12 | Eye-D Development Ii Ltd. | Rotating beam ocular identification apparatus and method |
US4523820A (en) * | 1981-05-19 | 1985-06-18 | Kaakinen Kari A | Procedure and means for establishing and recording errors of the eye |
Non-Patent Citations (1)
Title |
---|
See also references of EP0215818A4 * |
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1989004139A1 (en) * | 1987-11-10 | 1989-05-18 | Francis Giet | Apparatus for examining the eye by iriscopy and/or irigraphy |
FR2622786A1 (en) * | 1987-11-10 | 1989-05-12 | Giet Francis | APPARATUS FOR EYE EXAMINATION BY IRISCOPY AND IRIGRAPHY |
US5495338A (en) * | 1992-05-06 | 1996-02-27 | Parfums Christian Dior | Process for making up the face, particularly the eyes, and device for carrying it out |
EP0569291A1 (en) * | 1992-05-06 | 1993-11-10 | Parfums Christian Dior | Procedure for the making-up of a face, in particular the eyes and a device for carrying out the same |
WO1993021794A1 (en) * | 1992-05-06 | 1993-11-11 | Parfums Christian Dior | Method for making up the face, especially the eyes, and device for carrying out said method |
FR2690822A1 (en) * | 1992-05-06 | 1993-11-12 | Dior Christian Parfums | Process for making up the face, in particular the eyes, and device for its implementation. |
US5954583A (en) * | 1992-11-05 | 1999-09-21 | Com21 Limited | Secure access control system |
WO1994010658A1 (en) * | 1992-11-05 | 1994-05-11 | Coms21 Limited | Secure access control system |
DE4414216C1 (en) * | 1994-04-23 | 1995-04-06 | Daimler Benz Ag | Device for protecting a motor vehicle against use by third parties, with personalisation of the driving authorisation |
US6252978B1 (en) | 1994-04-23 | 2001-06-26 | Daimlerchrysler Ag | Device for protecting a motor vehicle against use by third parties, with individual driving authorization |
WO1997046979A1 (en) * | 1996-06-06 | 1997-12-11 | British Telecommunications Public Limited Company | Personal identification |
WO1997046980A1 (en) * | 1996-06-06 | 1997-12-11 | British Telecommunications Public Limited Company | Personal identification |
WO1997046978A1 (en) * | 1996-06-06 | 1997-12-11 | British Telecommunications Public Limited Company | Personal identification |
US6333988B1 (en) | 1996-06-06 | 2001-12-25 | British Telecommunications Plc | Personal identification |
US6081607A (en) * | 1996-07-25 | 2000-06-27 | Oki Electric Industry Co. | Animal body identifying device and body identifying system |
EP0821912A2 (en) * | 1996-07-25 | 1998-02-04 | Oki Electric Industry Co., Ltd. | Animal body identifying device and body identifying system |
EP0821912A3 (en) * | 1996-07-25 | 1998-04-08 | Oki Electric Industry Co., Ltd. | Animal body identifying device and body identifying system |
EP0872814A1 (en) * | 1997-04-15 | 1998-10-21 | BRITISH TELECOMMUNICATIONS public limited company | Optical apparatus |
US6252977B1 (en) | 1997-12-01 | 2001-06-26 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
US6055322A (en) * | 1997-12-01 | 2000-04-25 | Sensor, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
WO1999027844A1 (en) * | 1997-12-01 | 1999-06-10 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
DE10194154B4 (en) * | 2000-10-07 | 2005-12-22 | Qritek Co., Ltd. | An iris identification system and method and computer readable storage medium having computer executable instructions stored therein for implementing the iris identification method |
WO2002087442A1 (en) * | 2001-04-27 | 2002-11-07 | Bausch & Lomb Incorporated | Iris pattern recognition and alignment |
KR100897365B1 (en) * | 2001-04-27 | 2009-05-15 | 바슈 앤드 롬 인코포레이티드 | Iris Pattern Recognition And Alignment |
US8186830B2 (en) | 2001-04-27 | 2012-05-29 | Bausch & Lomb Incorporated | Iris pattern recognition and alignment |
US7796784B2 (en) | 2002-11-07 | 2010-09-14 | Panasonic Corporation | Personal authentication method for certificating individual iris |
US7623686B2 (en) | 2004-05-10 | 2009-11-24 | Panasonic Corporation | Techniques and apparatus for increasing accuracy of iris authentication by utilizing a plurality of iris images |
US8064649B2 (en) | 2004-05-10 | 2011-11-22 | Panasonic Corporation | Techniques and apparatus for increasing accuracy of iris authentication by utilizing a plurality of iris images having different pupil diameter/iris diameter ratios |
US8208693B2 (en) | 2004-05-10 | 2012-06-26 | Panasonic Corporation | Techniques and apparatus for increasing accuracy of iris authentication by utilizing a plurality of iris images having different pupil diameter/iris diameter ratios |
EP2265231B1 (en) * | 2008-04-02 | 2015-08-12 | Refocus Group, Inc. | System and method for identifying a position to insert a scleral prosthesis into an eye |
US9002053B2 (en) | 2011-10-07 | 2015-04-07 | Irisguard Inc. | Iris recognition systems |
US9008375B2 (en) | 2011-10-07 | 2015-04-14 | Irisguard Inc. | Security improvements for iris recognition systems |
EP2916262A1 (en) * | 2014-03-06 | 2015-09-09 | Sony Corporation | Information processing apparatus, information processing method, eyewear terminal, and authentication system |
US10460164B2 (en) | 2014-03-06 | 2019-10-29 | Sony Corporation | Information processing apparatus, information processing method, eyewear terminal, and authentication system |
US9977977B2 (en) | 2014-09-30 | 2018-05-22 | Qualcomm Incorporated | Apparatus and method for low-power object-detection in images using computer vision feature computation hardware |
US9986211B2 (en) | 2014-09-30 | 2018-05-29 | Qualcomm Incorporated | Low-power always-on face detection, tracking, recognition and/or analysis using events-based vision sensor |
US10515284B2 (en) | 2014-09-30 | 2019-12-24 | Qualcomm Incorporated | Single-processor computer vision hardware control and application execution |
US11068712B2 (en) | 2014-09-30 | 2021-07-20 | Qualcomm Incorporated | Low-power iris scan initialization |
EP3223232A4 (en) * | 2014-11-20 | 2018-07-18 | Sony Corporation | Control system, information processing device, control method, and program |
WO2017062235A3 (en) * | 2015-10-08 | 2017-05-26 | Microsoft Technology Licensing, Llc | Iris recognition |
CN108140113A (en) * | 2015-10-08 | 2018-06-08 | 微软技术许可有限责任公司 | Iris recognition |
US10157312B2 (en) | 2015-10-08 | 2018-12-18 | Microsoft Technology Licensing, Llc | Iris recognition |
WO2018111528A1 (en) * | 2016-12-16 | 2018-06-21 | Qualcomm Incorporated | Light source modulation for iris size adjustment |
US10614332B2 (en) | 2016-12-16 | 2020-04-07 | Qualcomm Incorportaed | Light source modulation for iris size adjustment |
US10984235B2 (en) | 2016-12-16 | 2021-04-20 | Qualcomm Incorporated | Low power data generation for iris-related detection and authentication |
Also Published As
Publication number | Publication date |
---|---|
CA1244552A (en) | 1988-11-08 |
BR8605561A (en) | 1987-04-22 |
DE3680618D1 (en) | 1991-09-05 |
JPH0584166B2 (en) | 1993-12-01 |
EP0215818B1 (en) | 1991-07-31 |
US4641349A (en) | 1987-02-03 |
ATE65851T1 (en) | 1991-08-15 |
MX163339A (en) | 1992-04-24 |
IL77920A (en) | 1989-06-30 |
JPS62501889A (en) | 1987-07-30 |
EP0215818A4 (en) | 1988-01-28 |
EP0215818A1 (en) | 1987-04-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US4641349A (en) | Iris recognition system | |
US8644562B2 (en) | Multimodal ocular biometric system and methods | |
KR100342159B1 (en) | Apparatus and method for acquiring iris images | |
US8382285B2 (en) | Device and method for determining the orientation of an eye | |
Hill | Retina identification | |
EP1093633B1 (en) | Iris identification system and method of identifying a person through iris recognition | |
da Costa et al. | Dynamic features for iris recognition | |
US20160198952A1 (en) | Imaging Device | |
EP1642527B1 (en) | Organism eye judgment method and device | |
US20030012413A1 (en) | Iris identification apparatus and iris image pickup apparatus | |
JPH0218850B2 (en) | ||
JP3453911B2 (en) | Gaze recognition device | |
EP1587414B1 (en) | A pupilometer | |
US7226166B2 (en) | Optimizing the properties of electromagnetic energy in a medium using stochastic parallel perturbation gradient descent optimization adaptive optics | |
KR20010006975A (en) | A method for identifying the iris of persons based on the reaction of the pupil and autonomous nervous wreath | |
US20040165147A1 (en) | Determining iris biometric and spatial orientation of an iris in accordance with same | |
CN109063674A (en) | A kind of living iris detection method and detection device based on hot spot on eyeball | |
KR20010006976A (en) | A system for identifying the iris of persons | |
Drahanský et al. | Recognition of Eye Characteristics | |
JP2002056389A (en) | Iris identification system and method for identifying person by iris recognition | |
DellaVecchia et al. | Methodology and apparatus for using the human iris as a robust biometric | |
CN114495247A (en) | Iris positioning method, device and equipment | |
JPH1147117A (en) | Iris photographing device and iris pattern recognition device | |
KR100356600B1 (en) | A Method For Identifying The Iris Of Persons Based On The Shape Of Lacuna And/Or Autonomous Nervous Wreath | |
JP2001017410A (en) | Eye image pickup device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): BR JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH DE FR GB IT LU NL SE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1986901250 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1986901250 Country of ref document: EP |
|
WWG | Wipo information: grant in national office |
Ref document number: 1986901250 Country of ref document: EP |