CA1244552A - Iris recognition system - Google Patents
Iris recognition systemInfo
- Publication number
- CA1244552A CA1244552A CA000502207A CA502207A CA1244552A CA 1244552 A CA1244552 A CA 1244552A CA 000502207 A CA000502207 A CA 000502207A CA 502207 A CA502207 A CA 502207A CA 1244552 A CA1244552 A CA 1244552A
- Authority
- CA
- Canada
- Prior art keywords
- pupil
- iris
- descriptors
- eye
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired
Links
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/12—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for looking at the eye fundus, e.g. ophthalmoscopes
- A61B3/1216—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for looking at the eye fundus, e.g. ophthalmoscopes for diagnostics of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/141—Control of illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C9/00—Individual registration on entry or exit
- G07C9/30—Individual registration on entry or exit not involving the use of a pass
- G07C9/32—Individual registration on entry or exit not involving the use of a pass in combination with an identity check
- G07C9/37—Individual registration on entry or exit not involving the use of a pass in combination with an identity check using biometric data, e.g. fingerprints, iris scans or voice recognition
Abstract
ABSTRACT OF THE DISCLOSURE
Methods and apparatus are disclosed for identi-fying an eye, especially a human eye, on the basis of the visible features of the iris and pupil. The eye is first illuminated until the pupil reaches a predetermined size, at which an image of the iris and pupil is obtained.
This image is then compared with stored image information for identification. The stored image information is pre-viously obtained from an eye, the pupil of which was sim-ilarly brought to the same predetermined size. The illu-mination of the iris may include oblique illumination from several positions around the circumference of the iris. The illumination from each position may be rela-tively monochromatic, so that the resulting shadow will lack the color of the light source at that position, pro-viding better contrast for elevation-dependent features.
A system for performing iris recognition may include a processor which controls an illumination control circuit and a camera to obtain images at several predetermined sizes of the pupil.
Methods and apparatus are disclosed for identi-fying an eye, especially a human eye, on the basis of the visible features of the iris and pupil. The eye is first illuminated until the pupil reaches a predetermined size, at which an image of the iris and pupil is obtained.
This image is then compared with stored image information for identification. The stored image information is pre-viously obtained from an eye, the pupil of which was sim-ilarly brought to the same predetermined size. The illu-mination of the iris may include oblique illumination from several positions around the circumference of the iris. The illumination from each position may be rela-tively monochromatic, so that the resulting shadow will lack the color of the light source at that position, pro-viding better contrast for elevation-dependent features.
A system for performing iris recognition may include a processor which controls an illumination control circuit and a camera to obtain images at several predetermined sizes of the pupil.
Description
IRIS REC~GNITION SYSTEM
BACKGROUND OF THE IN~ENTION
1. Field of the Invention The invention relates to methods and apparatus for identification of physical characteristics of a human being or other animal. ~ore specifically, the invention relates to the recognition of the physical characteris-tics of the eyes.
BACKGROUND OF THE IN~ENTION
1. Field of the Invention The invention relates to methods and apparatus for identification of physical characteristics of a human being or other animal. ~ore specifically, the invention relates to the recognition of the physical characteris-tics of the eyes.
2 Description of the Prior Art Several known techniques are traditionally used to identi~y a human being from physical characteristics.
Fingerprints are one of the best known because of their e~tensive use by law enforcement agencies. Identifica-tion by fingerprint is especially helpful where an indi-vidual has left his fingerprint during the commission of a crime, but in general, the technology of fingèrprint identification relies heavily on human judgment. In ad-dition, the collection and detection of fingerprints is typically difficult.
More recently, U.S. Patent 4,10~,237 has dis-closed an identification technique making use of the ret-inal vasculature patterns of the human eye. Amon~ the advantages of this technique are that it can be automa-ted, and is simpler than detecting, collecting, and iden-tifying fingerprints. Furthermore, although some indivi-duals may attempt to alter their fingerprints in order to avoid identification, altering the retinal vasculature pattern would be difficult and dangerous, and might im-pair or destroy the vision in that eye.
It would be advantageous to have an identifica tion technique making use of the unique characteristics of the eye but not requiring pupil dilation a~d the com-plex alignment procedure of retinal identification, in-cluding focusing on the retina and maintaining alignment with the direction of gaze of the subject. It would also be advantageous to have such an identification technique which could be used with minimal cooperation of the sub-ject.
SUMMARY OF THE INVENTION
The present invention is based on the discovery that the iris and pupil of the eye, especially the human eye, may be used for identification. Therefore, the present invention provides an identification technique based upon the recognition of the unique features of the iris and pupil, referred to herein as "iris identifica-tion".
Iris identification shares the advantages of retinal identification, but can be performed more easily because the iris and pupil are more readily visible. In addition, some of the features of the iris and pupil are highly regular, providing a geometric pattern which fa-cilitates automatic identification. An important feature of the iris, however, is that its visible features change substantially as the pupil of the eye expands or con-tracts. Therefore, an important aspect of effective iris identification is to provide a technique which takes into account these changing features.
~%~5~
The method of identi~ication according to the invention includes illuminating an eye, obtaining an im-age of the iris ~nd pupil of the eye, and comparing the obtained image with stored image information to identify the eye. Illuminating the eye may include driving the pupil of the eye to a predetermined size, which may be done by directing light onto the retina of the eye and adjusting the intensity of the light until the pupil reaches the predetermined size. The pupil may be driven to a plurality o~ predetermined sizes, and a respective image of the iris and the pupil may be obtained at each of the predetermined si~es. The image obtained at each predetermined pupil size may be compared with stored im-age information from an eye with the same pupil size.
In another embodiment, illuminating the eye in-cludes obliquely illuminating the iris to provide shadows of elevation-dependent features. The iris may be illumi~
nated by a plurality of spatially separated, relatively monochromatic light sources, so that each shadow of one of the elevation-dependent features will lack a color corr~sponding to one of the light sources.
The comparing of the obtained image with stored image information may be a poin-t-by-point comparison with a stored re~erence image~ In an alternative form of com-parison, a set of descriptors may be derived from the just-obtained image and these descriptors may be compared with reference descriptors derived from a previous image.
The system accordin~ to the invention includes means for illuminating an eye, means for obtaining an image of the iris and pupil of the eye, and means for com-paring the obtained image with stored image in~ormation to make an identification. The illuminating means may in-clude an adjustable light source for illuminating the eye at a selected value. In addi-tion, the illuminating means ~24~t~
may include control means which receive the obtained im-age of the eye and detect the size of the pupil, the con trol means being operable for adjusting the adjustable light source to a selected light intensity for control-ling the size of the pupil. The control means may fur-ther be operable to control the size of the pupil to ob-tain a plurality of predetermined pupil sizes.
The illuminating means may also include a p:Lur-ality o-f relatively monochromatic, spatially separated light sources for providing shadows of elevation-depen-dent features. Each resulting shadow of an elevation-dependent feature will lack a color corresponding to the color of one of the light sources.
The image-obtalning means may include a camera, and the comparing means may include a processor respon-sive to the output from the camera. The processor may include storage means for storing reference descriptors derived from a previous image, so that the processor may be operable to derive a set oP descriptors from the just-obtained image and to compare these descriptors with the reference descriptors stored in the storage means.
Other obiects, features and advantages of the invention will be apparent from the following descrip-tion, together with the accompanying drawings and the appended claims.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 is a schematic view of the iris and pupil of an eye, showing the major visible features.
Fig. 2 is a general flow chart showin~ the major functions performed in the method of Che invention.
Fig. 3 is a simplified block diagram showin~ the flow of information through a system according to the invention.
~.2~5iS'~
Fig. 4 is a flow chart showing in greater detail an embodiment of the method of Fig. 2~
Fig. 5 is a flow chart showing a method of mea-suring rate of pupil size change according to the inven-tio~.
Fig~ 6 is a simplified flow chart showing a method of side illumination according to the invention.
Figo 7 is a general flow chart showing a method of comparing accordin~ to the invention.
Fig. 8 is a schematic diagram showing the elec-trical and optical characteristics o~ a system according to the invention.
Fig. 9 is a plan view of a viewing screen for use in the system of Fig. 8.
Fig. 10 is a functional block diagram of one embodiment of the system of Fig~ 3.
Fig. 11 is a diagram illustrating the illumina-tion control circuit of Fig. 10, Fig. 12 is a flow chart of the operation of the processor of Fig. 10 according to the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
I. General Description Fig. 1 is a simplified drawing of an eye 10, showing in detai} the features of the iris 20 and pupil 30. Eye 10, as shown, is a human eye, but some other animals have a similar eye structure, permitting identi-fication by the method of the invention. As shown, pupil 30 is a black, light receptive orifice, while iris 20 may have any of a broad ran~e of hues and may be irregularly pigmented.
The iris 20 and pupil 30 are closely related in function. More precisely, one function of iris 20 is to control the size o~ pupil 30. Therefore, in addition to s~
the visible features of iris 20 relating to pignentation, a number of visible features relate to the movements m~de by iris 20 in order to contract and dilate pupil 30. In general, iris 20 may be divided i~ltO the ciliary area 22, which is an annular region at its periphery, and the pupillary area 24, which is an annular area bordering pupil 30. In general, the tissue o~ iris 20 is soft and loosely woven, and the illumination which enters the pupil 30 and falls on the retina of eye 10 controls mus-cles in iris 20 causing the size of pupil 30 to grow lar-ger or smaller to regulate the amount of light entering pupil 30. The chan~e in the size of pupil 30 results from involuntary reflexes and is not under conscious con-trol.
Among the pigment-related features of iris 20 are pigment spots 26, some types of which are known as moles, freckles, or nevi, and crypts 2~. Pigment spots 26 are random concentrations of pigment-bearing cells in the visible surface of the iris 20, and generally appear in the ciliary area 22. The color of the pigment spots 26 may be nearly black -- darker than the darkest brown iris 20. In general, pigment spots 26 are quite stable in number, position and colorO Crypts 28, unlike pigment spots 26, relate both to pigmentation and to the sur~ace structure of the iris 20. Iris 20 includes an anterior layer which is directly visible and a posterior layer behind it, which is very darkly pigmented. The anterior layer ranges in pigment from a very light, almost white color to a dark brown color. If the anterior layer is very light or white, the iris 20 appears blue due to the pigment in the posterior layer, and if iris 20 is rela-tively li~ht in color, the areas in which it is rela-tively thin will appear somewhat darker, due to the dark color of the posterior layer. These darkened areas may occur, for example, in the -features re~erred to as crypts ., 5~Z
28, which resemble sharply demarcated excavations and are typically located adjacent to the colla;rette 32, the boundary between ciliary area 22 and pupillary area 24.
In addition to the pigment-related features of the iris 20, several other visible features relate to its function of controlling the size of pupil 30. Collarette 32, discussed briefly above, is the boundary bet~een the ciliary area 22 and the pupillary area 24 J and is typic-ally a sinuous li~e which forms an elevated ridge running roughly parallel with the margin of the pupil 30. The collarette 32 is the thickest part of iris 20.
Extending radially in relation to the center of pupil 30 are radial furrows 34. A typical radial furrow 34 may begin near pupil 30 and extend through collarette 32. Radial furrows 34 are creases in the anterior layer of iris 20, from which loose tissue may bulge outward, and it is this loose tissue which permits the iris 20 to expand or contract, changing the size of pupil 30. The positions of radial furrows 34 in relation to each other are stable.
Similarly, concentric ~urrows 36 are creases from which loose tissue may bulge outward, but their shape is generally circular, concentric with the pupil J
so that they permit e~pansion and contraction of the iris in a different direction than radial furrows 34. Concen-tric furrows 36 typically appear in the ciliary area 22, near the periphery of iris 20.
Eye 10 may also have any of several rare anoma-lous visible features (not shown) unrelated to pigment or to controlling the size of pupil 30. Due to aging or trauma, atrophic areas may appear on iris 20, resulting in a "moth-eaten" texture. Tumors may grow on the iris.
Con~enital filaments may occur connecting the iris to the lens of -the eye.
s~
Finally, the most striking visible feature of eye 10 is typically pupil 300 As shown in Fig. 1, pupil 30 may not be exactly circular in shape, and its devia-tion from a circle is a visible characteristic. At the margin of pupil 3~ is a protruding portion of the poster-ior layer of iris ~0, the pigment frill 38. Pigment frill 38 typically has a very dark brown color.
The present invention, as discussed above, is based on the discovery that the iris of the eye may be used for identification. This discovery involves both the uniqueness of the iris aud the stability o-f the iris over time. It has been discovered that every iris is unique, particularly in the detailed structure of the front or anterior layer. Not only are the irises of the eyes of identical twins different, but the iris of each eye of any person is different from that of his other eye. Furthermore, although specific details of the ap-pearance of an iris vary dramatically, depending on level and direction of illumination, the basic, significant features of the iris remain e~tremely stable and do not change over a period of many years. Even featllres which do develop over time, such as the atrophic areas dis-cussed above, usually develop rather slowly, so that an updated iris image will permit identification ~or a sub-stan~ial length of time. A sudden or rapid change in such a feature may result in a failure to identify an individual, but this may alert the individual to the pos-sibility of pathology of the eye. In general, the visi-ble features of the iris and pupil may be used to iden-tify an eye.
The method of the invention may be generally understood from Fig. 2, a flow chart showinu the basic functions performed in identifying an eye from the visi-ble features of the iris and pupil. As shown in block 40, the first function is to illuminate the eye. Because 55~
the visible features of iris 20 and pupil 30 vary depend-ing on illumination, the illumination oE the eye will be performed to obtain a desired set of features. Then, in block 50, an image of iris 20 and pupil 30 is obtained, including the desired features. In block 60, this image is compared with stored image information in order to identify the eye.
Fig. 3-is a block diagram showing the major functional components of the system according to the in-vention. Illuminating means 70 provides light havi~g desired characteristics to the eye 10 under observation.
Eye 10 responds to the illumination provided, so that iris 20 and pupil 30 take a characteristic shape. In addition, iris 20 reflects light to imaging means 80, which obtains an image of iris 20 and pupil 30. This image is then compared with stored irnage information by comparing means 90 in order to identify the eye.
From these basic principles, a variety of meth-ods and systems may be provided for identifying an eye from the iris and pupil.
II. Methods of Identification A number of methods for implementing the inven-tion are shown in Figs. 4-7. Each of these figures is a flow chart showing a sequence of steps performed in a process of identifying an eye from the iris and pupil.
Fig. 4 shows a basic method of performing the invention in which the pupil is driven to at least one predetermined size at which an image is taken. In step 42, the pupil is illuminated at one extreme o the ran~e of illumination, either an extreme of darkness or of brightness. The extreme may be determined by the capa-bilities of the system or may be arrived at by a straightforward series of steps, For example, the illu~
mination may gradually be increased from a moderate level 55'~
of illumination to an illumination beyo~d which further illumination does not decrease the size of the pupil. A
more efficient approach, however, is to drive the pupil uutil its size is larger or smaller than the largest or smallest of the predetermined sizes at ~hich an image will be obtained. From this extreme illumination, it will then be possible to drive the pupil to the predeter-mined size for imaging.
From the size which it has after step 42, the pupil is driven to the first predetermined size in step 44. Assuming that the extreme of illumination was dark-ness, the illumination will be increased in step 44 until the pupil contracts to a predetermined size. When the pupil reaches this size, an image i9 obtained and stored in step 52 for subsequent comparison. Then, in step 54, a test is run to determine whether the number n, a count of the images obtained, is greater than or equal to the number N of predetermined sizes at which images are to be obtained. If not, the count of the ima~es obtained, n, is incremented in step 46, and steps 44, 52 and 54 are repeated for the next predetermined size. When all of the predetermined sizes have been imaged, step 54 will determine that n is equal to N, and image comparison will be done in step 62. In step 62, the image at each of the N predetermined sizes is compared with stored ima~e in-formation obtained from an eye illuminated to have the same predetermined pupil size.
Many variations may be made in the basic method of Fig. 4. For example, the largest size which the pupil attains as illumination is decreased could be treated as the first predetermined size, so that an ima~e would be obtained after step 42. If the maximum size of the pupil remains relatively constant over time, this size will be valuable as a descriptor of the eye.
-- 11 ~
Fig, 5 lllustrates another variation which can be made in the basic method of the invention. This vari-ation includes measurin~ the rate at which the pupil si2e changes between two predetermined sizes. In step 142, light from a light source is directed onto the retina of the eye in order to control the size of the pupil through refIex action. Then, in step 1~4> the position or inten-sity of the li~ht source is adjusted in order to obtain the first predetermined pupil size. In other ~ords, as an alternative to varying the light intensity, the light source may be moved from a central position in which it directs light onto the central area of the retina toward an oblique position in which it directs light at an angle through the pupil onto the more peripheral parts of the retina, or vice versa. After each adjustment, in step 144, a test is performed in step 146 to determine whether the first predetermined size has been obtained. If ns~, the adjustment is a~ain performed in step 144, and this loop is repeated until the first predetermined size is obtained.
After the first predetermined size is obtained, the illumination of the retina is changed at a predeter-mined rate in step 148. The changing of the illumination is continued until the test in step 150 determines that a second predetermined size has been attained. At that point, the time necessary to go from the first predster-mined size to the second predetermined size at the prede-termined rate will provide a measure of the rate of change, in step 152. This measure could also be related to the chan~e in size between the first and second prede-termined size and to the specific predetermined rate at which the illumination is changed in step 148. The re-sulting value may then be used to identify the eye. It should be borne in mind, however, that the rate of change of pupil size may vary for individuals whose refle~Yes are 55;~
affected by emotion, drugs, and so forth, so that the rate of change of pupil diameter should generally provide only a secondary indication of identity.
The methods shown in Figs. 4 and 5 could be com-bined in a simple manner by making minor modifications to the steps performed in Fig. 4. Step 44 could first be performed to obtain the first predetermined size, corres-ponding to steps 144 and i46 in Fig. 5. After the per-formance of steps 52, 54 and 46, step 44 could again be performed by changing the illumination at the predeter-mined rate to obtain the second predetermined size, cor-responding to steps 148 and lS0 in Fig. 5. Then, in ad-dition to obtaining the image in step 52, the rate of change of pupil size could also be determined in step 152.
Fi~. 6 illustrates a routine which may similarly be included in step 52 of Fig. 4, to obtain an image of the iris of an eye in which the elevation-dependent fea-tures are shown with greater contrast. This is done by illuminating the iris from a plurality of oblique posi-tions. The illumination from each position is provided by a relatively monochromatic light source to produce shadows of elevation-dependent features lacking the color of that light source. In step 250, illumination from position m is performed using the color m of the corres-ponding relatively monochromatic light source. This il-lumination will produce shadows lacking the color m, and an image of these shadows is obtained in step 252. Then, in step 254, a test is performed to determine whether m has reached M, the total number of positions from which the iris is to be illuminated. If not, _ is incremented in step 256, and the loop including steps 250, 252 and 254 is repeated until m equals ~1. When m reaches ~, the shadow images may be processed in step 258 to prepare tllem for subsequent comparison with re~erence images.
, ~ , ~4~
The method o~ Fig. 6 is especially beneficial for obtaining images of elevation-dependent features such as radial furrows 34. As can be seen from Fig. 1, how-ever, radial furrows 34 may e~tend radially in any direc-tion, so that illumination from any one direction will provide an excellent shadow image of radial furrows 34 which extend perpendicular to that direction of illumina-tion. Radial furrows which e~tend parallel to the direc-tion of illumination, however, will produce poor shadow images, Therefore, to produce the best images of radial furrows 34, it is desirable to obtain a number of shadow images and to combine them in processing step 258.
The comparison of images having the same pupil sizes in step 62 of Fig. 4 may be performed in many ways~
One straightforward method would be to store the refer-ence image as a pixel-by-pixel image resulting from a scanning camera such as a conventional video camera. The image obtained would then be obtained with a similar video camera, as discussed below in relation to Fig. 8, to obtain a pi~el-by-pixel image of the eye. The two images could then be compared, pixel-by-pixel, to deter-mine whether they were the same image. The comparison could be made in a way which would ta~e into account dif-ferences in the overall intensity of the two ima~es.
This method of comparison, although simple to implement, is e~tremely sensitive to re~istration. In other words, the test would only result in a conclusion that the two images were of the same eye if the eyes shown in the two images were of the same size and were positioned identi-cally. Therefore, it will be preferable to perform a comparison algorithm which is not so sensitive to regis-tration.
One comparison algorithm which would not be so sensitive to registration is the correlation algorithm, which has many variations. ~he ef~ect of an offset in the positioning of the two images could be eliminated based on an algorithm for determining the center of grav-ity of the pupil in each ima~a. These algorithms could thus be combined to ~educe registration sensitivity.
Another comparison method which would not be highly sensitive to registration would be a method using optical signal processin~, Hecht, J., "Light Modulators Help Crunch Image Data", High Technolo~y, Vol. 5, No. 1 (January, 1985), p. 69-70, 72, discusses techniques for optical comparison which would be readily applicable to the present invention.
Fig. 7 illustrates another method of comparison which could be used in the present invention. In step 64, descriptors are extracted from the ima~e obtained, which may be the image obtained in step 52 of Fig. 4.
One e~ample of descriptors which could be extracted would be the angular positions and lengths of the radia} fur-rows 34. Another set of descriptors would be the shape of the pupil 30 or of ~he pigment frill 38. Other de-scriptor sets could be obtained relating to the positions and ~izes o~ pigment spots 26, crypts 28 and anomalous features such as atrophic areas, tumors and congenital filaments. Similarly, generally circular features such as collarettes 32 and concentric furrows 36 could be de-scribed by respective descriptors. All of these descrip-tors would then be compared with the descriptors of a reference image in step 66. Because several features of the iris 20 and pupil 30 lend themselves readily to math-ematical description, this method of comparison should be effective and should not be sensitive to re~istration.
A number of al~orithms for extracting descriptors are discussed below.
~ Many types of equipment could be used to imple-ment the inven-tion and to perform the methods set forth above, One example of circuitry for performin~ these ~unctions is described below.
i~
5~i~
III. Identi~ication Systems As noted above, the necessary equipment ~or per-forming iris recognition according to the invention could take many forms. The major functional components of any such system, however, will be those shown in Fig. 3, in-cluding illuminating means 70, imaging means 80 and com-paring means 90.
Fig. 8 shows a system for illuminating and ob-taining an image of the eye. The system includes a sim-plified electrical circuit for controlling illumination and an optical arrangement in which the image of fi~ation target 71 seen by eye 10 and the reflected image of iris 20 and pupil 30 to camera 84 are approximately coaxial.
It is presently preferred to provide a target image, upon which eye 10 may be fixated, concentric with visual axis 16 for reference purposes. It is also preferred for cam-era 84 to capture the reflected image concentric or nearly concentric with visual axis 16, because if the image were obtained far from the axis 16, it would be distorted, which would then have to be compensated by appropriate processing. Additional illumination of iris 20 is provided obliquely to create good shadows of eleva-tion-dependent features.
In the system of Fig. 8, the source of light is a filament 72a which receives power from power source 72b connected in series through variable resistor 72c, which may be an electrically controlled potentiometer. The light emitted by filament 72a is collimated by lens 73 and illuminates target 71, which may have any appro~riate shape.
The image of target 71 is then focused by mov-able lens 74 and transmitted through monitor beam-split-ter 76 onto axial beam-splitting plane 82. The image is then reflected from the central mirrored area 82a through pupil 30 onto retina 12.
:, ~L2~L5~
~ he optical system shown in Fig. 8 includes means for placing target 71 at any desired optical dis-tance from eye 10. Target 71 can thus be made visible to subjects with a wide variety of refractive states. The refraction or bending of light rays at the air interface of cornea 14, for e~ample, is substantial and varies for different subjects. The necessary adjustment of optical distance is obtained by moving lens 74 toward or away from target 71 until target 71 is optically conjugate with retina 12. If necessary to compensate astigmatism, lens 74 may also be a variable sphero-cylindrical lens system of any of the well-known typesO Target 71 will be optically conjugate with retina 12 when, as shown in Fig.
8, the light from each point on target 71 comes to a focus at a corresponding point on retina 12. The conver-gence or divergence of light from target 71 due to the adjustment of lens 74 will thus compensate for the optics of eye 10.
Oblique light sources 78a and 78b provide illu-mination to the iris 20 and may ~e controlled similarly to filament 72a by using a variable resistor (not shown).
The resulting reflection will depend on the positioning of oblique light sources 78a and 78b, because the iris is a diffuse reflector, reflecting light according to a co-sine law in which the amount of reflected light in the direction nor~al to the iris varies as the cosine of the incident angle of the light.
Axial beam-splitting plane 82 serves the two functions of directing the image of target 71 froln lens 74 toward eye 10 and also transmitting the reflected im-age of iris 20 to camera 84. In the arrangement shown in Fig. 8, which is not to scale, beam splitting plane 82 is a thin transparent sheet such as glass on which mirror 82a is formed for reflecting the image. Mirror ~2a may be a microscopically thin, reflective film or coating, 5~
such as a metal, on the central part of plane 82 and on the side toward eye 10 7 to provide a first surface mir-ror. It may, for e~ample, have an elliptical shape. The reflected image of target 71 is directed through pupil 30 onto the retina 12 of eye 10 to control the size of pupil 30. The reflected light ~rom iris 20,,which will emerge from eye 10 quite divergently, may pass through the clearj transparent area of beam splittia~ plane 82 which surrounds mirror 82a. Mirror 82a, however, will create a non-transparent area in the center of plane 82, preven~-ing an intense reflection of the target image -from cornea 1~ from reaching camera 84. Since the reflected target image may be the bri~htest object in the field of view of camera 84, this may si~nificantly reduce the noise in the image of iris 20 and pupil 30. Monitor beam-splitter 76 will provide an image of the intense reflection from cor-nea 1~, as reflected by mirror 82a, and this reflection image may be used to position the eye according to known eye-positioning techniques. Proper positioning of the eye will help to provide a standardized image of iris 20 and pupil 30 and will ensure that the intense reflection of the target image does not reach camera 84.
Fig. 9 shows viewing screen 86 of camera 84, upon which are marked peripheral points 88a and central point 88b. Peripheral points 88a may be used to measure the size of pupil 30, as shown. When the perimeter of pupil 30 touches each of the peripheral points 88a, pupil 30 has been driven to a predetermined size at which an image should be obtained. Central point 88b may be use-ful in aligning pupil 30 so that size may be properly tested. As shown in Fi~. 9, a grid may also be provided on viewing screen 86 to assist in the determination of the size of pupil 30 when it does not meet peripheral ~2~55~
points 88a. Alternatively, concentric markings could be provided on screen 86, as disclosed in U.S. Patent
Fingerprints are one of the best known because of their e~tensive use by law enforcement agencies. Identifica-tion by fingerprint is especially helpful where an indi-vidual has left his fingerprint during the commission of a crime, but in general, the technology of fingèrprint identification relies heavily on human judgment. In ad-dition, the collection and detection of fingerprints is typically difficult.
More recently, U.S. Patent 4,10~,237 has dis-closed an identification technique making use of the ret-inal vasculature patterns of the human eye. Amon~ the advantages of this technique are that it can be automa-ted, and is simpler than detecting, collecting, and iden-tifying fingerprints. Furthermore, although some indivi-duals may attempt to alter their fingerprints in order to avoid identification, altering the retinal vasculature pattern would be difficult and dangerous, and might im-pair or destroy the vision in that eye.
It would be advantageous to have an identifica tion technique making use of the unique characteristics of the eye but not requiring pupil dilation a~d the com-plex alignment procedure of retinal identification, in-cluding focusing on the retina and maintaining alignment with the direction of gaze of the subject. It would also be advantageous to have such an identification technique which could be used with minimal cooperation of the sub-ject.
SUMMARY OF THE INVENTION
The present invention is based on the discovery that the iris and pupil of the eye, especially the human eye, may be used for identification. Therefore, the present invention provides an identification technique based upon the recognition of the unique features of the iris and pupil, referred to herein as "iris identifica-tion".
Iris identification shares the advantages of retinal identification, but can be performed more easily because the iris and pupil are more readily visible. In addition, some of the features of the iris and pupil are highly regular, providing a geometric pattern which fa-cilitates automatic identification. An important feature of the iris, however, is that its visible features change substantially as the pupil of the eye expands or con-tracts. Therefore, an important aspect of effective iris identification is to provide a technique which takes into account these changing features.
~%~5~
The method of identi~ication according to the invention includes illuminating an eye, obtaining an im-age of the iris ~nd pupil of the eye, and comparing the obtained image with stored image information to identify the eye. Illuminating the eye may include driving the pupil of the eye to a predetermined size, which may be done by directing light onto the retina of the eye and adjusting the intensity of the light until the pupil reaches the predetermined size. The pupil may be driven to a plurality o~ predetermined sizes, and a respective image of the iris and the pupil may be obtained at each of the predetermined si~es. The image obtained at each predetermined pupil size may be compared with stored im-age information from an eye with the same pupil size.
In another embodiment, illuminating the eye in-cludes obliquely illuminating the iris to provide shadows of elevation-dependent features. The iris may be illumi~
nated by a plurality of spatially separated, relatively monochromatic light sources, so that each shadow of one of the elevation-dependent features will lack a color corr~sponding to one of the light sources.
The comparing of the obtained image with stored image information may be a poin-t-by-point comparison with a stored re~erence image~ In an alternative form of com-parison, a set of descriptors may be derived from the just-obtained image and these descriptors may be compared with reference descriptors derived from a previous image.
The system accordin~ to the invention includes means for illuminating an eye, means for obtaining an image of the iris and pupil of the eye, and means for com-paring the obtained image with stored image in~ormation to make an identification. The illuminating means may in-clude an adjustable light source for illuminating the eye at a selected value. In addi-tion, the illuminating means ~24~t~
may include control means which receive the obtained im-age of the eye and detect the size of the pupil, the con trol means being operable for adjusting the adjustable light source to a selected light intensity for control-ling the size of the pupil. The control means may fur-ther be operable to control the size of the pupil to ob-tain a plurality of predetermined pupil sizes.
The illuminating means may also include a p:Lur-ality o-f relatively monochromatic, spatially separated light sources for providing shadows of elevation-depen-dent features. Each resulting shadow of an elevation-dependent feature will lack a color corresponding to the color of one of the light sources.
The image-obtalning means may include a camera, and the comparing means may include a processor respon-sive to the output from the camera. The processor may include storage means for storing reference descriptors derived from a previous image, so that the processor may be operable to derive a set oP descriptors from the just-obtained image and to compare these descriptors with the reference descriptors stored in the storage means.
Other obiects, features and advantages of the invention will be apparent from the following descrip-tion, together with the accompanying drawings and the appended claims.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 is a schematic view of the iris and pupil of an eye, showing the major visible features.
Fig. 2 is a general flow chart showin~ the major functions performed in the method of Che invention.
Fig. 3 is a simplified block diagram showin~ the flow of information through a system according to the invention.
~.2~5iS'~
Fig. 4 is a flow chart showing in greater detail an embodiment of the method of Fig. 2~
Fig. 5 is a flow chart showing a method of mea-suring rate of pupil size change according to the inven-tio~.
Fig~ 6 is a simplified flow chart showing a method of side illumination according to the invention.
Figo 7 is a general flow chart showing a method of comparing accordin~ to the invention.
Fig. 8 is a schematic diagram showing the elec-trical and optical characteristics o~ a system according to the invention.
Fig. 9 is a plan view of a viewing screen for use in the system of Fig. 8.
Fig. 10 is a functional block diagram of one embodiment of the system of Fig~ 3.
Fig. 11 is a diagram illustrating the illumina-tion control circuit of Fig. 10, Fig. 12 is a flow chart of the operation of the processor of Fig. 10 according to the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
I. General Description Fig. 1 is a simplified drawing of an eye 10, showing in detai} the features of the iris 20 and pupil 30. Eye 10, as shown, is a human eye, but some other animals have a similar eye structure, permitting identi-fication by the method of the invention. As shown, pupil 30 is a black, light receptive orifice, while iris 20 may have any of a broad ran~e of hues and may be irregularly pigmented.
The iris 20 and pupil 30 are closely related in function. More precisely, one function of iris 20 is to control the size o~ pupil 30. Therefore, in addition to s~
the visible features of iris 20 relating to pignentation, a number of visible features relate to the movements m~de by iris 20 in order to contract and dilate pupil 30. In general, iris 20 may be divided i~ltO the ciliary area 22, which is an annular region at its periphery, and the pupillary area 24, which is an annular area bordering pupil 30. In general, the tissue o~ iris 20 is soft and loosely woven, and the illumination which enters the pupil 30 and falls on the retina of eye 10 controls mus-cles in iris 20 causing the size of pupil 30 to grow lar-ger or smaller to regulate the amount of light entering pupil 30. The chan~e in the size of pupil 30 results from involuntary reflexes and is not under conscious con-trol.
Among the pigment-related features of iris 20 are pigment spots 26, some types of which are known as moles, freckles, or nevi, and crypts 2~. Pigment spots 26 are random concentrations of pigment-bearing cells in the visible surface of the iris 20, and generally appear in the ciliary area 22. The color of the pigment spots 26 may be nearly black -- darker than the darkest brown iris 20. In general, pigment spots 26 are quite stable in number, position and colorO Crypts 28, unlike pigment spots 26, relate both to pigmentation and to the sur~ace structure of the iris 20. Iris 20 includes an anterior layer which is directly visible and a posterior layer behind it, which is very darkly pigmented. The anterior layer ranges in pigment from a very light, almost white color to a dark brown color. If the anterior layer is very light or white, the iris 20 appears blue due to the pigment in the posterior layer, and if iris 20 is rela-tively li~ht in color, the areas in which it is rela-tively thin will appear somewhat darker, due to the dark color of the posterior layer. These darkened areas may occur, for example, in the -features re~erred to as crypts ., 5~Z
28, which resemble sharply demarcated excavations and are typically located adjacent to the colla;rette 32, the boundary between ciliary area 22 and pupillary area 24.
In addition to the pigment-related features of the iris 20, several other visible features relate to its function of controlling the size of pupil 30. Collarette 32, discussed briefly above, is the boundary bet~een the ciliary area 22 and the pupillary area 24 J and is typic-ally a sinuous li~e which forms an elevated ridge running roughly parallel with the margin of the pupil 30. The collarette 32 is the thickest part of iris 20.
Extending radially in relation to the center of pupil 30 are radial furrows 34. A typical radial furrow 34 may begin near pupil 30 and extend through collarette 32. Radial furrows 34 are creases in the anterior layer of iris 20, from which loose tissue may bulge outward, and it is this loose tissue which permits the iris 20 to expand or contract, changing the size of pupil 30. The positions of radial furrows 34 in relation to each other are stable.
Similarly, concentric ~urrows 36 are creases from which loose tissue may bulge outward, but their shape is generally circular, concentric with the pupil J
so that they permit e~pansion and contraction of the iris in a different direction than radial furrows 34. Concen-tric furrows 36 typically appear in the ciliary area 22, near the periphery of iris 20.
Eye 10 may also have any of several rare anoma-lous visible features (not shown) unrelated to pigment or to controlling the size of pupil 30. Due to aging or trauma, atrophic areas may appear on iris 20, resulting in a "moth-eaten" texture. Tumors may grow on the iris.
Con~enital filaments may occur connecting the iris to the lens of -the eye.
s~
Finally, the most striking visible feature of eye 10 is typically pupil 300 As shown in Fig. 1, pupil 30 may not be exactly circular in shape, and its devia-tion from a circle is a visible characteristic. At the margin of pupil 3~ is a protruding portion of the poster-ior layer of iris ~0, the pigment frill 38. Pigment frill 38 typically has a very dark brown color.
The present invention, as discussed above, is based on the discovery that the iris of the eye may be used for identification. This discovery involves both the uniqueness of the iris aud the stability o-f the iris over time. It has been discovered that every iris is unique, particularly in the detailed structure of the front or anterior layer. Not only are the irises of the eyes of identical twins different, but the iris of each eye of any person is different from that of his other eye. Furthermore, although specific details of the ap-pearance of an iris vary dramatically, depending on level and direction of illumination, the basic, significant features of the iris remain e~tremely stable and do not change over a period of many years. Even featllres which do develop over time, such as the atrophic areas dis-cussed above, usually develop rather slowly, so that an updated iris image will permit identification ~or a sub-stan~ial length of time. A sudden or rapid change in such a feature may result in a failure to identify an individual, but this may alert the individual to the pos-sibility of pathology of the eye. In general, the visi-ble features of the iris and pupil may be used to iden-tify an eye.
The method of the invention may be generally understood from Fig. 2, a flow chart showinu the basic functions performed in identifying an eye from the visi-ble features of the iris and pupil. As shown in block 40, the first function is to illuminate the eye. Because 55~
the visible features of iris 20 and pupil 30 vary depend-ing on illumination, the illumination oE the eye will be performed to obtain a desired set of features. Then, in block 50, an image of iris 20 and pupil 30 is obtained, including the desired features. In block 60, this image is compared with stored image information in order to identify the eye.
Fig. 3-is a block diagram showing the major functional components of the system according to the in-vention. Illuminating means 70 provides light havi~g desired characteristics to the eye 10 under observation.
Eye 10 responds to the illumination provided, so that iris 20 and pupil 30 take a characteristic shape. In addition, iris 20 reflects light to imaging means 80, which obtains an image of iris 20 and pupil 30. This image is then compared with stored irnage information by comparing means 90 in order to identify the eye.
From these basic principles, a variety of meth-ods and systems may be provided for identifying an eye from the iris and pupil.
II. Methods of Identification A number of methods for implementing the inven-tion are shown in Figs. 4-7. Each of these figures is a flow chart showing a sequence of steps performed in a process of identifying an eye from the iris and pupil.
Fig. 4 shows a basic method of performing the invention in which the pupil is driven to at least one predetermined size at which an image is taken. In step 42, the pupil is illuminated at one extreme o the ran~e of illumination, either an extreme of darkness or of brightness. The extreme may be determined by the capa-bilities of the system or may be arrived at by a straightforward series of steps, For example, the illu~
mination may gradually be increased from a moderate level 55'~
of illumination to an illumination beyo~d which further illumination does not decrease the size of the pupil. A
more efficient approach, however, is to drive the pupil uutil its size is larger or smaller than the largest or smallest of the predetermined sizes at ~hich an image will be obtained. From this extreme illumination, it will then be possible to drive the pupil to the predeter-mined size for imaging.
From the size which it has after step 42, the pupil is driven to the first predetermined size in step 44. Assuming that the extreme of illumination was dark-ness, the illumination will be increased in step 44 until the pupil contracts to a predetermined size. When the pupil reaches this size, an image i9 obtained and stored in step 52 for subsequent comparison. Then, in step 54, a test is run to determine whether the number n, a count of the images obtained, is greater than or equal to the number N of predetermined sizes at which images are to be obtained. If not, the count of the ima~es obtained, n, is incremented in step 46, and steps 44, 52 and 54 are repeated for the next predetermined size. When all of the predetermined sizes have been imaged, step 54 will determine that n is equal to N, and image comparison will be done in step 62. In step 62, the image at each of the N predetermined sizes is compared with stored ima~e in-formation obtained from an eye illuminated to have the same predetermined pupil size.
Many variations may be made in the basic method of Fig. 4. For example, the largest size which the pupil attains as illumination is decreased could be treated as the first predetermined size, so that an ima~e would be obtained after step 42. If the maximum size of the pupil remains relatively constant over time, this size will be valuable as a descriptor of the eye.
-- 11 ~
Fig, 5 lllustrates another variation which can be made in the basic method of the invention. This vari-ation includes measurin~ the rate at which the pupil si2e changes between two predetermined sizes. In step 142, light from a light source is directed onto the retina of the eye in order to control the size of the pupil through refIex action. Then, in step 1~4> the position or inten-sity of the li~ht source is adjusted in order to obtain the first predetermined pupil size. In other ~ords, as an alternative to varying the light intensity, the light source may be moved from a central position in which it directs light onto the central area of the retina toward an oblique position in which it directs light at an angle through the pupil onto the more peripheral parts of the retina, or vice versa. After each adjustment, in step 144, a test is performed in step 146 to determine whether the first predetermined size has been obtained. If ns~, the adjustment is a~ain performed in step 144, and this loop is repeated until the first predetermined size is obtained.
After the first predetermined size is obtained, the illumination of the retina is changed at a predeter-mined rate in step 148. The changing of the illumination is continued until the test in step 150 determines that a second predetermined size has been attained. At that point, the time necessary to go from the first predster-mined size to the second predetermined size at the prede-termined rate will provide a measure of the rate of change, in step 152. This measure could also be related to the chan~e in size between the first and second prede-termined size and to the specific predetermined rate at which the illumination is changed in step 148. The re-sulting value may then be used to identify the eye. It should be borne in mind, however, that the rate of change of pupil size may vary for individuals whose refle~Yes are 55;~
affected by emotion, drugs, and so forth, so that the rate of change of pupil diameter should generally provide only a secondary indication of identity.
The methods shown in Figs. 4 and 5 could be com-bined in a simple manner by making minor modifications to the steps performed in Fig. 4. Step 44 could first be performed to obtain the first predetermined size, corres-ponding to steps 144 and i46 in Fig. 5. After the per-formance of steps 52, 54 and 46, step 44 could again be performed by changing the illumination at the predeter-mined rate to obtain the second predetermined size, cor-responding to steps 148 and lS0 in Fig. 5. Then, in ad-dition to obtaining the image in step 52, the rate of change of pupil size could also be determined in step 152.
Fi~. 6 illustrates a routine which may similarly be included in step 52 of Fig. 4, to obtain an image of the iris of an eye in which the elevation-dependent fea-tures are shown with greater contrast. This is done by illuminating the iris from a plurality of oblique posi-tions. The illumination from each position is provided by a relatively monochromatic light source to produce shadows of elevation-dependent features lacking the color of that light source. In step 250, illumination from position m is performed using the color m of the corres-ponding relatively monochromatic light source. This il-lumination will produce shadows lacking the color m, and an image of these shadows is obtained in step 252. Then, in step 254, a test is performed to determine whether m has reached M, the total number of positions from which the iris is to be illuminated. If not, _ is incremented in step 256, and the loop including steps 250, 252 and 254 is repeated until m equals ~1. When m reaches ~, the shadow images may be processed in step 258 to prepare tllem for subsequent comparison with re~erence images.
, ~ , ~4~
The method o~ Fig. 6 is especially beneficial for obtaining images of elevation-dependent features such as radial furrows 34. As can be seen from Fig. 1, how-ever, radial furrows 34 may e~tend radially in any direc-tion, so that illumination from any one direction will provide an excellent shadow image of radial furrows 34 which extend perpendicular to that direction of illumina-tion. Radial furrows which e~tend parallel to the direc-tion of illumination, however, will produce poor shadow images, Therefore, to produce the best images of radial furrows 34, it is desirable to obtain a number of shadow images and to combine them in processing step 258.
The comparison of images having the same pupil sizes in step 62 of Fig. 4 may be performed in many ways~
One straightforward method would be to store the refer-ence image as a pixel-by-pixel image resulting from a scanning camera such as a conventional video camera. The image obtained would then be obtained with a similar video camera, as discussed below in relation to Fig. 8, to obtain a pi~el-by-pixel image of the eye. The two images could then be compared, pixel-by-pixel, to deter-mine whether they were the same image. The comparison could be made in a way which would ta~e into account dif-ferences in the overall intensity of the two ima~es.
This method of comparison, although simple to implement, is e~tremely sensitive to re~istration. In other words, the test would only result in a conclusion that the two images were of the same eye if the eyes shown in the two images were of the same size and were positioned identi-cally. Therefore, it will be preferable to perform a comparison algorithm which is not so sensitive to regis-tration.
One comparison algorithm which would not be so sensitive to registration is the correlation algorithm, which has many variations. ~he ef~ect of an offset in the positioning of the two images could be eliminated based on an algorithm for determining the center of grav-ity of the pupil in each ima~a. These algorithms could thus be combined to ~educe registration sensitivity.
Another comparison method which would not be highly sensitive to registration would be a method using optical signal processin~, Hecht, J., "Light Modulators Help Crunch Image Data", High Technolo~y, Vol. 5, No. 1 (January, 1985), p. 69-70, 72, discusses techniques for optical comparison which would be readily applicable to the present invention.
Fig. 7 illustrates another method of comparison which could be used in the present invention. In step 64, descriptors are extracted from the ima~e obtained, which may be the image obtained in step 52 of Fig. 4.
One e~ample of descriptors which could be extracted would be the angular positions and lengths of the radia} fur-rows 34. Another set of descriptors would be the shape of the pupil 30 or of ~he pigment frill 38. Other de-scriptor sets could be obtained relating to the positions and ~izes o~ pigment spots 26, crypts 28 and anomalous features such as atrophic areas, tumors and congenital filaments. Similarly, generally circular features such as collarettes 32 and concentric furrows 36 could be de-scribed by respective descriptors. All of these descrip-tors would then be compared with the descriptors of a reference image in step 66. Because several features of the iris 20 and pupil 30 lend themselves readily to math-ematical description, this method of comparison should be effective and should not be sensitive to re~istration.
A number of al~orithms for extracting descriptors are discussed below.
~ Many types of equipment could be used to imple-ment the inven-tion and to perform the methods set forth above, One example of circuitry for performin~ these ~unctions is described below.
i~
5~i~
III. Identi~ication Systems As noted above, the necessary equipment ~or per-forming iris recognition according to the invention could take many forms. The major functional components of any such system, however, will be those shown in Fig. 3, in-cluding illuminating means 70, imaging means 80 and com-paring means 90.
Fig. 8 shows a system for illuminating and ob-taining an image of the eye. The system includes a sim-plified electrical circuit for controlling illumination and an optical arrangement in which the image of fi~ation target 71 seen by eye 10 and the reflected image of iris 20 and pupil 30 to camera 84 are approximately coaxial.
It is presently preferred to provide a target image, upon which eye 10 may be fixated, concentric with visual axis 16 for reference purposes. It is also preferred for cam-era 84 to capture the reflected image concentric or nearly concentric with visual axis 16, because if the image were obtained far from the axis 16, it would be distorted, which would then have to be compensated by appropriate processing. Additional illumination of iris 20 is provided obliquely to create good shadows of eleva-tion-dependent features.
In the system of Fig. 8, the source of light is a filament 72a which receives power from power source 72b connected in series through variable resistor 72c, which may be an electrically controlled potentiometer. The light emitted by filament 72a is collimated by lens 73 and illuminates target 71, which may have any appro~riate shape.
The image of target 71 is then focused by mov-able lens 74 and transmitted through monitor beam-split-ter 76 onto axial beam-splitting plane 82. The image is then reflected from the central mirrored area 82a through pupil 30 onto retina 12.
:, ~L2~L5~
~ he optical system shown in Fig. 8 includes means for placing target 71 at any desired optical dis-tance from eye 10. Target 71 can thus be made visible to subjects with a wide variety of refractive states. The refraction or bending of light rays at the air interface of cornea 14, for e~ample, is substantial and varies for different subjects. The necessary adjustment of optical distance is obtained by moving lens 74 toward or away from target 71 until target 71 is optically conjugate with retina 12. If necessary to compensate astigmatism, lens 74 may also be a variable sphero-cylindrical lens system of any of the well-known typesO Target 71 will be optically conjugate with retina 12 when, as shown in Fig.
8, the light from each point on target 71 comes to a focus at a corresponding point on retina 12. The conver-gence or divergence of light from target 71 due to the adjustment of lens 74 will thus compensate for the optics of eye 10.
Oblique light sources 78a and 78b provide illu-mination to the iris 20 and may ~e controlled similarly to filament 72a by using a variable resistor (not shown).
The resulting reflection will depend on the positioning of oblique light sources 78a and 78b, because the iris is a diffuse reflector, reflecting light according to a co-sine law in which the amount of reflected light in the direction nor~al to the iris varies as the cosine of the incident angle of the light.
Axial beam-splitting plane 82 serves the two functions of directing the image of target 71 froln lens 74 toward eye 10 and also transmitting the reflected im-age of iris 20 to camera 84. In the arrangement shown in Fig. 8, which is not to scale, beam splitting plane 82 is a thin transparent sheet such as glass on which mirror 82a is formed for reflecting the image. Mirror ~2a may be a microscopically thin, reflective film or coating, 5~
such as a metal, on the central part of plane 82 and on the side toward eye 10 7 to provide a first surface mir-ror. It may, for e~ample, have an elliptical shape. The reflected image of target 71 is directed through pupil 30 onto the retina 12 of eye 10 to control the size of pupil 30. The reflected light ~rom iris 20,,which will emerge from eye 10 quite divergently, may pass through the clearj transparent area of beam splittia~ plane 82 which surrounds mirror 82a. Mirror 82a, however, will create a non-transparent area in the center of plane 82, preven~-ing an intense reflection of the target image -from cornea 1~ from reaching camera 84. Since the reflected target image may be the bri~htest object in the field of view of camera 84, this may si~nificantly reduce the noise in the image of iris 20 and pupil 30. Monitor beam-splitter 76 will provide an image of the intense reflection from cor-nea 1~, as reflected by mirror 82a, and this reflection image may be used to position the eye according to known eye-positioning techniques. Proper positioning of the eye will help to provide a standardized image of iris 20 and pupil 30 and will ensure that the intense reflection of the target image does not reach camera 84.
Fig. 9 shows viewing screen 86 of camera 84, upon which are marked peripheral points 88a and central point 88b. Peripheral points 88a may be used to measure the size of pupil 30, as shown. When the perimeter of pupil 30 touches each of the peripheral points 88a, pupil 30 has been driven to a predetermined size at which an image should be obtained. Central point 88b may be use-ful in aligning pupil 30 so that size may be properly tested. As shown in Fi~. 9, a grid may also be provided on viewing screen 86 to assist in the determination of the size of pupil 30 when it does not meet peripheral ~2~55~
points 88a. Alternatively, concentric markings could be provided on screen 86, as disclosed in U.S. Patent
3,966,310, issued to Larson.
It will be apparent ~rom Figs. 8 and 9 that the distance between camera 84 and eye 10 must be i'i~ed in order to provide a standardized image. This can be done by providing a headrest (not shown) -against which the forehead above eye 10 may be positioned. The headrest may also be equipped with appropriate additional surfaces for darkening a space around eye 10. This permits the system to provide a condition of nearly zero illumination to retina 12.
The system shown in Figs. 8 and 9 may be manu-ally operated to obtain the predetermined pupil size touching peripheral points 88a. First, variable resistor 72c is manually adjusted to its ma~imum resistance, ef-fectively turning off filament 72a. Oblique light sour-ces 78a and 78b are activated to provide a desired level of illumination upon iris 20. Then, variable resistor 72c is gradually turned to a lower resistance until the image of target 71 begins to illuminate retina 12. The operator, looking at viewing screen 86, will see the pu-pil beginning to decrease in size as the illumination from filament 72a increases. When the perimeter of pupil 30, as seen through viewing screen 86, just touches per-ipheral points 88a, the operator may activate camera 84 to obtain an image of iris 20 and pupil 30. If further, predetermined sizes are marked by points in viewing screen 86, the operator may continue to adjust variable resistor 72c until those additional predetermined sizes of pupil 30 are shown in viewin~ screen 8~ and are ima~ed when camera 84 is activated.
The optics of Fig. 8 cold be modified in many ways within the scope of the invention. For example, rather than providing a transparent beam-splittinv plane . ~ .~
5~
82 with a centered mirror 82a, the beam--splitting plane could be defined by two prisms positioned against each other or by a half-silvered mirror or o1ther partially reflecting surfaçe. Such arrangements, however, will result in a greater loss oi intensity o~ the collimated light and of the reflected image than the arrangement of Fig. 8. I~ such an arrangement is desirable, therefore, it may be appropriate to provide additional illumination of the iris 20, as discussed below in relation to Fig.
11 .
Fig. 10 shows a block diagram of electrical com ponents performing the functions set forth in Figo 3.
Iris recognition system 100 in Fig. 10 includes illumina-tion control circuit 170, camera 180 and processor 190.
Illumination control circuit 170 and camera 180 operate under the control of processor 190. In response to con-trol signals, illumination control circuit 170 illumi-nates eye 10 both ~o control the size of pupil 30 and to provide a reflected image of iris 20 and pupil 30. This reflected image is received by camera 180, which may be a video camera which converts the reflected image into an image si~nal for processor 190. Processor 190 then com-pares the image signal with a reference image in order to identify the human bein~.
Fig. 11 shows in greater detail the illumination system of iris recognition system 100. As shown in Fig.
11, illumination control circuit 170 controls light sour-ces 172, 174, 176 and 178. In general, these light sour-ces should be as small as possible to minimize noise in the reflected image from eye 10. Light source 172, which illuminates retina 12, as discussed in relation to Fig.
8, may do so through a fixation target 71 and a beam-splitting plane 82. Light source 172 will thus provide a target image upon which eye 10 will fix and focus, thereby becoming aligned for ima~in~. The target ima~e ~'Z4~LSS'~
should pre~erably be a distant object to discourage accommodation which would reduce pupil size. Rays from the target image may be relatively tightly collimated and subtend a small angle in order to provide a more criti-cal, centralized alignment, or may be less tightly col-limated if additional iris illumination is desired. The target image could also be compound, such as cross hairs in one plane which must be aligned on a design~ted part o-E a background at another plane to obtain alignment of the pupil, although this would depend on the acuity of the eye 10. Alternatively, a polarizing ring-sight de-vice could be used to create concentric circles on the retina independent o~ the refraction of eye 10. The tar-get image should not be extremely bright, because that would also reduce the size of pupil 30.
As mentioned above, the target image may be used to cause the subject to align pupil 30 for optimal imag-ing. In addition, however, it would be possible to pro-vide an automatic alignment system using a feedback cir-cuit and a servo-control mechanism to adjust the relative positions of the iris 20 and the camera 180. This ad-justment of position could be accomplished with rotatable mirrors, plano-parallel plates, or prismatic devices such as Risley prisms. Alternatively, the head rest (not shown) can be moved in order to orient the head.
In addition to pupil-driving light source 172, the illumination system shown in Fig. 11 includes oblique light sources 174, 176 and 178. As discussed above in relation to Fig. 6, each of these oblique light sources 174, 176 and 178 may provide a relatively monochromatic beam of light, so that the shadows resulting frorn the oblique positioning of these light sources will lack the corresponding color, facilitating the subsequent proces-~4~55~
sing of the shadow images. Each source 174, 176 and 173 could be a relatively monochromatic source such as an LED
or a solid state laser.
Although Fig. 11 shows an arra~gement in which three light sources are provided at appro~imately 120 angles from each other, the number of light sources and their relative angular positions could be varied in many ways. In addition, other arran~ements could be used to obtain additional information about elevation-dependent features of iris 20. For example, more than one camera 180 could be provided, or a single camera 180 could be provided which could be moved between a plurality of po-sitions in order to obtain three-dimensional information.
Any other means for obtaining an ima~e could be used, in-cluding laser-based techniques such as holography and in-cluding flyi~g spot scanning by mechanical or electro-optical techniques.
Fig. 12 shows a flow chart o~ the operation of processor 190 in ~ig. 10. Processor 190 will at first be initialized in step 342, in order to provide appropriate memory contents and in order to perform the necessary loadin~ of programs. In addition, as a part of initial-ization, processor 190 may provide control signals caus-ing camera 180 to begin to operate and providing power to illumination control circuit 170. In step 344, processor 190 provides control signals to illumination control cir-cuit 170 causing pupil-drivin~ light source 172 to illu-minate the retina 12 at the extreme illumination of the range available. In addition, oblique light sources 174, 176 and 178 may be activated to provide sufficient illu-mination so that camera 180 is able to provide an image of pupil 30 to processor 190.
When the extreme size of the pupil has been ob-tained in step 344, the pupil is then driven across the ran~e of its size in step 346. This is performecl by i5'~
gradually increasing the illumination from pupil-driving light source 17~. After each increase in intensity, the pupil size is measured in step 348 and tested in step 352 to determine whether it equals the desired predetermined size xn. For the first iteration r the first predeter-mined size x1 will be obtained after the necessary reit-teration of the loop including steps 346, 348 and 352.
Oblique light sources 374, 376 and 378 will also be acti-vated, either at the same time or in sequence according to a method like that sho~n in Fig. 6, in order to obtain an image of the iris 20 and pupil 3~ of eye 10. Oblique light sources 374, 376 and 378 should be activated in a way which minimizes the constriction of pupil 30. There-fore, it may be necessary to activate them before or while the illumination from light source 172 is in-creased.
From this image, processor 190 will then extract the nth descriptor set in step 362. For the image ob-tained when the pupil size is x1, the first descriptor set will be obtained. Then, in step 364, n is tested to determined if it is equal to N, the total number of de-scriptor sets to be obtained. If not, n is incremented in step 366, and the loop including steps 346, 348, 352, 354, 362 and 36~ is repeated for the ne~t value of n.
This loop will be repeated until step 364 determines that n equals N.
When n equals N, the resulting descriptor sets are compared with one or more reference descriptor sets in step 368. If the objective is to determine whether the eye 10 is that of one of a population of people, each descriptor set resultin~ from step 362 will be compared with the correspondin~ descriptor sets of the me~bers of that population. On the other hand, if the objective is sirnply to confirm that an eye 10 is that of a specific individual, it is only necessary to compare the descrip-5S~
tor sets resulting ~rom step 362 with the corresponding descriptor set of that individual, which may, for exam-ple, be stored on a credit card, identification card, or o~her identi~ying document, or in the memory of a compu-ter system to which the processor 190 has access.
Upon the completion of step 3&8, an appropriate indication may be provided to the operator of iris recog~
nition system 100 that the eye 10 has been identifiecl or has not been identified, completing the operation shown in Fig. 12. At this point, the operator may choose to initiate further iris recognition by repeating the pro-cess sho~n in Fig. 12.
Processor 190, in extracting descriptors in step 362, may perform a wide variety of algorithms. A number o-f such algorithms are set forth in Ballard, D.H. and Brown, C.M., Com~uter Vision, Prentice-Hall, Englewood Cliffs, New Jersey, 1982; Duda; RØ and Hart, P.E., Pattern Recognition and Scene Analysis, Wiley, New York, 1973; and Rosenfeld, A., Digital Picture Processing, Aca-demic Press, New York, 1982. Many of these algorithms use classical mathematical techrliques such as difference operators, taking a differential equation of selected points and testing a rate of change, and mask matchin~ or surface fitting, fitting to a know~ shape at various ori-entations~ Although these and other such algorithms are known, they have not previously been applied to the spe-cific features of the iris and pupil.
The location, size and deviation from circular-ity of the pupil could provide a number of descriptors.
The pupil could -first be detected using a boundary detec-tion al~orith~ or an edge detection algorithm, which may, for example, detect an abrupt gray level change. Systems yerformin~ pupil size measuring ~lgorithms are disclosed, for example, in U.S. Patent 3,598,1~7, issued to Ishikawa et al, and U.S. Patent 3,533,683, issued to Star~ et al.
55;~
A Hough transform algorithm for detecting circles, which mapæ curves into the transform space according to charac-teristics such as curvature, could be used on a boundary or edge which is detected. Alternatively, the received image could be subjected to a thresholding algorithm af-ter which a region-growing or aggregation algorithm, such as the blob coloring algorithm discussed by Ballard and Brown, could be performed to find the largest connected region of pixels with intensity values below threshold.
Since the pupil is central, the region could be grown outward from a central dark pi~el in the image, progres-sively aggregating the adjacent dark pixels until the pupil boundary is reached, beyond which adjacent pixels will not be dark. This will provide a measure both of pupil size and of location, as the center of the pupil may be determined from its boundary~ This may be done, for example, by taking slices through the dark region in a number of radial directions to select a largest diame-ter. To eliminate the chance that the largest diameter is not the true diameter, as, for e~ample, if it extended to a notch or other irregularity in the boundary, the least sum of squares from its center to the boundary could be taken and compared with that of the second lar-gest diameter. The center and radius giving the least sum of squares would define the circle most closely ap-proximating the pupil boundary, and the residual sum of squares would be a quantitative measure of deviation of the pupil from circularity, a "figure of merit" for the pupil. This figure of merit would be a useful descriptor of the pupil.
The characteristics of the radial furrows could also be used to obtain descriptors. From the center of the pupil, the precision of radiality of the radial fur-rows can be obtained. Line detection algorithms and streak or thick line detection algorithms could be used ...
55~
to locate the radial furrows. The radial furrows may then be compared wit~l true stralght lines radiatiog from the center by obtaining a sum of squares of differences, indicating radialityO Some furrows, rather than being entirely straight, include, for example, two offset parts, each of which is generally radial, the two being connected by a non-radial jog. A measure of thickness could be obtained for such a furrow by determining how wide a streak is necessary to cover the furrow. Also, the Hough transform algorithm for detecting straight lines could be applied to obtain such descriptors as the length, angular location and direction of radial furrows.
The concentric furrows may be detected and de-scribed using edge detection algorithms and the Hough transform algorithms for detecting curved lines. Simi-larly, the collarette and the contours of the crypts along it could be described using edge detection algor-ithms and chain coding algorithms. In addition, points of high curvature along these contours could be located and represented for matching purposes, which may be more efficient. Isolated crypts could be described by a thresholding algorithm, followed by a region-growing al-gorithm and then algorithms obtaining the center of grav-ity and the lower order central moments, as described by Duda and Hart.
Spot detection algorithms, detecting hi~h con-trast spikes in areas of relatively constant gray level, could detect small pigment spots. More generally, algor-ithms li~e those for measuring the pupil could be used to obtain location and lower-order central moments of pig-ement spots, and other such features. Also, high-curva-ture points alon~ the contour of a feature could be de-tected and described.
il~;2~455~
Finally, the color of the iris generally could be found by an algorithm obtaining a histogram in a three-dimensional RGB (Red-Green-Blue) color space, as described by Ballard and Brown at page 153. The peak in the histo~ram will provide a descriptor o~ color.
As noted above, the embodiment o~ Figs. 8-12 could be modified in many ways within the scope of the invention, Processor 190 may control the operation of illumination control circuit 170 and camera 180 in any appropriate manner to obtain images for processing. For example, the image from camera 180 may be constantly re-ceived and processed by processor 190 to determine the appropriate level of illumination by illumination control circuit 170. At the same time, when it is desired to obtain an image at one of the predetermined sizes of pu-pil 30, processor 190 may control illumination means 170 to maintain a temporarily constant level o~ illumination while camera 180 obtains a complete image for processing.
Although the present invention has been de-scribed in connection with a plurality of preferred em-bodiments thereof, many other variations and modifica-tions will now become apparent to those skilled in the art. It is preferred, therefore, that the present inven-tion be limited not by the specific disclosure herein, but only by the appended clalms.
.~
It will be apparent ~rom Figs. 8 and 9 that the distance between camera 84 and eye 10 must be i'i~ed in order to provide a standardized image. This can be done by providing a headrest (not shown) -against which the forehead above eye 10 may be positioned. The headrest may also be equipped with appropriate additional surfaces for darkening a space around eye 10. This permits the system to provide a condition of nearly zero illumination to retina 12.
The system shown in Figs. 8 and 9 may be manu-ally operated to obtain the predetermined pupil size touching peripheral points 88a. First, variable resistor 72c is manually adjusted to its ma~imum resistance, ef-fectively turning off filament 72a. Oblique light sour-ces 78a and 78b are activated to provide a desired level of illumination upon iris 20. Then, variable resistor 72c is gradually turned to a lower resistance until the image of target 71 begins to illuminate retina 12. The operator, looking at viewing screen 86, will see the pu-pil beginning to decrease in size as the illumination from filament 72a increases. When the perimeter of pupil 30, as seen through viewing screen 86, just touches per-ipheral points 88a, the operator may activate camera 84 to obtain an image of iris 20 and pupil 30. If further, predetermined sizes are marked by points in viewing screen 86, the operator may continue to adjust variable resistor 72c until those additional predetermined sizes of pupil 30 are shown in viewin~ screen 8~ and are ima~ed when camera 84 is activated.
The optics of Fig. 8 cold be modified in many ways within the scope of the invention. For example, rather than providing a transparent beam-splittinv plane . ~ .~
5~
82 with a centered mirror 82a, the beam--splitting plane could be defined by two prisms positioned against each other or by a half-silvered mirror or o1ther partially reflecting surfaçe. Such arrangements, however, will result in a greater loss oi intensity o~ the collimated light and of the reflected image than the arrangement of Fig. 8. I~ such an arrangement is desirable, therefore, it may be appropriate to provide additional illumination of the iris 20, as discussed below in relation to Fig.
11 .
Fig. 10 shows a block diagram of electrical com ponents performing the functions set forth in Figo 3.
Iris recognition system 100 in Fig. 10 includes illumina-tion control circuit 170, camera 180 and processor 190.
Illumination control circuit 170 and camera 180 operate under the control of processor 190. In response to con-trol signals, illumination control circuit 170 illumi-nates eye 10 both ~o control the size of pupil 30 and to provide a reflected image of iris 20 and pupil 30. This reflected image is received by camera 180, which may be a video camera which converts the reflected image into an image si~nal for processor 190. Processor 190 then com-pares the image signal with a reference image in order to identify the human bein~.
Fig. 11 shows in greater detail the illumination system of iris recognition system 100. As shown in Fig.
11, illumination control circuit 170 controls light sour-ces 172, 174, 176 and 178. In general, these light sour-ces should be as small as possible to minimize noise in the reflected image from eye 10. Light source 172, which illuminates retina 12, as discussed in relation to Fig.
8, may do so through a fixation target 71 and a beam-splitting plane 82. Light source 172 will thus provide a target image upon which eye 10 will fix and focus, thereby becoming aligned for ima~in~. The target ima~e ~'Z4~LSS'~
should pre~erably be a distant object to discourage accommodation which would reduce pupil size. Rays from the target image may be relatively tightly collimated and subtend a small angle in order to provide a more criti-cal, centralized alignment, or may be less tightly col-limated if additional iris illumination is desired. The target image could also be compound, such as cross hairs in one plane which must be aligned on a design~ted part o-E a background at another plane to obtain alignment of the pupil, although this would depend on the acuity of the eye 10. Alternatively, a polarizing ring-sight de-vice could be used to create concentric circles on the retina independent o~ the refraction of eye 10. The tar-get image should not be extremely bright, because that would also reduce the size of pupil 30.
As mentioned above, the target image may be used to cause the subject to align pupil 30 for optimal imag-ing. In addition, however, it would be possible to pro-vide an automatic alignment system using a feedback cir-cuit and a servo-control mechanism to adjust the relative positions of the iris 20 and the camera 180. This ad-justment of position could be accomplished with rotatable mirrors, plano-parallel plates, or prismatic devices such as Risley prisms. Alternatively, the head rest (not shown) can be moved in order to orient the head.
In addition to pupil-driving light source 172, the illumination system shown in Fig. 11 includes oblique light sources 174, 176 and 178. As discussed above in relation to Fig. 6, each of these oblique light sources 174, 176 and 178 may provide a relatively monochromatic beam of light, so that the shadows resulting frorn the oblique positioning of these light sources will lack the corresponding color, facilitating the subsequent proces-~4~55~
sing of the shadow images. Each source 174, 176 and 173 could be a relatively monochromatic source such as an LED
or a solid state laser.
Although Fig. 11 shows an arra~gement in which three light sources are provided at appro~imately 120 angles from each other, the number of light sources and their relative angular positions could be varied in many ways. In addition, other arran~ements could be used to obtain additional information about elevation-dependent features of iris 20. For example, more than one camera 180 could be provided, or a single camera 180 could be provided which could be moved between a plurality of po-sitions in order to obtain three-dimensional information.
Any other means for obtaining an ima~e could be used, in-cluding laser-based techniques such as holography and in-cluding flyi~g spot scanning by mechanical or electro-optical techniques.
Fig. 12 shows a flow chart o~ the operation of processor 190 in ~ig. 10. Processor 190 will at first be initialized in step 342, in order to provide appropriate memory contents and in order to perform the necessary loadin~ of programs. In addition, as a part of initial-ization, processor 190 may provide control signals caus-ing camera 180 to begin to operate and providing power to illumination control circuit 170. In step 344, processor 190 provides control signals to illumination control cir-cuit 170 causing pupil-drivin~ light source 172 to illu-minate the retina 12 at the extreme illumination of the range available. In addition, oblique light sources 174, 176 and 178 may be activated to provide sufficient illu-mination so that camera 180 is able to provide an image of pupil 30 to processor 190.
When the extreme size of the pupil has been ob-tained in step 344, the pupil is then driven across the ran~e of its size in step 346. This is performecl by i5'~
gradually increasing the illumination from pupil-driving light source 17~. After each increase in intensity, the pupil size is measured in step 348 and tested in step 352 to determine whether it equals the desired predetermined size xn. For the first iteration r the first predeter-mined size x1 will be obtained after the necessary reit-teration of the loop including steps 346, 348 and 352.
Oblique light sources 374, 376 and 378 will also be acti-vated, either at the same time or in sequence according to a method like that sho~n in Fig. 6, in order to obtain an image of the iris 20 and pupil 3~ of eye 10. Oblique light sources 374, 376 and 378 should be activated in a way which minimizes the constriction of pupil 30. There-fore, it may be necessary to activate them before or while the illumination from light source 172 is in-creased.
From this image, processor 190 will then extract the nth descriptor set in step 362. For the image ob-tained when the pupil size is x1, the first descriptor set will be obtained. Then, in step 364, n is tested to determined if it is equal to N, the total number of de-scriptor sets to be obtained. If not, n is incremented in step 366, and the loop including steps 346, 348, 352, 354, 362 and 36~ is repeated for the ne~t value of n.
This loop will be repeated until step 364 determines that n equals N.
When n equals N, the resulting descriptor sets are compared with one or more reference descriptor sets in step 368. If the objective is to determine whether the eye 10 is that of one of a population of people, each descriptor set resultin~ from step 362 will be compared with the correspondin~ descriptor sets of the me~bers of that population. On the other hand, if the objective is sirnply to confirm that an eye 10 is that of a specific individual, it is only necessary to compare the descrip-5S~
tor sets resulting ~rom step 362 with the corresponding descriptor set of that individual, which may, for exam-ple, be stored on a credit card, identification card, or o~her identi~ying document, or in the memory of a compu-ter system to which the processor 190 has access.
Upon the completion of step 3&8, an appropriate indication may be provided to the operator of iris recog~
nition system 100 that the eye 10 has been identifiecl or has not been identified, completing the operation shown in Fig. 12. At this point, the operator may choose to initiate further iris recognition by repeating the pro-cess sho~n in Fig. 12.
Processor 190, in extracting descriptors in step 362, may perform a wide variety of algorithms. A number o-f such algorithms are set forth in Ballard, D.H. and Brown, C.M., Com~uter Vision, Prentice-Hall, Englewood Cliffs, New Jersey, 1982; Duda; RØ and Hart, P.E., Pattern Recognition and Scene Analysis, Wiley, New York, 1973; and Rosenfeld, A., Digital Picture Processing, Aca-demic Press, New York, 1982. Many of these algorithms use classical mathematical techrliques such as difference operators, taking a differential equation of selected points and testing a rate of change, and mask matchin~ or surface fitting, fitting to a know~ shape at various ori-entations~ Although these and other such algorithms are known, they have not previously been applied to the spe-cific features of the iris and pupil.
The location, size and deviation from circular-ity of the pupil could provide a number of descriptors.
The pupil could -first be detected using a boundary detec-tion al~orith~ or an edge detection algorithm, which may, for example, detect an abrupt gray level change. Systems yerformin~ pupil size measuring ~lgorithms are disclosed, for example, in U.S. Patent 3,598,1~7, issued to Ishikawa et al, and U.S. Patent 3,533,683, issued to Star~ et al.
55;~
A Hough transform algorithm for detecting circles, which mapæ curves into the transform space according to charac-teristics such as curvature, could be used on a boundary or edge which is detected. Alternatively, the received image could be subjected to a thresholding algorithm af-ter which a region-growing or aggregation algorithm, such as the blob coloring algorithm discussed by Ballard and Brown, could be performed to find the largest connected region of pixels with intensity values below threshold.
Since the pupil is central, the region could be grown outward from a central dark pi~el in the image, progres-sively aggregating the adjacent dark pixels until the pupil boundary is reached, beyond which adjacent pixels will not be dark. This will provide a measure both of pupil size and of location, as the center of the pupil may be determined from its boundary~ This may be done, for example, by taking slices through the dark region in a number of radial directions to select a largest diame-ter. To eliminate the chance that the largest diameter is not the true diameter, as, for e~ample, if it extended to a notch or other irregularity in the boundary, the least sum of squares from its center to the boundary could be taken and compared with that of the second lar-gest diameter. The center and radius giving the least sum of squares would define the circle most closely ap-proximating the pupil boundary, and the residual sum of squares would be a quantitative measure of deviation of the pupil from circularity, a "figure of merit" for the pupil. This figure of merit would be a useful descriptor of the pupil.
The characteristics of the radial furrows could also be used to obtain descriptors. From the center of the pupil, the precision of radiality of the radial fur-rows can be obtained. Line detection algorithms and streak or thick line detection algorithms could be used ...
55~
to locate the radial furrows. The radial furrows may then be compared wit~l true stralght lines radiatiog from the center by obtaining a sum of squares of differences, indicating radialityO Some furrows, rather than being entirely straight, include, for example, two offset parts, each of which is generally radial, the two being connected by a non-radial jog. A measure of thickness could be obtained for such a furrow by determining how wide a streak is necessary to cover the furrow. Also, the Hough transform algorithm for detecting straight lines could be applied to obtain such descriptors as the length, angular location and direction of radial furrows.
The concentric furrows may be detected and de-scribed using edge detection algorithms and the Hough transform algorithms for detecting curved lines. Simi-larly, the collarette and the contours of the crypts along it could be described using edge detection algor-ithms and chain coding algorithms. In addition, points of high curvature along these contours could be located and represented for matching purposes, which may be more efficient. Isolated crypts could be described by a thresholding algorithm, followed by a region-growing al-gorithm and then algorithms obtaining the center of grav-ity and the lower order central moments, as described by Duda and Hart.
Spot detection algorithms, detecting hi~h con-trast spikes in areas of relatively constant gray level, could detect small pigment spots. More generally, algor-ithms li~e those for measuring the pupil could be used to obtain location and lower-order central moments of pig-ement spots, and other such features. Also, high-curva-ture points alon~ the contour of a feature could be de-tected and described.
il~;2~455~
Finally, the color of the iris generally could be found by an algorithm obtaining a histogram in a three-dimensional RGB (Red-Green-Blue) color space, as described by Ballard and Brown at page 153. The peak in the histo~ram will provide a descriptor o~ color.
As noted above, the embodiment o~ Figs. 8-12 could be modified in many ways within the scope of the invention, Processor 190 may control the operation of illumination control circuit 170 and camera 180 in any appropriate manner to obtain images for processing. For example, the image from camera 180 may be constantly re-ceived and processed by processor 190 to determine the appropriate level of illumination by illumination control circuit 170. At the same time, when it is desired to obtain an image at one of the predetermined sizes of pu-pil 30, processor 190 may control illumination means 170 to maintain a temporarily constant level o~ illumination while camera 180 obtains a complete image for processing.
Although the present invention has been de-scribed in connection with a plurality of preferred em-bodiments thereof, many other variations and modifica-tions will now become apparent to those skilled in the art. It is preferred, therefore, that the present inven-tion be limited not by the specific disclosure herein, but only by the appended clalms.
.~
Claims (32)
PROPERTY OR PRIVILEGE IS CLAIMED ARE DEFINED AS FOLLOWS:
1. A method of identification of a person, comprising storing image information of at least a portion of the iris and pupil of the person's eye; illuminating an eye, of an unidenti-fied person having an iris and a pupil; obtaining at least one image of at least the same portion of the iris and pupil of the eye of the unidentified person; and comparing at least the iris portion of the obtained image with the stored image information to identify the unidentified person.
2. The method of claim 1 in which illuminating com-prises driving the pupil of the eye to at least one predetermined size, comparing at least the iris portion of the obtained image with stored image information obtained from an eye with its pupil having the same predetermined size.
3. The method of claim 2 in which driving the pupil comprises directing light onto the retina of the eye and adjust-ing the intensity of the light until the pupil reaches the prede-termined size.
4. The method of claim 2 in which driving the pupil comprises driving the pupil to a plurality of the predetermined sizes; the step of obtaining at least one image comprising obtaining a respective image at each of the predetermined sizes of the pupil.
5. The method of claim 4 in which illuminating com-prises changing the intensity of illumination at a predetermined rate between at least two of the predetermined pupil sizes and measuring a rate of change of pupil size as the illumination intensity changes.
6. The method of claim 1 in which illuminating com-prises illuminating the iris obliquely for providing shadows of elevation-dependent features.
7. The method of claim 6 in which illuminating the iris obliquely comprises illuminating the iris from a plurality of spatially separated light sources for providing shadows of elevation-dependent feature extending in a plurality of direc-tions.
8. The method of claim 7 in which illuminating the iris from a plurality of spatially separated light sources com-prises illuminating the iris from a plurality of relatively monochromatic light sources, each light source providing shadows lacking a corresponding color of light.
9. The method of claim 1 in which comparing the obtained image with the stored image information comprises com-paring at least the iris portion of the obtained image with a stored reference image for identifying the person.
10. The method of claim 1 in which comparing the obtained image with stored image information comprises deriving a set of descriptors of at least the iris portion of the obtained image and comparing the derived descriptors with stored reference descriptors derived from a previous image for identifying the person.
11. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of the pupil.
12. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of a pigment frill around the pupil.
13. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of a collarette in the iris.
14. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one con-centric furrow in the iris.
15. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one radial furrow in the iris.
16. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one crypt in the iris.
17. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one pig-ment spot in the iris.
18. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one atrophic area in the iris.
19. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one tumor in the iris.
20. The method of claim 10 in which deriving a set of descriptors comprises deriving descriptors of at least one con-genital filament in the iris.
21. A system of identification, comprising: means for storing an image of at least a portion of the iris and pupil of a person's eye; means for illuminating an eye having an iris and pupil; means for obtaining an image of at least the same portion of the iris and pupil of the eye; and means for comparing at least the iris portion of the obtained image with the stored image information to identify the person.
22. The system of claim 21 in which the means for illu-minating comprises an adjustable light source for illuminating the eye at a selected value in a range of light intensity.
23. The system of claim 22 in which the means for illu-minating further comprises control means responsive to at least the iris portion of the obtained image of the eye for detecting the size of the pupil and operable for adjusting the adjustable light source to a selected light intensity for controlling the size of the pupil.
24. The system of claim 23 in which the control means is further operable for controlling the size of the pupil to obtain a plurality of predetermined pupil sizes; the control means being further operable for adjusting the light source at a predetermined rate across the range of light intensity between at least two of the predetermined pupil sizes for measuring the rate of change of pupil sizes as the light intensity is adjusted.
25. The system of claim 21 in which the illuminating means further comprises a plurality of spatially separated oblique light sources for illuminating the iris obliquely.
26. The system of claim 25 in which each of the oblique light sources comprises a relatively monochromatic light source for providing shadows lacking a corresponding color of light.
27. The system of claim 26 in which each relatively monochromatic light source comprises a laser.
28. The system of claim 21 in which the image obtaining means comprises a camera, the comparing means comprising a pro-cessor responsive to an output from the camera for comparing the obtained image with stored image information.
29. The system of claim 28 in which the processor com-prises storage means for storing reference descriptors derived from a previous image, the processor being operable for deriving a set of descriptors of the obtained image and for comparing the derived descriptors with the reference descriptors for identify-ing the person.
30. The system of claim 21 in which the image obtaining means comprises a laser.
31. The system of claim 21 further including target means for providing an object upon which the eye fixates.
32. The system of claim 31 further including focusing means for focusing the target means.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US703,312 | 1985-02-20 | ||
US06/703,312 US4641349A (en) | 1985-02-20 | 1985-02-20 | Iris recognition system |
Publications (1)
Publication Number | Publication Date |
---|---|
CA1244552A true CA1244552A (en) | 1988-11-08 |
Family
ID=24824891
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA000502207A Expired CA1244552A (en) | 1985-02-20 | 1986-02-19 | Iris recognition system |
Country Status (10)
Country | Link |
---|---|
US (1) | US4641349A (en) |
EP (1) | EP0215818B1 (en) |
JP (1) | JPS62501889A (en) |
AT (1) | ATE65851T1 (en) |
BR (1) | BR8605561A (en) |
CA (1) | CA1244552A (en) |
DE (1) | DE3680618D1 (en) |
IL (1) | IL77920A (en) |
MX (1) | MX163339A (en) |
WO (1) | WO1986005018A1 (en) |
Families Citing this family (368)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4854329A (en) * | 1987-07-21 | 1989-08-08 | Walruff James C | Apparatus and method for noninvasive testing of voluntary and involuntary motor response patterns |
FR2622786A1 (en) * | 1987-11-10 | 1989-05-12 | Giet Francis | APPARATUS FOR EYE EXAMINATION BY IRISCOPY AND IRIGRAPHY |
US4902123A (en) * | 1987-11-25 | 1990-02-20 | Taunton Technologies, Inc. | Topography measuring apparatus |
US5106183A (en) * | 1987-11-25 | 1992-04-21 | Taunton Technologies, Inc. | Topography measuring apparatus |
US4998819A (en) * | 1987-11-25 | 1991-03-12 | Taunton Technologies, Inc. | Topography measuring apparatus |
US5016282A (en) * | 1988-07-14 | 1991-05-14 | Atr Communication Systems Research Laboratories | Eye tracking image pickup apparatus for separating noise from feature portions |
US4931865A (en) * | 1988-08-24 | 1990-06-05 | Sebastiano Scarampi | Apparatus and methods for monitoring television viewers |
US5016173A (en) * | 1989-04-13 | 1991-05-14 | Vanguard Imaging Ltd. | Apparatus and method for monitoring visually accessible surfaces of the body |
DE4102895C1 (en) * | 1991-01-31 | 1992-01-30 | Siemens Ag, 8000 Muenchen, De | |
US5231580A (en) * | 1991-04-01 | 1993-07-27 | The United States Of America As Represented By The Secretary Of The Department Of Health And Human Services | Automated method and apparatus for determining characteristics of nerve fibers |
US5291560A (en) * | 1991-07-15 | 1994-03-01 | Iri Scan Incorporated | Biometric personal identification system based on iris analysis |
US5360005A (en) * | 1992-01-10 | 1994-11-01 | Wilk Peter J | Medical diagnosis device for sensing cardiac activity and blood flow |
US5437278A (en) * | 1992-01-10 | 1995-08-01 | Wilk; Peter J. | Medical diagnosis system and method |
WO1993014454A1 (en) * | 1992-01-10 | 1993-07-22 | Foster-Miller, Inc. | A sensory integrated data interface |
US5359669A (en) * | 1992-04-13 | 1994-10-25 | Motorola, Inc. | Remote retinal scan identifier |
US5331544A (en) * | 1992-04-23 | 1994-07-19 | A. C. Nielsen Company | Market research method and system for collecting retail store and shopper market research data |
FR2690822B1 (en) * | 1992-05-06 | 1994-08-05 | Dior Christian Parfums | METHOD FOR MAKING UP THE FACE, ESPECIALLY EYES, AND DEVICE FOR IMPLEMENTING SAME. |
JPH0694979A (en) * | 1992-09-14 | 1994-04-08 | Nikon Corp | Camera provided with device for detecting |
KR100307792B1 (en) * | 1992-10-10 | 2001-11-30 | 존 이. 시드라즈 | Biometric measurement based on iris analysis |
RU2121162C1 (en) * | 1992-11-05 | 1998-10-27 | Комс 21 Лимитед | Device for control of safe access |
US5954583A (en) * | 1992-11-05 | 1999-09-21 | Com21 Limited | Secure access control system |
US5526436A (en) * | 1993-06-17 | 1996-06-11 | Asahi Kogaku Kogyo Kabushiki Kaisha | Image detecting apparatus for an individual identifying system |
US5471542A (en) * | 1993-09-27 | 1995-11-28 | Ragland; Richard R. | Point-of-gaze tracker |
DE69501327T3 (en) * | 1994-03-24 | 2005-12-22 | Minnesota Mining And Mfg. Co., Saint Paul | Biometric personal authentication system |
DE4414216C1 (en) * | 1994-04-23 | 1995-04-06 | Daimler Benz Ag | Device for protecting a motor vehicle against use by third parties, with personalisation of the driving authorisation |
US6902108B1 (en) * | 1994-08-25 | 2005-06-07 | Bryan P. Chapman | Method and apparatus for providing identification |
AU719428B2 (en) * | 1994-09-02 | 2000-05-11 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US5572596A (en) * | 1994-09-02 | 1996-11-05 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US6714665B1 (en) * | 1994-09-02 | 2004-03-30 | Sarnoff Corporation | Fully automated iris recognition system utilizing wide and narrow fields of view |
US5842194A (en) * | 1995-07-28 | 1998-11-24 | Mitsubishi Denki Kabushiki Kaisha | Method of recognizing images of faces or general images using fuzzy combination of multiple resolutions |
US5729619A (en) * | 1995-08-08 | 1998-03-17 | Northrop Grumman Corporation | Operator identity, intoxication and drowsiness monitoring system and method |
JPH09134430A (en) * | 1995-11-08 | 1997-05-20 | Oki Electric Ind Co Ltd | Iris data collating system |
JP3427873B2 (en) * | 1995-11-17 | 2003-07-22 | 沖電気工業株式会社 | Automatic transaction device and automatic transaction system |
JP3625941B2 (en) * | 1996-01-30 | 2005-03-02 | 沖電気工業株式会社 | Iris recognition system |
JPH09212644A (en) * | 1996-02-07 | 1997-08-15 | Oki Electric Ind Co Ltd | Iris recognition device and iris recognition method |
US6259827B1 (en) | 1996-03-21 | 2001-07-10 | Cognex Corporation | Machine vision methods for enhancing the contrast between an object and its background using multiple on-axis images |
JP3580634B2 (en) * | 1996-04-19 | 2004-10-27 | 沖電気工業株式会社 | Automatic transaction equipment |
US5717512A (en) * | 1996-05-15 | 1998-02-10 | Chmielewski, Jr.; Thomas A. | Compact image steering and focusing device |
JP3575915B2 (en) * | 1996-05-15 | 2004-10-13 | 沖電気工業株式会社 | Automatic transaction equipment |
US6320610B1 (en) | 1998-12-31 | 2001-11-20 | Sensar, Inc. | Compact imaging device incorporating rotatably mounted cameras |
DE69717826T2 (en) * | 1996-06-06 | 2003-09-04 | British Telecomm Public Ltd Co | IDENTIFICATION OF PERSONS |
EP0872814A1 (en) * | 1997-04-15 | 1998-10-21 | BRITISH TELECOMMUNICATIONS public limited company | Optical apparatus |
JP3337913B2 (en) * | 1996-06-19 | 2002-10-28 | 沖電気工業株式会社 | Iris imaging method and imaging device thereof |
JP3751368B2 (en) * | 1996-06-28 | 2006-03-01 | 沖電気工業株式会社 | Iris recognition system and iris recognition device |
US5876926A (en) | 1996-07-23 | 1999-03-02 | Beecham; James E. | Method, apparatus and system for verification of human medical data |
JP3436293B2 (en) * | 1996-07-25 | 2003-08-11 | 沖電気工業株式会社 | Animal individual identification device and individual identification system |
EP0959769A1 (en) | 1996-08-25 | 1999-12-01 | Sensar, Inc. | Apparatus for the iris acquiring images |
JP3587635B2 (en) * | 1996-11-15 | 2004-11-10 | 沖電気工業株式会社 | Personal recognition device using iris and automatic transaction system using this personal recognition device |
US6075881A (en) * | 1997-03-18 | 2000-06-13 | Cognex Corporation | Machine vision methods for identifying collinear sets of points from an image |
US5974169A (en) * | 1997-03-20 | 1999-10-26 | Cognex Corporation | Machine vision methods for determining characteristics of an object using boundary points and bounding regions |
US6229905B1 (en) | 1997-03-26 | 2001-05-08 | Oki Electric Industry Co., Ltd. | Animal identification based on irial granule analysis |
US6215891B1 (en) * | 1997-03-26 | 2001-04-10 | Oki Electric Industry Co., Ltd. | Eye image recognition method eye image selection method and system therefor |
US6285780B1 (en) * | 1997-03-28 | 2001-09-04 | Oki Electric Industry Co., Ltd. | Apparatus for identifying individual animals and image processing method |
US6144754A (en) * | 1997-03-28 | 2000-11-07 | Oki Electric Industry Co., Ltd. | Method and apparatus for identifying individuals |
US6202151B1 (en) | 1997-05-09 | 2001-03-13 | Gte Service Corporation | System and method for authenticating electronic transactions using biometric certificates |
CN1139894C (en) | 1997-05-09 | 2004-02-25 | Gte服务公司 | Biometric certificates |
US5867587A (en) * | 1997-05-19 | 1999-02-02 | Northrop Grumman Corporation | Impaired operator detection and warning system employing eyeblink analysis |
US6373968B2 (en) * | 1997-06-06 | 2002-04-16 | Oki Electric Industry Co., Ltd. | System for identifying individuals |
US6608647B1 (en) | 1997-06-24 | 2003-08-19 | Cognex Corporation | Methods and apparatus for charge coupled device image acquisition with independent integration and readout |
US6119096A (en) * | 1997-07-31 | 2000-09-12 | Eyeticket Corporation | System and method for aircraft passenger check-in and boarding using iris recognition |
JP2985839B2 (en) * | 1997-08-04 | 1999-12-06 | 日本電気株式会社 | Biometric matching method and apparatus, information storage medium |
JPH1169219A (en) * | 1997-08-21 | 1999-03-09 | Victor Co Of Japan Ltd | Video camera for iris photographing |
US6185318B1 (en) * | 1997-08-22 | 2001-02-06 | International Business Machines Corporation | System and method for matching (fingerprint) images an aligned string-based representation |
US5892837A (en) * | 1997-08-29 | 1999-04-06 | Eastman Kodak Company | Computer program product for locating objects in an image |
GB9721506D0 (en) * | 1997-10-10 | 1997-12-10 | Virulite Limited | Treatment of diseases |
US6031934A (en) * | 1997-10-15 | 2000-02-29 | Electric Planet, Inc. | Computer vision system for subject characterization |
US6532022B1 (en) * | 1997-10-15 | 2003-03-11 | Electric Planet, Inc. | Method and apparatus for model-based compositing |
US6072494A (en) | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
US6007202A (en) | 1997-10-23 | 1999-12-28 | Lasersight Technologies, Inc. | Eye illumination system and method |
EP0910986A1 (en) | 1997-10-24 | 1999-04-28 | BRITISH TELECOMMUNICATIONS public limited company | Imaging apparatus |
US6064752A (en) * | 1997-11-04 | 2000-05-16 | Sensar, Inc. | Method and apparatus for positioning subjects before a single camera |
US6069967A (en) * | 1997-11-04 | 2000-05-30 | Sensar, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses |
US6021210A (en) | 1997-12-01 | 2000-02-01 | Sensar, Inc. | Image subtraction to remove ambient illumination |
US6055322A (en) * | 1997-12-01 | 2000-04-25 | Sensor, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
US6028949A (en) * | 1997-12-02 | 2000-02-22 | Mckendall; Raymond A. | Method of verifying the presence of an eye in a close-up image |
US5953440A (en) | 1997-12-02 | 1999-09-14 | Sensar, Inc. | Method of measuring the focus of close-up images of eyes |
US6381375B1 (en) | 1998-02-20 | 2002-04-30 | Cognex Corporation | Methods and apparatus for generating a projection of an image |
US6850631B1 (en) | 1998-02-20 | 2005-02-01 | Oki Electric Industry Co., Ltd. | Photographing device, iris input device and iris image input method |
US5978494A (en) * | 1998-03-04 | 1999-11-02 | Sensar, Inc. | Method of selecting the best enroll image for personal identification |
JP3271750B2 (en) | 1998-03-05 | 2002-04-08 | 沖電気工業株式会社 | Iris identification code extraction method and device, iris recognition method and device, data encryption device |
JP3868621B2 (en) * | 1998-03-17 | 2007-01-17 | 株式会社東芝 | Image acquisition apparatus, image acquisition method, and recording medium |
US6199985B1 (en) | 1998-05-15 | 2001-03-13 | Christopher Scott Anderson | Pupilometer methods and apparatus |
US5956122A (en) * | 1998-06-26 | 1999-09-21 | Litton Systems, Inc | Iris recognition apparatus and method |
US20040208343A1 (en) * | 1998-07-09 | 2004-10-21 | Colorado State University Research Foundation | Apparatus and method for creating a record using biometric information |
JP3315648B2 (en) | 1998-07-17 | 2002-08-19 | 沖電気工業株式会社 | Iris code generation device and iris recognition system |
JP3610234B2 (en) | 1998-07-17 | 2005-01-12 | 株式会社メディア・テクノロジー | Iris information acquisition device and iris identification device |
JP3813023B2 (en) * | 1998-08-17 | 2006-08-23 | 沖電気工業株式会社 | Iris recognition device |
AU6243799A (en) | 1998-09-04 | 2000-03-27 | James E. Beecham | Verification of human medical data |
US6522772B1 (en) * | 1998-09-30 | 2003-02-18 | Ncr Corporation | Self-service checkout terminal having a biometric sensing device for verifying identity of a user and associated method |
US6289113B1 (en) | 1998-11-25 | 2001-09-11 | Iridian Technologies, Inc. | Handheld iris imaging apparatus and method |
US6532298B1 (en) | 1998-11-25 | 2003-03-11 | Iridian Technologies, Inc. | Portable authentication device and method using iris patterns |
US6377699B1 (en) | 1998-11-25 | 2002-04-23 | Iridian Technologies, Inc. | Iris imaging telephone security module and method |
US6424727B1 (en) * | 1998-11-25 | 2002-07-23 | Iridian Technologies, Inc. | System and method of animal identification and animal transaction authorization using iris patterns |
US6753919B1 (en) | 1998-11-25 | 2004-06-22 | Iridian Technologies, Inc. | Fast focus assessment system and method for imaging |
US6687402B1 (en) | 1998-12-18 | 2004-02-03 | Cognex Corporation | Machine vision methods and systems for boundary feature comparison of patterns and images |
US6381366B1 (en) | 1998-12-18 | 2002-04-30 | Cognex Corporation | Machine vision methods and system for boundary point-based comparison of patterns and images |
NL1011509C2 (en) * | 1999-01-05 | 2000-07-06 | Nedap Nv | Method for biometric identification of animals. |
US6944318B1 (en) | 1999-01-15 | 2005-09-13 | Citicorp Development Center, Inc. | Fast matching systems and methods for personal identification |
US6332193B1 (en) | 1999-01-18 | 2001-12-18 | Sensar, Inc. | Method and apparatus for securely transmitting and authenticating biometric data over a network |
US20020056043A1 (en) * | 1999-01-18 | 2002-05-09 | Sensar, Inc. | Method and apparatus for securely transmitting and authenticating biometric data over a network |
US6183086B1 (en) * | 1999-03-12 | 2001-02-06 | Bausch & Lomb Surgical, Inc. | Variable multiple color LED illumination system |
KR100320188B1 (en) * | 1999-03-23 | 2002-01-10 | 구자홍 | Forgery judgment method for iris recognition system |
GB9907515D0 (en) | 1999-04-01 | 1999-05-26 | Ncr Int Inc | Self service terminal |
JP4619384B2 (en) * | 1999-04-09 | 2011-01-26 | アイリテック株式会社 | Iris identification system |
US6247813B1 (en) * | 1999-04-09 | 2001-06-19 | Iritech, Inc. | Iris identification system and method of identifying a person through iris recognition |
KR100356600B1 (en) * | 1999-04-09 | 2002-10-19 | 아이리텍 잉크 | A Method For Identifying The Iris Of Persons Based On The Shape Of Lacuna And/Or Autonomous Nervous Wreath |
JP4006192B2 (en) * | 1999-04-09 | 2007-11-14 | アイリテック株式会社 | Iris recognition system and method |
US6700998B1 (en) * | 1999-04-23 | 2004-03-02 | Oki Electric Industry Co, Ltd. | Iris registration unit |
US6820979B1 (en) * | 1999-04-23 | 2004-11-23 | Neuroptics, Inc. | Pupilometer with pupil irregularity detection, pupil tracking, and pupil response detection capability, glaucoma screening capability, intracranial pressure detection capability, and ocular aberration measurement capability |
JP4290278B2 (en) | 1999-06-03 | 2009-07-01 | 株式会社ニデック | Ophthalmic equipment |
JP2001034754A (en) * | 1999-07-19 | 2001-02-09 | Sony Corp | Iris authentication device |
US6553494B1 (en) | 1999-07-21 | 2003-04-22 | Sensar, Inc. | Method and apparatus for applying and verifying a biometric-based digital signature to an electronic document |
JP2003511183A (en) | 1999-10-21 | 2003-03-25 | テクノラス ゲーエムベーハー オフタルモロギッシェ システム | Personalized corneal profile |
CN100362975C (en) | 1999-10-21 | 2008-01-23 | 泰思诺拉斯眼科系统公司 | Iris recognition and tracking for optical treatment |
WO2001035349A1 (en) | 1999-11-09 | 2001-05-17 | Iridian Technologies, Inc. | System and method of biometric authentication of electronic signatures using iris patterns |
WO2001035348A1 (en) | 1999-11-09 | 2001-05-17 | Iridian Technologies, Inc. | System and method for authentication of shipping transactions using printable and readable biometric data |
US6684402B1 (en) | 1999-12-01 | 2004-01-27 | Cognex Technology And Investment Corporation | Control methods and apparatus for coupling multiple image acquisition devices to a digital data processor |
US6505193B1 (en) | 1999-12-01 | 2003-01-07 | Iridian Technologies, Inc. | System and method of fast biometric database searching using digital certificates |
AU2001291157B2 (en) * | 2000-02-03 | 2003-12-18 | Gregg S. Homer | Method for alteration of iris pigment |
AU2003241638B2 (en) * | 2000-02-03 | 2004-01-22 | Gregg S. Homer | Method for alteration of iris pigment |
WO2001059580A1 (en) * | 2000-02-09 | 2001-08-16 | Nobuyoshi Ochiai | Personal authentication system |
JP2001331799A (en) * | 2000-03-16 | 2001-11-30 | Toshiba Corp | Image processor and image processing method |
DE60119418T2 (en) * | 2000-03-22 | 2007-05-24 | Kabushiki Kaisha Toshiba, Kawasaki | Face-capturing recognition device and passport verification device |
JP3825222B2 (en) * | 2000-03-24 | 2006-09-27 | 松下電器産業株式会社 | Personal authentication device, personal authentication system, and electronic payment system |
US6748104B1 (en) | 2000-03-24 | 2004-06-08 | Cognex Corporation | Methods and apparatus for machine vision inspection using single and multiple templates or patterns |
US6299306B1 (en) | 2000-03-31 | 2001-10-09 | Sensar, Inc. | Method and apparatus for positioning subjects using a holographic optical element |
US7044602B2 (en) | 2002-05-30 | 2006-05-16 | Visx, Incorporated | Methods and systems for tracking a torsional orientation and position of an eye |
JP2003533801A (en) * | 2000-05-16 | 2003-11-11 | スイスコム・モバイル・アクチエンゲゼルシヤフト | Biometric identification and authentication methods |
JP4228520B2 (en) * | 2000-06-12 | 2009-02-25 | 沖電気工業株式会社 | Iris photography device |
JP2002101322A (en) * | 2000-07-10 | 2002-04-05 | Matsushita Electric Ind Co Ltd | Iris camera module |
KR100373850B1 (en) * | 2000-10-07 | 2003-02-26 | 주식회사 큐리텍 | Identification system and method using iris, and media that can record computer program sources thereof |
JP3401502B2 (en) * | 2000-07-13 | 2003-04-28 | 松下電器産業株式会社 | Eye imaging device |
KR100400975B1 (en) * | 2000-09-01 | 2003-10-10 | 주승기 | Digital Processing System of Microscopic Image for Ophthalmology |
JP4529263B2 (en) * | 2000-09-18 | 2010-08-25 | 沖電気工業株式会社 | Iris recognition device |
KR100357441B1 (en) * | 2000-09-22 | 2002-10-19 | 에버미디어 주식회사 | Human iris recognition method using harr wavelet transform and lvq |
US7277561B2 (en) * | 2000-10-07 | 2007-10-02 | Qritek Co., Ltd. | Iris identification |
DE10052201B8 (en) * | 2000-10-20 | 2005-06-30 | Carl Zeiss Meditec Ag | Method and device for identifying a patient and an operating area |
US6453057B1 (en) | 2000-11-02 | 2002-09-17 | Retinal Technologies, L.L.C. | Method for generating a unique consistent signal pattern for identification of an individual |
US7224822B2 (en) * | 2000-11-02 | 2007-05-29 | Retinal Technologies, L.L.C. | System for capturing an image of the retina for identification |
US6920237B2 (en) * | 2000-12-19 | 2005-07-19 | Eastman Kodak Company | Digital image processing method and computer program product for detecting human irises in an image |
US7006669B1 (en) | 2000-12-31 | 2006-02-28 | Cognex Corporation | Machine vision method and apparatus for thresholding images of non-uniform materials |
MXPA03006923A (en) * | 2001-02-02 | 2003-11-18 | Homer Gregg S | Method for alteration of iris pigment. |
GB0113255D0 (en) * | 2001-05-31 | 2001-07-25 | Scient Generics Ltd | Number generator |
EP1366595B1 (en) * | 2001-02-14 | 2018-11-14 | GenKey Netherlands B.V. | Data processing apparatus and method |
KR100374708B1 (en) * | 2001-03-06 | 2003-03-04 | 에버미디어 주식회사 | Non-contact type human iris recognition method by correction of rotated iris image |
KR100897365B1 (en) * | 2001-04-27 | 2009-05-15 | 바슈 앤드 롬 인코포레이티드 | Iris Pattern Recognition And Alignment |
JP2002341406A (en) * | 2001-05-11 | 2002-11-27 | Matsushita Electric Ind Co Ltd | Method and device for imaging object to be authenticated |
US20040193893A1 (en) * | 2001-05-18 | 2004-09-30 | Michael Braithwaite | Application-specific biometric templates |
JP2003016434A (en) * | 2001-06-27 | 2003-01-17 | Matsushita Electric Ind Co Ltd | Individual authenticating device |
JP2003030659A (en) * | 2001-07-16 | 2003-01-31 | Matsushita Electric Ind Co Ltd | Iris authentication device and iris image pickup device |
KR20030034258A (en) * | 2001-08-04 | 2003-05-09 | 주식회사 큐리텍 | Identification system and method using iris and retina, and media that can record computer program sources thereof |
GB0119859D0 (en) * | 2001-08-15 | 2001-10-10 | Qinetiq Ltd | Eye tracking system |
US7416302B2 (en) * | 2001-09-14 | 2008-08-26 | Stmicroelectronics S.A. | Ring location |
US7203343B2 (en) * | 2001-09-21 | 2007-04-10 | Hewlett-Packard Development Company, L.P. | System and method for determining likely identity in a biometric database |
NO316489B1 (en) | 2001-10-01 | 2004-01-26 | Genkey As | System, portable device and method for digital authentication, encryption and signing by generating volatile but consistent and repeatable crypton keys |
JP4172930B2 (en) * | 2001-10-31 | 2008-10-29 | 松下電器産業株式会社 | Eye imaging device and entrance / exit management system |
US7775665B2 (en) * | 2001-11-13 | 2010-08-17 | Dellavecchia Michael A | Method for optically scanning objects |
US20040165147A1 (en) * | 2001-11-13 | 2004-08-26 | Della Vecchia Michael A. | Determining iris biometric and spatial orientation of an iris in accordance with same |
US7377647B2 (en) * | 2001-11-13 | 2008-05-27 | Philadelphia Retina Endowment Fund | Clarifying an image of an object to perform a procedure on the object |
US7226166B2 (en) | 2001-11-13 | 2007-06-05 | Philadelphia Retina Endowment Fund | Optimizing the properties of electromagnetic energy in a medium using stochastic parallel perturbation gradient descent optimization adaptive optics |
AUPR912101A0 (en) * | 2001-11-26 | 2001-12-20 | Argus Solutions Pty Ltd | Computerised identity matching management |
US6665426B1 (en) * | 2002-01-29 | 2003-12-16 | West Virginia University Research Corporation | Method of biometric identification of an individual and associated apparatus |
US7155035B2 (en) | 2002-02-05 | 2006-12-26 | Matsushita Electric Industrial Co., Ltd. | Personal authentication method, personal authentication apparatus and image capturing device |
US20030165259A1 (en) * | 2002-02-15 | 2003-09-04 | Balent James S. | Signal analysis using image processing techniques |
US20040052418A1 (en) * | 2002-04-05 | 2004-03-18 | Bruno Delean | Method and apparatus for probabilistic image analysis |
US7369685B2 (en) * | 2002-04-05 | 2008-05-06 | Identix Corporation | Vision-based operating method and system |
GB0228434D0 (en) * | 2002-12-05 | 2003-01-08 | Scient Generics Ltd | Error correction |
AUPS254302A0 (en) * | 2002-05-24 | 2002-06-13 | Resmed Limited | A sleepiness test |
EP1520369B1 (en) * | 2002-05-31 | 2006-10-18 | Scientific Generics Limited | Biometric authentication system |
WO2004042658A1 (en) * | 2002-11-07 | 2004-05-21 | Matsushita Electric Industrial Co., Ltd. | Method for cerficating individual, iris registering device, system for certificating iris, and program for cerficating individual |
JP3639291B2 (en) * | 2002-11-07 | 2005-04-20 | 松下電器産業株式会社 | Personal authentication method, iris registration device, iris authentication device, and personal authentication program |
AU2003900586A0 (en) | 2003-02-11 | 2003-02-27 | Argus Solutions Pty Ltd | Delivery of metered doses |
US7436986B2 (en) * | 2003-03-25 | 2008-10-14 | Bausch & Lomb Incorporated | Positive patient identification |
US7599524B2 (en) * | 2003-04-04 | 2009-10-06 | Sarnoff Corporation | Method and apparatus for providing a robust object finder |
JP4584912B2 (en) * | 2003-04-11 | 2010-11-24 | ボシュ・アンド・ロム・インコーポレイテッド | System and method for eye data acquisition and alignment and tracking |
US7660443B2 (en) * | 2003-04-28 | 2010-02-09 | Panasonic Corporation | Artificial eye distinguishing method and device, artificial eye distinguishing program, iris recognition method, false printed matter distinguishing method, and image distinguishing method |
KR200321670Y1 (en) * | 2003-05-07 | 2003-07-31 | 주식회사 큐리텍 | Iris identification camera |
US20040233192A1 (en) * | 2003-05-22 | 2004-11-25 | Hopper Stephen A. | Focally-controlled imaging system and method |
US7458683B2 (en) * | 2003-06-16 | 2008-12-02 | Amo Manufacturing Usa, Llc | Methods and devices for registering optical measurement datasets of an optical system |
AU2003903392A0 (en) | 2003-07-03 | 2003-07-17 | Argus Solutions Pty Ltd | Management control of assets |
JP3848953B2 (en) * | 2003-07-04 | 2006-11-22 | 松下電器産業株式会社 | Living body eye determination method and living body eye determination device |
CN100350877C (en) * | 2003-07-04 | 2007-11-28 | 松下电器产业株式会社 | Organism eye judgment method and organism eye judgment device |
GB0316631D0 (en) * | 2003-07-16 | 2003-08-20 | Omniperception Ltd | Facial liveness assessment system |
JP3879719B2 (en) * | 2003-08-22 | 2007-02-14 | 松下電器産業株式会社 | Image input device and authentication device using the same |
US8049812B2 (en) | 2006-03-03 | 2011-11-01 | Honeywell International Inc. | Camera with auto focus capability |
US8705808B2 (en) | 2003-09-05 | 2014-04-22 | Honeywell International Inc. | Combined face and iris recognition system |
US8064647B2 (en) | 2006-03-03 | 2011-11-22 | Honeywell International Inc. | System for iris detection tracking and recognition at a distance |
US8098901B2 (en) * | 2005-01-26 | 2012-01-17 | Honeywell International Inc. | Standoff iris recognition system |
US7756301B2 (en) * | 2005-01-26 | 2010-07-13 | Honeywell International Inc. | Iris recognition system and method |
US8090157B2 (en) * | 2005-01-26 | 2012-01-03 | Honeywell International Inc. | Approaches and apparatus for eye detection in a digital image |
US7593550B2 (en) * | 2005-01-26 | 2009-09-22 | Honeywell International Inc. | Distance iris recognition |
US8442276B2 (en) | 2006-03-03 | 2013-05-14 | Honeywell International Inc. | Invariant radial iris segmentation |
US8085993B2 (en) | 2006-03-03 | 2011-12-27 | Honeywell International Inc. | Modular biometrics collection system architecture |
KR20050026347A (en) * | 2003-09-09 | 2005-03-15 | 유웅덕 | Apparatus for photographing iris pattern |
US20050111704A1 (en) * | 2003-11-24 | 2005-05-26 | Verghis James S. | Iris mapping and compatibility and personality evaluation |
WO2005055125A1 (en) * | 2003-12-01 | 2005-06-16 | Sung-Bok Shin | Identification system and method having iris, and media that can record computer program sources thereof |
FR2864290B1 (en) * | 2003-12-18 | 2006-05-26 | Sagem | METHOD AND DEVICE FOR RECOGNIZING IRIS |
US7481536B2 (en) * | 2004-02-19 | 2009-01-27 | Amo Manufacturing Usa, Llc | Methods and systems for differentiating left and right eye images |
JP2005304809A (en) * | 2004-04-22 | 2005-11-04 | Matsushita Electric Ind Co Ltd | Eye image pickup device with lighting system |
JP3940759B2 (en) | 2004-05-10 | 2007-07-04 | 松下電器産業株式会社 | Iris registration method, iris registration device, and iris registration program |
GB0412175D0 (en) * | 2004-06-01 | 2004-06-30 | Smart Sensors Ltd | Identification of image characteristics |
GB0413034D0 (en) * | 2004-06-10 | 2004-07-14 | Scient Generics Ltd | Secure workflow engine |
US20050281440A1 (en) * | 2004-06-18 | 2005-12-22 | Pemer Frederick A | Iris feature detection and sensor-based edge detection |
GB0414113D0 (en) * | 2004-06-24 | 2004-07-28 | Virulite Distrib Ltd | Cosmetic uses of electromagnetic radiation |
JP3966309B2 (en) * | 2004-07-14 | 2007-08-29 | 松下電器産業株式会社 | Pupil detection device and iris authentication device |
JP4599110B2 (en) * | 2004-07-30 | 2010-12-15 | キヤノン株式会社 | Image processing apparatus and method, imaging apparatus, and program |
JP4455216B2 (en) * | 2004-08-06 | 2010-04-21 | キヤノン株式会社 | Detection device |
US7248720B2 (en) * | 2004-10-21 | 2007-07-24 | Retica Systems, Inc. | Method and system for generating a combined retina/iris pattern biometric |
US7167736B2 (en) * | 2004-11-04 | 2007-01-23 | Q Step Technologies, Inc. | Non-invasive measurement system and method for measuring the concentration of an optically-active substance |
US20060115130A1 (en) * | 2004-11-29 | 2006-06-01 | Douglas Kozlay | Eyewear with biometrics to protect displayed data |
IL165586A0 (en) * | 2004-12-06 | 2006-01-15 | Daphna Palti Wasserman | Multivariate dynamic biometrics system |
US20060147095A1 (en) * | 2005-01-03 | 2006-07-06 | Usher David B | Method and system for automatically capturing an image of a retina |
US20060206724A1 (en) * | 2005-02-16 | 2006-09-14 | David Schaufele | Biometric-based systems and methods for identity verification |
WO2006101943A2 (en) * | 2005-03-16 | 2006-09-28 | Lc Technologies, Inc. | System and method for eyeball surface topography as a biometric discriminator |
JP4498224B2 (en) * | 2005-06-14 | 2010-07-07 | キヤノン株式会社 | Image processing apparatus and method |
GB0512038D0 (en) * | 2005-06-14 | 2005-07-20 | Dougal Gordon | Therapeutic and cosmetic uses of electromagnetic radiation |
US7639861B2 (en) | 2005-09-14 | 2009-12-29 | Cognex Technology And Investment Corporation | Method and apparatus for backlighting a wafer during alignment |
US8111904B2 (en) * | 2005-10-07 | 2012-02-07 | Cognex Technology And Investment Corp. | Methods and apparatus for practical 3D vision system |
US8260008B2 (en) | 2005-11-11 | 2012-09-04 | Eyelock, Inc. | Methods for performing biometric recognition of a human eye and corroboration of same |
US7801335B2 (en) * | 2005-11-11 | 2010-09-21 | Global Rainmakers Inc. | Apparatus and methods for detecting the presence of a human eye |
JP2009520178A (en) * | 2005-12-19 | 2009-05-21 | ハネウェル・インターナショナル・インコーポレーテッド | Multi-fluid refrigerant system |
US7744216B1 (en) | 2006-01-06 | 2010-06-29 | Lockheed Martin Corporation | Display system intensity adjustment based on pupil dilation |
GB0603411D0 (en) * | 2006-02-21 | 2006-03-29 | Xvista Ltd | Method of processing an image of an eye |
KR101299074B1 (en) * | 2006-03-03 | 2013-08-30 | 허니웰 인터내셔널 인코포레이티드 | Iris encoding system |
EP1991947B1 (en) * | 2006-03-03 | 2020-04-29 | Gentex Corporation | Indexing and database search system |
DE602007007062D1 (en) * | 2006-03-03 | 2010-07-22 | Honeywell Int Inc | IRISER IDENTIFICATION SYSTEM WITH IMAGE QUALITY METERING |
US8364646B2 (en) * | 2006-03-03 | 2013-01-29 | Eyelock, Inc. | Scalable searching of biometric databases using dynamic selection of data subsets |
AU2007220010B2 (en) * | 2006-03-03 | 2011-02-17 | Gentex Corporation | Single lens splitter camera |
DE102007016397B4 (en) * | 2006-04-03 | 2011-02-03 | Sabine Raveau | Apparatus and method for image acquisition of eyes |
WO2008039252A2 (en) * | 2006-05-15 | 2008-04-03 | Retica Systems, Inc. | Multimodal ocular biometric system |
US8604901B2 (en) | 2006-06-27 | 2013-12-10 | Eyelock, Inc. | Ensuring the provenance of passengers at a transportation facility |
US8162584B2 (en) * | 2006-08-23 | 2012-04-24 | Cognex Corporation | Method and apparatus for semiconductor wafer alignment |
PL380581A1 (en) * | 2006-09-07 | 2008-03-17 | Naukowa I Akademicka Sieć Komputerowa | Testing method of eye vitality and device for thesting of eye vitality |
US8170293B2 (en) * | 2006-09-15 | 2012-05-01 | Identix Incorporated | Multimodal ocular biometric system and methods |
US8121356B2 (en) | 2006-09-15 | 2012-02-21 | Identix Incorporated | Long distance multimodal biometric system and method |
WO2008033784A2 (en) * | 2006-09-15 | 2008-03-20 | Retica Systems, Inc. | Long distance multimodal biometric system and method |
WO2008036897A1 (en) | 2006-09-22 | 2008-03-27 | Global Rainmakers, Inc. | Compact biometric acquisition system and method |
US7970179B2 (en) * | 2006-09-25 | 2011-06-28 | Identix Incorporated | Iris data extraction |
TW200816060A (en) * | 2006-09-29 | 2008-04-01 | Univ Nat Chiao Tung | Iris recognition method |
US8280120B2 (en) | 2006-10-02 | 2012-10-02 | Eyelock Inc. | Fraud resistant biometric financial transaction system and method |
US9846739B2 (en) | 2006-10-23 | 2017-12-19 | Fotonation Limited | Fast database matching |
US7809747B2 (en) * | 2006-10-23 | 2010-10-05 | Donald Martin Monro | Fuzzy database matching |
ES2276637B1 (en) * | 2006-11-03 | 2008-11-16 | Jose Antonio Gil Soldevilla | COLOR OBTAINING PROCEDURE FROM IRIS. |
RU2318438C1 (en) * | 2006-11-28 | 2008-03-10 | Дмитрий Евгеньевич Антонов | Device for getting image of iris |
US8953849B2 (en) | 2007-04-19 | 2015-02-10 | Eyelock, Inc. | Method and system for biometric recognition |
WO2008131201A1 (en) | 2007-04-19 | 2008-10-30 | Global Rainmakers, Inc. | Method and system for biometric recognition |
US8063889B2 (en) * | 2007-04-25 | 2011-11-22 | Honeywell International Inc. | Biometric data collection system |
US20120239458A9 (en) * | 2007-05-18 | 2012-09-20 | Global Rainmakers, Inc. | Measuring Effectiveness of Advertisements and Linking Certain Consumer Activities Including Purchases to Other Activities of the Consumer |
US7652716B2 (en) * | 2007-05-31 | 2010-01-26 | Microsoft Corporation | Computer-controlled lighting for video communication |
IL184399A0 (en) * | 2007-07-03 | 2007-10-31 | Yossi Tsuria | Content delivery system |
US20090060348A1 (en) * | 2007-08-28 | 2009-03-05 | Donald Martin Monro | Determination of Image Similarity |
US8212870B2 (en) | 2007-09-01 | 2012-07-03 | Hanna Keith J | Mirror system and method for acquiring biometric data |
US9002073B2 (en) | 2007-09-01 | 2015-04-07 | Eyelock, Inc. | Mobile identity platform |
US9117119B2 (en) | 2007-09-01 | 2015-08-25 | Eyelock, Inc. | Mobile identity platform |
US9036871B2 (en) | 2007-09-01 | 2015-05-19 | Eyelock, Inc. | Mobility identity platform |
WO2009029765A1 (en) * | 2007-09-01 | 2009-03-05 | Global Rainmakers, Inc. | Mirror system and method for acquiring biometric data |
US8393734B2 (en) | 2007-09-14 | 2013-03-12 | Neuroptics, Inc. | Pupilary screening method and system |
US7824034B2 (en) * | 2007-09-19 | 2010-11-02 | Utc Fire & Security Americas Corporation, Inc. | Iris imaging system and method for the same |
US10398599B2 (en) | 2007-10-05 | 2019-09-03 | Topcon Medical Laser Systems Inc. | Semi-automated ophthalmic photocoagulation method and apparatus |
JP5674473B2 (en) * | 2007-11-27 | 2015-02-25 | ウェイヴフロント・バイオメトリック・テクノロジーズ・ピーティーワイ・リミテッド | Biometric authentication using eyes |
US20090252382A1 (en) * | 2007-12-06 | 2009-10-08 | University Of Notre Dame Du Lac | Segmentation of iris images using active contour processing |
US8189879B2 (en) * | 2008-02-14 | 2012-05-29 | Iristrac, Llc | System and method for animal identification using IRIS images |
US8201942B2 (en) * | 2008-04-02 | 2012-06-19 | Refocus Group, Inc. | System and method for identifying a position to insert a scleral prosthesis into an eye |
US8345936B2 (en) * | 2008-05-09 | 2013-01-01 | Noblis, Inc. | Multispectral iris fusion for enhancement and interoperability |
US8436907B2 (en) * | 2008-05-09 | 2013-05-07 | Honeywell International Inc. | Heterogeneous video capturing system |
DE602008003019D1 (en) | 2008-06-25 | 2010-11-25 | Deutsche Telekom Ag | System for extraction, identification and verification of iris features based on directionlets |
WO2009158662A2 (en) | 2008-06-26 | 2009-12-30 | Global Rainmakers, Inc. | Method of reducing visibility of illimination while acquiring high quality imagery |
GB0812753D0 (en) * | 2008-07-14 | 2008-08-20 | Dougal Gordon R P | Electromagnetic radiation and its therapeutic effect |
EP2321765B1 (en) * | 2008-07-24 | 2013-10-16 | Kynen Llc | Mobile biometric identification system and method |
US8213782B2 (en) | 2008-08-07 | 2012-07-03 | Honeywell International Inc. | Predictive autofocusing system |
US8090246B2 (en) * | 2008-08-08 | 2012-01-03 | Honeywell International Inc. | Image acquisition system |
US8290219B2 (en) * | 2008-08-22 | 2012-10-16 | International Business Machines Corporation | Registration-free transforms for cancelable iris biometrics |
US9633261B2 (en) * | 2008-08-22 | 2017-04-25 | International Business Machines Corporation | Salting system and method for cancelable iris biometric |
US8306279B2 (en) * | 2008-09-15 | 2012-11-06 | Eyelock, Inc. | Operator interface for face and iris recognition devices |
CN101727574A (en) * | 2008-10-17 | 2010-06-09 | 深圳富泰宏精密工业有限公司 | Iris recognition system and method |
US20100278394A1 (en) * | 2008-10-29 | 2010-11-04 | Raguin Daniel H | Apparatus for Iris Capture |
US8317325B2 (en) | 2008-10-31 | 2012-11-27 | Cross Match Technologies, Inc. | Apparatus and method for two eye imaging for iris identification |
US7967442B2 (en) * | 2008-11-28 | 2011-06-28 | Neuroptics, Inc. | Methods, systems, and devices for monitoring anisocoria and asymmetry of pupillary reaction to stimulus |
US8280119B2 (en) | 2008-12-05 | 2012-10-02 | Honeywell International Inc. | Iris recognition system using quality metrics |
US8374404B2 (en) * | 2009-02-13 | 2013-02-12 | Raytheon Company | Iris recognition using hyper-spectral signatures |
WO2010099475A1 (en) * | 2009-02-26 | 2010-09-02 | Kynen Llc | User authentication system and method |
US8195044B2 (en) | 2009-03-30 | 2012-06-05 | Eyelock Inc. | Biometric camera mount system |
US8630464B2 (en) | 2009-06-15 | 2014-01-14 | Honeywell International Inc. | Adaptive iris matching using database indexing |
US8472681B2 (en) | 2009-06-15 | 2013-06-25 | Honeywell International Inc. | Iris and ocular recognition system using trace transforms |
US8750575B2 (en) * | 2009-08-04 | 2014-06-10 | International Business Machines Corporation | Reflexive iris template |
US20110119141A1 (en) * | 2009-11-16 | 2011-05-19 | Hoyos Corporation | Siccolla Identity Verification Architecture and Tool |
US8577094B2 (en) | 2010-04-09 | 2013-11-05 | Donald Martin Monro | Image template masking |
WO2011151812A1 (en) | 2010-05-10 | 2011-12-08 | Ramot At Tel-Aviv University Ltd. | System for treating glaucoma by directing electromagnetic energy to the limbal area of an eye |
US8742887B2 (en) | 2010-09-03 | 2014-06-03 | Honeywell International Inc. | Biometric visitor check system |
WO2012040196A1 (en) * | 2010-09-20 | 2012-03-29 | Amo Development Llc | System and methods for mitigating changes in pupil size during laser refractive surgery to maintain ablation centration |
US9753025B2 (en) | 2010-10-26 | 2017-09-05 | Bi2 Technologies, LLC | Mobile wireless hand-held identification system and breathalyzer |
US10068080B2 (en) | 2010-10-26 | 2018-09-04 | Bi2 Technologies, LLC | Mobile wireless hand-held biometric identification system |
US9507926B2 (en) | 2010-10-26 | 2016-11-29 | Bi2 Technologies, LLC | Mobile wireless hand-held identification system and method for identification |
US8719584B2 (en) | 2010-10-26 | 2014-05-06 | Bi2 Technologies, LLC | Mobile, wireless hand-held biometric capture, processing and communication system and method for biometric identification |
US8254768B2 (en) * | 2010-12-22 | 2012-08-28 | Michael Braithwaite | System and method for illuminating and imaging the iris of a person |
US8831416B2 (en) * | 2010-12-22 | 2014-09-09 | Michael Braithwaite | System and method for illuminating and identifying a person |
US10043229B2 (en) | 2011-01-26 | 2018-08-07 | Eyelock Llc | Method for confirming the identity of an individual while shielding that individual's personal data |
RU2589859C2 (en) | 2011-02-17 | 2016-07-10 | АЙЛОК ЭлЭлСи | Efficient method and system for obtaining image data of scene and iris image using one sensor |
CN103797495A (en) | 2011-04-19 | 2014-05-14 | 眼锁股份有限公司 | Biometric chain of provenance |
US9124798B2 (en) * | 2011-05-17 | 2015-09-01 | Eyelock Inc. | Systems and methods for illuminating an iris with visible light for biometric acquisition |
US8885882B1 (en) | 2011-07-14 | 2014-11-11 | The Research Foundation For The State University Of New York | Real time eye tracking for human computer interaction |
EP2748768A4 (en) | 2011-08-22 | 2016-05-11 | Eyelock Llc | Systems and methods for capturing artifact free images |
GB2495324B (en) | 2011-10-07 | 2018-05-30 | Irisguard Inc | Security improvements for Iris recognition systems |
GB2495328B (en) * | 2011-10-07 | 2018-05-30 | Irisguard Inc | Improvements relating to Iris cameras |
GB2495323B (en) | 2011-10-07 | 2018-05-30 | Irisguard Inc | Improvements for iris recognition systems |
US9412022B2 (en) | 2012-09-06 | 2016-08-09 | Leonard Flom | Iris identification system and method |
US9101297B2 (en) | 2012-12-11 | 2015-08-11 | Elwha Llc | Time-based unobtrusive active eye interrogation |
US9039180B2 (en) | 2012-12-11 | 2015-05-26 | Elwah LLC | Self-aligning unobtrusive active eye interrogation |
US9039179B2 (en) | 2012-12-11 | 2015-05-26 | Elwha Llc | Unobtrusive active eye interrogation |
CN105050556B (en) | 2013-02-26 | 2017-06-06 | 贝尔金激光有限公司 | For the system of glaucoma treatment |
US9495526B2 (en) | 2013-03-15 | 2016-11-15 | Eyelock Llc | Efficient prevention of fraud |
JP2014206932A (en) * | 2013-04-15 | 2014-10-30 | オムロン株式会社 | Authentication device, authentication method, control program, and recording medium |
US8958608B2 (en) | 2013-06-04 | 2015-02-17 | Ut-Battelle, Llc | Frontal view reconstruction for iris recognition |
US10032075B2 (en) | 2013-12-23 | 2018-07-24 | Eyelock Llc | Methods and apparatus for power-efficient iris recognition |
WO2015103595A1 (en) | 2014-01-06 | 2015-07-09 | Eyelock, Inc. | Methods and apparatus for repetitive iris recognition |
WO2015108911A1 (en) * | 2014-01-16 | 2015-07-23 | Delta ID Inc. | Method and apparatus for controlling intensity of illumination in eye based biometric systems |
JP6417676B2 (en) | 2014-03-06 | 2018-11-07 | ソニー株式会社 | Information processing apparatus, information processing method, eyewear terminal, and authentication system |
US9818114B2 (en) | 2014-08-11 | 2017-11-14 | Mastercard International Incorporated | Systems and methods for performing payment card transactions using a wearable computing device |
US9870457B2 (en) | 2014-08-15 | 2018-01-16 | California Institute Of Technology | HERMA—heartbeat microwave authentication |
CN106796655A (en) | 2014-09-12 | 2017-05-31 | 眼锁有限责任公司 | Method and apparatus for guiding sight line of the user in iris authentication system |
US10425814B2 (en) | 2014-09-24 | 2019-09-24 | Princeton Identity, Inc. | Control of wireless communication device capability in a mobile device with a biometric key |
US20170132466A1 (en) | 2014-09-30 | 2017-05-11 | Qualcomm Incorporated | Low-power iris scan initialization |
US9554100B2 (en) | 2014-09-30 | 2017-01-24 | Qualcomm Incorporated | Low-power always-on face detection, tracking, recognition and/or analysis using events-based vision sensor |
US10515284B2 (en) | 2014-09-30 | 2019-12-24 | Qualcomm Incorporated | Single-processor computer vision hardware control and application execution |
US9940533B2 (en) | 2014-09-30 | 2018-04-10 | Qualcomm Incorporated | Scanning window for isolating pixel values in hardware for computer vision operations |
US9585616B2 (en) | 2014-11-17 | 2017-03-07 | Elwha Llc | Determining treatment compliance using speech patterns passively captured from a patient environment |
US10430557B2 (en) | 2014-11-17 | 2019-10-01 | Elwha Llc | Monitoring treatment compliance using patient activity patterns |
US9589107B2 (en) | 2014-11-17 | 2017-03-07 | Elwha Llc | Monitoring treatment compliance using speech patterns passively captured from a patient environment |
WO2016081609A1 (en) | 2014-11-19 | 2016-05-26 | Eyelock Llc | Model-based prediction of an optimal convenience metric for authorizing transactions |
EP3223232B1 (en) * | 2014-11-20 | 2022-05-25 | Sony Group Corporation | Glasses-type wearable device for capturing an image of a wearer's iris |
CA2969331A1 (en) | 2014-12-03 | 2016-06-09 | Princeton Identity, Inc. | System and method for mobile device biometric add-on |
WO2016118473A1 (en) | 2015-01-20 | 2016-07-28 | Eyelock Llc | Lens system for high quality visible image acquisition and infra-red iris image acquisition |
US9509690B2 (en) | 2015-03-12 | 2016-11-29 | Eyelock Llc | Methods and systems for managing network activity using biometrics |
US10589105B2 (en) | 2015-03-27 | 2020-03-17 | The Invention Science Fund Ii, Llc | Method and system for controlling ear stimulation |
US10406376B2 (en) | 2015-03-27 | 2019-09-10 | Equility Llc | Multi-factor control of ear stimulation |
US10398902B2 (en) | 2015-03-27 | 2019-09-03 | Equility Llc | Neural stimulation method and system with audio output |
US10512783B2 (en) | 2015-03-27 | 2019-12-24 | Equility Llc | User interface method and system for ear stimulation |
US11364380B2 (en) | 2015-03-27 | 2022-06-21 | Elwha Llc | Nerve stimulation system, subsystem, headset, and earpiece |
US10327984B2 (en) | 2015-03-27 | 2019-06-25 | Equility Llc | Controlling ear stimulation in response to image analysis |
US9987489B2 (en) | 2015-03-27 | 2018-06-05 | Elwha Llc | Controlling ear stimulation in response to electrical contact sensing |
US10039928B2 (en) | 2015-03-27 | 2018-08-07 | Equility Llc | Ear stimulation with neural feedback sensing |
US9811729B2 (en) | 2015-05-12 | 2017-11-07 | Ut-Battelle, Llc | Iris recognition via plenoptic imaging |
CN104899583A (en) * | 2015-06-30 | 2015-09-09 | 成都点石创想科技有限公司 | Iris identifying method for door access monitoring system |
US10832426B2 (en) | 2015-09-24 | 2020-11-10 | Apple Inc. | Systems and methods for surface monitoring |
US11100673B2 (en) | 2015-09-24 | 2021-08-24 | Apple Inc. | Systems and methods for localization using surface imaging |
US10157312B2 (en) * | 2015-10-08 | 2018-12-18 | Microsoft Technology Licensing, Llc | Iris recognition |
US10311299B2 (en) | 2015-12-21 | 2019-06-04 | Eyelock Llc | Reflected optic camera module for iris recognition in a computing device |
EP3403217A4 (en) | 2016-01-12 | 2019-08-21 | Princeton Identity, Inc. | Systems and methods of biometric analysis |
US10373008B2 (en) | 2016-03-31 | 2019-08-06 | Princeton Identity, Inc. | Systems and methods of biometric analysis with adaptive trigger |
US10366296B2 (en) | 2016-03-31 | 2019-07-30 | Princeton Identity, Inc. | Biometric enrollment systems and methods |
JP6409088B2 (en) * | 2016-04-28 | 2018-10-17 | シャープ株式会社 | Image processing method and image processing apparatus |
EP3458997A2 (en) | 2016-05-18 | 2019-03-27 | Eyelock, LLC | Iris recognition methods and systems based on an iris stochastic texture model |
US11138741B2 (en) | 2016-05-27 | 2021-10-05 | Rochester Institute Of Technology | System and method for eye tracking |
US10442439B1 (en) | 2016-08-18 | 2019-10-15 | Apple Inc. | System and method for road friction coefficient estimation |
JP6798233B2 (en) * | 2016-10-06 | 2020-12-09 | 富士ゼロックス株式会社 | Eye light measuring device and eye light measuring method |
US10614332B2 (en) * | 2016-12-16 | 2020-04-07 | Qualcomm Incorportaed | Light source modulation for iris size adjustment |
US10984235B2 (en) | 2016-12-16 | 2021-04-20 | Qualcomm Incorporated | Low power data generation for iris-related detection and authentication |
WO2018156726A1 (en) | 2017-02-24 | 2018-08-30 | Eyelock, Llc | Systems and methods for providing illumination for iris biometric acquisition |
US10607096B2 (en) | 2017-04-04 | 2020-03-31 | Princeton Identity, Inc. | Z-dimension user feedback biometric system |
US10657401B2 (en) * | 2017-06-06 | 2020-05-19 | Microsoft Technology Licensing, Llc | Biometric object spoof detection based on image intensity variations |
US10902104B2 (en) | 2017-07-26 | 2021-01-26 | Princeton Identity, Inc. | Biometric security systems and methods |
CA3015802C (en) | 2017-08-31 | 2021-06-22 | Eyelock, Llc | Systems and methods of biometric acquistion using positive optical distortion |
IL308110A (en) | 2018-07-02 | 2023-12-01 | Belkin Vision Ltd | Direct selective laser trabeculoplasty |
CN109657531A (en) * | 2018-09-18 | 2019-04-19 | 深圳先牛信息技术有限公司 | A kind of human face in-vivo detection method and detection device based on hot spot on eyeball |
EP3764272A1 (en) | 2019-07-08 | 2021-01-13 | Fraunhofer Gesellschaft zur Förderung der Angewand | System and method for identifying a test object |
DE102019213931B4 (en) | 2019-09-12 | 2021-10-28 | Zf Friedrichshafen Ag | Method and computer program product for identifying a vehicle user and control device for automated driving functions |
US20240013606A1 (en) * | 2022-07-08 | 2024-01-11 | EZ Vending Machines, Inc. | Verification vending machine |
US11762969B1 (en) | 2023-01-12 | 2023-09-19 | King Saud University | Systems and methods for facilitating biometric recognition |
Family Cites Families (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1192512A (en) * | 1912-02-10 | 1916-07-25 | George W Fitz | Shadow-pupillometer. |
US1887115A (en) * | 1931-04-23 | 1932-11-08 | Zeiss Carl Fa | Instrument for the examination of the reaction of the eye pupils when different parts of the retina are illuminated |
US2434890A (en) * | 1938-03-28 | 1948-01-20 | James B Saunders | Pupillometric exposure meter |
US3136839A (en) * | 1958-09-16 | 1964-06-09 | Safir Aran | Apparatus for objectively testing an optical system |
US3366786A (en) * | 1965-04-15 | 1968-01-30 | Richard P. Delano | Apparatus for producing color effects |
US3473868A (en) * | 1967-04-27 | 1969-10-21 | Space Sciences Inc | Eye position and movement monitor |
US3533683A (en) * | 1967-06-13 | 1970-10-13 | Whittaker Corp | Dynamic pupillometers using television camera system |
US3533684A (en) * | 1967-06-26 | 1970-10-13 | Lawrence Stark | Display of measurement adequacy marker system for pupillometers |
US3482904A (en) * | 1967-12-08 | 1969-12-09 | David Volk | Method and apparatus for determining the eccentricity of a conicoid surface |
US3598107A (en) * | 1968-07-25 | 1971-08-10 | Hamamatsu T V Co Ltd | Pupillary motion observing apparatus |
US3600098A (en) * | 1969-12-29 | 1971-08-17 | Bausch & Lomb | Optical alignment method and apparatus |
US3778135A (en) * | 1972-11-29 | 1973-12-11 | American Optical Corp | Ophthalmometer having alternative viewing and measuring systems and including an improved illumination system |
JPS49136227U (en) * | 1973-03-22 | 1974-11-22 | ||
US3966310A (en) * | 1974-02-15 | 1976-06-29 | Larson Merlin D | Pupillometer and method of use thereof |
DE2414322C3 (en) * | 1974-03-25 | 1980-01-17 | Philips Patentverwaltung Gmbh, 2000 Hamburg | Method and arrangement for decoding overlay images of three-dimensional objects |
CH581986A5 (en) * | 1974-08-23 | 1976-11-30 | Bracher Damiel | |
US3915564A (en) * | 1974-09-12 | 1975-10-28 | Zeiss Stiftung | Retinal image-display system |
US4008606A (en) * | 1975-10-20 | 1977-02-22 | The United States Of America As Represented By The Secretary Of The Navy | Ship's bottom inspection apparatus |
US4172632A (en) * | 1976-01-21 | 1979-10-30 | Holmes Lawrence Jr | Method and apparatus producing three-dimensional shadow images |
US4189215A (en) * | 1976-03-22 | 1980-02-19 | Humphrey Instruments Inc. | Method and apparatus for the corneal positioning of a patient's eye |
JPS52144174A (en) * | 1976-05-25 | 1977-12-01 | Mitsubishi Electric Corp | Composite illumination |
JPS53125379A (en) * | 1976-10-25 | 1978-11-01 | Mitsubishi Electric Corp | Mixed lights illumination method |
US4109237A (en) * | 1977-01-17 | 1978-08-22 | Hill Robert B | Apparatus and method for identifying individuals through their retinal vasculature patterns |
JPS53126791A (en) * | 1977-04-12 | 1978-11-06 | Canon Kk | Ophthalmolgic decice |
US4253743A (en) * | 1977-05-17 | 1981-03-03 | Canon Kabushiki Kaisha | Eye testing instrument |
JPS6054053B2 (en) * | 1977-11-15 | 1985-11-28 | ミノルタ株式会社 | Fundus camera for easy pupil alignment |
US4309085A (en) * | 1979-07-12 | 1982-01-05 | Morrison Robert J | Method for measuring eye features with a contact lens |
JPS5663330A (en) * | 1979-10-25 | 1981-05-29 | Canon Kk | Inspecting machine for eye |
US4375320A (en) * | 1980-09-05 | 1983-03-01 | Smirmaul Heinz J | Dual image corneal radius measurement |
US4393366A (en) * | 1981-02-17 | 1983-07-12 | Eye-D Development Ii Ltd. | Rotating beam ocular identification apparatus and method |
FI61992C (en) * | 1981-05-19 | 1982-11-10 | Kaakinen Kari A | FOERFARANDE OCH ANORDNING FOER BEFINNING OCH REGISTRERING AV OEGAS FEL |
-
1985
- 1985-02-20 US US06/703,312 patent/US4641349A/en not_active Expired - Lifetime
-
1986
- 1986-02-04 WO PCT/US1986/000227 patent/WO1986005018A1/en active IP Right Grant
- 1986-02-04 EP EP86901250A patent/EP0215818B1/en not_active Expired - Lifetime
- 1986-02-04 DE DE8686901250T patent/DE3680618D1/en not_active Expired - Lifetime
- 1986-02-04 AT AT86901250T patent/ATE65851T1/en not_active IP Right Cessation
- 1986-02-04 BR BR8605561A patent/BR8605561A/en not_active IP Right Cessation
- 1986-02-04 JP JP61501122A patent/JPS62501889A/en active Granted
- 1986-02-19 MX MX1587A patent/MX163339A/en unknown
- 1986-02-19 CA CA000502207A patent/CA1244552A/en not_active Expired
- 1986-02-19 IL IL77920A patent/IL77920A/en not_active IP Right Cessation
Also Published As
Publication number | Publication date |
---|---|
ATE65851T1 (en) | 1991-08-15 |
DE3680618D1 (en) | 1991-09-05 |
EP0215818A4 (en) | 1988-01-28 |
EP0215818B1 (en) | 1991-07-31 |
JPS62501889A (en) | 1987-07-30 |
BR8605561A (en) | 1987-04-22 |
EP0215818A1 (en) | 1987-04-01 |
MX163339A (en) | 1992-04-24 |
IL77920A (en) | 1989-06-30 |
JPH0584166B2 (en) | 1993-12-01 |
WO1986005018A1 (en) | 1986-08-28 |
US4641349A (en) | 1987-02-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA1244552A (en) | Iris recognition system | |
US8644562B2 (en) | Multimodal ocular biometric system and methods | |
US7248720B2 (en) | Method and system for generating a combined retina/iris pattern biometric | |
US8718335B2 (en) | Biometric authentication using the eye | |
da Costa et al. | Dynamic features for iris recognition | |
US6247813B1 (en) | Iris identification system and method of identifying a person through iris recognition | |
KR100342159B1 (en) | Apparatus and method for acquiring iris images | |
CA1211218A (en) | Fovea-centered eye fundus scanner | |
US20030012413A1 (en) | Iris identification apparatus and iris image pickup apparatus | |
JPH0218850B2 (en) | ||
JP3453911B2 (en) | Gaze recognition device | |
US9704039B2 (en) | Biometric authentication using the eye | |
US11769345B2 (en) | Presentation attack detection | |
KR20010006975A (en) | A method for identifying the iris of persons based on the reaction of the pupil and autonomous nervous wreath | |
US7824034B2 (en) | Iris imaging system and method for the same | |
JP3848953B2 (en) | Living body eye determination method and living body eye determination device | |
JP2001195594A (en) | Iris identifying system and method of identifying person by iris recognition | |
KR100362484B1 (en) | The iris image acquisition method and device using fixed focus camera | |
KR20010006976A (en) | A system for identifying the iris of persons | |
JP2002056389A (en) | Iris identification system and method for identifying person by iris recognition | |
JP2004288221A (en) | Iris identification system | |
AU2014280908A1 (en) | Biometric Authentication using the Eye |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
MKEX | Expiry |