US20140336460A1 - Stereoscopic endoscope - Google Patents

Stereoscopic endoscope Download PDF

Info

Publication number
US20140336460A1
US20140336460A1 US14/341,098 US201414341098A US2014336460A1 US 20140336460 A1 US20140336460 A1 US 20140336460A1 US 201414341098 A US201414341098 A US 201414341098A US 2014336460 A1 US2014336460 A1 US 2014336460A1
Authority
US
United States
Prior art keywords
image
lens
prism
images
optical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/341,098
Inventor
David Christopher Shafer
Dennis C. Leiner
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intuitive Surgical Operations Inc
Original Assignee
Intuitive Surgical Operations Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intuitive Surgical Operations Inc filed Critical Intuitive Surgical Operations Inc
Priority to US14/341,098 priority Critical patent/US20140336460A1/en
Publication of US20140336460A1 publication Critical patent/US20140336460A1/en
Priority to US16/051,592 priority patent/US11382496B2/en
Priority to US17/751,227 priority patent/US20220280031A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00163Optical arrangements
    • A61B1/00193Optical arrangements adapted for stereoscopic vision
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00002Operational features of endoscopes
    • A61B1/00004Operational features of endoscopes characterised by electronic signal processing
    • A61B1/00009Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00002Operational features of endoscopes
    • A61B1/00043Operational features of endoscopes provided with output arrangements
    • A61B1/00045Display arrangement
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00163Optical arrangements
    • A61B1/00194Optical arrangements adapted for three-dimensional imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/05Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by the image sensor, e.g. camera, being in the distal end portion
    • A61B1/051Details of CCD assembly
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B23/00Telescopes, e.g. binoculars; Periscopes; Instruments for viewing the inside of hollow bodies; Viewfinders; Optical aiming or sighting devices
    • G02B23/24Instruments or systems for viewing the inside of hollow bodies, e.g. fibrescopes
    • G02B23/2407Optical details
    • G02B23/2415Stereoscopic endoscopes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00002Operational features of endoscopes
    • A61B1/00004Operational features of endoscopes characterised by electronic signal processing
    • A61B1/00009Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
    • A61B1/000095Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope for image enhancement
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00064Constructional details of the endoscope body
    • A61B1/00071Insertion part of the endoscope body
    • A61B1/0008Insertion part of the endoscope body characterised by distal tip features
    • A61B1/00096Optical elements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/044Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances for absorption imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/05Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by the image sensor, e.g. camera, being in the distal end portion
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B5/00Optical elements other than lenses
    • G02B5/04Prisms
    • G02B5/045Prism arrays

Definitions

  • the invention pertains to stereoscopic imaging systems, and more particularly to stereoscopic image capture endoscopes.
  • Stereoscopic endoscopes are typically mounted at the distal end of rigid shafts that extend through a cannula so as to image a surgical site during, e.g., robot-assisted surgery.
  • the images of the target object viewed by the two imaging systems should match in at least the following alignment parameters: (1) image location along the horizontal axis; (2) image location along the vertical axis; (3) image rotation; (4) image scale; (5) geometric distortion; (5) focus at the image center; (6) focal shift along the horizontal axis; and (7) focal shift along the vertical axis.
  • the tolerable errors in the matching between the two images in a stereoscopic pair depend to some extent upon the display and viewer, but in general are much more stringent requirements than exist for monocular viewing.
  • mismatches in the other parameters are difficult to correct for in image post-processing without introducing imaging artifacts.
  • the position of a second, separate, sensor has six degrees of freedom in its mounting: three of translation and three of rotation.
  • Errors in two of the translation axes between the sensors affect the relative horizontal and vertical positions of the viewed images, while errors in the third translation axis, the axis perpendicular to the sensor surface, affects both the image scale (if the objective is not telecentric in image space) and focus.
  • Errors in rotation between the two image sensors, around the axis perpendicular to the sensor surface directly affect image rotation and cannot always be corrected by alignment of the optics, while rotation errors about the other two axes affect the focal plane shifts across the imaging field.
  • a rigid body e.g., an optical image sensor chip
  • moving up and down heaving
  • moving left and right swaying
  • moving forward and backward surging
  • tilting up and down pitching
  • turning left and right yawing
  • tilting side to side rolling.
  • two separate image sensors there are a total of 12 degrees of freedom that must be controlled when mounting the two sensors to the optical train. For example, if two physically separate sensors are used, then each sensor must be aligned with its respective optical train and additional image processing (e.g., to compensate for rotation) is required in order to align the captured left and right images with each other to present the stereoscopic view to a person viewing the images.
  • two side-by-side optical paths transmit stereoscopic right side and left side images onto the surface of a single image sensing chip.
  • the single image sensing chip may be placed at various orientations (e.g., perpendicular, parallel) with respect to the lens trains in the optical paths.
  • a single prism is used to turn the light for both the right side and left side images onto the single image sensing chip.
  • one prism is used to turn the light for the right side image and another prism is used to turn the light for the left side image, and the reflective surfaces of the two prisms are substantially coplanar such that the right side and left side images are incident on the single image sensor chip.
  • FIG. 1 is an illustrative, diagrammatic plan view of the active image sensing area of an image sensing semiconductor integrated circuit.
  • FIG. 2 is a diagrammatic view of image data that is captured by an image sensor.
  • FIG. 3 is a diagrammatic view of a stereoscopic imaging and viewing system.
  • FIG. 4 is a schematic plan view of imaging optics.
  • FIG. 5 is a schematic side elevation view of imaging optics.
  • FIG. 6 is a perspective view of a prism.
  • FIG. 7 is a perspective view of an infrared filter.
  • FIG. 8 is a cross-sectional schematic side elevation view of an image capture device.
  • FIG. 9 is a schematic plan view of an image capture device.
  • FIGS. 10A , 10 B, and 10 C are schematic front elevation views of an image capture device.
  • FIGS. 11A , 11 B, and 11 C are perspective views of an image capture device at the distal end of an endoscope.
  • FIG. 12 is a schematic partial cutaway plan view of another embodiment of imaging optics.
  • FIG. 13 is a schematic plan view of a relay lens and image sensor configuration.
  • spatially relative terms such as “beneath”, “below”, “lower”, “above”, “upper”, “underlying”, “outside”, “between”, and the like may be used to describe one element's or feature's relationship to another as illustrated in the figures.
  • the spatially relative terms are for ease of description and are intended to encompass different orientations of the device in use or operation in addition to the orientation shown in the figures. For example, if the device in the figures is turned over, elements described as “below” or “beneath” other elements or features would then appear to be oriented “above” the other elements or features. Nevertheless, the exemplary term “below” can encompass a real world orientation of above and below.
  • the device may be otherwise oriented (e.g., rotated 90 degrees or other orientations) and the spatially relative descriptors used in this description should be interpreted accordingly.
  • FIG. 1 is an illustrative, diagrammatic plan view of the active image sensing area 100 of an image sensing semiconductor integrated circuit (e.g., CCD, CMOS). Additional active circuits and packaging are omitted from the drawing for clarity.
  • the width:height dimensional aspect ratio of image sensing area 100 is approximately 5:4 (Micron Technology, Inc. MT9M131). In other embodiments, other standard aspect ratios such as 4:3 or 16:9 may be used.
  • a custom sized image sensing area may be fabricated to eliminate portions of the sensing area that are not used to capture images. Such custom sizing helps save space in small endoscope distal ends. Such custom sizing may also accommodate desired optical configurations for desired stereoscopic viewing, as described in detail below.
  • Two optical paths cause stereoscopic images to be incident on the surface of image sensing area 100 .
  • the right side image 101 is incident on one side of image sensing area 100 .
  • the left side image 102 is incident on the other side of image sensing area 100 .
  • Skilled artisans will understand that various optical train configurations may cause the right and left side images 101 , 102 to be incident on either side of image sensing area 100 .
  • the use of a single, planar substrate for sensing stereoscopic images makes alignment during construction easier and the imaging device more compact than if two physically separate image sensor chips are used.
  • the optics in the optical paths are configured to provide a minimally distorted image within a particular circular area. As shown in FIG. 1 , this particular area is represented by the dashed line circle areas 104 for the right side image 101 and 106 for the left side image 102 .
  • the optimized area is an approximately 25-degree radius from the center of the image, as represented by radius arrow 108 in FIG. 1 .
  • Other embodiments may optimize images for other fields of view, as measured on the diagonal at a viewing system, or optimize over an image area of non-circular shape. Image information outside the optimized areas 104 , 106 is usable despite being of slightly less optical quality.
  • one or more field separators and/or field stops prevent most of the right side image 101 from being incident on the side of image sensing area 100 that receives the left side image 102 , and vice versa.
  • a small amount of image cross-talk does exist, however, and so the optical paths position circle areas 104 , 106 with some space between them to avoid being in cross-talk region 110 .
  • image sensing area 100 may have a higher pixel pitch (number of pixels per unit area) since unusable image data is not being read out and sent to downstream image processing circuits.
  • a custom sized image sensing area 100 may be fabricated to match the dimensions of horizontal portion 112 .
  • the size and position of circle areas 104 , 106 are determined by design considerations described below.
  • FIG. 2 is a diagrammatic view of image data that is captured by image sensing area 100 .
  • Data 212 corresponds to data captured by horizontal portion 112 .
  • Right optimum image data in circle 204 corresponds to the optimum image data captured in circle area 104
  • left optimum image data in circle 206 corresponds to the optimum image data captured in circle area 106 .
  • Cross-talk data 210 corresponds to image data captured in cross-talk region 110 . As described below, in some cases the captured image data is stored in a memory for subsequent image processing.
  • FIG. 2 shows two illustrative display image data areas 214 , 216 .
  • the data used for displaying the right side image is in display image data area 214 .
  • the data used for displaying the left side image is in display image data area 216 .
  • display image data areas 214 , 216 are rectangular with an approximately 4:3 width:height aspect ratio, which corresponds to the width:height aspect ratio of the displays (e.g., CRTs) used to output the images in one embodiment.
  • display image data areas 214 , 216 have different aspect ratios such as, e.g., 16:9 or 1:1, which may or may not correspond to the width:height aspect ratio of the output displays, depending on the desired stereoscopic image for viewing.
  • display image data areas 214 , 216 are non-rectangular (e.g., circular, octagonal, etc.). As shown in FIG. 2 , in order to capture most of the optimum right and left image data, a small amount of non-optimum image data exists at the corners of rectangular display image data areas 214 , 216 .
  • Display image data areas 214 , 216 are defined to not include cross-talk data 210 .
  • the data in display image data areas 214 , 216 of data 212 are further processed using well-known image data processing methods.
  • FIG. 3 is a diagrammatic view of a stereoscopic imaging system.
  • Image capture subsystem 302 described in more detail below, captures stereoscopic image data as described above with reference to FIG. 1 .
  • Image processing subsystem 304 includes memory 306 and graphics processor 308 .
  • Memory 306 stores captured image data as described above with reference to FIG. 2 .
  • Right and left display image data (formatted as, e.g., S-Video, VGA, DVI, SDI and shown as representative arrows 310 , 312 ) are sent from graphics processor 308 to image display subsystem 314 .
  • image processing subsystem 304 is a personal computer
  • graphics processor 308 is a dual-head graphics processing card.
  • image processing subsystem 304 is an image processing system that is dedicated for use in a surgical operating room. Because the right and left images in the stereoscopic pair are read out of the sensor together, the scanning of the two images is inherently synchronized, and therefore the image processing subsystem 304 can be designed to operate on the images in a flow-through fashion, without requiring the storage of a full frame of image data in memory 306 . This flow-through design minimizes the latency in the presentation of the stereoscopic image to the viewer, which is important in many applications and is not possible if the scanning of the right and left images is not synchronized.
  • Image display subsystem 314 includes right display 316 and left display 318 . Optics (not shown) allow the viewer to simultaneously view displays 316 , 318 and perceive a stereoscopic image.
  • image display subsystem 314 is the surgeon's console on a advance® surgical robotic system manufactured by Intuitive Surgical, Inc. of Sunnyvale, Calif., which displays the stereo image at about an 18-inch working distance.
  • FIG. 4 is an illustrative, schematic plan view of an embodiment of imaging optics used in image capture subsystem 302 .
  • FIG. 4 shows two parallel optical lens trains 402 R and 402 L positioned side-by-side in front of an imaging integrated circuit 404 .
  • Light e.g., visible, infrared, ultraviolet
  • ray 408 R for the right side image
  • ray 408 L for the left side image
  • image sensing area 100 of imaging integrated circuit 404 as described above.
  • rays 408 R and 408 L pass from object 405 to image capture area 100 without turning.
  • the optical paths from object 405 to image capture area 100 do not include any turns other than the refraction from the lenses, filter, and windows.
  • Right lens train 402 R includes objective negative lens 406 R, positive lens 408 R positioned behind lens 406 R, and doublet 410 R positioned behind lens 408 R.
  • Left lens train 402 L includes similar counterpart lenses 406 L, 408 L, and 410 L.
  • the lenses shown in FIG. 4 are illustrative of various lens configurations that may be used in the optical trains. In one embodiment lenses 406 R, 408 R, and 410 R are held in position within a tube (not shown) so that the entire lens train can be moved for focusing, and left lens train 402 L is similarly configured.
  • FIG. 4 shows various other components of this embodiment of image capture subsystem 302 .
  • Window 412 R is shown positioned in front of lens 406 R
  • window 412 L is shown positioned in front of lens 406 L.
  • the window or windows are made from, e.g., sapphire and protect the lenses.
  • infrared (IR) filter color balancing filter 414 positioned between lenses 410 R, 410 L and image capture area 100 .
  • IR filters color balancing filter
  • FIG. 4 further shows illustrative aperture stop 416 R positioned between lens 406 R and lens 408 R.
  • aperture stop 416 L is shown positioned between lens 406 L and 408 L.
  • the aperture stop positions are illustrative of various positions. In one embodiment aperture stops 416 R, 416 L have fixed apertures, and in other embodiments the apertures may be variable.
  • FIG. 4 shows illustrative field separator 418 positioned between right and left lens trains 402 R, 402 L.
  • Field separator 418 is made of non-reflective material and helps eliminate image cross talk at image sensor area 100 . As shown, field separator 418 stops at IR filter 414 . In other embodiments, illustrated below, field separator 418 may extend closer to image sensor area 100 .
  • the tubes surrounding and holding the right and left lens trains 402 R, 402 L may function as field separator 418 , and/or field separator 418 may include other structures.
  • object 405 is a finite distance from lenses 412 R and 412 L
  • the distance between the centers of the right side and left side images on the surface 100 of imaging integrated circuit 404 is slightly more than the interpupilary distance between lenses 412 R and 412 L.
  • the stereoscopic right and left side optical paths are spaced apart generally on the scale of the image sensor chip upon which the right and left side images are incident.
  • FIG. 5 is a schematic side elevation view of another embodiment of imaging optics used in image capture subsystem 302 .
  • the configuration of the imaging optics is similar to the configuration shown in FIG. 4 , but the optical path between object 405 and image capture area 100 is turned once by reflective prism 502 .
  • Lens train 402 L is shown in FIG. 5 .
  • Lens train 402 R is directly behind and obscured by lens train 402 L in this view.
  • left side image ray 504 L passes through window 412 L, through optical train 402 L, and through IR filter 414 in a manner similar to that described for FIG. 4 .
  • left side image ray 504 L passes through side face 506 of prism 502 , is internally reflected by side face 508 , and then exits side face 510 to be incident on image capture area 100 .
  • a counterpart right side image ray similarly passes through lens train 402 R, is turned by prism 502 , and is incident on image capture area 100 .
  • Prism 502 is a single prism that turns both the right side and the left side images, as described in more detail below.
  • side face 508 acts as a single reflective plane for both the right and left side images of object 405 .
  • the image will be inverted on the sensor with respect to the image formed by the optical train shown in FIG. 4 .
  • the image can be displayed correctly on the display by inverting it in image processing subsystem 304 ( FIG. 3 ), or preferably by altering the scanning pattern of image sensor 404 to match the inverted image, which avoids the processing delay associated with performing the inversion after scanning.
  • IR filter 414 is coupled directly to side face 506 of prism 502 . In other embodiments, IR filter 414 may be at various other positions.
  • the associated integrated circuit 404 occupies a relatively large cross-sectional area compared to the cross-sectional area of the optical trains. Accordingly, the cross-sectional area of image capture subsystem 302 is relatively large so as to accommodate integrated circuit 404 and its associated hardware and electronic interface circuitry Turning the optical paths as shown in FIG. 5 allows integrated circuit 404 to be positioned so that the plane of sensor area 100 is substantially parallel to rays 408 R, 408 L passing through the optical trains. Consequently, the cross-sectional area of image capture subsystem 302 is relatively smaller.
  • Prism 502 is shown as a right isosceles triangular prism, although in other embodiments other prism configurations can be used.
  • the plane of sensor area 100 may be tilted from the substantially parallel position illustrated in FIG. 5 .
  • FIG. 5 illustrates embodiments in which object 405 is generally directly in line with the longitudinal axes of lens trains 402 R, 402 L.
  • one or more reflecting prisms or other optical components can be placed, e.g., between windows 412 R, 412 L and objective lenses 406 R, 406 L in accordance with well-known optical design principles.
  • FIG. 6 is a perspective view showing details of one embodiment of prism 502 .
  • prism 502 is a single prism that includes field separator 602 half-way between the two base faces 604 a, 604 b.
  • the right half 606 of prism 502 reflects the right side image
  • the left half 608 of prism 502 reflects the left side image.
  • Field separator 602 prevents light from the right and left side images from crossing into the sensor areas for the opposite side image.
  • Field separator 602 is a non-reflective coating.
  • the non-reflective coating is applied to the material (e.g., glass) used for one half of prism 502 , and then the material used for the other half of prism 502 is attached to the first half so that the non-reflective coating is between the two halves. Then, prism 502 is ground and polished as a single prism having a field separator sandwiched between the two halves. Since prism 502 is a single prism having a single side face 508 that is used to reflect both right and left images in the stereoscopic system, alignment operations with the underlying image capture area 100 and the right and left lens trains are made easier than if, e.g., two separate prisms were to be used.
  • the material e.g., glass
  • prism 502 may be composed of two separate parts, each of which has the field separator 602 applied, and the two parts may be aligned at the time the lenses are mounted.
  • FIG. 7 is a perspective view showing details of one embodiment of IR filter 414 .
  • IR filter 414 is a rectangular prism configured with a field separator 702 in a manner similar to the embodiment of prism 502 illustrated by FIG. 6 .
  • a separate field separator may be used between the two filters.
  • the field separator may be eliminated.
  • the IR filter or filter is glued to the end of tubes holding the lens train, as described below.
  • the IR filter or filters are glued to side face 506 of prism 502 .
  • FIG. 8 is a cross-sectional side elevation schematic view of an embodiment of image capture device 800 .
  • FIG. 9 is a schematic plan view of an embodiment of image capture device 800 .
  • Image capture device 800 functions as an endoscope for, e.g., minimally invasive surgical procedures and is part of image capture subsystem 302 .
  • an image capture integrated circuit 802 is positioned over and electrically coupled to printed circuit board 804 .
  • Image capture integrated circuit 802 and printed circuit board 804 are mounted in image capture device 800 so that they are substantially parallel to longitudinal axis 806 of image capture device 800 .
  • Cable 808 is also electrically coupled to printed circuit board 804 .
  • Printed circuit board 804 includes electronic components that support the transfer of captured image data from integrated circuit 802 via cable 808 to subsequent graphics processing operations, as described above.
  • Components of image capture integrated circuit 802 include package 810 , the semiconductor chip 812 that includes active image sensor array 814 (rows and columns of imaging pixels; the array generally defines the boundaries of image sensing area 100 ), and glass 816 positioned over image sensor array 814 .
  • Optically clear right triangular prism 818 is mounted (e.g., glued) so that one side face is flush with top surface 820 of glass 816 .
  • An adhesive with an index of refraction matched to the prism and cover glass material may be used to prevent reflections from occurring at this interface.
  • side face 822 of prism 818 acts a reflective plane that reflects light exiting the stereoscopic optical lens trains onto image sensor array 814 .
  • FIGS. 8 and 9 further show lens mount 824 that, in this illustrative embodiment, holds two lens tubes 826 R and 826 L, each containing a lens train, one (e.g., 402 R) for the right side image and one (e.g., 402 L) for the left side image, as described above.
  • the longitudinal axes 828 R and 828 L of the lens tubes are substantially parallel with each other, with longitudinal axis 806 of image capture device 800 , and with the plane of image sensor array 814 .
  • Windows 412 R, 412 L are positioned in front of lens tubes 826 , and an IR filter as described above (not shown) is positioned between the lens tubes 826 and prism 818 .
  • the lens tubes 826 R and 826 L act as field separators.
  • a black epoxy is used to fill the gap between the end of lens tubes 826 R, 826 L and the input at the IR filter or prism 818 to further act as a continuous field separator.
  • Other gaps in the field separator from the windows 412 R, 412 L to the image sensor array 814 may be similarly filled.
  • a housing surrounds the optical and electronic components of image capture device 800 .
  • the one or more windows 412 are joined with housing 830 so as to protect the internal components of image capture device 800 from an external environment such as a surgical site, chemical sterilization conditions, or the interior of an autoclave.
  • housing 830 is about 15 mm high, 15 mm wide, and 25 mm long.
  • housing 830 is about 5 mm high, 10 mm wide, and 20 mm long.
  • housing 830 is about 25 mm long and has a cylindrical cross section with an approximately 12 mm diameter, which will allow it to be compatible with the form factor of endoscopes used with the da Vinci® surgical robotic systems.
  • spaces 832 next to and under lens mount 824 may exist.
  • Other components such as components used to illuminate an object being imaged (e.g., optical fibers piping light from a remote source, a light generating source), may be placed in these spaces 832 .
  • lens mount 824 includes several small ports 902 that allow glue to be applied to keep lens tubes 826 R and 826 L in place in lens mount 824 .
  • jigs hold one assembly of integrated circuit 802 and printed circuit board 804 near another assembly of lens mount 824 and prism 818 .
  • lens mount 824 has features for prism 818 alignment, and lens mount 824 and prism 818 are glued together.
  • a UV-cured glue with a matched index of refraction as described above is placed between prism 818 and glass 816 , and then the jigs are adjusted so that prism 818 reflects the left and right images onto the surface of the underlying image sensor array in proper alignment. Then, the right and left lens tubes 826 R, 826 L are moved along longitudinal axes 828 R, 828 L within lens mount 824 so that the left and right images are properly focused on the surface of image sensor array 814 . If adjustment of the optical elements or spacing is required to match the image magnification to the required degree, this adjustment is also done at the same time.
  • glue e.g., 5-minute cure time epoxy
  • glue is applied via ports 902 to hold the lens tubes 826 R, 826 L in place.
  • any final adjustment of the image position and rotation can be made by moving the lens mount 824 and prism 818 assembly, and then UV light is applied to cure the UV-cured glue between prism 818 and glass 816 .
  • the assemblies are removed from the jigs and the right and left optical paths from the objective lens to the image sensor remain in alignment with proper focus.
  • FIGS. 10A , 10 B, and 10 C are front schematic elevation views of device 800 that show windows 412 , behind which lens tubes 826 R, 826 L are positioned.
  • FIGS. 10A-C illustrate various configurations and cross sections of housing 830 , windows 412 , and illumination ports.
  • FIG. 10A shows two D-shaped illumination ports 1002 positioned on either side of windows 412 R, 412 L. Light to illuminate the object to be imaged is output through ports 1002 .
  • FIG. 10B shows a single window 412 used for both the right and left images, and also the illumination ports 1002 having a rounded rectangular shape.
  • FIG. 10C shows circular illumination ports 1002 . The shape and position of the one or more illumination ports 1002 shown in FIGS.
  • the illumination port or ports may be positioned between the left and right optical trains.
  • the illumination source may be inside housing 830 or may be separate from housing 830 and routed via, e.g., optical fiber.
  • Housing 830 may have various cross-sectional shapes.
  • FIG. 10A shows one illustrative embodiment of housing 830 having a rounded rectangle cross section.
  • FIG. 10B illustrates housing 830 having another rounded rectangle cross section.
  • FIG. 10C illustrates housing 830 having a rounded trapezoid cross-sectional shape.
  • Other cross-sectional geometric shapes e.g., regular and irregular polygons having rounded or sharp corners, continuous curves may be used.
  • the nominal working distance for lens trains 402 R, 402 L is about 37 mm, and the depth of field is in the range from about 25-75 mm. In one embodiment, the field of view is about 60 degrees diagonally across the image area.
  • FIGS. 11A-11C are illustrative perspective views showing various embodiments of image capture device 800 mounted at the distal end of an endoscope for use during minimally invasive surgery (e.g., robot-assisted surgery).
  • the distal end is inserted, e.g., through a cannula, into a patient and is positioned to allow image capture device 800 to image a desired area in a surgical site.
  • endoscope body 1102 is shown having a cross section smaller than the cross section of image capture device 800 .
  • endoscope body 1102 is about 500 mm long and is about 5 mm in diameter or more.
  • Endoscope body 1102 contains, e.g., power and data cables for the imaging components of image capture device and optical fibers for illumination.
  • the distal end of endoscope body 1102 may be rigid, movable (discrete sections or continuum-type), or may have both a rigid and a movable portion.
  • Various movement mechanisms are known for minimally invasive surgical applications. If the distal end is movable (the one or more joints are not shown), then in one embodiment endoscope body 1102 contains mechanical control components (e.g., actuation cables).
  • Components e.g., video data interface to image processing subsystem 304 , electrical power connections, illumination connections, connections to motor actuators, etc.
  • Components associated with the features at the distal end of endoscope body are positioned at the proximal end (not shown) of endoscope body 1102 .
  • the outer covering of endoscope body 1102 protects the internal components from the surgical environment and from various sterilization methods.
  • FIG. 11B illustrates embodiments in which the cross-sectional shape of endoscope body 1102 is the same as the cross-sectional shape of image capture device 800 .
  • FIG. 11C illustrates embodiments in which image capture device is configured with optics that allow viewing at an angle (e.g., downwards), as described above.
  • image capture device 800 is shown positioned at the distal end of an endoscope, in other embodiments image capture device may be mounted on other mechanisms used during surgery (e.g., used as a vision system for robots designed to move inside body cavities).
  • the inter-pupil distance between the left and right lens trains is about 2.4 mm, based on the use of the MT9M131 sensor. In another embodiment the inter-pupil distance is about 3.3 mm, based on stereoscopic considerations to match a given optimum working distance at the surgical site with the surgeon's perceived working distance at the stereoscopic display. Inter-pupil distance of the lens trains may be constrained by, e.g., the size of the sensor array or the size of the housing. Custom sensor designs, however, allow the designer to choose any value for the stereo separation rather than being constrained by the layout of a particular commercially available image sensor chip width.
  • the ratio of the inter-pupil distance to the endoscope's working distance should be equal to the ratio of human inter-pupil distance to the perceived viewing distance in the fused display images.
  • the optimum inter-pupil distance for the left and right objective lenses is about 4.3 mm.
  • the endoscope's desired working distance is about 37 mm
  • the desired inter-pupil distance of the lenses is about 5.3 mm.
  • Embodiments of the invention are constructed (e.g., using a custom image sensor array of sufficient width) to come as close as possible to a desired inter-pupil distance of the lens trains for specified endoscope working distances and perceived viewing distances and still stay within the size constraints specified for the image sensing device.
  • FIG. 12 is a schematic partial cutaway plan view of another embodiment of imaging optics.
  • two separate prisms are used to turn the right side and left side images onto the underlying image sensor area.
  • the advantages of using a single prism are described above. Using two prisms, however, still retains some advantages such as the compact configuration of the optical path components and the underlying image sensor chip and the simplified alignment of the stereo images during assembly.
  • FIG. 12 shows lens mount 1202 .
  • Two channels 1204 R and 1204 L are aligned with lens mount 1202 's longitudinal axis.
  • channels 1204 R, 1204 L are cylindrical, and other shapes may be used.
  • Right lens tube 1206 R slides in channel 1204 R.
  • left lens tube 1206 L slides in channel 1204 L.
  • sliding the lens tubes enables the image to be focused on the image sensing area.
  • Other lens configurations, with or without sliding tubes may be used.
  • the lens tubes are fixed in position within the channels as, e.g., described above.
  • the shape of lens mount 1202 shown in FIG. 12 is illustrative of many variations.
  • Two reflecting prisms 1208 R and 1208 L are mounted within recess 1210 of lens mount 1202 such that the reflecting planes of each prism 1208 R, 1208 L are substantially coplanar.
  • the front faces of prisms 1208 R, 1208 L are glued against back face 1212 of recess 1210 so as to be precisely aligned with the optical paths through lens tubes 1206 R and 1206 L.
  • the prisms may be secured against side faces 1214 of recess 1210 .
  • Various suitable ways to mount the prisms may be used.
  • a single IR filter or two separate IR filters may be positioned between the prisms 1208 R, 1208 L and lens mount 1202 or at some other suitable location in the right and left optical paths.
  • the embodiment shown in FIG. 12 is illustrative also of one way in which one single prism that turns both right side and left side images, as described above, may be mounted.
  • the internal reflection of the inner end faces 1216 R, 1216 L of prisms 1208 R, 1208 L may be sufficient to keep light from one optical path from reaching the image sensing area for the other optical path.
  • Other ways of separating the optical fields may be used, such as placing a non-reflective coating on end faces 1216 R, 1216 L, placing a field separating piece between prisms 1208 R, 1208 L, or extending a piece of lens mount 1202 between prisms 1208 R, 1208 L.
  • FIG. 12 is illustrative of embodiments, such as those described above, in which the right side and left side images first come to focus at the surface of the image sensor.
  • Other optical geometries are possible in which an image exists prior to the surface of the image sensor.
  • a field stop may be placed at the image point to prevent optical cross talk.
  • Alignment of the assembly that includes lens mount 1202 , lens tubes 1206 R, 1206 L, and prisms 1208 R, 1208 L may be performed in a manner similar to the one described above.
  • the precise mounting geometry of lens mount 1202 allows prisms 1208 R, 1208 L to be set in position against it for subsequent alignment with the underlying image sensor area. That is, once mounted the prisms 1208 R, 1208 L lose their individual degrees of freedom and may be treated as a single prism for alignment with the image sensor. Once alignment is complete, the lens and prism assembly is fixed to the optical sensor.
  • FIG. 13 is a schematic plan view of a relay lens system in accordance with the invention.
  • Light for the right side image is transmitted via right lens train 1302 R
  • light for the left side image is transmitted via left lens train 1302 L.
  • Each lens train 1302 R, 1302 L is shown with two illustrative rod lenses 1304 in addition to other lenses.
  • the right and left side images are focused on the surface of image sensing array 1306 (e.g., a 16:9 (1920 ⁇ 1080 pixels) width:height aspect ratio image sensor).
  • the lens trains 1302 R, 1302 L are spaced apart to provide a desired interpupilary distance within the constraints of the width of the image sensing chip that receives the right side and left side images.
  • a field stop may be positioned at any of these foci.
  • Other optical components such as aperture stops and various filter types, may be positioned at various points in the lens trains.
  • the optics shown in FIG. 13 are illustrative of many embodiments in which a single image sensor area is positioned at a location other than at the distal end 1308 of an endoscope.
  • the single image sensing array 1306 may be placed at a proximate end 1310 of an endoscope, outside the patient's body.
  • Such placement offers the simplicity of the single-chip being placed to receive both right side and left side images without the need for any turns (reflections) in the optical path from the object to the image sensor, and may also allow the distal end 1308 of the endoscope to have a relatively smaller size because the image sensing chip 1306 and its supporting circuitry are not located at the distal end 1308 inside the patient.

Abstract

Two side-by-side optical paths transmit stereoscopic right side and left side images onto the surface of a single image sensing chip. The single image sensing chip may be placed at various orientations with respect to the lens trains in the optical paths. In some embodiments a single prism is used to turn the light for both the right side and left side images onto the single image sensing chip. In other embodiments one prism is used to turn the light for the right side image and another prism is used to turn the light for the left side image, and the reflective surfaces of the two prisms are substantially coplanar such that the right side and left side images are incident on the single image sensor chip.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • Not applicable.
  • BACKGROUND
  • 1. Field of Invention
  • The invention pertains to stereoscopic imaging systems, and more particularly to stereoscopic image capture endoscopes.
  • 2. Background Art
  • The field of minimally invasive surgery (e.g., laparoscopic surgery) requires increasingly smaller, increasingly mobile stereoscopic imaging systems. Stereoscopic endoscopes are typically mounted at the distal end of rigid shafts that extend through a cannula so as to image a surgical site during, e.g., robot-assisted surgery.
  • To get acceptable stereoscopic imaging without causing viewer fatigue or eyestrain, the images of the target object viewed by the two imaging systems should match in at least the following alignment parameters: (1) image location along the horizontal axis; (2) image location along the vertical axis; (3) image rotation; (4) image scale; (5) geometric distortion; (5) focus at the image center; (6) focal shift along the horizontal axis; and (7) focal shift along the vertical axis. The tolerable errors in the matching between the two images in a stereoscopic pair depend to some extent upon the display and viewer, but in general are much more stringent requirements than exist for monocular viewing. In addition, except for the image location, mismatches in the other parameters are difficult to correct for in image post-processing without introducing imaging artifacts.
  • While these parameters are all affected to some degree by the positions of the optical elements in the imaging system, they are also affected by the accuracy of the mounting of the two image sensors conventionally used in a stereoscopic endoscope with respect to each other. Taking one of the sensors as a reference, the position of a second, separate, sensor has six degrees of freedom in its mounting: three of translation and three of rotation. Errors in two of the translation axes between the sensors affect the relative horizontal and vertical positions of the viewed images, while errors in the third translation axis, the axis perpendicular to the sensor surface, affects both the image scale (if the objective is not telecentric in image space) and focus. Errors in rotation between the two image sensors, around the axis perpendicular to the sensor surface, directly affect image rotation and cannot always be corrected by alignment of the optics, while rotation errors about the other two axes affect the focal plane shifts across the imaging field.
  • In three dimensions, a rigid body (e.g., an optical image sensor chip) has six degrees of freedom: moving up and down (heaving), moving left and right (swaying), moving forward and backward (surging), tilting up and down (pitching), turning left and right (yawing), and tilting side to side (rolling). With two separate image sensors there are a total of 12 degrees of freedom that must be controlled when mounting the two sensors to the optical train. For example, if two physically separate sensors are used, then each sensor must be aligned with its respective optical train and additional image processing (e.g., to compensate for rotation) is required in order to align the captured left and right images with each other to present the stereoscopic view to a person viewing the images.
  • SUMMARY
  • In aspects of the invention, two side-by-side optical paths transmit stereoscopic right side and left side images onto the surface of a single image sensing chip. The single image sensing chip may be placed at various orientations (e.g., perpendicular, parallel) with respect to the lens trains in the optical paths.
  • In some aspects of the invention, a single prism is used to turn the light for both the right side and left side images onto the single image sensing chip.
  • In other aspects of the invention, one prism is used to turn the light for the right side image and another prism is used to turn the light for the left side image, and the reflective surfaces of the two prisms are substantially coplanar such that the right side and left side images are incident on the single image sensor chip.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is an illustrative, diagrammatic plan view of the active image sensing area of an image sensing semiconductor integrated circuit.
  • FIG. 2 is a diagrammatic view of image data that is captured by an image sensor.
  • FIG. 3 is a diagrammatic view of a stereoscopic imaging and viewing system.
  • FIG. 4 is a schematic plan view of imaging optics.
  • FIG. 5 is a schematic side elevation view of imaging optics.
  • FIG. 6 is a perspective view of a prism.
  • FIG. 7 is a perspective view of an infrared filter.
  • FIG. 8 is a cross-sectional schematic side elevation view of an image capture device.
  • FIG. 9 is a schematic plan view of an image capture device.
  • FIGS. 10A, 10B, and 10C are schematic front elevation views of an image capture device.
  • FIGS. 11A, 11B, and 11C are perspective views of an image capture device at the distal end of an endoscope.
  • FIG. 12 is a schematic partial cutaway plan view of another embodiment of imaging optics.
  • FIG. 13 is a schematic plan view of a relay lens and image sensor configuration.
  • DETAILED DESCRIPTION
  • In the following description, reference is made to the accompanying drawings, which illustrate embodiments of the invention. It is understood that other embodiments may be used, and mechanical, compositional, structural, electrical, and operational changes may be made, without departing from the spirit and scope of this disclosure. The following detailed description is not to be taken in a limiting sense. The scope of the embodiments of the invention is defined only by the claims of the issued patent.
  • The terminology used in this description is to describe particular embodiments only and is not intended to limit the invention. Spatially relative terms, such as “beneath”, “below”, “lower”, “above”, “upper”, “underlying”, “outside”, “between”, and the like may be used to describe one element's or feature's relationship to another as illustrated in the figures. The spatially relative terms are for ease of description and are intended to encompass different orientations of the device in use or operation in addition to the orientation shown in the figures. For example, if the device in the figures is turned over, elements described as “below” or “beneath” other elements or features would then appear to be oriented “above” the other elements or features. Nevertheless, the exemplary term “below” can encompass a real world orientation of above and below. The device may be otherwise oriented (e.g., rotated 90 degrees or other orientations) and the spatially relative descriptors used in this description should be interpreted accordingly.
  • As used in this description, the singular forms “a”, “an”, and “the” are intended to include the plural forms as well, unless the context indicates otherwise. The terms “comprises” and/or “comprising”, “includes” and/or “including”, and the like specify the presence of stated features, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, steps, operations, elements, components, and/or groups.
  • Numerous specific details are described to provide a thorough understanding of the present invention. In certain instances, however, well known or conventional details are not described in order to avoid obscuring the description of the present invention. References to one or an embodiment in this disclosure are not necessarily to the same embodiment, and such references mean at least one. Like numbers in the drawings refer to the same or similar elements.
  • Consider an image sensor positioned flat with reference to the ground (two horizontal axes at 90 degrees with respect to each other, and a vertical axis). If this single, larger image sensor is used to record the images from the two side by side stereoscopic optical trains, only 6 degrees of freedom exist in mounting the sensor with reference to the optical trains as opposed to 12 if two separate sensors are used. Furthermore, of these 6 degrees of freedom, the three translational degrees (heaving, swaying, surging) cause imaging errors that affect both the right and the left side images in an equal way, which maintains good stereoscopic viewing. Rotation around the axis perpendicular to the sensor surface (yawing) requires cropping off a small amount of the field along the edges, but both images rotate by exactly the same amount. In addition, although the entire image appears slightly rotated, the two views do not experience any differential rotation, to which the viewer is much more sensitive. Tilt of the sensor around an axis oriented horizontally (pitching) causes tilt of the focal plane, but again to the same degree in both images. The final axis, which is tilt of the sensor about the other horizontal axis (rolling), does affect the two images differentially, but an optical adjustment may be made to correct for this by moving one of the optical trains back and forth to match the focus point of the other optical train.
  • FIG. 1 is an illustrative, diagrammatic plan view of the active image sensing area 100 of an image sensing semiconductor integrated circuit (e.g., CCD, CMOS). Additional active circuits and packaging are omitted from the drawing for clarity. In one embodiment, the width:height dimensional aspect ratio of image sensing area 100 is approximately 5:4 (Micron Technology, Inc. MT9M131). In other embodiments, other standard aspect ratios such as 4:3 or 16:9 may be used. In still other embodiments, a custom sized image sensing area may be fabricated to eliminate portions of the sensing area that are not used to capture images. Such custom sizing helps save space in small endoscope distal ends. Such custom sizing may also accommodate desired optical configurations for desired stereoscopic viewing, as described in detail below.
  • Two optical paths, described below, cause stereoscopic images to be incident on the surface of image sensing area 100. The right side image 101 is incident on one side of image sensing area 100. Likewise, the left side image 102 is incident on the other side of image sensing area 100. Skilled artisans will understand that various optical train configurations may cause the right and left side images 101,102 to be incident on either side of image sensing area 100. As described above and in more detail below, the use of a single, planar substrate for sensing stereoscopic images makes alignment during construction easier and the imaging device more compact than if two physically separate image sensor chips are used.
  • With a single, planar substrate some image processing may be required to move the right and left images up or down with respect to each other so as to correct for position errors of the optical trains with respect to the sensor surface, but processing for rotation or scaling is not necessary, and the focal planes of the two images are automatically in alignment. By using a single sensor to record both images of a stereoscopic pair, the problems of alignment in manufacture are greatly simplified, and the resulting imaging system is simpler and more resistant to misalignment due to shocks, thermal cycles, and other effects occurring over the image capture system's (e.g., endoscope's) lifetime.
  • The optics in the optical paths are configured to provide a minimally distorted image within a particular circular area. As shown in FIG. 1, this particular area is represented by the dashed line circle areas 104 for the right side image 101 and 106 for the left side image 102. In one embodiment the optimized area is an approximately 25-degree radius from the center of the image, as represented by radius arrow 108 in FIG. 1. Other embodiments may optimize images for other fields of view, as measured on the diagonal at a viewing system, or optimize over an image area of non-circular shape. Image information outside the optimized areas 104,106 is usable despite being of slightly less optical quality.
  • Also shown in FIG. 1 is an image cross-talk region 110 between the right and left side images 101,102 that are incident on image sensing area 100. As described in more detail below, one or more field separators and/or field stops prevent most of the right side image 101 from being incident on the side of image sensing area 100 that receives the left side image 102, and vice versa. A small amount of image cross-talk does exist, however, and so the optical paths position circle areas 104,106 with some space between them to avoid being in cross-talk region 110.
  • In the embodiment depicted in FIG. 1, only a horizontal portion 112 that includes circle areas 104 and 106, shown between the two heavy dashed lines, of image sensing area 100 is read during each image frame capture. Areas generally above and below circle areas 104,106 are not read out since the image data in these areas will not be used in the image display. This limited read out of data helps speed image capture and processing time. Consequently, frame rate may be increased above the rate required if all of image sensing area 100 data were read. One embodiment uses a 24-30 fps progressive scan. Other embodiments may use higher rate, e.g., 60 Hz, progressive or interlaced scanning. Alternatively, or in combination with the increased frame rate, image sensing area 100 may have a higher pixel pitch (number of pixels per unit area) since unusable image data is not being read out and sent to downstream image processing circuits. A custom sized image sensing area 100 may be fabricated to match the dimensions of horizontal portion 112. The size and position of circle areas 104,106 are determined by design considerations described below.
  • FIG. 2 is a diagrammatic view of image data that is captured by image sensing area 100. Data 212 corresponds to data captured by horizontal portion 112. Right optimum image data in circle 204 corresponds to the optimum image data captured in circle area 104, and left optimum image data in circle 206 corresponds to the optimum image data captured in circle area 106. Cross-talk data 210 corresponds to image data captured in cross-talk region 110. As described below, in some cases the captured image data is stored in a memory for subsequent image processing.
  • FIG. 2 shows two illustrative display image data areas 214,216. The data used for displaying the right side image is in display image data area 214. Similarly, the data used for displaying the left side image is in display image data area 216. As shown in FIG. 2, display image data areas 214,216 are rectangular with an approximately 4:3 width:height aspect ratio, which corresponds to the width:height aspect ratio of the displays (e.g., CRTs) used to output the images in one embodiment. In other embodiments, display image data areas 214,216 have different aspect ratios such as, e.g., 16:9 or 1:1, which may or may not correspond to the width:height aspect ratio of the output displays, depending on the desired stereoscopic image for viewing. In still other embodiments, display image data areas 214,216 are non-rectangular (e.g., circular, octagonal, etc.). As shown in FIG. 2, in order to capture most of the optimum right and left image data, a small amount of non-optimum image data exists at the corners of rectangular display image data areas 214,216. Display image data areas 214,216 are defined to not include cross-talk data 210. The data in display image data areas 214,216 of data 212 are further processed using well-known image data processing methods.
  • FIG. 3 is a diagrammatic view of a stereoscopic imaging system. Image capture subsystem 302, described in more detail below, captures stereoscopic image data as described above with reference to FIG. 1. Image processing subsystem 304 includes memory 306 and graphics processor 308. Memory 306 stores captured image data as described above with reference to FIG. 2. Right and left display image data (formatted as, e.g., S-Video, VGA, DVI, SDI and shown as representative arrows 310,312) are sent from graphics processor 308 to image display subsystem 314. In one embodiment image processing subsystem 304 is a personal computer, and graphics processor 308 is a dual-head graphics processing card. In another embodiment, image processing subsystem 304 is an image processing system that is dedicated for use in a surgical operating room. Because the right and left images in the stereoscopic pair are read out of the sensor together, the scanning of the two images is inherently synchronized, and therefore the image processing subsystem 304 can be designed to operate on the images in a flow-through fashion, without requiring the storage of a full frame of image data in memory 306. This flow-through design minimizes the latency in the presentation of the stereoscopic image to the viewer, which is important in many applications and is not possible if the scanning of the right and left images is not synchronized.
  • Image display subsystem 314 includes right display 316 and left display 318. Optics (not shown) allow the viewer to simultaneously view displays 316,318 and perceive a stereoscopic image. In one embodiment image display subsystem 314 is the surgeon's console on a advance® surgical robotic system manufactured by Intuitive Surgical, Inc. of Sunnyvale, Calif., which displays the stereo image at about an 18-inch working distance.
  • FIG. 4 is an illustrative, schematic plan view of an embodiment of imaging optics used in image capture subsystem 302. FIG. 4 shows two parallel optical lens trains 402R and 402L positioned side-by-side in front of an imaging integrated circuit 404. Light (e.g., visible, infrared, ultraviolet), shown in FIG. 4 as ray 408R for the right side image and as ray 408L for the left side image, from object 405 passes through the two lens trains 402R,402L and is incident on image sensing area 100 of imaging integrated circuit 404, as described above. As shown in FIG. 4, in one embodiment rays 408R and 408L pass from object 405 to image capture area 100 without turning. The optical paths from object 405 to image capture area 100 do not include any turns other than the refraction from the lenses, filter, and windows.
  • Right lens train 402R includes objective negative lens 406R, positive lens 408R positioned behind lens 406R, and doublet 410R positioned behind lens 408R. Left lens train 402L includes similar counterpart lenses 406L, 408L, and 410L. The lenses shown in FIG. 4 are illustrative of various lens configurations that may be used in the optical trains. In one embodiment lenses 406R, 408R, and 410R are held in position within a tube (not shown) so that the entire lens train can be moved for focusing, and left lens train 402L is similarly configured.
  • In addition to the lenses, FIG. 4 shows various other components of this embodiment of image capture subsystem 302. Window 412R is shown positioned in front of lens 406R, and window 412L is shown positioned in front of lens 406L. In other embodiments a single window may be used. The window or windows are made from, e.g., sapphire and protect the lenses. Also shown in FIG. 4 is infrared (IR) filter (color balancing filter) 414 positioned between lenses 410R,410L and image capture area 100. In other embodiments separate IR filters may be used for each lens train. The IR filter or filters may be at other positions along the optical path between object 406 and image sensor area 100. FIG. 4 further shows illustrative aperture stop 416R positioned between lens 406R and lens 408R. Likewise, aperture stop 416L is shown positioned between lens 406L and 408L. The aperture stop positions are illustrative of various positions. In one embodiment aperture stops 416R,416L have fixed apertures, and in other embodiments the apertures may be variable. Finally, FIG. 4 shows illustrative field separator 418 positioned between right and left lens trains 402R,402L. Field separator 418 is made of non-reflective material and helps eliminate image cross talk at image sensor area 100. As shown, field separator 418 stops at IR filter 414. In other embodiments, illustrated below, field separator 418 may extend closer to image sensor area 100. The tubes surrounding and holding the right and left lens trains 402R,402L may function as field separator 418, and/or field separator 418 may include other structures.
  • Since object 405 is a finite distance from lenses 412R and 412L, persons skilled in the art will understand that the distance between the centers of the right side and left side images on the surface 100 of imaging integrated circuit 404 is slightly more than the interpupilary distance between lenses 412R and 412L. Thus it can be seen that in embodiments of the invention the stereoscopic right and left side optical paths are spaced apart generally on the scale of the image sensor chip upon which the right and left side images are incident.
  • FIG. 5 is a schematic side elevation view of another embodiment of imaging optics used in image capture subsystem 302. As shown in FIG. 5, the configuration of the imaging optics is similar to the configuration shown in FIG. 4, but the optical path between object 405 and image capture area 100 is turned once by reflective prism 502. Lens train 402L is shown in FIG. 5. Lens train 402R is directly behind and obscured by lens train 402L in this view.
  • As shown in FIG. 5, left side image ray 504L passes through window 412L, through optical train 402L, and through IR filter 414 in a manner similar to that described for FIG. 4. After exiting IR filter 414, however, left side image ray 504L passes through side face 506 of prism 502, is internally reflected by side face 508, and then exits side face 510 to be incident on image capture area 100. A counterpart right side image ray (not shown) similarly passes through lens train 402R, is turned by prism 502, and is incident on image capture area 100. Prism 502 is a single prism that turns both the right side and the left side images, as described in more detail below. Thus side face 508 acts as a single reflective plane for both the right and left side images of object 405. As a consequence of this reflection, the image will be inverted on the sensor with respect to the image formed by the optical train shown in FIG. 4. The image can be displayed correctly on the display by inverting it in image processing subsystem 304 (FIG. 3), or preferably by altering the scanning pattern of image sensor 404 to match the inverted image, which avoids the processing delay associated with performing the inversion after scanning.
  • In one embodiment, illustrated by FIG. 5, IR filter 414 is coupled directly to side face 506 of prism 502. In other embodiments, IR filter 414 may be at various other positions.
  • Since the plane of image capture area 100 is shown as being substantially perpendicular to rays 408R,408L in FIG. 4, the associated integrated circuit 404 occupies a relatively large cross-sectional area compared to the cross-sectional area of the optical trains. Accordingly, the cross-sectional area of image capture subsystem 302 is relatively large so as to accommodate integrated circuit 404 and its associated hardware and electronic interface circuitry Turning the optical paths as shown in FIG. 5 allows integrated circuit 404 to be positioned so that the plane of sensor area 100 is substantially parallel to rays 408R,408L passing through the optical trains. Consequently, the cross-sectional area of image capture subsystem 302 is relatively smaller. Prism 502 is shown as a right isosceles triangular prism, although in other embodiments other prism configurations can be used. In addition, the plane of sensor area 100 may be tilted from the substantially parallel position illustrated in FIG. 5.
  • FIG. 5 illustrates embodiments in which object 405 is generally directly in line with the longitudinal axes of lens trains 402R,402L. In other embodiments, in order to give the image capture system an upward, downward, or sideward viewing angle, one or more reflecting prisms or other optical components can be placed, e.g., between windows 412R,412L and objective lenses 406R,406L in accordance with well-known optical design principles.
  • FIG. 6 is a perspective view showing details of one embodiment of prism 502. As shown in FIG. 6, prism 502 is a single prism that includes field separator 602 half-way between the two base faces 604 a, 604 b. The right half 606 of prism 502 reflects the right side image, and the left half 608 of prism 502 reflects the left side image. Field separator 602 prevents light from the right and left side images from crossing into the sensor areas for the opposite side image. Field separator 602 is a non-reflective coating. In one embodiment, the non-reflective coating is applied to the material (e.g., glass) used for one half of prism 502, and then the material used for the other half of prism 502 is attached to the first half so that the non-reflective coating is between the two halves. Then, prism 502 is ground and polished as a single prism having a field separator sandwiched between the two halves. Since prism 502 is a single prism having a single side face 508 that is used to reflect both right and left images in the stereoscopic system, alignment operations with the underlying image capture area 100 and the right and left lens trains are made easier than if, e.g., two separate prisms were to be used. It can be seen that side face 508 internally reflects both right and left images onto the underlying image capture area 100. In an alternative embodiment, prism 502 may be composed of two separate parts, each of which has the field separator 602 applied, and the two parts may be aligned at the time the lenses are mounted.
  • FIG. 7 is a perspective view showing details of one embodiment of IR filter 414. In the embodiment illustrated by FIG. 7, IR filter 414 is a rectangular prism configured with a field separator 702 in a manner similar to the embodiment of prism 502 illustrated by FIG. 6. In other embodiments in which separate IR filters are used for each right and left side image, a separate field separator may be used between the two filters. In still other embodiments in which the IR filter is very thin, the field separator may be eliminated. In some embodiments the IR filter or filter is glued to the end of tubes holding the lens train, as described below. In some embodiments the IR filter or filters are glued to side face 506 of prism 502.
  • FIG. 8 is a cross-sectional side elevation schematic view of an embodiment of image capture device 800. FIG. 9 is a schematic plan view of an embodiment of image capture device 800. Image capture device 800 functions as an endoscope for, e.g., minimally invasive surgical procedures and is part of image capture subsystem 302.
  • As shown in FIGS. 8 and 9, an image capture integrated circuit 802 is positioned over and electrically coupled to printed circuit board 804. Image capture integrated circuit 802 and printed circuit board 804 are mounted in image capture device 800 so that they are substantially parallel to longitudinal axis 806 of image capture device 800. Cable 808 is also electrically coupled to printed circuit board 804. Printed circuit board 804 includes electronic components that support the transfer of captured image data from integrated circuit 802 via cable 808 to subsequent graphics processing operations, as described above. Components of image capture integrated circuit 802 include package 810, the semiconductor chip 812 that includes active image sensor array 814 (rows and columns of imaging pixels; the array generally defines the boundaries of image sensing area 100), and glass 816 positioned over image sensor array 814.
  • Optically clear right triangular prism 818 is mounted (e.g., glued) so that one side face is flush with top surface 820 of glass 816. An adhesive with an index of refraction matched to the prism and cover glass material may be used to prevent reflections from occurring at this interface. As described above with reference to FIGS. 5 and 6, side face 822 of prism 818 acts a reflective plane that reflects light exiting the stereoscopic optical lens trains onto image sensor array 814.
  • Since there is a slight gap between the top surface 820 of glass 816 and image sensor array 814, however, a small amount of right and left image light may cross over on the surface of sensor array 814 because the field separator in prism 818 does not extend to the surface of image sensor array 814. This cross over is illustrated as cross talk area 110 in FIG. 1. In other embodiments the field separator may be positioned closer to the image sensor array, thus further minimizing or eliminating the image cross over.
  • FIGS. 8 and 9 further show lens mount 824 that, in this illustrative embodiment, holds two lens tubes 826R and 826L, each containing a lens train, one (e.g., 402R) for the right side image and one (e.g., 402L) for the left side image, as described above. As shown in the FIGS. 8 and 9 embodiments, the longitudinal axes 828R and 828L of the lens tubes are substantially parallel with each other, with longitudinal axis 806 of image capture device 800, and with the plane of image sensor array 814. Windows 412R,412L are positioned in front of lens tubes 826, and an IR filter as described above (not shown) is positioned between the lens tubes 826 and prism 818. As described above, the lens tubes 826R and 826L act as field separators. In one embodiment a black epoxy is used to fill the gap between the end of lens tubes 826R,826L and the input at the IR filter or prism 818 to further act as a continuous field separator. Other gaps in the field separator from the windows 412R,412L to the image sensor array 814 may be similarly filled.
  • A housing, represented as dashed line 830, surrounds the optical and electronic components of image capture device 800. The one or more windows 412 are joined with housing 830 so as to protect the internal components of image capture device 800 from an external environment such as a surgical site, chemical sterilization conditions, or the interior of an autoclave. In one embodiment housing 830 is about 15 mm high, 15 mm wide, and 25 mm long. In another embodiment, housing 830 is about 5 mm high, 10 mm wide, and 20 mm long. In yet another embodiment, housing 830 is about 25 mm long and has a cylindrical cross section with an approximately 12 mm diameter, which will allow it to be compatible with the form factor of endoscopes used with the da Vinci® surgical robotic systems.
  • Depending on the shape of housing 830, spaces 832 next to and under lens mount 824 may exist. Other components, such as components used to illuminate an object being imaged (e.g., optical fibers piping light from a remote source, a light generating source), may be placed in these spaces 832.
  • As shown in FIG. 9, lens mount 824 includes several small ports 902 that allow glue to be applied to keep lens tubes 826R and 826L in place in lens mount 824. During one construction embodiment jigs hold one assembly of integrated circuit 802 and printed circuit board 804 near another assembly of lens mount 824 and prism 818. In one embodiment lens mount 824 has features for prism 818 alignment, and lens mount 824 and prism 818 are glued together.
  • To align and attach the two assemblies, a UV-cured glue with a matched index of refraction as described above is placed between prism 818 and glass 816, and then the jigs are adjusted so that prism 818 reflects the left and right images onto the surface of the underlying image sensor array in proper alignment. Then, the right and left lens tubes 826R,826L are moved along longitudinal axes 828R,828L within lens mount 824 so that the left and right images are properly focused on the surface of image sensor array 814. If adjustment of the optical elements or spacing is required to match the image magnification to the required degree, this adjustment is also done at the same time. When the left and right images are properly adjusted, glue (e.g., 5-minute cure time epoxy) is applied via ports 902 to hold the lens tubes 826R,826L in place. Once the glue is cured, any final adjustment of the image position and rotation can be made by moving the lens mount 824 and prism 818 assembly, and then UV light is applied to cure the UV-cured glue between prism 818 and glass 816. When the glues have cured, the assemblies are removed from the jigs and the right and left optical paths from the objective lens to the image sensor remain in alignment with proper focus.
  • FIGS. 10A, 10B, and 10C are front schematic elevation views of device 800 that show windows 412, behind which lens tubes 826R,826L are positioned. FIGS. 10A-C illustrate various configurations and cross sections of housing 830, windows 412, and illumination ports. For example, FIG. 10A shows two D-shaped illumination ports 1002 positioned on either side of windows 412R,412L. Light to illuminate the object to be imaged is output through ports 1002. FIG. 10B shows a single window 412 used for both the right and left images, and also the illumination ports 1002 having a rounded rectangular shape. FIG. 10C shows circular illumination ports 1002. The shape and position of the one or more illumination ports 1002 shown in FIGS. 10A-C is illustrative of various configurations and arrangements of one or more illuminating structures. For example, the illumination port or ports may be positioned between the left and right optical trains. The illumination source may be inside housing 830 or may be separate from housing 830 and routed via, e.g., optical fiber.
  • Housing 830 may have various cross-sectional shapes. For example, FIG. 10A shows one illustrative embodiment of housing 830 having a rounded rectangle cross section. FIG. 10B illustrates housing 830 having another rounded rectangle cross section. FIG. 10C illustrates housing 830 having a rounded trapezoid cross-sectional shape. Other cross-sectional geometric shapes (e.g., regular and irregular polygons having rounded or sharp corners, continuous curves) may be used.
  • In one embodiment, the nominal working distance for lens trains 402R,402L is about 37 mm, and the depth of field is in the range from about 25-75 mm. In one embodiment, the field of view is about 60 degrees diagonally across the image area.
  • FIGS. 11A-11C are illustrative perspective views showing various embodiments of image capture device 800 mounted at the distal end of an endoscope for use during minimally invasive surgery (e.g., robot-assisted surgery). The distal end is inserted, e.g., through a cannula, into a patient and is positioned to allow image capture device 800 to image a desired area in a surgical site. In FIG. 11A, endoscope body 1102 is shown having a cross section smaller than the cross section of image capture device 800. In one embodiment endoscope body 1102 is about 500 mm long and is about 5 mm in diameter or more. Endoscope body 1102 contains, e.g., power and data cables for the imaging components of image capture device and optical fibers for illumination. The distal end of endoscope body 1102 may be rigid, movable (discrete sections or continuum-type), or may have both a rigid and a movable portion. Various movement mechanisms are known for minimally invasive surgical applications. If the distal end is movable (the one or more joints are not shown), then in one embodiment endoscope body 1102 contains mechanical control components (e.g., actuation cables). Components (e.g., video data interface to image processing subsystem 304, electrical power connections, illumination connections, connections to motor actuators, etc.) associated with the features at the distal end of endoscope body are positioned at the proximal end (not shown) of endoscope body 1102. In one embodiment the outer covering of endoscope body 1102 protects the internal components from the surgical environment and from various sterilization methods.
  • FIG. 11B illustrates embodiments in which the cross-sectional shape of endoscope body 1102 is the same as the cross-sectional shape of image capture device 800. FIG. 11C illustrates embodiments in which image capture device is configured with optics that allow viewing at an angle (e.g., downwards), as described above. Although image capture device 800 is shown positioned at the distal end of an endoscope, in other embodiments image capture device may be mounted on other mechanisms used during surgery (e.g., used as a vision system for robots designed to move inside body cavities).
  • Referring to FIGS. 4, 5, 8, and 9, in one embodiment the inter-pupil distance between the left and right lens trains is about 2.4 mm, based on the use of the MT9M131 sensor. In another embodiment the inter-pupil distance is about 3.3 mm, based on stereoscopic considerations to match a given optimum working distance at the surgical site with the surgeon's perceived working distance at the stereoscopic display. Inter-pupil distance of the lens trains may be constrained by, e.g., the size of the sensor array or the size of the housing. Custom sensor designs, however, allow the designer to choose any value for the stereo separation rather than being constrained by the layout of a particular commercially available image sensor chip width.
  • Persons of skill in the art will understand that for optimum stereoscopic display, the ratio of the inter-pupil distance to the endoscope's working distance should be equal to the ratio of human inter-pupil distance to the perceived viewing distance in the fused display images. For example, for an average human inter-pupil distance of about 65 mm, a desired perceived viewing distance of about 18 inches (about 457 mm), and a desired 30 mm working distance of the endoscope, then the optimum inter-pupil distance for the left and right objective lenses is about 4.3 mm. In another exemplary embodiment, if the endoscope's desired working distance is about 37 mm, then the desired inter-pupil distance of the lenses is about 5.3 mm. Embodiments of the invention are constructed (e.g., using a custom image sensor array of sufficient width) to come as close as possible to a desired inter-pupil distance of the lens trains for specified endoscope working distances and perceived viewing distances and still stay within the size constraints specified for the image sensing device.
  • FIG. 12 is a schematic partial cutaway plan view of another embodiment of imaging optics. In the embodiments illustrated by FIG. 12, two separate prisms are used to turn the right side and left side images onto the underlying image sensor area. The advantages of using a single prism are described above. Using two prisms, however, still retains some advantages such as the compact configuration of the optical path components and the underlying image sensor chip and the simplified alignment of the stereo images during assembly.
  • FIG. 12 shows lens mount 1202. Two channels 1204R and 1204L are aligned with lens mount 1202's longitudinal axis. In one instance channels 1204R,1204L are cylindrical, and other shapes may be used. Right lens tube 1206R slides in channel 1204R. Likewise, left lens tube 1206L slides in channel 1204L. As described above, sliding the lens tubes enables the image to be focused on the image sensing area. Other lens configurations, with or without sliding tubes, may be used. After the right and left side images are focused, the lens tubes are fixed in position within the channels as, e.g., described above. The shape of lens mount 1202 shown in FIG. 12 is illustrative of many variations.
  • Two reflecting prisms 1208R and 1208L are mounted within recess 1210 of lens mount 1202 such that the reflecting planes of each prism 1208R,1208L are substantially coplanar. In one embodiment the front faces of prisms 1208R,1208L are glued against back face 1212 of recess 1210 so as to be precisely aligned with the optical paths through lens tubes 1206R and 1206L. In addition, or alternatively, the prisms may be secured against side faces 1214 of recess 1210. Various suitable ways to mount the prisms may be used. Also, a single IR filter or two separate IR filters—one each for the right and left sides (not shown)—may be positioned between the prisms 1208R,1208L and lens mount 1202 or at some other suitable location in the right and left optical paths. The embodiment shown in FIG. 12 is illustrative also of one way in which one single prism that turns both right side and left side images, as described above, may be mounted.
  • Depending on the geometry of certain embodiments, the internal reflection of the inner end faces 1216R,1216L of prisms 1208R,1208L may be sufficient to keep light from one optical path from reaching the image sensing area for the other optical path. Other ways of separating the optical fields may be used, such as placing a non-reflective coating on end faces 1216R,1216L, placing a field separating piece between prisms 1208R,1208L, or extending a piece of lens mount 1202 between prisms 1208R,1208L.
  • FIG. 12 is illustrative of embodiments, such as those described above, in which the right side and left side images first come to focus at the surface of the image sensor. Other optical geometries are possible in which an image exists prior to the surface of the image sensor. In such embodiments, a field stop may be placed at the image point to prevent optical cross talk.
  • Alignment of the assembly that includes lens mount 1202, lens tubes 1206R,1206L, and prisms 1208R,1208L may be performed in a manner similar to the one described above. The precise mounting geometry of lens mount 1202 allows prisms 1208R,1208L to be set in position against it for subsequent alignment with the underlying image sensor area. That is, once mounted the prisms 1208R,1208L lose their individual degrees of freedom and may be treated as a single prism for alignment with the image sensor. Once alignment is complete, the lens and prism assembly is fixed to the optical sensor.
  • As described above, various optical configurations may be used. FIG. 13 is a schematic plan view of a relay lens system in accordance with the invention. Light for the right side image is transmitted via right lens train 1302R, and light for the left side image is transmitted via left lens train 1302L. Each lens train 1302R,1302L is shown with two illustrative rod lenses 1304 in addition to other lenses. The right and left side images are focused on the surface of image sensing array 1306 (e.g., a 16:9 (1920×1080 pixels) width:height aspect ratio image sensor). As described above, the lens trains 1302R,1302L are spaced apart to provide a desired interpupilary distance within the constraints of the width of the image sensing chip that receives the right side and left side images.
  • Since the right side and left side images come to a focus at various positions in the lens trains 1302R,1302L, a field stop may be positioned at any of these foci. Other optical components, such as aperture stops and various filter types, may be positioned at various points in the lens trains.
  • The optics shown in FIG. 13 are illustrative of many embodiments in which a single image sensor area is positioned at a location other than at the distal end 1308 of an endoscope. For example, the single image sensing array 1306 may be placed at a proximate end 1310 of an endoscope, outside the patient's body. Such placement offers the simplicity of the single-chip being placed to receive both right side and left side images without the need for any turns (reflections) in the optical path from the object to the image sensor, and may also allow the distal end 1308 of the endoscope to have a relatively smaller size because the image sensing chip 1306 and its supporting circuitry are not located at the distal end 1308 inside the patient.

Claims (2)

1. A stereoscopic endoscope comprising:
a first lens train;
a second lens train;
a prism; and
an image sensor array;
wherein light for a right side image of an object passes through the first lens train and through a first portion of the prism to be incident on a right side image area of the image sensor array; and
wherein light for a left side image of the object passes through the second lens train and through a second portion of the prism to be incident on a left side image area of the image sensor array.
2-17. (canceled)
US14/341,098 2006-12-21 2014-07-25 Stereoscopic endoscope Abandoned US20140336460A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US14/341,098 US20140336460A1 (en) 2006-12-21 2014-07-25 Stereoscopic endoscope
US16/051,592 US11382496B2 (en) 2006-12-21 2018-08-01 Stereoscopic endoscope
US17/751,227 US20220280031A1 (en) 2006-12-21 2022-05-23 Stereoscopic endoscope

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/614,661 US8814779B2 (en) 2006-12-21 2006-12-21 Stereoscopic endoscope
US14/341,098 US20140336460A1 (en) 2006-12-21 2014-07-25 Stereoscopic endoscope

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/614,661 Continuation US8814779B2 (en) 2006-12-21 2006-12-21 Stereoscopic endoscope

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/051,592 Continuation US11382496B2 (en) 2006-12-21 2018-08-01 Stereoscopic endoscope

Publications (1)

Publication Number Publication Date
US20140336460A1 true US20140336460A1 (en) 2014-11-13

Family

ID=39468826

Family Applications (4)

Application Number Title Priority Date Filing Date
US11/614,661 Active 2031-01-04 US8814779B2 (en) 2006-12-21 2006-12-21 Stereoscopic endoscope
US14/341,098 Abandoned US20140336460A1 (en) 2006-12-21 2014-07-25 Stereoscopic endoscope
US16/051,592 Active 2027-04-24 US11382496B2 (en) 2006-12-21 2018-08-01 Stereoscopic endoscope
US17/751,227 Pending US20220280031A1 (en) 2006-12-21 2022-05-23 Stereoscopic endoscope

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/614,661 Active 2031-01-04 US8814779B2 (en) 2006-12-21 2006-12-21 Stereoscopic endoscope

Family Applications After (2)

Application Number Title Priority Date Filing Date
US16/051,592 Active 2027-04-24 US11382496B2 (en) 2006-12-21 2018-08-01 Stereoscopic endoscope
US17/751,227 Pending US20220280031A1 (en) 2006-12-21 2022-05-23 Stereoscopic endoscope

Country Status (2)

Country Link
US (4) US8814779B2 (en)
WO (1) WO2008079578A2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9005113B2 (en) 2006-12-21 2015-04-14 Intuitive Surgical Operations, Inc. Hermetically sealed endoscope
CN106667418A (en) * 2016-11-22 2017-05-17 珠海维尔康生物科技有限公司 Endoscope
US11375884B2 (en) * 2011-09-27 2022-07-05 California Institute Of Technology Multi-angle rear-viewing endoscope and method of operation thereof
US11382496B2 (en) 2006-12-21 2022-07-12 Intuitive Surgical Operations, Inc. Stereoscopic endoscope

Families Citing this family (72)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9789608B2 (en) 2006-06-29 2017-10-17 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
US9962066B2 (en) * 2005-12-30 2018-05-08 Intuitive Surgical Operations, Inc. Methods and apparatus to shape flexible entry guides for minimally invasive surgery
US7930065B2 (en) 2005-12-30 2011-04-19 Intuitive Surgical Operations, Inc. Robotic surgery system including position sensors using fiber bragg gratings
US20080064931A1 (en) * 2006-06-13 2008-03-13 Intuitive Surgical, Inc. Minimally invasive surgical illumination
US8784435B2 (en) 2006-06-13 2014-07-22 Intuitive Surgical Operations, Inc. Surgical system entry guide
US10258425B2 (en) 2008-06-27 2019-04-16 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide
US10008017B2 (en) 2006-06-29 2018-06-26 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US9718190B2 (en) 2006-06-29 2017-08-01 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US9469034B2 (en) 2007-06-13 2016-10-18 Intuitive Surgical Operations, Inc. Method and system for switching modes of a robotic system
WO2010047463A1 (en) * 2008-10-24 2010-04-29 Meerecompany Laparoscope and setting method thereof
JP5321009B2 (en) * 2008-11-21 2013-10-23 ソニー株式会社 Image signal processing apparatus, image signal processing method, and image projection apparatus
CN101588513B (en) * 2009-01-07 2011-05-18 深圳市掌网立体时代视讯技术有限公司 Device and method of stereo camera
US9706903B2 (en) 2009-06-18 2017-07-18 Endochoice, Inc. Multiple viewing elements endoscope system with modular imaging units
US9492927B2 (en) 2009-08-15 2016-11-15 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
JP5834177B2 (en) * 2010-02-17 2015-12-16 パナソニックIpマネジメント株式会社 Stereoscopic image display system and stereoscopic glasses
US9486189B2 (en) 2010-12-02 2016-11-08 Hitachi Aloka Medical, Ltd. Assembly for use with surgery system
CN102256151B (en) * 2011-07-14 2014-03-12 深圳市掌网立体时代视讯技术有限公司 Double-optical path single-sensor synthesis module and three-dimensional imaging device
US8684914B2 (en) 2011-08-12 2014-04-01 Intuitive Surgical Operations, Inc. Image capture unit and an imaging pipeline with enhanced color performance in a surgical instrument and method
US8764633B2 (en) 2011-08-12 2014-07-01 Intuitive Surgical Operations, Inc. Feature differentiation image capture unit and method in a surgical instrument
US8734328B2 (en) 2011-08-12 2014-05-27 Intuitive Surgical Operations, Inc. Increased resolution and dynamic range image capture unit in a surgical instrument and method
US8672838B2 (en) 2011-08-12 2014-03-18 Intuitive Surgical Operations, Inc. Image capture unit in a surgical instrument
JP6258201B2 (en) * 2011-08-12 2018-01-10 インテュイティブ サージカル オペレーションズ, インコーポレイテッド Surgical instrument image capture device
US8784301B2 (en) * 2011-08-12 2014-07-22 Intuitive Surgical Operations, Inc. Image capture unit and method with an extended depth of field
EP2747410B1 (en) * 2011-08-16 2018-08-29 Fujifilm Corporation Imaging apparatus
US9452276B2 (en) 2011-10-14 2016-09-27 Intuitive Surgical Operations, Inc. Catheter with removable vision probe
US9387048B2 (en) 2011-10-14 2016-07-12 Intuitive Surgical Operations, Inc. Catheter sensor systems
US20130303944A1 (en) 2012-05-14 2013-11-14 Intuitive Surgical Operations, Inc. Off-axis electromagnetic sensor
JP5973707B2 (en) * 2011-10-14 2016-08-23 オリンパス株式会社 3D endoscope device
US10238837B2 (en) 2011-10-14 2019-03-26 Intuitive Surgical Operations, Inc. Catheters with control modes for interchangeable probes
KR101699597B1 (en) * 2012-01-13 2017-01-24 스테로페스 테크놀로지스, 엘엘씨 Single optical path anamorphic stereoscopic imager
JP5931528B2 (en) * 2012-03-21 2016-06-08 オリンパス株式会社 Surgical video system and method of operating surgical video system
US9642606B2 (en) 2012-06-27 2017-05-09 Camplex, Inc. Surgical visualization system
US9492065B2 (en) 2012-06-27 2016-11-15 Camplex, Inc. Surgical retractor with video cameras
JP2014073143A (en) * 2012-10-02 2014-04-24 Canon Inc Endoscope system
CN102871638B (en) * 2012-10-16 2014-11-05 广州市盛光微电子有限公司 Medical short-distance imaging method, system and probe
US9167160B2 (en) * 2012-11-14 2015-10-20 Karl Storz Imaging, Inc. Image capture stabilization
CN103070660A (en) * 2013-01-18 2013-05-01 浙江大学 Three-dimensional electronic endoscope image pick-up device
US10507066B2 (en) 2013-02-15 2019-12-17 Intuitive Surgical Operations, Inc. Providing information of tools by filtering image areas adjacent to or on displayed images of the tools
WO2014189969A1 (en) 2013-05-21 2014-11-27 Camplex, Inc. Surgical visualization systems
CN105592767B (en) * 2013-06-28 2018-04-03 恩多巧爱思股份有限公司 More observation element endoscopic systems with modularization imaging unit
EP3046458B1 (en) 2013-09-20 2020-10-21 Camplex, Inc. Surgical visualization systems
JP6521982B2 (en) 2013-09-20 2019-05-29 キャンプレックス インコーポレイテッド Surgical visualization system and display
CN104717415B (en) * 2013-12-12 2019-03-01 华为技术有限公司 A kind of photographic device
JP5889495B2 (en) * 2014-02-14 2016-03-22 オリンパス株式会社 Endoscope system
US9983384B2 (en) 2014-04-20 2018-05-29 Lenny Lipton Stereoscopic lens for digital cameras
WO2015198981A1 (en) * 2014-06-27 2015-12-30 オリンパス株式会社 Endoscopy system
WO2016090336A1 (en) 2014-12-05 2016-06-09 Camplex, Inc. Surgical visualization systems and displays
EP3138468A4 (en) * 2014-12-15 2018-01-17 Olympus Corporation Image capturing system
US10932882B2 (en) * 2015-03-09 2021-03-02 Synaptive Medical (Barbados) Inc. Surgical camera system with automatic alternation between two depths of field
US11154378B2 (en) 2015-03-25 2021-10-26 Camplex, Inc. Surgical visualization systems and displays
EP3079131B1 (en) * 2015-04-08 2017-03-01 Axis AB Monitoring camera
US9357116B1 (en) * 2015-07-22 2016-05-31 Ic Real Tech, Inc. Isolating opposing lenses from each other for an assembly that produces concurrent non-overlapping image circles on a common image sensor
WO2017091704A1 (en) 2015-11-25 2017-06-01 Camplex, Inc. Surgical visualization systems and displays
CN105635720A (en) * 2016-01-19 2016-06-01 浙江大学 Stereo vision camera with double-lens single sensor
US11141071B2 (en) * 2016-06-16 2021-10-12 Stryker European Operations Limited Closed cavity adjustable sensor mount systems and methods
CN106361255B (en) * 2016-11-10 2020-07-14 微创(上海)医疗机器人有限公司 3D electronic endoscope
CA3057162C (en) 2016-12-01 2024-02-06 Synaptive Medical (Barbados) Inc. A camera system for providing images with simultaneous high resolution and large depth of field
WO2018143218A1 (en) * 2017-02-02 2018-08-09 オリンパス株式会社 Endoscope
US10477190B2 (en) * 2017-03-14 2019-11-12 Karl Storz Imaging, Inc. Constant horizon 3D imaging system and related method
JP6422621B1 (en) * 2017-05-01 2018-11-14 オリンパス株式会社 Endoscopic imaging device
US10918455B2 (en) 2017-05-08 2021-02-16 Camplex, Inc. Variable light source
US10524643B2 (en) 2017-11-06 2020-01-07 Karl Storz Endovision, Inc. Image sensor module with turning prism
DE102018105845A1 (en) * 2018-03-14 2019-09-19 Olympus Winter & Ibe Gmbh Holder for an optical system of an endoscope and method for producing a holder for an optical system of an endoscope
CN109259717B (en) * 2018-08-27 2020-08-14 彭波 Stereoscopic endoscope and endoscope measuring method
CN110393499B (en) * 2018-08-31 2021-12-07 上海微创医疗机器人(集团)股份有限公司 Electronic endoscope and electronic endoscope system
US11278360B2 (en) * 2018-11-16 2022-03-22 Globus Medical, Inc. End-effectors for surgical robotic systems having sealed optical components
CN110680264A (en) * 2019-11-08 2020-01-14 中国科学院长春光学精密机械与物理研究所 3D optical endoscope system based on dual-optical-path design
US20230056943A1 (en) * 2019-12-13 2023-02-23 Dinesh Vyas Stapler apparatus and methods for use
US20220008068A1 (en) * 2019-12-13 2022-01-13 Dinesh Vyas Stapler apparatus and methods for use
US11925347B2 (en) 2019-12-13 2024-03-12 Dinesh Vyas Stapler apparatus and methods for use
CN217960050U (en) * 2022-02-24 2022-12-06 深圳迈瑞生物医疗电子股份有限公司 3D electronic endoscope and camera system thereof
CN116784772A (en) * 2023-08-23 2023-09-22 福建福特科光电股份有限公司 3D hard tube endoscope

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4682219A (en) * 1985-08-16 1987-07-21 Fuji Photo Optical Co., Ltd. Endoscope
US5237446A (en) * 1987-04-30 1993-08-17 Olympus Optical Co., Ltd. Optical low-pass filter
US5557454A (en) * 1992-12-25 1996-09-17 Olympus Optical Co., Ltd. Stereoscopic endoscope
US6144762A (en) * 1998-02-23 2000-11-07 Olympus America Inc. Stereo video microscope
US20030125608A1 (en) * 1999-11-19 2003-07-03 Olympus Optical Co., Ltd. Endoscope apparatus
US20030233024A1 (en) * 2002-06-14 2003-12-18 Fuji Photo Optical Co., Ltd. Electronic endoscope for stereoscopic endoscope system
US7170677B1 (en) * 2002-01-25 2007-01-30 Everest Vit Stereo-measurement borescope with 3-D viewing
US8144409B2 (en) * 2007-09-07 2012-03-27 Hon Hai Precision Industry Co., Ltd. Prism system and method for producing the same

Family Cites Families (191)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3499107A (en) 1954-03-11 1970-03-03 Sheldon Edward E Light transfer devices using light conducting members of multilayered construction and photoelectric means
US3520587A (en) * 1967-03-29 1970-07-14 Olympus Optical Co Stereoscopic endoscope
US3534729A (en) 1967-07-07 1970-10-20 Olympus Optical Co Elongated stereooptical system
US3788303A (en) 1972-01-24 1974-01-29 American Cystoscope Makers Inc Orthogonally deflectable endoscope
DE3214615C2 (en) 1981-04-21 1985-04-25 Kabushiki Kaisha Medos Kenkyusho, Tokio/Tokyo Curvature control arrangement for an endoscope
JPS58141135A (en) 1981-10-20 1983-08-22 富士写真フイルム株式会社 Image transmitting system of endoscope using solid image sensor
US4491865A (en) * 1982-09-29 1985-01-01 Welch Allyn, Inc. Image sensor assembly
JPS60104915A (en) * 1983-11-11 1985-06-10 Fuji Photo Optical Co Ltd Endoscope
JPS60137342A (en) 1983-12-27 1985-07-20 オリンパス光学工業株式会社 Electronic scope
US4692608A (en) * 1984-05-18 1987-09-08 Fuji Photo Optical Company, Ltd. Compact optical imaging system
JPS6150546A (en) 1984-08-20 1986-03-12 富士写真光機株式会社 Endoscope
JPH0644105B2 (en) 1985-01-14 1994-06-08 オリンパス光学工業株式会社 Endoscope
US4779613A (en) 1986-03-13 1988-10-25 Olympus Optical Co., Ltd. Endoscope with means for preventing an observing optical system from being fogged
US4832003A (en) * 1986-09-12 1989-05-23 Olympus Optical Co., Ltd. Electronic endoscope tip
US4918521A (en) 1987-01-20 1990-04-17 Olympus Optical Co., Ltd. Solid state imaging apparatus
US4873572A (en) * 1987-02-27 1989-10-10 Olympus Optical Co., Ltd. Electronic endoscope apparatus
US4867138A (en) 1987-05-13 1989-09-19 Olympus Optical Co., Ltd. Rigid electronic endoscope
JPH045126Y2 (en) 1987-09-03 1992-02-14
US5096292A (en) 1987-09-08 1992-03-17 Olympus Optical Co., Ltd. Borescope apparatus
US5051823A (en) * 1988-01-28 1991-09-24 Fuji Optical Systems, Inc. Dental instrument including laser device and electronic video dental camera
JPH0772092B2 (en) 1988-02-10 1995-08-02 日本特殊陶業株式会社 Low temperature firing substrate
US5419313A (en) 1988-10-27 1995-05-30 Rosemarie Lemke Microwave sterilizable and reusable endoscope
US5124838A (en) 1989-01-31 1992-06-23 Precision Optics Corporation Optical coupler
US4878485A (en) 1989-02-03 1989-11-07 Adair Edwin Lloyd Rigid video endoscope with heat sterilizable sheath
USRE33854E (en) 1989-02-03 1992-03-24 sterilizable sheathpe with .[.heat.].
US4924853A (en) * 1989-05-22 1990-05-15 Medical Dimensions, Inc. Stereoscopic medical viewing device
DE3921233A1 (en) * 1989-06-28 1991-02-14 Storz Karl Gmbh & Co ENDOSCOPE WITH A VIDEO DEVICE AT THE DISTAL END
JP2991299B2 (en) 1989-08-04 1999-12-20 株式会社東芝 Endoscope device
JP3081628B2 (en) 1989-10-30 2000-08-28 オリンパス光学工業株式会社 Electronic endoscope
US5430475A (en) 1990-06-29 1995-07-04 Olympus Optical Co., Ltd. Electronic endoscope apparatus having micro array on photoelectric conversion surface
US5212595A (en) 1990-12-21 1993-05-18 Mp Video, Inc. Sterilizable focusing lens assembly
US5122650A (en) * 1991-04-18 1992-06-16 Mckinley Optics, Inc. Stereo video endoscope objective lens system
US5191203A (en) * 1991-04-18 1993-03-02 Mckinley Optics, Inc. Stereo video endoscope objective lens system
US5188094A (en) 1991-09-30 1993-02-23 Adair Edwin Lloyd Heat sterilizable electronic video endoscope
JP3583777B2 (en) 1992-01-21 2004-11-04 エス・アール・アイ・インターナシヨナル Teleoperator system and telepresence method
US5299559A (en) 1992-03-13 1994-04-05 Acuson Corporation Endoscope with overload protective device
DE4211547C2 (en) 1992-04-06 1994-08-11 Henke Sass Wolf Gmbh Protective cover for the distal end of endoscopes
US5305121A (en) 1992-06-08 1994-04-19 Origin Medsystems, Inc. Stereoscopic endoscope system
JP3257640B2 (en) * 1992-06-09 2002-02-18 オリンパス光学工業株式会社 Stereoscopic endoscope device
US5609561A (en) 1992-06-09 1997-03-11 Olympus Optical Co., Ltd Electronic type endoscope in which image pickup unit is dismounted to execute disinfection/sterilization processing
US5381784A (en) 1992-09-30 1995-01-17 Adair; Edwin L. Stereoscopic endoscope
US5603687A (en) * 1992-10-28 1997-02-18 Oktas General Partnership Asymmetric stereo-optic endoscope
US6450950B2 (en) * 1992-11-12 2002-09-17 Karl Storz Gmbh & Co. Kg Endoscope having stereo-lateral-view optics
US5587736A (en) 1993-02-16 1996-12-24 Envision Medical Corporation Sterilizable CCD video camera
US5588948A (en) * 1993-02-17 1996-12-31 Olympus Optical Co. Ltd. Stereoscopic endoscope
DE4305376C1 (en) 1993-02-22 1994-09-29 Wolf Gmbh Richard Medical instrument shaft
US5349137A (en) 1993-05-17 1994-09-20 W. L. Gore & Associates, Inc. Sterilizable cable assemblies
DE9309545U1 (en) 1993-06-26 1993-08-19 Wolf Gmbh Richard Endoscopic instrument
US5949483A (en) 1994-01-28 1999-09-07 California Institute Of Technology Active pixel sensor array with multiresolution readout
US5841126A (en) 1994-01-28 1998-11-24 California Institute Of Technology CMOS active pixel sensor type imaging system on a chip
US6166768A (en) 1994-01-28 2000-12-26 California Institute Of Technology Active pixel sensor array with simple floating gate pixels
WO1995024857A1 (en) 1994-03-15 1995-09-21 Hibbard Erich M N Autoclavable rigid endoscope
US5743846A (en) * 1994-03-17 1998-04-28 Olympus Optical Co., Ltd. Stereoscopic endoscope objective lens system having a plurality of front lens groups and one common rear lens group
US5454827A (en) 1994-05-24 1995-10-03 Aust; Gilbert M. Surgical instrument
JPH0829701A (en) * 1994-07-18 1996-02-02 Olympus Optical Co Ltd Stereoscopic viewing endoscope system
US5940126A (en) * 1994-10-25 1999-08-17 Kabushiki Kaisha Toshiba Multiple image video camera apparatus
US5673147A (en) * 1995-04-18 1997-09-30 Mckinley Optics, Inc. Stereo video endoscope objective lens systems
US6606113B2 (en) * 1995-05-24 2003-08-12 Olympus Optical Co., Ltd. Stereoscopic endocsope system and TV imaging system for endoscope
US5864359A (en) * 1995-05-30 1999-01-26 Smith & Nephew, Inc. Stereoscopic autofocusing based on comparing the left and right eye images
US5835133A (en) * 1996-01-23 1998-11-10 Silicon Graphics, Inc. Optical system for single camera stereo video
US5944656A (en) 1996-01-31 1999-08-31 Three E Laboratories, Iec. Endoscope
WO1997027798A1 (en) 1996-02-01 1997-08-07 Heartport, Inc. Stereoscopic endoscope
US6436107B1 (en) 1996-02-20 2002-08-20 Computer Motion, Inc. Method and apparatus for performing minimally invasive surgical procedures
US6139490A (en) * 1996-02-22 2000-10-31 Precision Optics Corporation Stereoscopic endoscope with virtual reality viewing
US5868664A (en) 1996-02-23 1999-02-09 Envision Medical Corporation Electrically isolated sterilizable endoscopic video camera head
US6104426A (en) * 1996-03-23 2000-08-15 Street; Graham S. B. Stereo-endoscope
US5810713A (en) 1996-07-10 1998-09-22 Valquest Medical, Inc. Autoclavable endoscope
US6328691B1 (en) 1996-07-26 2001-12-11 Karl Storz Gmbh & Co. Kg Endoscope with at least one glued and additionally welded end window
US6228468B1 (en) 1996-07-26 2001-05-08 Paul L. Hickman High density ceramic BGA package and method for making same
DE19743431B4 (en) 1997-10-01 2011-02-17 Karl Storz Gmbh & Co. Kg Endoscope with composite window
JP3771988B2 (en) * 1997-03-12 2006-05-10 オリンパス株式会社 Measuring endoscope device
JPH10258034A (en) 1997-03-19 1998-09-29 Olympus Optical Co Ltd Photographing apparatus for endoscope
US5990469A (en) * 1997-04-02 1999-11-23 Gentex Corporation Control circuit for image array sensors
JP4093503B2 (en) 1997-06-13 2008-06-04 フジノン株式会社 Stereoscopic endoscope
US6211904B1 (en) 1997-09-11 2001-04-03 Edwin L. Adair Surgical devices incorporating reduced area imaging devices
US6043839A (en) 1997-10-06 2000-03-28 Adair; Edwin L. Reduced area imaging devices
US5929901A (en) 1997-10-06 1999-07-27 Adair; Edwin L. Reduced area imaging devices incorporated within surgical instruments
US6424369B1 (en) 1997-10-06 2002-07-23 Edwin L. Adair Hand-held computers incorporating reduced area imaging devices
US6452626B1 (en) 1997-10-06 2002-09-17 Edwin L. Adair Communication devices incorporating reduced area imaging devices
US6310642B1 (en) 1997-11-24 2001-10-30 Micro-Medical Devices, Inc. Reduced area imaging devices incorporated within surgical instruments
US6982742B2 (en) 1997-10-06 2006-01-03 Adair Edwin L Hand-held computers incorporating reduced area imaging devices
US5986693A (en) 1997-10-06 1999-11-16 Adair; Edwin L. Reduced area imaging devices incorporated within surgical instruments
US20110034769A1 (en) 1997-10-06 2011-02-10 Micro-Imaging Solutions Llc Reduced area imaging device incorporated within wireless endoscopic devices
US7030904B2 (en) 1997-10-06 2006-04-18 Micro-Medical Devices, Inc. Reduced area imaging device incorporated within wireless endoscopic devices
US6982740B2 (en) 1997-11-24 2006-01-03 Micro-Medical Devices, Inc. Reduced area imaging devices utilizing selected charge integration periods
US6191809B1 (en) * 1998-01-15 2001-02-20 Vista Medical Technologies, Inc. Method and apparatus for aligning stereo images
US6721008B2 (en) 1998-01-22 2004-04-13 Eastman Kodak Company Integrated CMOS active pixel digital camera
DE19804234C1 (en) 1998-02-04 1999-11-04 Storz Karl Gmbh & Co Endoscope, especially video endoscope
US6149582A (en) 1998-05-29 2000-11-21 Fuji Photo Optical Co., Ltd. Front end structure of stereoscopic endoscope
JP3526531B2 (en) * 1998-06-19 2004-05-17 富士写真光機株式会社 Advanced structure of stereoscopic endoscope
US6419626B1 (en) 1998-08-12 2002-07-16 Inbae Yoon Surgical instrument endoscope with CMOS image sensor and physical parameter sensor
US6080101A (en) 1998-08-26 2000-06-27 Olympus Optical Co. Ltd. Endoscope video camera head which can be autoclaved
DE19941320B4 (en) 1998-09-01 2005-07-21 Olympus Optical Co., Ltd. Endoscope suitable for autoclave sterilization
US8527094B2 (en) 1998-11-20 2013-09-03 Intuitive Surgical Operations, Inc. Multi-user medical robotic system for collaboration or training in minimally invasive surgical procedures
US6720988B1 (en) * 1998-12-08 2004-04-13 Intuitive Surgical, Inc. Stereo imaging system and method for use in telerobotic systems
JP2000193893A (en) 1998-12-28 2000-07-14 Suzuki Motor Corp Bending device of insertion tube for inspection
US7116352B2 (en) * 1999-02-25 2006-10-03 Visionsense Ltd. Capsule
US8248457B2 (en) * 1999-02-25 2012-08-21 Visionsense, Ltd. Optical device
US7683926B2 (en) * 1999-02-25 2010-03-23 Visionsense Ltd. Optical device
JP3742532B2 (en) 1999-06-08 2006-02-08 オリンパス株式会社 Endoscope
US6547722B1 (en) 1999-07-13 2003-04-15 Olympus Optical Co., Ltd. Endoscope having resistance to high-temperature and high-pressure steam
JP2001108916A (en) * 1999-10-08 2001-04-20 Olympus Optical Co Ltd Solid mirror optical system
US6572536B1 (en) 1999-11-05 2003-06-03 Visionary Biomedical, Inc. Autoclavable flexible fiberscope
US6817975B1 (en) * 2000-01-14 2004-11-16 Intuitive Surgical, Inc. Endoscope
JP3297033B2 (en) 2000-02-02 2002-07-02 オリンパス光学工業株式会社 Endoscope
IL135571A0 (en) * 2000-04-10 2001-05-20 Doron Adler Minimal invasive surgery imaging system
US6692430B2 (en) * 2000-04-10 2004-02-17 C2Cure Inc. Intra vascular imaging apparatus
JP2002077945A (en) * 2000-06-07 2002-03-15 Canon Inc Picture recorder, imaging apparatus, imaging system, method for processing signal, method for controlling recording and storage medium
JP2002065577A (en) 2000-08-28 2002-03-05 Olympus Optical Co Ltd Endoscope
US6614595B2 (en) * 2001-02-16 2003-09-02 Olympus Optical Co., Ltd. Stereo endoscope
JP4632577B2 (en) * 2001-05-30 2011-02-16 オリンパス株式会社 Measuring endoscope device
JP4610799B2 (en) * 2001-06-25 2011-01-12 オリンパス株式会社 Stereoscopic observation system and endoscope apparatus
US6558316B2 (en) 2001-06-27 2003-05-06 Olympus Optical Co., Ltd. Endoscope optical system including composition having durability to a sterilization treatment
US6817974B2 (en) 2001-06-29 2004-11-16 Intuitive Surgical, Inc. Surgical tool having positively positionable tendon-actuated multi-disk wrist joint
US6692431B2 (en) 2001-09-07 2004-02-17 Smith & Nephew, Inc. Endoscopic system with a solid-state light source
US6648817B2 (en) * 2001-11-15 2003-11-18 Endactive, Inc. Apparatus and method for stereo viewing in variable direction-of-view endoscopy
US6891266B2 (en) 2002-02-14 2005-05-10 Mia-Com RF transition for an area array package
JP2005525896A (en) * 2002-05-16 2005-09-02 シー2キュア インコーポレイティド Small camera head
US7083752B2 (en) 2002-05-20 2006-08-01 Eastman Kodak Company Cellulose acetate films prepared by coating methods
US6898022B2 (en) * 2002-06-20 2005-05-24 Olympus Corporation Stereo optical system pair for stereo endoscope system
US6932760B1 (en) 2002-11-21 2005-08-23 Stryker Corporation Autoclavable coupler for endoscopic camera system
US6955644B2 (en) 2003-03-06 2005-10-18 Precision Optics Corporation Autoclavable endoscope
JP3668480B2 (en) 2003-03-06 2005-07-06 オリンパス株式会社 Imaging device
US7428997B2 (en) 2003-07-29 2008-09-30 Microvision, Inc. Method and apparatus for illuminating a field-of-view and capturing an image
JP4418202B2 (en) * 2003-10-06 2010-02-17 オリンパス株式会社 Endoscope
DE102005018047A1 (en) 2004-04-19 2005-11-17 Acmi Corp., Southborough Electronic video camera for endoscope, has assembly of lens cell, lens cell holder, image sensor and motorized focusing system that are covered within cavity of enclosure
CN2728025Y (en) * 2004-07-22 2005-09-21 富士康(昆山)电脑接插件有限公司 Electric connector
US6997868B1 (en) 2004-07-27 2006-02-14 Martin Uram Autoclavable endoscope
US8602971B2 (en) 2004-09-24 2013-12-10 Vivid Medical. Inc. Opto-Electronic illumination and vision module for endoscopy
US8480566B2 (en) * 2004-09-24 2013-07-09 Vivid Medical, Inc. Solid state illumination for endoscopy
EP1646249A1 (en) 2004-10-08 2006-04-12 Dialog Semiconductor GmbH Single chip stereo image pick-up system with dual array design
DE112005002972B4 (en) 2004-12-13 2019-05-16 Gyrus ACMI, Inc. (n.d.Ges.d. Staates Delaware) Hermetic endoscope assembly
US8971597B2 (en) 2005-05-16 2015-03-03 Intuitive Surgical Operations, Inc. Efficient vision and kinematic data fusion for robotic surgical instruments and other applications
US9492240B2 (en) 2009-06-16 2016-11-15 Intuitive Surgical Operations, Inc. Virtual measurement tool for minimally invasive surgery
JP4987257B2 (en) * 2005-06-29 2012-07-25 オリンパス株式会社 Endoscope
DE102005045729A1 (en) * 2005-09-23 2007-03-29 Karl Storz Gmbh & Co. Kg Illumination system for endoscopic examinations
US7819859B2 (en) 2005-12-20 2010-10-26 Intuitive Surgical Operations, Inc. Control system for reducing internally generated frictional and inertial resistance to manual positioning of a surgical manipulator
US9962066B2 (en) 2005-12-30 2018-05-08 Intuitive Surgical Operations, Inc. Methods and apparatus to shape flexible entry guides for minimally invasive surgery
US7835823B2 (en) 2006-01-05 2010-11-16 Intuitive Surgical Operations, Inc. Method for tracking and reporting usage events to determine when preventive maintenance is due for a medical robotic system
TW200744518A (en) * 2006-01-06 2007-12-16 Olympus Medical Systems Corp Medical system conducted percutaneous or using naturally ocurring body orifice
US8219178B2 (en) 2007-02-16 2012-07-10 Catholic Healthcare West Method and system for performing invasive medical procedures using a surgical robot
JP2007260066A (en) 2006-03-28 2007-10-11 Pentax Corp Endoscope apparatus
US20080064931A1 (en) 2006-06-13 2008-03-13 Intuitive Surgical, Inc. Minimally invasive surgical illumination
US8814779B2 (en) 2006-12-21 2014-08-26 Intuitive Surgical Operations, Inc. Stereoscopic endoscope
US8556807B2 (en) 2006-12-21 2013-10-15 Intuitive Surgical Operations, Inc. Hermetically sealed distal sensor endoscope
US8013350B2 (en) 2007-02-05 2011-09-06 Panasonic Corporation Optical device and method for manufacturing optical device, and camera module and endoscope module equipped with optical device
US20090076329A1 (en) 2007-09-17 2009-03-19 Wei Su Disposable Stereoscopic Endoscope System
US8081380B2 (en) * 2007-10-10 2011-12-20 Mckinley Harry R Stereoscopic zoom endoscope
US9118850B2 (en) 2007-11-27 2015-08-25 Capso Vision, Inc. Camera system with multiple pixel arrays on a chip
US8633975B2 (en) 2008-01-16 2014-01-21 Karl Storz Imaging, Inc. Network based endoscopic surgical system
US8808164B2 (en) 2008-03-28 2014-08-19 Intuitive Surgical Operations, Inc. Controlling a robotic surgical tool with a display monitor
US8155479B2 (en) 2008-03-28 2012-04-10 Intuitive Surgical Operations Inc. Automated panning and digital zooming for robotic surgical systems
US8723922B2 (en) 2008-09-25 2014-05-13 Sony Corporation Single camera device and method for 3D video imaging using a refracting lens
US9621825B2 (en) 2008-11-25 2017-04-11 Capsovision Inc Camera system with multiple pixel arrays on a chip
US8594841B2 (en) 2008-12-31 2013-11-26 Intuitive Surgical Operations, Inc. Visual force feedback in a minimally invasive surgical procedure
US8374723B2 (en) 2008-12-31 2013-02-12 Intuitive Surgical Operations, Inc. Obtaining force information in a minimally invasive surgical procedure
EP2409286B1 (en) 2009-03-20 2018-09-05 The Johns Hopkins University Method and system for quantifying technical skill
WO2010129324A2 (en) 2009-05-08 2010-11-11 Boston Scientific Scimed, Inc. Endoscope with distal tip having encased optical components and display orientation capabilities
US9155592B2 (en) 2009-06-16 2015-10-13 Intuitive Surgical Operations, Inc. Virtual measurement tool for minimally invasive surgery
WO2011014687A2 (en) 2009-07-31 2011-02-03 Inneroptic Technology, Inc. Dual-tube stereoscope
US8888789B2 (en) 2009-09-23 2014-11-18 Intuitive Surgical Operations, Inc. Curved cannula surgical system control
US20110071541A1 (en) 2009-09-23 2011-03-24 Intuitive Surgical, Inc. Curved cannula
US8551115B2 (en) 2009-09-23 2013-10-08 Intuitive Surgical Operations, Inc. Curved cannula instrument
US8465476B2 (en) 2009-09-23 2013-06-18 Intuitive Surgical Operations, Inc. Cannula mounting fixture
KR20120101450A (en) 2009-11-13 2012-09-13 더 캘리포니아 인스티튜트 오브 테크놀로지 Stereo imaging miniature endoscope with single imaging chip and conjugated multi-bandpass filters
US20140253684A1 (en) 2010-09-10 2014-09-11 The Johns Hopkins University Visualization of registered subsurface anatomy
US9486189B2 (en) 2010-12-02 2016-11-08 Hitachi Aloka Medical, Ltd. Assembly for use with surgery system
DE102010053881A1 (en) 2010-12-09 2012-06-14 Schölly Fiberoptic GmbH Endoscope and method for recording at least one stereoscopic image with an endoscope
JP6113666B2 (en) 2011-02-15 2017-04-12 インテュイティブ サージカル オペレーションズ, インコーポレイテッド Indicator of knife position in a staple or vascular seal device
US9044138B2 (en) 2011-10-21 2015-06-02 Viking Systems, Inc. Steerable electronic stereoscopic endoscope
US10136954B2 (en) 2012-06-21 2018-11-27 Globus Medical, Inc. Surgical tool systems and method
US10231791B2 (en) 2012-06-21 2019-03-19 Globus Medical, Inc. Infrared signal based position recognition system for use with a robot-assisted surgery
EP3957229A1 (en) 2012-07-13 2022-02-23 Steris Instrument Management Services, Inc. Stereo endoscope system
US9572626B2 (en) 2013-02-15 2017-02-21 Intuitive Surgical Operations, Inc. Actuated cannula seal
WO2014144947A1 (en) 2013-03-15 2014-09-18 Olive Medical Corporation Super resolution and color motion artifact correction in a pulsed color imaging system
CA2906821A1 (en) 2013-03-15 2014-09-18 Olive Medical Corporation Scope sensing in a light controlled environment
US10188411B2 (en) 2013-04-16 2019-01-29 Calcula Technologies, Inc. Everting balloon for medical devices
US10219864B2 (en) 2013-04-16 2019-03-05 Calcula Technologies, Inc. Basket and everting balloon with simplified design and control
US9232956B2 (en) 2013-04-16 2016-01-12 Calcula Technologies, Inc. Device for removing kidney stones
US9387045B2 (en) 2013-05-14 2016-07-12 Intuitive Surgical Operations, Inc. Grip force normalization for surgical instrument
CN108836486B (en) 2013-08-09 2020-11-06 直观外科手术操作公司 Medical robotic system having remote current controller for controlling a plurality of motors housed at a distal side
US9965856B2 (en) 2013-10-22 2018-05-08 Seegrid Corporation Ranging cameras using a common substrate
EP3107476A4 (en) 2014-02-21 2017-11-22 The University of Akron Imaging and display system for guiding medical interventions
KR102450087B1 (en) 2014-03-17 2022-10-06 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 Automated structure with pre-established arm positions in a teleoperated medical system
CN106659541B (en) 2014-03-19 2019-08-16 直观外科手术操作公司 Integrated eyeball stares medical device, the system and method that tracking is used for stereoscopic viewer
JP6644699B2 (en) 2014-03-19 2020-02-12 インテュイティブ サージカル オペレーションズ, インコーポレイテッド Medical devices, systems and methods using gaze tracking
US10084944B2 (en) 2014-03-21 2018-09-25 DePuy Synthes Products, Inc. Card edge connector for an imaging sensor
US20150305602A1 (en) 2014-04-23 2015-10-29 Calcula Technologies, Inc. Integrated medical imaging system
CN107690319B (en) 2015-07-07 2021-01-15 直观外科手术操作公司 Control of multiple devices
US9835821B1 (en) 2016-07-26 2017-12-05 Omnivision Technologies, Inc. Five-surface wide field-of-view compound lens and associated camera module
CN110025338B (en) 2018-01-12 2021-02-12 杭州术创机器人有限公司 Sterile adapter assembly for robotic surgical system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4682219A (en) * 1985-08-16 1987-07-21 Fuji Photo Optical Co., Ltd. Endoscope
US5237446A (en) * 1987-04-30 1993-08-17 Olympus Optical Co., Ltd. Optical low-pass filter
US5557454A (en) * 1992-12-25 1996-09-17 Olympus Optical Co., Ltd. Stereoscopic endoscope
US6144762A (en) * 1998-02-23 2000-11-07 Olympus America Inc. Stereo video microscope
US20030125608A1 (en) * 1999-11-19 2003-07-03 Olympus Optical Co., Ltd. Endoscope apparatus
US7170677B1 (en) * 2002-01-25 2007-01-30 Everest Vit Stereo-measurement borescope with 3-D viewing
US20030233024A1 (en) * 2002-06-14 2003-12-18 Fuji Photo Optical Co., Ltd. Electronic endoscope for stereoscopic endoscope system
US8144409B2 (en) * 2007-09-07 2012-03-27 Hon Hai Precision Industry Co., Ltd. Prism system and method for producing the same

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9005113B2 (en) 2006-12-21 2015-04-14 Intuitive Surgical Operations, Inc. Hermetically sealed endoscope
US9271633B2 (en) 2006-12-21 2016-03-01 Intuitive Surgical Operations, Inc. Stereo camera for hermetically sealed endoscope
US9565997B2 (en) 2006-12-21 2017-02-14 Intuitive Surgical Operations, Inc. Hermetically sealed endoscope with optical component attached to inner protective window
US9962069B2 (en) 2006-12-21 2018-05-08 Intuitive Surgical Operations, Inc. Endoscope with distal hermetically sealed sensor
US10682046B2 (en) 2006-12-21 2020-06-16 Intuitive Surgical Operations, Inc. Surgical system with hermetically sealed endoscope
US11039738B2 (en) 2006-12-21 2021-06-22 Intuitive Surgical Operations, Inc. Methods for a hermetically sealed endoscope
US11382496B2 (en) 2006-12-21 2022-07-12 Intuitive Surgical Operations, Inc. Stereoscopic endoscope
US11716455B2 (en) 2006-12-21 2023-08-01 Intuitive Surgical Operations, Inc. Hermetically sealed stereo endoscope of a minimally invasive surgical system
US11375884B2 (en) * 2011-09-27 2022-07-05 California Institute Of Technology Multi-angle rear-viewing endoscope and method of operation thereof
CN106667418A (en) * 2016-11-22 2017-05-17 珠海维尔康生物科技有限公司 Endoscope

Also Published As

Publication number Publication date
WO2008079578A3 (en) 2008-09-04
US11382496B2 (en) 2022-07-12
US20080151041A1 (en) 2008-06-26
US20220280031A1 (en) 2022-09-08
US20190029501A1 (en) 2019-01-31
WO2008079578A2 (en) 2008-07-03
US8814779B2 (en) 2014-08-26

Similar Documents

Publication Publication Date Title
US20220280031A1 (en) Stereoscopic endoscope
US11147443B2 (en) Surgical visualization systems and displays
US20230122367A1 (en) Surgical visualization systems and displays
US10028651B2 (en) Surgical visualization systems and displays
US7621868B2 (en) Convergence optics for stereoscopic imaging systems
US5381784A (en) Stereoscopic endoscope
US6671099B2 (en) Image-forming optical system
US10334225B2 (en) Stereoscopic camera
US8390675B1 (en) Stereoscopic camera and system
US11723524B2 (en) Angled endoscope tip image capture unit
KR20140139985A (en) Video Endoscopic Device
US10674136B2 (en) Device for capturing a stereoscopic image
US20010055150A1 (en) Video-type stereoscopic microscope
US6388827B2 (en) Image display apparatus having three-dimensionally decentered optical path

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION