US8982261B2 - Imaging with interleaved detection accumulations - Google Patents

Imaging with interleaved detection accumulations Download PDF

Info

Publication number
US8982261B2
US8982261B2 US13/651,387 US201213651387A US8982261B2 US 8982261 B2 US8982261 B2 US 8982261B2 US 201213651387 A US201213651387 A US 201213651387A US 8982261 B2 US8982261 B2 US 8982261B2
Authority
US
United States
Prior art keywords
recited
detections
accumulators
infrared light
accumulations
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
US13/651,387
Other versions
US20140104463A1 (en
Inventor
Kurt Eugene Spears
Otto K Sievert
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hewlett Packard Development Co LP
Original Assignee
Hewlett Packard Development Co LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hewlett Packard Development Co LP filed Critical Hewlett Packard Development Co LP
Priority to US13/651,387 priority Critical patent/US8982261B2/en
Assigned to HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. reassignment HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SPEARS, KURT EUGENE, SIEVERT, OTTO K.
Priority to TW102135546A priority patent/TWI502449B/en
Publication of US20140104463A1 publication Critical patent/US20140104463A1/en
Application granted granted Critical
Publication of US8982261B2 publication Critical patent/US8982261B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03542Light pens for emitting or receiving light
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03545Pens or stylus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means

Definitions

  • remote collaborators can interact with and modify a shared virtual object.
  • the shared virtual object can be in the form of visible images, instances of which are presented locally to respective collaborators. Interactions with the virtual object can be effected using human input devices.
  • the virtual object can be a virtual document page; a collaborator can annotate the virtual document page using an IR pen (a stylus with a tip that emits infra-red light).
  • the annotations can then be presented to remote collaborators.
  • the images of the document pages at the different locations can be reconciled to the effect that they all represent the same object.
  • FIG. 1 is a schematic diagram of an imaging system in accordance with an example.
  • FIG. 2 is a flow chart of a process in accordance with an example.
  • FIG. 3 is a schematic diagram of a collaborative environment in accordance with an example.
  • FIG. 4 is a schematic diagram of an imaging system of the collaborative environment of FIG. 3 .
  • FIG. 5 is a flow chart of a collaborative process in accordance with an example.
  • PCT Patent Application number PCT/US11/58896 filed 2011 Nov. 2 and entitled “Projection Capture System, Programming And Method” discloses a virtual-object collaborative environment (e.g., a mixed reality system) with respective imaging systems provided for each collaborator.
  • Each imaging system includes a projector for projecting a visible image representing a virtual object, e.g., a document page.
  • Each imaging system includes a visible-light camera to track changes in the virtual object, e.g., due to actions by the local collaborator using an IR pen. To this end, an IR camera is used to track the position and motion of the IR pen.
  • Such a system must address the challenge of color registration.
  • the images taken by the IR camera must be aligned (registered) with the visible light image.
  • the visible-light camera uses separate sensor elements for color components (e.g., red, green, and blue) of the visible light
  • the resulting monochrome images must be registered to avoid undesirable artifacts in the composite image.
  • Calibration and post-processing procedures may be available to address registration problems, but they can incur undesirable penalties in time and processing power.
  • the post-processing can increase the latency required to update a projected visible image based on a captured digital image.
  • Examples disclosed herein address registration challenges in part by using full-range sensor elements to capture all colors, albeit at different times.
  • detections output by each sensor element are routed to at least three different color-dedicated accumulators as a function of time so that each accumulator accumulates detections for a respective color.
  • the inclusion of an accumulator dedicated to IR detections addresses the problem of registration of the IR light to the visible light and provides for interleaving of the accumulations to minimize time-based registration problems even in the case of scenes with moving elements, e.g., a moving IR pen.
  • An imaging system 100 shown in FIG. 1 , includes a sensor array 102 of sensor elements 104 .
  • Each sensor element 104 is to convert incident light 106 to detections 108 .
  • detections 108 may take the form of electric charges or some other form.
  • imaging system 100 includes detection accumulators 110 .
  • detection accumulators 110 Associated with each sensor element 104 is a set of n detection accumulators, where n is an integer greater than or equal to three (n ⁇ 3) so that, for example, at least a first detection accumulator 110 R can be dedicated to accumulating detections of red light, at least a second detection accumulator 110 G can be dedicated to accumulating detections of green light, and at least a third detection accumulator 110 B can be dedicated to detections of blue light.
  • each of three or more detection accumulators can be dedicated to a respective dimension of a color space other than red-green-blue (RGB).
  • Imaging system 100 further includes a switch set 112 of switches 114 . Each switch 114 is to route detections from a sensor element 104 to detection accumulators 110 as a function of the color of the incident light being converted to the detections.
  • An imaging process 200 that can be implemented using imaging system 100 or another imaging system is flow charted in FIG. 2 .
  • incident light e.g., incident a sensor element 104
  • detections are routed (e.g., by switch set 114 ) to “color-dedicated” detection accumulators 100 as a function of the color of the incident light being converted to the detections.
  • plural color-dedicated accumulators advantageously decouples the process of inputting detection to the accumulators from the process of outputting accumulations from accumulators. If only one accumulator is used for plural colors, the accumulator must be read out and reset whenever the color to be detected is changed. When plural dedicated-color accumulators are used, one color can be read while another is accumulated. Also, when plural accumulators are used (per sensor element), accumulations can be “interleaved”, i.e., a first color can be accumulated, then another color can be accumulated, and then the first color can be accumulated again, all without an intervening readout or reset.
  • the following pattern could be used: 10 ms of red, readout and reset, 10 ms of green, readout and reset, 10 ms of blue, readout and reset, and repeat. If there are moving elements in the scene or object being imaged, there will be an average of about 15 ms of opportunity for color misregistration due to motion in the scene being imaged. The examples described herein reduce this opportunity for color misregistration due to scene motion.
  • Plural color-dedicated accumulators would permit an interleaving pattern such as 5 ms red, 5 ms green, 5 ms blue, 5 ms red, 5 ms green, 5 ms blue, readout and reset, and repeat.
  • the average time available for misregistration is reduced by half to about 7.5 ms.
  • the time available for misregistration due to scene movement can be reduced to less than 2 ms by cycling through 1 ms phases for each color. In that case, the average misregistration between colors is limited by the amount of scene movement that occurs with 2 ms.
  • the interleavings provided for by using plural dedicated color accumulators can readily reduce color misregistration by an order of magnitude or more. Further advantages are attainable when infra-red capabilities are integrated into the camera, as in the following example.
  • a collaborative environment 300 is illustrated schematically in FIG. 3 including imaging systems 310 and 320 coupled by a network 330 . While two imaging systems are represented in FIG. 3 , other examples include other numbers of imaging systems.
  • Imaging system 310 includes an RGB projector 312 , an RGB-IR camera 314 , and an IR pen 316 .
  • Imaging system 320 is essentially similar to imaging system 310 ; imaging system 320 includes an RGB projector 322 , an RGB-IR camera 324 , and an IR pen 326 .
  • Imaging system 310 can be used to generate a machine-readable digital image 332 of a physical object 334 (which can be, for example, a three-dimensional object or a page of a document).
  • RGB projector 312 can be used to illuminate physical object 334
  • RGB-IR camera 314 can capture light reflected by physical object 334 to produce digital image 332 .
  • Digital image 332 can then be input to RGB projector 312 so that RGB projector 312 projects (e.g., generates) a human-perceptible visible image 336 .
  • Visible image 336 can serve as a virtual object that can be manipulated/modified by a user, e.g., using IR pen 316 as well as other human-input devices.
  • IR pen 316 is a stylus with a tip that emits IR light; the tip may also emit visible light to let a user known when the IR light is active.
  • a local collaborator (or other user) can use IR pen 316 , for example, to control the position of a cursor in visible image 336 and to manipulate or annotate visible image 336 by gesturing or “writing” with IR pen 316 .
  • RGB-IR camera 314 detects the position of IR pen 316 so that the position of IR pen 316 can be tracked. Commands can be implemented and cursor position adjusted in digital image 332 ; as it is updated, digital image 332 can be input to RGB projector 312 to update visible image 336 .
  • digital image 332 can be communicated over network 330 to imaging system 320 .
  • digital image 332 can be input to RGB projector 322 to generate visible image 338 .
  • visible image 338 can be manipulated by a remote collaborator (or other user). e.g., using IR pen 326 .
  • RGB-IR camera 324 can track the position of IR pen 326 and update digital image 332 .
  • each imaging system 310 , 320 maintains an instance of digital image 332 ; programs on the imaging systems ensure that both instances are updated so that visible images 336 and 338 are reconciled (i.e., synchronized) in near real time.
  • local and remote collaborators can work together to dialog about and edit the virtual object represented by visible images 336 and 338 .
  • Collaborative environment 300 and its imaging systems 310 and 320 have some applications and some elements in common with their counterparts disclosed in PCT Patent Application number PCT/US11/58896 entitled “Projection Capture System, Programming And Method”.
  • imaging systems disclosed in that application used separate cameras for visible colors and for IR.
  • RGB-IR cameras 314 and 324 combine visible color capture and IR capture for simplicity, economy, and better registration of color and IR image components.
  • RGB-IR camera 314 includes a channel array 400 of detection channels 402 .
  • Each detection channel 402 includes a respective sensor element 404 , a respective switch 406 , and a respective accumulator set 408 .
  • channel array 400 includes a two-dimensional sensor array 410 of sensor elements 404 , a two-dimensional switch array 412 of switches 406 , and a two-dimensional accumulator-set array 414 of accumulator sets 408 .
  • Sensor array may be formed on the back side of a backside-illuminated CMOS sensor, while switch array 412 and accumulator-set array 414 may be formed on the front side of the backside-illuminated CMOS sensor.
  • Each accumulator set 408 includes plural color-dedicated accumulators 416 , in this case, a red-dedicated accumulator 416 R, a green-dedicated accumulator 416 G, a blue-dedicated accumulator 416 B, and an IR-dedicated accumulator 416 J.
  • Accumulators 416 can be implemented as integrating capacitors.
  • Sensor elements 404 are “full-range” in that each sensor element 404 can detect incident red, green, blue, and infra-red light. In response to detection of incident light, e.g., in the form of photons, each sensor element 404 outputs “detections”, in this case, in the form of electrical charges. All accumulators 416 are essentially similar; the differing labels reflect the fact that, in use, each accumulator is dedicated to a respective color (e.g., red. green, blue, and infra-red).
  • Each switch 406 has outputs coupled to inputs of respective accumulators 416 of a respective accumulator set 408 .
  • Accumulator outputs are coupled to readout circuitry 418 , which includes an analog-to-digital converter (ADC) for converting analog accumulator values to digital values.
  • ADCs are located between accumulators and readout circuitry.
  • the output of readout circuitry 418 is received by an image-data handler 420 , e.g., a computer, which can use the received digital data to construct and update digital image 332 .
  • Digital image 332 is communicated from image-data handler 420 to RGB projector 312 to update visible image 336 .
  • image-data handler 420 communicates via network 330 with its counterparts in other imaging systems, e.g., imaging system 320 , to reconcile (i.e., resolve differences between, synchronize, equalize) instances of digital image 332 .
  • a timing controller 422 controls and synchronizes (i.e., coordinates the timings for) switches 406 , readout circuitry 418 , and red, green, and blue emitters 424 R, 424 G, 424 B of RGB projector 312 .
  • Each switch 406 is synchronized with projector 312 so that, at any given time, detections are routed to the accumulator corresponding to the color being emitted by projector 312 .
  • IR detections are routed to IR accumulators 416 J.
  • Timing controller 422 can include circuitry within RGB projector 312 , within RGB-IR camera 314 , and/or external to both RGB projector 312 and RGB-IR camera.
  • An imaging process 500 can be implemented using collaborative environment 300 or another environment.
  • timings of an imaging system are controlled so that emitting by a projector, switching within a camera, and readout from the camera are synchronized (i.e., their timings are coordinated).
  • timing controller 422 is configured to synchronize timings used for operation of projector 312 , camera switches 408 , and readout circuitry 418 . For example, while timing controller 422 is causing RGB projector 312 is emitting red, switch 406 is directing (presumably red) detections to red accumulator 416 R.
  • timing controller 422 while timing controller 422 is causing RGB projector 312 to not emit any color, timing controller 422 causes switch 406 to direct (presumably IR) detections to IR accumulator 416 J. Also, timing controller 422 causes readout circuitry 418 to read out respectively from accumulators 416 R, 416 G, 416 B, and 416 J, only while they are not receiving detections.
  • RGB projector 312 can emit colors sequentially to yield emitted visible light 428 , resulting in visible light 430 to be incident to camera 314 due to reflections of emitted visible light 428 .
  • incident visible light 430 can consist of cyclical phases, e.g., a red phase 430 R, a green phase 430 G, a blue phase 430 B, and a gap phase 430 X.
  • Switches 406 route detections during red phases 430 R to red accumulator 416 R, detections during green phases 430 G to green accumulator 416 G, detections during blue phases 430 B to blue accumulator 416 B, and detections during gap phases 430 X to IR accumulator 416 J.
  • the phases can have different durations and frequencies. For example, a pattern RGBXRGBXRGBX, etc., allows the position of IR PEN 316 (which may be moving) to be sampled more frequently than visible image 336 (which may be a stationary document).
  • colors are emitted sequentially, i.e., one at a time.
  • the sequence can be RGBRGBRGB . . . in which a sequence RGB is repeated; alternatively, a sequence may not have a repeating pattern.
  • the pattern can include gaps, e.g., RGBXRGBX or RXGXBXRXGXBX, e.g., to allow a camera to detect light from a different source (such as infra-red from an IR pen).
  • the number m of different colors represented is greater than or equal to three so that a full color image can be obtained by integrating over the different colors.
  • the light emitted over each color phase of each cycle can be uniform (e.g., to illuminate a physical object) or image-bearing (e.g., when projecting an image).
  • a collaborator or other entity interacts with a scene using a human interface device such as IR pen 316 .
  • the scene can include a physical object, e.g., illuminated by RGB projector 312 , and/or a visible image, e.g., projected by projector 312 .
  • a collaborator may use IR pen 316 to point to a portion of visible image 336 ; collaborative environment 300 can then reconcile visible image 338 so that collaborators using imaging systems 310 and 320 , respectively, can focus on the same area of the common virtual object.
  • IR pen 316 can be used to annotate a document image (or even a physical object) or issue commands (e.g., “rotate”, “zoom”, etc.).
  • incident light (visible and IR) is detected by sensor array 410 , resulting in detections.
  • switches are operated to direct detections to n ⁇ m color-dedicated accumulations.
  • Switches 406 can be controlled by timing controller 422 in synchronization with RGB projector 312 so that the detections are routed to the accumulator 416 R, 416 G, 416 B, or 416 J corresponding to the phase of the incident light causing the detections. For example, while RGB projector 312 is emitting red light, switches 406 are set so that detections are routed to “red” accumulator 416 R.
  • switches 406 are set so that detections are routed to IR accumulator 416 J.
  • the color phases can vary in number and/or duration by color. For example, there can be two green cycles or a double-length green cycle to take advantage of the fact that green is perceived as most closely related to intensity (brightness), to which human eyes are more sensitive, than hue.
  • accumulations are interleaved.
  • accumulations are “interleaved” when an accumulation within one accumulator coupled to a switch to receive detections from a sensor includes detections acquired both before and after detections accumulated within another accumulator coupled to the switch to receive detections from the sensor.
  • “Interleaving” can include single-color interleaving, e.g., RGBG, in which only one accumulator (in this case, green accumulator ( 416 G) is interleaved, and all-color interleaving, e.g., RGBRGB in which all accumulations are interleaved.
  • Interleaving is made possible by the presence of dedicated-color accumulators (as opposed to using a single accumulator that must be emptied before it can be used to store detections associated with the next color). As explained below, interleaving makes possible dramatic reductions in problems with color misregistration due to moving elements in a scene.
  • accumulations are read out; once its contents have been read, an accumulator is reset (e.g., to zero).
  • accumulators can be read out and reset each color cycle, e.g., RGB, or RGBJ.
  • readouts/resets can occur after plural color cycles, e.g., RGBRGB or RGBJRGBJ (where J can correspond to switch settings in which projector 312 is not emitting a color and detections routed are to IR accumulator 416 J.
  • the number of color-phase cycles can be large, e.g., tens or hundreds between readouts so that the accumulations in a set of accumulators are highly overlapped in time so as to minimize misregistration of colors in a digital image (without requiring color registration post processing, which can be time consuming and, thus, delay collaborative image updating).
  • readout circuitry 418 can convert analog values stored in the accumulators to digital values for use by image-data handler 420 .
  • a digital image 332 is created/updated based on the data read out from the accumulators.
  • the image data is interpreted, e.g., by image data handler 420 to track IR pen position and motion. Note that, if the IR pen is to be tracked against a stationary object or projected image, the pen position can be sampled more frequently than the object or projected image.
  • projector 312 can emit repetitions of RXGXBX and detections can be routed with repetitions of the pattern RJGJBJ.
  • instances of a digital image at different imaging systems are reconciled. For example, copies of digital image 332 ( FIG. 3 ) stored by respective imaging systems 310 and 320 can be reconciled.
  • the reconciled digital images can be input to respective projectors to generate/update visible images. Since digital images are reconciled, the visible images generated from them are also reconciled. For example, visible images 336 and 338 ( FIG. 3 ) are reconciled so that collaborators using respective imaging systems 310 and 320 can operate on the same virtual objects.
  • IR accumulators can be used to detect phase and thus depth of an IR pen.
  • m ⁇ n i.e., the number m of emitters in the projector involved in time-sequential emissions is less than or equal to the number n of accumulators per accumulator set, i.e., per sensor channel.
  • three accumulators can be used with a projector that emits three colors (RGB), or four accumulators can be used with a projector that emits four colors (RGB-IR), or four accumulators can be used with a projector that emits three colors (RGB), with the fourth accumulator used to detect IR from a source other than the projector.
  • a projector can have m>n emitters, but limit the number used to n.
  • a projector may have emitters for R, G, B, IR, and UV, but use only one of IR and UV at a time.
  • a projector can have six emitters: red, green, blue, magenta, cyan, and yellow, and three accumulators can be used during even cycles for red, green, and blue and during odd cycles for magenta, cyan, and yellow.
  • each accumulator is dedicated to accumulating detections only for a single color (R, G, B, or IR).
  • the color accumulator by an accumulator may be changed, e.g., for different readout cycles.
  • a “system” is a set of interacting non-transitory tangible elements, wherein the elements can be, by way of example and not of limitation, mechanical components, electrical elements, atoms, physical encodings of instructions, and process segments.
  • device refers to a hardware or hardware+software system.
  • process refers to a sequence of actions resulting in or involving a physical transformation.
  • imaging process is a process for creating visible and/or digital images.
  • imaging system is a system that creates visible and/or digital images.
  • image refers to a (uniform or non-uniform) spatial distribution of light or a digital representation of such a distribution of light.
  • visible image is an image that a human can perceive;
  • digital image is a non-transitory tangible encoding of data in digital format that represents a visible image (and may include other data).
  • a “virtual object” is a digitally-defined object that represents a human-manipulable object and that can be manipulated by a human as if it were that manipulable object.
  • a projected image of a document can represent a hardcopy document and can be manipulated, e.g., annotated using an IR pen, (more or less) as if it were the hardcopy document.
  • light is electromagnetic radiation.
  • Light encompasses “visible light”, which consists of light within a wavelength range perceptible to the human eye, and “invisible light”, which is light outside the wavelength range perceptible to the human eye and encompasses “infra-red light” and “ultra-violet light”.
  • a “sensor” is a hardware device for converting incident light (i.e., light that reaches the sensor) into detections.
  • a “sensor array” is a sensor constituted by an array, typically two-dimensional) of “sensor elements”, each of which is a sensor in its own right.
  • a “detection” is a tangible entity produced in response to incident light.
  • a detection can be, for example, an electrical charge or an set of electrical charges.
  • a detection can be a voltage level or a light intensity level (where the sensor effectively generates amplified light in response to incident light), or a detection can take another form.
  • a “detection accumulator” or just “accumulator” is a device that accumulates or counts detections.
  • an accumulator can be an integrating capacitor that increases its charge level as detections in the form of electrical charges are received.
  • an accumulator can take another form such as a counter that counts light pulses generated in response to incident light.
  • switch refers to a device with an input, plural outputs, and a control port for receiving a signal that selects one of the outputs to be connected to the input.
  • a switch input is connected to an output of a sensor element for receiving a detection therefrom; each switch output is coupled to a respective color-dedicated accumulator to, when coupled to the switch input, direct the detection to the respective accumulator; and each control port is coupled to the timing controller so that the switch settings can be synchronized with projector emitters.
  • partitioned means “allocated”, in the sense that each detection is allocated to a respective accumulator according to the switch setting at the time the detection is made.
  • set requires at least two elements.
  • a “readout system” and “readout subsystem” refer to systems for reading out values from other devices, such as accumulators, e.g., to determine the number or amount of detections accumulated by an accumulator.
  • “Reset” herein refers to initializing a device, e.g., setting an accumulator so that the amount of detections it represents is zero.
  • “computer” refers to a hardware machine for processing physically encoded data in accordance with physically encoded instructions.
  • a “server” is a computer that performs services for other computers. Depending on context, reference to a computer or server may or may not include software installed on the computer.
  • “storage medium” and “storage media” refer to a system including non-transitory tangible material in or on which information is or can be encoded with information including data and instructions.
  • “Computer-readable” refers to storage media in which information is encoded in computer-readable form.

Abstract

An imaging system includes a sensor array of sensor elements to convert incident light to detections. Plural detection accumulators are provided to store detections. Switches alternatively route detections from a sensor element selectively to n≧3 members of a set of said detection accumulators. An imaging process includes converting incident light to detections, and routing detections so as to partition them among n≧3 accumulators.

Description

BACKGROUND
In a virtual-object collaborative environment, remote collaborators can interact with and modify a shared virtual object. The shared virtual object can be in the form of visible images, instances of which are presented locally to respective collaborators. Interactions with the virtual object can be effected using human input devices. For example, the virtual object can be a virtual document page; a collaborator can annotate the virtual document page using an IR pen (a stylus with a tip that emits infra-red light). The annotations can then be presented to remote collaborators. In other words, the images of the document pages at the different locations can be reconciled to the effect that they all represent the same object.
BRIEF DESCRIPTION OF THE DRAWINGS
The following figures represent examples and not the invention itself.
FIG. 1 is a schematic diagram of an imaging system in accordance with an example.
FIG. 2 is a flow chart of a process in accordance with an example.
FIG. 3 is a schematic diagram of a collaborative environment in accordance with an example.
FIG. 4 is a schematic diagram of an imaging system of the collaborative environment of FIG. 3.
FIG. 5 is a flow chart of a collaborative process in accordance with an example.
DETAILED DESCRIPTION
PCT Patent Application number PCT/US11/58896 filed 2011 Nov. 2 and entitled “Projection Capture System, Programming And Method” discloses a virtual-object collaborative environment (e.g., a mixed reality system) with respective imaging systems provided for each collaborator. Each imaging system includes a projector for projecting a visible image representing a virtual object, e.g., a document page. Each imaging system includes a visible-light camera to track changes in the virtual object, e.g., due to actions by the local collaborator using an IR pen. To this end, an IR camera is used to track the position and motion of the IR pen.
Such a system must address the challenge of color registration. First of all, the images taken by the IR camera must be aligned (registered) with the visible light image. Furthermore, if the visible-light camera uses separate sensor elements for color components (e.g., red, green, and blue) of the visible light, the resulting monochrome images must be registered to avoid undesirable artifacts in the composite image. Calibration and post-processing procedures may be available to address registration problems, but they can incur undesirable penalties in time and processing power. For example, the post-processing can increase the latency required to update a projected visible image based on a captured digital image.
Examples disclosed herein address registration challenges in part by using full-range sensor elements to capture all colors, albeit at different times. As explained below with respect to imaging system 100 (FIG. 1), detections output by each sensor element are routed to at least three different color-dedicated accumulators as a function of time so that each accumulator accumulates detections for a respective color. As explained further below with reference to collaborative environment 300 (FIGS. 3 and 4), the inclusion of an accumulator dedicated to IR detections addresses the problem of registration of the IR light to the visible light and provides for interleaving of the accumulations to minimize time-based registration problems even in the case of scenes with moving elements, e.g., a moving IR pen.
An imaging system 100, shown in FIG. 1, includes a sensor array 102 of sensor elements 104. Each sensor element 104 is to convert incident light 106 to detections 108. Depending on the type of technology employed, detections 108 may take the form of electric charges or some other form.
In addition to sensor elements 104, imaging system 100 includes detection accumulators 110. Associated with each sensor element 104 is a set of n detection accumulators, where n is an integer greater than or equal to three (n≧3) so that, for example, at least a first detection accumulator 110R can be dedicated to accumulating detections of red light, at least a second detection accumulator 110G can be dedicated to accumulating detections of green light, and at least a third detection accumulator 110B can be dedicated to detections of blue light. Alternatively, each of three or more detection accumulators can be dedicated to a respective dimension of a color space other than red-green-blue (RGB). Imaging system 100 further includes a switch set 112 of switches 114. Each switch 114 is to route detections from a sensor element 104 to detection accumulators 110 as a function of the color of the incident light being converted to the detections.
An imaging process 200 that can be implemented using imaging system 100 or another imaging system is flow charted in FIG. 2. At 201, incident light (e.g., incident a sensor element 104) is converted to detections. At 202, the detections are routed (e.g., by switch set 114) to “color-dedicated” detection accumulators 100 as a function of the color of the incident light being converted to the detections.
The use of plural color-dedicated accumulators advantageously decouples the process of inputting detection to the accumulators from the process of outputting accumulations from accumulators. If only one accumulator is used for plural colors, the accumulator must be read out and reset whenever the color to be detected is changed. When plural dedicated-color accumulators are used, one color can be read while another is accumulated. Also, when plural accumulators are used (per sensor element), accumulations can be “interleaved”, i.e., a first color can be accumulated, then another color can be accumulated, and then the first color can be accumulated again, all without an intervening readout or reset.
For example, to achieve a frame duration of 30 milliseconds (ms) using a single accumulator, the following pattern could be used: 10 ms of red, readout and reset, 10 ms of green, readout and reset, 10 ms of blue, readout and reset, and repeat. If there are moving elements in the scene or object being imaged, there will be an average of about 15 ms of opportunity for color misregistration due to motion in the scene being imaged. The examples described herein reduce this opportunity for color misregistration due to scene motion.
Plural color-dedicated accumulators would permit an interleaving pattern such as 5 ms red, 5 ms green, 5 ms blue, 5 ms red, 5 ms green, 5 ms blue, readout and reset, and repeat. In this case, the average time available for misregistration is reduced by half to about 7.5 ms. The time available for misregistration due to scene movement can be reduced to less than 2 ms by cycling through 1 ms phases for each color. In that case, the average misregistration between colors is limited by the amount of scene movement that occurs with 2 ms. Thus, the interleavings provided for by using plural dedicated color accumulators can readily reduce color misregistration by an order of magnitude or more. Further advantages are attainable when infra-red capabilities are integrated into the camera, as in the following example.
A collaborative environment 300 is illustrated schematically in FIG. 3 including imaging systems 310 and 320 coupled by a network 330. While two imaging systems are represented in FIG. 3, other examples include other numbers of imaging systems. Imaging system 310 includes an RGB projector 312, an RGB-IR camera 314, and an IR pen 316. Imaging system 320 is essentially similar to imaging system 310; imaging system 320 includes an RGB projector 322, an RGB-IR camera 324, and an IR pen 326.
Imaging system 310 can be used to generate a machine-readable digital image 332 of a physical object 334 (which can be, for example, a three-dimensional object or a page of a document). To this end, RGB projector 312 can be used to illuminate physical object 334, and RGB-IR camera 314 can capture light reflected by physical object 334 to produce digital image 332. Digital image 332 can then be input to RGB projector 312 so that RGB projector 312 projects (e.g., generates) a human-perceptible visible image 336.
Visible image 336 can serve as a virtual object that can be manipulated/modified by a user, e.g., using IR pen 316 as well as other human-input devices. IR pen 316 is a stylus with a tip that emits IR light; the tip may also emit visible light to let a user known when the IR light is active. A local collaborator (or other user) can use IR pen 316, for example, to control the position of a cursor in visible image 336 and to manipulate or annotate visible image 336 by gesturing or “writing” with IR pen 316. RGB-IR camera 314 detects the position of IR pen 316 so that the position of IR pen 316 can be tracked. Commands can be implemented and cursor position adjusted in digital image 332; as it is updated, digital image 332 can be input to RGB projector 312 to update visible image 336.
In addition to its use locally with respect to imaging system 310, digital image 332 can be communicated over network 330 to imaging system 320. There, digital image 332 can be input to RGB projector 322 to generate visible image 338. Like visible image 336, visible image 338 can be manipulated by a remote collaborator (or other user). e.g., using IR pen 326. RGB-IR camera 324 can track the position of IR pen 326 and update digital image 332. In practice, each imaging system 310, 320 maintains an instance of digital image 332; programs on the imaging systems ensure that both instances are updated so that visible images 336 and 338 are reconciled (i.e., synchronized) in near real time. Thus, local and remote collaborators can work together to dialog about and edit the virtual object represented by visible images 336 and 338.
Collaborative environment 300 and its imaging systems 310 and 320 have some applications and some elements in common with their counterparts disclosed in PCT Patent Application number PCT/US11/58896 entitled “Projection Capture System, Programming And Method”. However, imaging systems disclosed in that application used separate cameras for visible colors and for IR. RGB- IR cameras 314 and 324 combine visible color capture and IR capture for simplicity, economy, and better registration of color and IR image components.
RGB-IR camera 314, as shown in FIG. 4, includes a channel array 400 of detection channels 402. Each detection channel 402 includes a respective sensor element 404, a respective switch 406, and a respective accumulator set 408. Accordingly, channel array 400 includes a two-dimensional sensor array 410 of sensor elements 404, a two-dimensional switch array 412 of switches 406, and a two-dimensional accumulator-set array 414 of accumulator sets 408. Sensor array may be formed on the back side of a backside-illuminated CMOS sensor, while switch array 412 and accumulator-set array 414 may be formed on the front side of the backside-illuminated CMOS sensor. Each accumulator set 408 includes plural color-dedicated accumulators 416, in this case, a red-dedicated accumulator 416R, a green-dedicated accumulator 416G, a blue-dedicated accumulator 416B, and an IR-dedicated accumulator 416J. Accumulators 416 can be implemented as integrating capacitors.
Sensor elements 404 are “full-range” in that each sensor element 404 can detect incident red, green, blue, and infra-red light. In response to detection of incident light, e.g., in the form of photons, each sensor element 404 outputs “detections”, in this case, in the form of electrical charges. All accumulators 416 are essentially similar; the differing labels reflect the fact that, in use, each accumulator is dedicated to a respective color (e.g., red. green, blue, and infra-red).
Each switch 406 has outputs coupled to inputs of respective accumulators 416 of a respective accumulator set 408. Accumulator outputs are coupled to readout circuitry 418, which includes an analog-to-digital converter (ADC) for converting analog accumulator values to digital values. In other examples, ADCs are located between accumulators and readout circuitry.
The output of readout circuitry 418 is received by an image-data handler 420, e.g., a computer, which can use the received digital data to construct and update digital image 332. Digital image 332 is communicated from image-data handler 420 to RGB projector 312 to update visible image 336. In addition, image-data handler 420 communicates via network 330 with its counterparts in other imaging systems, e.g., imaging system 320, to reconcile (i.e., resolve differences between, synchronize, equalize) instances of digital image 332.
A timing controller 422 controls and synchronizes (i.e., coordinates the timings for) switches 406, readout circuitry 418, and red, green, and blue emitters 424R, 424G, 424B of RGB projector 312. Each switch 406 is synchronized with projector 312 so that, at any given time, detections are routed to the accumulator corresponding to the color being emitted by projector 312. During gaps in emissions by projector 312, IR detections are routed to IR accumulators 416J. Thus detections are routed to accumulators as a function of the color of the incident light from which the detections resulted. Timing controller 422 can include circuitry within RGB projector 312, within RGB-IR camera 314, and/or external to both RGB projector 312 and RGB-IR camera.
An imaging process 500, flow charted in FIG. 5, can be implemented using collaborative environment 300 or another environment. At 501, timings of an imaging system are controlled so that emitting by a projector, switching within a camera, and readout from the camera are synchronized (i.e., their timings are coordinated). In the context of collaborative environment 300, timing controller 422 is configured to synchronize timings used for operation of projector 312, camera switches 408, and readout circuitry 418. For example, while timing controller 422 is causing RGB projector 312 is emitting red, switch 406 is directing (presumably red) detections to red accumulator 416R. For another example, while timing controller 422 is causing RGB projector 312 to not emit any color, timing controller 422 causes switch 406 to direct (presumably IR) detections to IR accumulator 416J. Also, timing controller 422 causes readout circuitry 418 to read out respectively from accumulators 416R, 416G, 416B, and 416J, only while they are not receiving detections.
For example, RGB projector 312 can emit colors sequentially to yield emitted visible light 428, resulting in visible light 430 to be incident to camera 314 due to reflections of emitted visible light 428. Due to control of RGB projector 312 by timing controller 412, incident visible light 430, like emitted visible light 428, can consist of cyclical phases, e.g., a red phase 430R, a green phase 430G, a blue phase 430B, and a gap phase 430X. Switches 406 route detections during red phases 430R to red accumulator 416R, detections during green phases 430G to green accumulator 416G, detections during blue phases 430B to blue accumulator 416B, and detections during gap phases 430X to IR accumulator 416J. The phases can have different durations and frequencies. For example, a pattern RGBXRGBXRGBX, etc., allows the position of IR PEN 316 (which may be moving) to be sampled more frequently than visible image 336 (which may be a stationary document).
At 502, colors are emitted sequentially, i.e., one at a time. For example, the sequence can be RGBRGBRGB . . . in which a sequence RGB is repeated; alternatively, a sequence may not have a repeating pattern. In a case such as imaging system 310 (in which RGB-IR camera 314 detects more colors than RGB projector 312 emits), the pattern can include gaps, e.g., RGBXRGBX or RXGXBXRXGXBX, e.g., to allow a camera to detect light from a different source (such as infra-red from an IR pen). The number m of different colors represented is greater than or equal to three so that a full color image can be obtained by integrating over the different colors. The light emitted over each color phase of each cycle can be uniform (e.g., to illuminate a physical object) or image-bearing (e.g., when projecting an image).
At 503, a collaborator or other entity interacts with a scene using a human interface device such as IR pen 316. The scene can include a physical object, e.g., illuminated by RGB projector 312, and/or a visible image, e.g., projected by projector 312. For example, a collaborator may use IR pen 316 to point to a portion of visible image 336; collaborative environment 300 can then reconcile visible image 338 so that collaborators using imaging systems 310 and 320, respectively, can focus on the same area of the common virtual object. Also, IR pen 316 can be used to annotate a document image (or even a physical object) or issue commands (e.g., “rotate”, “zoom”, etc.).
At 504, incident light (visible and IR) is detected by sensor array 410, resulting in detections. At 505, switches are operated to direct detections to n≧m color-dedicated accumulations. Switches 406 can be controlled by timing controller 422 in synchronization with RGB projector 312 so that the detections are routed to the accumulator 416R, 416G, 416B, or 416J corresponding to the phase of the incident light causing the detections. For example, while RGB projector 312 is emitting red light, switches 406 are set so that detections are routed to “red” accumulator 416R. For another example, during gaps between color emissions by RGB projector 312, switches 406 are set so that detections are routed to IR accumulator 416J. Note that the color phases can vary in number and/or duration by color. For example, there can be two green cycles or a double-length green cycle to take advantage of the fact that green is perceived as most closely related to intensity (brightness), to which human eyes are more sensitive, than hue.
At 506, accumulations are interleaved. Herein, accumulations are “interleaved” when an accumulation within one accumulator coupled to a switch to receive detections from a sensor includes detections acquired both before and after detections accumulated within another accumulator coupled to the switch to receive detections from the sensor. “Interleaving” can include single-color interleaving, e.g., RGBG, in which only one accumulator (in this case, green accumulator (416G) is interleaved, and all-color interleaving, e.g., RGBRGB in which all accumulations are interleaved. Interleaving is made possible by the presence of dedicated-color accumulators (as opposed to using a single accumulator that must be emptied before it can be used to store detections associated with the next color). As explained below, interleaving makes possible dramatic reductions in problems with color misregistration due to moving elements in a scene.
At 507, accumulations are read out; once its contents have been read, an accumulator is reset (e.g., to zero). For embodiments that do not employ interleaving, accumulators can be read out and reset each color cycle, e.g., RGB, or RGBJ. For embodiments that do employ interleaving, readouts/resets can occur after plural color cycles, e.g., RGBRGB or RGBJRGBJ (where J can correspond to switch settings in which projector 312 is not emitting a color and detections routed are to IR accumulator 416J.
In some examples, the number of color-phase cycles (e.g., RGBX instances) can be large, e.g., tens or hundreds between readouts so that the accumulations in a set of accumulators are highly overlapped in time so as to minimize misregistration of colors in a digital image (without requiring color registration post processing, which can be time consuming and, thus, delay collaborative image updating). In the course of 507, readout circuitry 418 can convert analog values stored in the accumulators to digital values for use by image-data handler 420.
At 508, a digital image 332 is created/updated based on the data read out from the accumulators. At 509 the image data is interpreted, e.g., by image data handler 420 to track IR pen position and motion. Note that, if the IR pen is to be tracked against a stationary object or projected image, the pen position can be sampled more frequently than the object or projected image. For example, projector 312 can emit repetitions of RXGXBX and detections can be routed with repetitions of the pattern RJGJBJ.
At 510, instances of a digital image at different imaging systems are reconciled. For example, copies of digital image 332 (FIG. 3) stored by respective imaging systems 310 and 320 can be reconciled. At 511, the reconciled digital images can be input to respective projectors to generate/update visible images. Since digital images are reconciled, the visible images generated from them are also reconciled. For example, visible images 336 and 338 (FIG. 3) are reconciled so that collaborators using respective imaging systems 310 and 320 can operate on the same virtual objects.
In other examples, other color sets are used. For example, ultra-violet light may be detected instead of or in addition to infra-red. Detection color phases may be longer and/or more frequent than others. In some examples, there is more than one accumulator per color per sensor. For example, plural IR accumulators can be used to detect phase and thus depth of an IR pen.
In the illustrated examples, m<n, i.e., the number m of emitters in the projector involved in time-sequential emissions is less than or equal to the number n of accumulators per accumulator set, i.e., per sensor channel. For example, three accumulators can be used with a projector that emits three colors (RGB), or four accumulators can be used with a projector that emits four colors (RGB-IR), or four accumulators can be used with a projector that emits three colors (RGB), with the fourth accumulator used to detect IR from a source other than the projector.
In alternative examples, m>n. For example, a projector can have m>n emitters, but limit the number used to n. Thus, a projector may have emitters for R, G, B, IR, and UV, but use only one of IR and UV at a time. For another example, a projector can have six emitters: red, green, blue, magenta, cyan, and yellow, and three accumulators can be used during even cycles for red, green, and blue and during odd cycles for magenta, cyan, and yellow. In the illustrated examples, each accumulator is dedicated to accumulating detections only for a single color (R, G, B, or IR). However, in some examples, e.g., where m>n, the color accumulator by an accumulator may be changed, e.g., for different readout cycles.
Herein, a “system” is a set of interacting non-transitory tangible elements, wherein the elements can be, by way of example and not of limitation, mechanical components, electrical elements, atoms, physical encodings of instructions, and process segments. Herein “device” refers to a hardware or hardware+software system. Herein, “process” refers to a sequence of actions resulting in or involving a physical transformation. An “imaging process” is a process for creating visible and/or digital images.
An “imaging system” is a system that creates visible and/or digital images. Herein, “image” refers to a (uniform or non-uniform) spatial distribution of light or a digital representation of such a distribution of light. A “visible image” is an image that a human can perceive; a “digital image” is a non-transitory tangible encoding of data in digital format that represents a visible image (and may include other data).
Herein, a “virtual object” is a digitally-defined object that represents a human-manipulable object and that can be manipulated by a human as if it were that manipulable object. For example, a projected image of a document can represent a hardcopy document and can be manipulated, e.g., annotated using an IR pen, (more or less) as if it were the hardcopy document.
Herein “light” is electromagnetic radiation. “Light” encompasses “visible light”, which consists of light within a wavelength range perceptible to the human eye, and “invisible light”, which is light outside the wavelength range perceptible to the human eye and encompasses “infra-red light” and “ultra-violet light”. A “sensor” is a hardware device for converting incident light (i.e., light that reaches the sensor) into detections. A “sensor array” is a sensor constituted by an array, typically two-dimensional) of “sensor elements”, each of which is a sensor in its own right.
Herein, a “detection” is a tangible entity produced in response to incident light. A detection can be, for example, an electrical charge or an set of electrical charges. Alternatively, a detection can be a voltage level or a light intensity level (where the sensor effectively generates amplified light in response to incident light), or a detection can take another form.
Herein, a “detection accumulator” or just “accumulator” is a device that accumulates or counts detections. For example, an accumulator can be an integrating capacitor that increases its charge level as detections in the form of electrical charges are received. In other examples, an accumulator can take another form such as a counter that counts light pulses generated in response to incident light.
Herein, “switch” refers to a device with an input, plural outputs, and a control port for receiving a signal that selects one of the outputs to be connected to the input. In the present case, a switch input is connected to an output of a sensor element for receiving a detection therefrom; each switch output is coupled to a respective color-dedicated accumulator to, when coupled to the switch input, direct the detection to the respective accumulator; and each control port is coupled to the timing controller so that the switch settings can be synchronized with projector emitters. Herein, “partitioned” means “allocated”, in the sense that each detection is allocated to a respective accumulator according to the switch setting at the time the detection is made. Herein, “set” requires at least two elements.
Herein, a “readout system” and “readout subsystem” refer to systems for reading out values from other devices, such as accumulators, e.g., to determine the number or amount of detections accumulated by an accumulator. “Reset” herein refers to initializing a device, e.g., setting an accumulator so that the amount of detections it represents is zero.
Herein, “computer” refers to a hardware machine for processing physically encoded data in accordance with physically encoded instructions. A “server” is a computer that performs services for other computers. Depending on context, reference to a computer or server may or may not include software installed on the computer. Herein, “storage medium” and “storage media” refer to a system including non-transitory tangible material in or on which information is or can be encoded with information including data and instructions. “Computer-readable” refers to storage media in which information is encoded in computer-readable form.
In this specification, related art is discussed for expository purposes. Related art labeled “prior art”, if any, is admitted prior art. Related art not labeled “prior art” is not admitted prior art. In the claims, “said” introduces elements for which there is explicit verbatim antecedent basis; “the” introduces elements for which the antecedent basis may be implicit. The illustrated and other described embodiments, as well as modifications thereto and variations thereupon are within the scope of the following claims.

Claims (20)

What is claimed is:
1. An imaging system comprising:
a human interface device to emit infrared light;
a sensor array of sensor elements to convert incident light to detections;
detection accumulators;
switches to couple said detection accumulators to said sensor elements, each switch selectively routing detections from a sensor element so that they are partitioned among n≧3 members of a set of said detection accumulators to yield n concurrent accumulations, wherein one of the n concurrent accumulations corresponds to the infrared light emitted by the human interface device; and
a readout subsystem to read out said accumulations after they become interleaved such that, at a given instant, a first accumulation in a first accumulator of said charge accumulators includes a first portion and a second portion, and a second accumulator of said charge accumulators includes a third portion accumulated after said first portion and before said second portion.
2. An imaging system as recited in claim 1 further comprising:
an emitter subsystem to emit m visible colors sequentially; and
a timing controller to synchronize said switches with the human interface device and said emitter subsystem so that each of said charge accumulators accumulates detections corresponding to at most one of the infrared light and the m visible colors, and so that different ones of said charge accumulators of said set accumulate detections for respective different ones of the infrared light and the m visible colors.
3. An imaging system as recited in claim 2 wherein n≧m.
4. An imaging system as recited in claim 3 further comprising a readout subsystem to read out said accumulations after they become interleaved.
5. An image system as recited in claim 3 wherein n>m.
6. An imaging system as recited in claim 5 wherein said m colors include red, green, and blue, and said n charge accumulators are used for separately storing red, green, blue, and infra-red detections.
7. An imaging system as recited in claim 2 wherein the emitter subsystem is to illuminate a physical object with a visible image to serve as a virtual object.
8. An imaging system as recited in claim 7 wherein the incident light detected by the sensor array of sensor elements comprises reflection of the infrared light and the m visible colors from the physical object.
9. An imaging system as recited in claim 7 wherein the human interface device is manipulated by a user to control the virtual object using the infrared light.
10. An imaging system as recited in claim 7 further comprising an image-data handler to synchronize the virtual object with a second virtual object presented to a remote collaborator.
11. An imaging process comprising:
emitting infrared light using a human interface;
detecting incident light using a sensor array of sensor elements to yield detections;
operating switches to route said detections to detection accumulators so that detections made by a sensor element are partitioned for storage among a set of n≧3 detection accumulators to yield n concurrent accumulations, wherein one of the n concurrent accumulations corresponds to the infrared light emitted by the human interface device; and
reading out said n concurrent accumulations after they become interleaved such that, at a given instant, a first accumulation in a first accumulator of said charge accumulators includes a first portion and a second portion, and a second accumulator of said charge accumulators includes a third portion accumulated after said first portion and before said second portion.
12. An imaging process as recited in claim 11 further comprising:
sequentially emitting light of m visible colors with the infrared light; and
synchronizing said switches with said emitting of the infrared light and the m visible colors so that each of the n concurrent accumulations corresponds to at most one of the infrared light and the at least three visible colors.
13. An imaging process as recited in claim 12 wherein n≧m.
14. An imaging process as recited in claim 13 further comprising reading out said n concurrent accumulations after they become interleaved.
15. An imaging process as recited in claim 13 wherein n>m.
16. An imaging process as recited in claim 15 wherein said m colors include red, green, and blue, and said n charge accumulators are used for separately storing red, green, blue, and infra-red detections.
17. An imaging process as recited in claim 12 wherein the emitting light of m visible colors is to illuminate a physical object with a visible image to serve as a virtual object.
18. An imaging process as recited in claim 17 wherein the incident light detected by the sensor array of sensor elements comprises reflections of the infrared light and the m visible colors from the physical object.
19. An imaging process as recited in claim 17 wherein the human interface device is manipulated by a user to control the virtual object using the infrared light.
20. An imaging process as recited in claim 17 further comprising synchronizing the virtual object with a second virtual object presented to a remote collaborator.
US13/651,387 2012-10-13 2012-10-13 Imaging with interleaved detection accumulations Expired - Fee Related US8982261B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/651,387 US8982261B2 (en) 2012-10-13 2012-10-13 Imaging with interleaved detection accumulations
TW102135546A TWI502449B (en) 2012-10-13 2013-10-01 Imaging with detection routing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/651,387 US8982261B2 (en) 2012-10-13 2012-10-13 Imaging with interleaved detection accumulations

Publications (2)

Publication Number Publication Date
US20140104463A1 US20140104463A1 (en) 2014-04-17
US8982261B2 true US8982261B2 (en) 2015-03-17

Family

ID=50475016

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/651,387 Expired - Fee Related US8982261B2 (en) 2012-10-13 2012-10-13 Imaging with interleaved detection accumulations

Country Status (2)

Country Link
US (1) US8982261B2 (en)
TW (1) TWI502449B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140333814A1 (en) * 2013-05-10 2014-11-13 Canon Kabushiki Kaisha Solid-state image sensor and camera
US10114465B2 (en) 2016-01-15 2018-10-30 Google Llc Virtual reality head-mounted devices having reduced numbers of cameras, and methods of operating the same

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6167511B2 (en) * 2012-12-04 2017-07-26 セイコーエプソン株式会社 Document camera and document camera control method
JP3194297U (en) 2014-08-15 2014-11-13 リープ モーション, インコーポレーテッドLeap Motion, Inc. Motion sensing control device for automobile and industrial use
DE102014224552A1 (en) * 2014-12-01 2016-06-02 Robert Bosch Gmbh Projection apparatus and method for pixel-by-pixel projecting of an image
JP2023518317A (en) * 2020-03-18 2023-04-28 アナロジック・カナダ・コーポレーション Multistage pixel architecture for synchronous readout/integrated radiological imaging and related systems, devices and methods

Citations (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050168437A1 (en) 2004-01-30 2005-08-04 Carl Stewart R. Processing pose data derived from the pose of an elongate object
US7023536B2 (en) 2004-03-08 2006-04-04 Electronic Scripting Products, Inc. Apparatus and method for determining orientation parameters of an elongate object
US7038846B2 (en) 2003-11-24 2006-05-02 Electronic Scripting Products, Inc. Solid catadioptric lens with a single viewpoint
US7088440B2 (en) 2003-12-22 2006-08-08 Electronic Scripting Products, Inc. Method and apparatus for determining absolute position of a tip of an elongate object on a plane surface with invariant features
US7098952B2 (en) * 1998-04-16 2006-08-29 Intel Corporation Imager having multiple storage locations for each pixel sensor
US7110100B2 (en) 2003-11-04 2006-09-19 Electronic Scripting Products, Inc. Apparatus and method for determining an inclination of an elongate object contacting a plane surface
US7113270B2 (en) 2004-06-18 2006-09-26 Electronics Scripting Products, Inc. Determination of an orientation parameter of an elongate object with a scan beam apparatus
US7161664B2 (en) 2004-04-13 2007-01-09 Electronic Scripting Products, Inc. Apparatus and method for optical determination of intermediate distances
US7203384B2 (en) 2003-02-24 2007-04-10 Electronic Scripting Products, Inc. Implement for optically inferring information from a planar jotting surface
US7268956B2 (en) 2003-11-24 2007-09-11 Electronic Scripting Products, Inc. Solid catadioptric lens with two viewpoints
US20070211013A1 (en) 2006-03-03 2007-09-13 Nec Corporation Light source apparatus, display apparatus, terminal apparatus, and control method thereof
US7349104B2 (en) * 2003-10-23 2008-03-25 Technest Holdings, Inc. System and a method for three-dimensional imaging systems
US20080285056A1 (en) * 2007-05-17 2008-11-20 Ilya Blayvas Compact 3D scanner with fixed pattern projector and dual band image sensor
US7460160B2 (en) 2004-09-24 2008-12-02 Microsoft Corporation Multispectral digital camera employing both visible light and non-visible light sensing on a single image sensor
US7586074B2 (en) * 2003-02-17 2009-09-08 Raytheon Company Multi-mode high capacity dual integration direct injection detector input circuit
US7697051B2 (en) * 2001-10-24 2010-04-13 Aptina Imaging Corporation Image sensor with pixels having multiple capacitive storage elements
US7729515B2 (en) 2006-03-08 2010-06-01 Electronic Scripting Products, Inc. Optical navigation apparatus using fixed beacons and a centroid sensing device
US7826641B2 (en) 2004-01-30 2010-11-02 Electronic Scripting Products, Inc. Apparatus and method for determining an absolute pose of a manipulated object in a real three-dimensional environment with invariant features
US20100289885A1 (en) 2007-10-04 2010-11-18 Yuesheng Lu Combined RGB and IR Imaging Sensor
US7961909B2 (en) 2006-03-08 2011-06-14 Electronic Scripting Products, Inc. Computer interface employing a manipulated object with absolute pose detection component and a display
US20110279414A1 (en) 2009-01-20 2011-11-17 Mikihiro Noma Area sensor and liquid crystal display device with area sensor
US8102435B2 (en) * 2007-09-18 2012-01-24 Stmicroelectronics S.R.L. Method for acquiring a digital image with a large dynamic range with a sensor of lesser dynamic range
US20120038549A1 (en) 2004-01-30 2012-02-16 Mandella Michael J Deriving input from six degrees of freedom interfaces
US8139055B2 (en) 2006-06-12 2012-03-20 Sharp Kabushiki Kaisha Combined image sensor and display device
US20130027596A1 (en) * 2011-07-27 2013-01-31 Chung Chun Wan Color imaging using time-multiplexed light sources and monochrome image sensors with multi-storage-node pixels
US20130194418A1 (en) 2013-03-13 2013-08-01 Electronic Scripting Products, Inc. Reduced Homography for Recovery of Pose Parameters of an Optical Apparatus producing Image Data with Structural Uncertainty
US8558929B2 (en) * 2006-12-20 2013-10-15 Carestream Health, Inc. Imaging array for multiple frame capture
US20140104464A1 (en) * 2012-10-13 2014-04-17 Hewlett-Packard Development Company, L.P. Imaging Using Offsetting Accumulations

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8736674B2 (en) * 2010-09-23 2014-05-27 Dolby Laboratories Licensing Corporation Method and system for 3D display calibration with feedback determined by a camera device
TWM419987U (en) * 2011-08-19 2012-01-01 Tpk Touch Solutions Inc An optical touch system

Patent Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7098952B2 (en) * 1998-04-16 2006-08-29 Intel Corporation Imager having multiple storage locations for each pixel sensor
US7697051B2 (en) * 2001-10-24 2010-04-13 Aptina Imaging Corporation Image sensor with pixels having multiple capacitive storage elements
US7586074B2 (en) * 2003-02-17 2009-09-08 Raytheon Company Multi-mode high capacity dual integration direct injection detector input circuit
US7474809B2 (en) 2003-02-24 2009-01-06 Electronic Scripting Products, Inc. Implement for optically inferring information from a jotting surface and environmental landmarks
US7203384B2 (en) 2003-02-24 2007-04-10 Electronic Scripting Products, Inc. Implement for optically inferring information from a planar jotting surface
US7349104B2 (en) * 2003-10-23 2008-03-25 Technest Holdings, Inc. System and a method for three-dimensional imaging systems
US7110100B2 (en) 2003-11-04 2006-09-19 Electronic Scripting Products, Inc. Apparatus and method for determining an inclination of an elongate object contacting a plane surface
US7268956B2 (en) 2003-11-24 2007-09-11 Electronic Scripting Products, Inc. Solid catadioptric lens with two viewpoints
US7038846B2 (en) 2003-11-24 2006-05-02 Electronic Scripting Products, Inc. Solid catadioptric lens with a single viewpoint
US7088440B2 (en) 2003-12-22 2006-08-08 Electronic Scripting Products, Inc. Method and apparatus for determining absolute position of a tip of an elongate object on a plane surface with invariant features
US7826641B2 (en) 2004-01-30 2010-11-02 Electronic Scripting Products, Inc. Apparatus and method for determining an absolute pose of a manipulated object in a real three-dimensional environment with invariant features
US20050168437A1 (en) 2004-01-30 2005-08-04 Carl Stewart R. Processing pose data derived from the pose of an elongate object
US20120038549A1 (en) 2004-01-30 2012-02-16 Mandella Michael J Deriving input from six degrees of freedom interfaces
US7023536B2 (en) 2004-03-08 2006-04-04 Electronic Scripting Products, Inc. Apparatus and method for determining orientation parameters of an elongate object
US7161664B2 (en) 2004-04-13 2007-01-09 Electronic Scripting Products, Inc. Apparatus and method for optical determination of intermediate distances
US7113270B2 (en) 2004-06-18 2006-09-26 Electronics Scripting Products, Inc. Determination of an orientation parameter of an elongate object with a scan beam apparatus
US7460160B2 (en) 2004-09-24 2008-12-02 Microsoft Corporation Multispectral digital camera employing both visible light and non-visible light sensing on a single image sensor
US20070211013A1 (en) 2006-03-03 2007-09-13 Nec Corporation Light source apparatus, display apparatus, terminal apparatus, and control method thereof
US7729515B2 (en) 2006-03-08 2010-06-01 Electronic Scripting Products, Inc. Optical navigation apparatus using fixed beacons and a centroid sensing device
US7961909B2 (en) 2006-03-08 2011-06-14 Electronic Scripting Products, Inc. Computer interface employing a manipulated object with absolute pose detection component and a display
US20110227915A1 (en) 2006-03-08 2011-09-22 Mandella Michael J Computer interface employing a manipulated object with absolute pose detection component and a display
US8139055B2 (en) 2006-06-12 2012-03-20 Sharp Kabushiki Kaisha Combined image sensor and display device
US8558929B2 (en) * 2006-12-20 2013-10-15 Carestream Health, Inc. Imaging array for multiple frame capture
US20080285056A1 (en) * 2007-05-17 2008-11-20 Ilya Blayvas Compact 3D scanner with fixed pattern projector and dual band image sensor
US8102435B2 (en) * 2007-09-18 2012-01-24 Stmicroelectronics S.R.L. Method for acquiring a digital image with a large dynamic range with a sensor of lesser dynamic range
US20100289885A1 (en) 2007-10-04 2010-11-18 Yuesheng Lu Combined RGB and IR Imaging Sensor
US20110279414A1 (en) 2009-01-20 2011-11-17 Mikihiro Noma Area sensor and liquid crystal display device with area sensor
US20130027596A1 (en) * 2011-07-27 2013-01-31 Chung Chun Wan Color imaging using time-multiplexed light sources and monochrome image sensors with multi-storage-node pixels
US20140104464A1 (en) * 2012-10-13 2014-04-17 Hewlett-Packard Development Company, L.P. Imaging Using Offsetting Accumulations
US20130194418A1 (en) 2013-03-13 2013-08-01 Electronic Scripting Products, Inc. Reduced Homography for Recovery of Pose Parameters of an Optical Apparatus producing Image Data with Structural Uncertainty

Non-Patent Citations (18)

* Cited by examiner, † Cited by third party
Title
Gao. Rui et al; Microsoft Research-Mobile Surface; Microsoft Research; 2010; http://research.microsoft.com/en-us/projects/mobilesurface/.
Hand, Randall; Infinite Z Launches zSpace Virtual Holographic 3D Display Designers; VizWorld.com; Dec. 13, 2011: http://www.vizworld.com/2011/12/infinite-launches-zspace-virtual-holographic-3d-display-designers/#sthash.j6Ysf31PX.dpbs.
Harrison, Beverly & Ziola, Ryder; Bringing Toys To Life: Intel Labs OASIS Project; Agumented Engineering; Jan. 26, 2011; http://augmentedengineering.wordpress.com/2011/01/26/bringing-toys-to-life-intel-labs-oasis-project/.
Harrison, Chris et al; OmniTouch: Wearable Multitouch Interaction Everywhere; USIT'11; Oct. 16, 2011; Santa Barbara, California; http://research.microsoft.com/en-us/um/people/awilson/publications/HarrisonUIST2011/HarrisonUIST2011.html.
Hartmann. Bjorn et al; Pictionaire: Supporting Collaborative Design Work by Integrating Physical and Digital Artifacts; CSCW 2010; Feb. 6, 2010; Savannah, Georgia; http://research.microsoft.com/en-us/um/people/awilson/publications/HartmannCSCW2010/HartmannCSCW2010.html.
Hinckley, Ken et al; Pen +Touch =New Tools; UIST'10; Oct. 3, 2010; New York, New York; http://research.microsoft.com/en-us/um/people/awilson/publications/HinckleyUIST2010/HinckleyUIST2010.html.
Izadi, Shahram et al; C-Slate: A Multi-Touch and Object Recognition System for Remote Collaboration Using Horizontal Surfaces; IEEE; 2007; http://firesearch.microsoft.com/pubs/132551/cslatel.pdf.
Junuzovic, Sasa et al; Microsoft Research-IllumiShare; Microsoft Research; 2012; http://delivery.acm.org/10.1145/2210000/2208333/p1919-junuzovic.pdf?ip=15.219.153.76&id=2208333&acc=ACTIVE%20SERVICE&key=C2716FEBFA981EF153FE223D54A46A411F9E1F05F1DA6F38&CIFD=27 1558808&CFTOKEN=96334302&-acm-=1387226323-00af7aa5e159c7c67d724c49ef01back.
Junuzovic, Sasa et al; Microsoft Research-IllumiShare; Microsoft Research; 2012; http://delivery.acm.org/10.1145/2210000/2208333/p1919-junuzovic.pdf?ip=15.219.153.76&id=2208333&acc=ACTIVE%20SERVICE&key=C2716FEBFA981EF153FE223D54A46A411F9E1F05F1DA6F38&CIFD=27 1558808&CFTOKEN=96334302&—acm—=1387226323—00af7aa5e159c7c67d724c49ef01back.
Kane, Shaun K. et al; Bonfire: A Nomadic System for Hybird Laptop-Tabletop Interaction; UIST'09; Oct. 4, 2009; Victoria, British Columbia, Canada; http://dub.washington.edu/djangosite/media/papers/uist09.pdf.
Kawada, S. et al. ; A Wide Dynamic Range Checkered-color CMOS Image Sensor with IR-cut RGB and Visible-to-near-IR Pixels; http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber.
Linder; Natan et al; LuminAR: Portable Robotic Augmented Reality Interface Design and Prototype; UIST'10, Oct. 3, 2010; New York, New York; http://fluid.media.mit.edu/sites/default/files/2010-10-03-luminar-uist10-demo.pdf.
Linder; Natan et al; LuminAR: Portable Robotic Augmented Reality Interface Design and Prototype; UIST'10, Oct. 3, 2010; New York, New York; http://fluid.media.mit.edu/sites/default/files/2010-10-03-luminar—uist10—demo.pdf.
Melanson, Donald; Microsoft Research Working on Portable Surface; Mar. 2, 2010; http://www.engadget.com/2010/03/02/microsoft-research-working-on-portable-surface/.
Melanson, Donald; Wiimote Repurposed for Multi-Point Interactive Whiteboard; Dec. 10, 2007; http://www.engadget.com/2007/12/10/wiimote-repurposed-for-multi-point-interactive-whiteboard/.
Simonite, Tom; A Kitchen Countertop With a Brain; MIT Technology Review; Jul. 2, 2010; http://www.technologyreview.com/news/419639/a-kitchen-countertop-with-a-brain/.
Wilson, Andrew D. et al; Combining Multiple Depth Cameras and Projectors for Interactions on, Above, and Between Surfaces; UIST'10; Oct. 3, 2010; New York, New York; http://research.microsoft.com/en-us/um/people/awilson/publication/WilsonUIST2010/WisonUIST2010.html.
Wilson, Andrew D.; Using a Depth Camera as a Touch Sensor; ITS 2010: Devices & Algorithms; Nov. 7, 2010; Saarbrucken, Germany; http://research.microsoft.com/en-us/um/people/awilson/publications/WilsonITS2010/WilsonITS2010.html.

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140333814A1 (en) * 2013-05-10 2014-11-13 Canon Kabushiki Kaisha Solid-state image sensor and camera
US9455289B2 (en) * 2013-05-10 2016-09-27 Canon Kabushiki Kaisha Solid-state image sensor and camera
US9978792B2 (en) 2013-05-10 2018-05-22 Canon Kabushiki Kaisha Solid-state image sensor and camera which can detect visible light and infrared light at a high S/N ratio
US10475833B2 (en) 2013-05-10 2019-11-12 Canon Kabushiki Kaisha Solid-state image sensor and camera which can detect visible light and infrared light at a high S/N ratio
US10114465B2 (en) 2016-01-15 2018-10-30 Google Llc Virtual reality head-mounted devices having reduced numbers of cameras, and methods of operating the same
US10990186B2 (en) 2016-01-15 2021-04-27 Google Llc Virtual reality head-mounted devices having reduced numbers of cameras, and methods of operating the same

Also Published As

Publication number Publication date
TWI502449B (en) 2015-10-01
TW201419094A (en) 2014-05-16
US20140104463A1 (en) 2014-04-17

Similar Documents

Publication Publication Date Title
US8982261B2 (en) Imaging with interleaved detection accumulations
JP2021060590A (en) Pass-through display of captured imagery
US9083905B2 (en) Structured light imaging system
CN107710415B (en) Interline charge coupled device
JP6377855B2 (en) Imaging method, imaging apparatus, and electronic apparatus
CN103425354B (en) The control method of data processing equipment, display device and data processing equipment
US20140240492A1 (en) Depth sensor using modulated light projector and image sensor with color and ir sensing
US20140028804A1 (en) 3d imaging apparatus
US9462236B2 (en) Imaging using offsetting accumulations
US11256372B1 (en) Method and apparatus for creating an adaptive Bayer pattern
KR102481774B1 (en) Image apparatus and operation method thereof
WO2006085580A1 (en) Pointer light tracking method, program, and recording medium thereof
JP2009289243A (en) Position detection device, position detection system, video display device and video display system
WO2015134961A1 (en) Method and system for unsynchronized structured lighting
JP6042674B2 (en) Image projection device with 3D information acquisition function
Moreno et al. Unsynchronized structured light
US11314334B2 (en) Gesture recognition apparatus, control method thereof, and display apparatus
CN106415450B (en) Module, system and method for generating an image matrix for gesture recognition
JP2018074311A (en) Imaging apparatus and driving method of imaging apparatus
US20140354546A1 (en) Interactive projection system and interactive image-detecting method
JP2010154088A (en) Image processing matching position and image
WO2016157920A1 (en) Information processing device, information processing method, and program
JP5208293B2 (en) POSITION DETECTION DEVICE, POSITION DETECTION SYSTEM, VIDEO DISPLAY DEVICE, AND VIDEO DISPLAY SYSTEM
CN112702586A (en) Projector virtual touch tracking method, device and system based on visible light
US10091441B1 (en) Image capture at multiple resolutions

Legal Events

Date Code Title Description
AS Assignment

Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SPEARS, KURT EUGENE;SIEVERT, OTTO K.;SIGNING DATES FROM 20121011 TO 20121012;REEL/FRAME:029217/0780

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction
MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20230317