US20070076099A1 - Device and method for hybrid resolution video frames - Google Patents
Device and method for hybrid resolution video frames Download PDFInfo
- Publication number
- US20070076099A1 US20070076099A1 US11/414,370 US41437006A US2007076099A1 US 20070076099 A1 US20070076099 A1 US 20070076099A1 US 41437006 A US41437006 A US 41437006A US 2007076099 A1 US2007076099 A1 US 2007076099A1
- Authority
- US
- United States
- Prior art keywords
- view
- segment
- image sensor
- image
- captured
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/951—Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/272—Means for inserting a foreground image in a background image, i.e. inlay, outlay
Definitions
- the present invention relates generally to the capture of images, and particularly to the processing and viewing of streams of images that includes different pixel resolutions densities at different areas of interest of a view.
- Pan, tilt zoom (PTZ) cameras that may zoom in on a particular area of interest of a view are also used.
- PTZ Pan, tilt zoom
- a user may lose some or part of the wide view as the camera focuses on a small area of the view.
- a first segment of a wide view may be captured at a first resolution
- a second segment of a wide view may be captured at a second resolution.
- the invention includes a system having more than one image sensor; and a processor to reference a group of pixels captured by a first of the image sensors at a first resolution to a segment of a model of a view, and to reference a group of pixels captured by a second of image sensors at a second resolution to the segment of the model of the view, and to display a first part of the segment of the view in a first scale, where such display of the first part of the segment has a first set of resolutions, and to display a second part of the segment of the view in a second scale, where the display of the second part of the segment has a second set of resolutions.
- the processor is to accept an instruction from an input device to alter a stitch of the view captured by the first image sensor and of the view captured by the second image sensor.
- a physical position of the first image sensor may not be calibrated to a position of the second image sensor.
- the processor may alter a second scale in response to a signal from an input device.
- an image sensor may be or include any or all of a digital video camera, a digital still camera, an analog video camera, an analog still camera, an infra red sensor, a radar sensor or an Xray sensor.
- an image sensor may be or include a: pan-tilt-zoom camera.
- a segment of an image may include less than all of the view in such image.
- the processor may define a size or area of a segment in response to a signal from an input device.
- Some embodiments of the invention may include a method of referencing to a segment of a model of a view, a group of pixels captured by a first of a group of image sensors at a first resolution, referencing to the segment of the model of the view, a group of pixels captured by a second of the group of image sensors at a second resolution, displaying a first part of the segment of the view in a first scale, such display of the first part of the segment having a first set of resolutions and displaying a second part of the segment of the view in a second scale, having a second set of resolutions.
- FIG. 1 is a conceptual illustration of a view captured by one or more image sensors having different resolutions, in accordance with an embodiment of the invention.
- FIG. 2 is a block diagram of a method in accordance with some embodiments of the invention.
- FIG. 1 a conceptual illustration of a view captured by one or more image sensors having different resolutions, in accordance with an embodiment of the invention.
- one or more images or streams of images may be captured of one or more objects, or parts of objects or of a group of objects in a view 100 of objects.
- images of view 100 may be captured by one or more images sensors 102 , 104 and 106 .
- images sensors 102 , 104 and 106 may capture images of for example view 100 at the same or different resolutions.
- image sensor 106 may be or include a low resolution video camera, that may capture images at a resolution of 1 million pixels per frame
- image sensor 102 may be or include a medium resolution camera, that may capture images at a resolution of 4 million pixels per frame
- image sensor 104 may be or include a high resolution video camera, that may capture images at a resolution of 10 million pixels per frame.
- Other numbers of cameras having other resolutions may be used.
- a lens on an image sensor 102 may influence or determine a resolution of an image captured with such image sensor 102 .
- one or more of image sensors 102 , 104 or 106 may be or include for example a digital video camera, a digital still camera, an analog video camera, an analog still camera, an infra red sensor, a radar sensor, an X-ray sensor or other device to capture an image or stream of images.
- an image sensor 102 may be or include for example a PTZ camera that may zoom a lens upon for example an instruction from a user.
- image sensor 104 may be focused on for example a particular object in view 100 , such as for example upon a face 108 of a person in view 100 .
- Other objects or sizes of objects may be the subject of a focus of image sensor 104 .
- Image sensor 102 may be focused on for example a body 110 of a person, and the images captured by image sensor 102 may include some, all or none of face 108 .
- Image sensor 106 may be focused on a wider area of view 100 and such wider area may include all, some or none of body 110 .
- a processor 120 such as for example a central processor unit that may be found in a personal computer, video console, or other electronic device, may generate a virtual map, matrix, model 122 or other set of multi-dimensional coordinates that may represent some or all of the area between some or all of the objects in view 100 and some or all of the image sensors 102 , 104 and 106 .
- model 122 may map view 100 , as it may be captured by for example image sensor 106 .
- a processor such as for example processor 120 may reference the pixels captured by one or more of image sensors 102 , 104 and 106 onto the model 122 .
- coordinates x and y of model 122 may indicate the location of a pixel or group of pixels representing face 108 in the image captured by image sensor 106 or in some other section or segment of view 100 .
- Processor 122 may then associate or reference the pixel or group of pixels that include face 108 as was captured by image sensor 102 over the same coordinates of model 122 that include face 108 , and may similarly map, reference or associate the pixels or group of pixels that include face 108 as were captured by image sensor 104 on those same coordinates.
- the higher density pixels, such as those captured by image sensor 104 may write over pixels from lower resolution images that may have been mapped to the same coordinates of model 122 .
- the segments of view 100 that are captured by the various images sensors 102 , 104 and 106 may not overlap, such that for example, only image sensor 1 . 04 capture an image of face 108 , and only image sensor 102 may capture an image of body 110 , and only image sensor 106 may capture an image of tree 111 .
- processor 120 may map or create a model 122 of the various parts of the view 100 that are captured by the respective image sensors 102 , 104 and 106 , and may stitch the images together in model 122 .
- a physical position, angle or location of one image sensor 102 may be moved or altered relative to a position of another image sensor 104 , and processor 120 may not be required to calibrate such positions or angles.
- a calibration may be accomplished at for example model 122 where the pixels from the image sensors 102 , 104 106 may be overlaid onto model 122 .
- mapping or referencing of pixels captured by different image sensors 102 , 104 , 106 may be performed by for example stitching of the images captured or by other means.
- the map or model 122 of view 100 may include pixels having different resolutions or pixel densities.
- pixels 130 mapped onto model 122 from image sensor 104 may have a density of 10 million pixel per frame
- pixels 132 mapped onto model 122 from image sensor 106 may have a density of 1 million pixel per frame.
- processor 120 may display an image that may include for example a wide or panoramic range of view 100 .
- the displayed image may include pixels from the various streams of image sensors 102 , 104 106 that may have been stitched together by processor 120 . Such stitching may in some embodiments be adjusted by a user by way of signals from input device 124 .
- the displayed image of view 100 may include parts or segments having pixels captured by some or all three image devices 102 , 104 and 106 , and having several resolutions. In such an image, a scale of the objects in view 100 may be preserved to offer a consistent size of objects in the image, even though the pixel resolutions of such objects may differ.
- a screen 126 or other display medium may not have sufficient pixels capacity to show the resolution of for example the area 134 in the image that was captured in high resolution.
- processor 120 may delete or not show some of the pixels that may be available from model 122 .
- a signal or instruction from for example a user or other operator may designate one or more areas of an image for display at a high resolution, and other areas of an image for display at a lower resolution.
- processor 120 may alter or adjust a scale of the objects displayed in for example a high resolution area. Such adjustment of scale may provide more room on display 26 to see the objects slated for high definition display so that more pixels on the display 26 can be included in the image of the object.
- an area designated for, for example, high definition viewing may include pixels at several resolution rates.
- a user or other operator may instruct a processor to display face 108 and an upper part of body 110 at a high resolution or pixel density rate.
- the segment of the displayed image of face 108 and part of body 110 may include at least two pixel resolution rates and a scale of face and upper part of body 110 may be increased to allow the higher resolution to be seen on a larger part of display 126 .
- a lower part of body 110 and tree 111 may be displayed at one or more lower resolution or pixel density rates at a scale similar to that of for example other parts of the displayed image.
- a processor may reference or map a group of pixels captured by a first of a group of image sensors to a segment of a model of a view at a first resolution.
- the same or another processor may reference or map a group of pixels captured by a second of the group of image sensors to such segment of such model of such view, at a second resolution.
- the same or another processor may display a first part of such segment of such view in a first scale, such display of such first part of such segment having a first set of pixel resolutions.
- the same or another processor may display a second part of such segment of such view in a second scale, having a second set of resolutions.
Abstract
Description
- This application claims the benefit of U.S. Provisional Patent Application No. 60/722,429 filed on Oct. 3, 2005, and entitled Apparatus and Method for Hybrid Resolution Video Frames, incorporated by reference herein in its entirety.
- The present invention relates generally to the capture of images, and particularly to the processing and viewing of streams of images that includes different pixel resolutions densities at different areas of interest of a view.
- Combining or stitching multiple video streams to create a wide view of an area of interest is used in fields such as for example security surveillance or industrial control. Pan, tilt zoom (PTZ) cameras that may zoom in on a particular area of interest of a view are also used. When using a PTZ camera, a user may lose some or part of the wide view as the camera focuses on a small area of the view. Furthermore, a first segment of a wide view may be captured at a first resolution, and a second segment of a wide view may be captured at a second resolution.
- In some embodiments, the invention includes a system having more than one image sensor; and a processor to reference a group of pixels captured by a first of the image sensors at a first resolution to a segment of a model of a view, and to reference a group of pixels captured by a second of image sensors at a second resolution to the segment of the model of the view, and to display a first part of the segment of the view in a first scale, where such display of the first part of the segment has a first set of resolutions, and to display a second part of the segment of the view in a second scale, where the display of the second part of the segment has a second set of resolutions.
- In some embodiments, the processor is to accept an instruction from an input device to alter a stitch of the view captured by the first image sensor and of the view captured by the second image sensor.
- In some embodiments a physical position of the first image sensor may not be calibrated to a position of the second image sensor.
- In some embodiments, the processor may alter a second scale in response to a signal from an input device.
- In some embodiments, an image sensor may be or include any or all of a digital video camera, a digital still camera, an analog video camera, an analog still camera, an infra red sensor, a radar sensor or an Xray sensor. In some embodiments, an image sensor may be or include a: pan-tilt-zoom camera. In some embodiments, a segment of an image may include less than all of the view in such image. In some embodiments the processor may define a size or area of a segment in response to a signal from an input device.
- Some embodiments of the invention may include a method of referencing to a segment of a model of a view, a group of pixels captured by a first of a group of image sensors at a first resolution, referencing to the segment of the model of the view, a group of pixels captured by a second of the group of image sensors at a second resolution, displaying a first part of the segment of the view in a first scale, such display of the first part of the segment having a first set of resolutions and displaying a second part of the segment of the view in a second scale, having a second set of resolutions.
- Embodiments of the invention are illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like reference numerals may indicate corresponding, analogous or similar elements, and in which:
-
FIG. 1 is a conceptual illustration of a view captured by one or more image sensors having different resolutions, in accordance with an embodiment of the invention; and -
FIG. 2 is a block diagram of a method in accordance with some embodiments of the invention. - It will be appreciated that for simplicity and clarity of illustration, elements shown in the figures have not necessarily been drawn to scale. For example, the dimensions of some of the elements may be exaggerated relative to other elements.
- In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of embodiments of the invention. However it will be understood by those of ordinary skill in the art that the embodiments of the invention may be practiced without these specific details. In other instances, well-known methods, procedures, and components have not been described in detail so as not to obscure the embodiments of the invention.
- Reference is made to
FIG. 1 , a conceptual illustration of a view captured by one or more image sensors having different resolutions, in accordance with an embodiment of the invention. In some embodiments, one or more images or streams of images may be captured of one or more objects, or parts of objects or of a group of objects in aview 100 of objects. In some embodiments, images ofview 100 may be captured by one ormore images sensors images sensors image sensor 106 may be or include a low resolution video camera, that may capture images at a resolution of 1 million pixels per frame,image sensor 102 may be or include a medium resolution camera, that may capture images at a resolution of 4 million pixels per frame, andimage sensor 104 may be or include a high resolution video camera, that may capture images at a resolution of 10 million pixels per frame. Other numbers of cameras having other resolutions may be used. In some embodiments, a lens on animage sensor 102 may influence or determine a resolution of an image captured withsuch image sensor 102. - In some embodiments, one or more of
image sensors image sensor 102 may be or include for example a PTZ camera that may zoom a lens upon for example an instruction from a user. - In some
embodiments image sensor 104 may be focused on for example a particular object inview 100, such as for example upon aface 108 of a person inview 100. Other objects or sizes of objects may be the subject of a focus ofimage sensor 104.Image sensor 102 may be focused on for example abody 110 of a person, and the images captured byimage sensor 102 may include some, all or none offace 108.Image sensor 106 may be focused on a wider area ofview 100 and such wider area may include all, some or none ofbody 110. - In some embodiments, a
processor 120, such as for example a central processor unit that may be found in a personal computer, video console, or other electronic device, may generate a virtual map, matrix,model 122 or other set of multi-dimensional coordinates that may represent some or all of the area between some or all of the objects inview 100 and some or all of theimage sensors model 122 maymap view 100, as it may be captured by forexample image sensor 106. In some embodiments, a processor such as forexample processor 120 may reference the pixels captured by one or more ofimage sensors model 122. For example, coordinates x and y ofmodel 122 may indicate the location of a pixel or group ofpixels representing face 108 in the image captured byimage sensor 106 or in some other section or segment ofview 100.Processor 122 may then associate or reference the pixel or group of pixels that includeface 108 as was captured byimage sensor 102 over the same coordinates ofmodel 122 that includeface 108, and may similarly map, reference or associate the pixels or group of pixels that includeface 108 as were captured byimage sensor 104 on those same coordinates. In some embodiments, the higher density pixels, such as those captured byimage sensor 104 may write over pixels from lower resolution images that may have been mapped to the same coordinates ofmodel 122. - In some embodiments, the segments of
view 100 that are captured by thevarious images sensors face 108, and onlyimage sensor 102 may capture an image ofbody 110, and onlyimage sensor 106 may capture an image oftree 111. In such case,processor 120 may map or create amodel 122 of the various parts of theview 100 that are captured by therespective image sensors model 122. - In some embodiments, a physical position, angle or location of one
image sensor 102, may be moved or altered relative to a position of anotherimage sensor 104, andprocessor 120 may not be required to calibrate such positions or angles. A calibration may be accomplished at forexample model 122 where the pixels from theimage sensors model 122. - In some embodiments, the mapping or referencing of pixels captured by
different image sensors - In some embodiments, the map or
model 122 ofview 100 may include pixels having different resolutions or pixel densities. For example,pixels 130 mapped ontomodel 122 fromimage sensor 104 may have a density of 10 million pixel per frame, whilepixels 132 mapped ontomodel 122 fromimage sensor 106 may have a density of 1 million pixel per frame. - In some embodiments,
processor 120 may display an image that may include for example a wide or panoramic range ofview 100. The displayed image may include pixels from the various streams ofimage sensors processor 120. Such stitching may in some embodiments be adjusted by a user by way of signals frominput device 124. In some embodiments, the displayed image ofview 100 may include parts or segments having pixels captured by some or all threeimage devices view 100 may be preserved to offer a consistent size of objects in the image, even though the pixel resolutions of such objects may differ. In some embodiments, ascreen 126 or other display medium may not have sufficient pixels capacity to show the resolution of for example the area 134 in the image that was captured in high resolution. To accommodate the lack of resolution available to display 126,processor 120 may delete or not show some of the pixels that may be available frommodel 122. - In some embodiments, a signal or instruction from for example a user or other operator may designate one or more areas of an image for display at a high resolution, and other areas of an image for display at a lower resolution. In some embodiments,
processor 120 may alter or adjust a scale of the objects displayed in for example a high resolution area. Such adjustment of scale may provide more room on display 26 to see the objects slated for high definition display so that more pixels on the display 26 can be included in the image of the object. In some embodiments, an area designated for, for example, high definition viewing may include pixels at several resolution rates. - For example, a user or other operator may instruct a processor to display
face 108 and an upper part ofbody 110 at a high resolution or pixel density rate. The segment of the displayed image offace 108 and part ofbody 110 may include at least two pixel resolution rates and a scale of face and upper part ofbody 110 may be increased to allow the higher resolution to be seen on a larger part ofdisplay 126. At, for example, a same or different time, a lower part ofbody 110 andtree 111 may be displayed at one or more lower resolution or pixel density rates at a scale similar to that of for example other parts of the displayed image. - Reference is made to
FIG. 2 , a flow diagram of a method in accordance with an embodiment of the invention. Inblock 200, a processor may reference or map a group of pixels captured by a first of a group of image sensors to a segment of a model of a view at a first resolution. Inblock 202, the same or another processor may reference or map a group of pixels captured by a second of the group of image sensors to such segment of such model of such view, at a second resolution. Inblock 204, the same or another processor may display a first part of such segment of such view in a first scale, such display of such first part of such segment having a first set of pixel resolutions. Inblock 206, the same or another processor may display a second part of such segment of such view in a second scale, having a second set of resolutions. - While certain features of the invention have been illustrated and described herein, many modifications, substitutions, changes, and equivalents will now occur to those of ordinary skill in the art. It is, therefore, to be understood that the appended claims are intended to cover all such modifications and changes as fall within the spirit of the invention.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/414,370 US20070076099A1 (en) | 2005-10-03 | 2006-05-01 | Device and method for hybrid resolution video frames |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US72242905P | 2005-10-03 | 2005-10-03 | |
US11/414,370 US20070076099A1 (en) | 2005-10-03 | 2006-05-01 | Device and method for hybrid resolution video frames |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070076099A1 true US20070076099A1 (en) | 2007-04-05 |
Family
ID=37901504
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/414,370 Abandoned US20070076099A1 (en) | 2005-10-03 | 2006-05-01 | Device and method for hybrid resolution video frames |
Country Status (1)
Country | Link |
---|---|
US (1) | US20070076099A1 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060290796A1 (en) * | 2005-06-23 | 2006-12-28 | Nokia Corporation | Digital image processing |
US20080170129A1 (en) * | 2007-01-17 | 2008-07-17 | Samsung Techwin Co., Ltd. | Digital photographing apparatus and method for controlling the same |
US20090040322A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090040323A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090041368A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090129693A1 (en) * | 2007-11-15 | 2009-05-21 | Bloebaum L Scott | System and method for generating a photograph with variable image quality |
US20090189993A1 (en) * | 2006-06-13 | 2009-07-30 | Panasonic Corporation | Imaging apparatus |
US20090320081A1 (en) * | 2008-06-24 | 2009-12-24 | Chui Charles K | Providing and Displaying Video at Multiple Resolution and Quality Levels |
US20100231734A1 (en) * | 2007-07-17 | 2010-09-16 | Yang Cai | Multiple resolution video network with context based control |
FR2953088A1 (en) * | 2009-11-26 | 2011-05-27 | Defiboat Technology | Method for transmitting image data e.g. video stream, involves transmitting image data to control room based on received request, and displaying image data transmitted by cameras simultaneously in distinct display windows |
US20110200009A1 (en) * | 2007-07-27 | 2011-08-18 | Sony Computer Entertainment Inc. | Nat traversal for mobile network devices |
US20130265311A1 (en) * | 2012-04-04 | 2013-10-10 | Samsung Electronics Co., Ltd. | Apparatus and method for improving quality of enlarged image |
WO2016107329A1 (en) * | 2014-12-29 | 2016-07-07 | Beijing Zhigu Rui Tuo Tech Co., Ltd. | Light field display control methods and apparatuses, light field display devices |
EP3107279A1 (en) | 2015-06-15 | 2016-12-21 | Coherent Synchro, S.L. | Method, device and installation for composing a video signal |
US9774823B1 (en) | 2016-10-04 | 2017-09-26 | Avaya Inc. | System and method for processing digital images during videoconference |
US20170294033A1 (en) * | 2016-04-06 | 2017-10-12 | Varex Imaging Corporation | Dose efficient x-ray detector and method |
US9936162B1 (en) * | 2016-10-04 | 2018-04-03 | Avaya Inc. | System and method for processing digital images during videoconference |
US9936129B2 (en) * | 2016-06-15 | 2018-04-03 | Obsidian Sensors, Inc. | Generating high resolution images |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020057279A1 (en) * | 1999-05-20 | 2002-05-16 | Compaq Computer Corporation | System and method for displaying images using foveal video |
US20020180759A1 (en) * | 1999-05-12 | 2002-12-05 | Imove Inc. | Camera system with both a wide angle view and a high resolution view |
US20030026588A1 (en) * | 2001-05-14 | 2003-02-06 | Elder James H. | Attentive panoramic visual sensor |
US7006950B1 (en) * | 2000-06-12 | 2006-02-28 | Siemens Corporate Research, Inc. | Statistical modeling and performance characterization of a real-time dual camera surveillance system |
US7298409B1 (en) * | 1999-08-02 | 2007-11-20 | Fujifilm Corporation | Imaging system |
-
2006
- 2006-05-01 US US11/414,370 patent/US20070076099A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020180759A1 (en) * | 1999-05-12 | 2002-12-05 | Imove Inc. | Camera system with both a wide angle view and a high resolution view |
US20020057279A1 (en) * | 1999-05-20 | 2002-05-16 | Compaq Computer Corporation | System and method for displaying images using foveal video |
US7298409B1 (en) * | 1999-08-02 | 2007-11-20 | Fujifilm Corporation | Imaging system |
US7006950B1 (en) * | 2000-06-12 | 2006-02-28 | Siemens Corporate Research, Inc. | Statistical modeling and performance characterization of a real-time dual camera surveillance system |
US20030026588A1 (en) * | 2001-05-14 | 2003-02-06 | Elder James H. | Attentive panoramic visual sensor |
Cited By (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8045047B2 (en) * | 2005-06-23 | 2011-10-25 | Nokia Corporation | Method and apparatus for digital image processing of an image having different scaling rates |
US20060290796A1 (en) * | 2005-06-23 | 2006-12-28 | Nokia Corporation | Digital image processing |
US8264586B2 (en) * | 2006-06-13 | 2012-09-11 | Panasonic Corporation | Imaging apparatus |
US20090189993A1 (en) * | 2006-06-13 | 2009-07-30 | Panasonic Corporation | Imaging apparatus |
US20080170129A1 (en) * | 2007-01-17 | 2008-07-17 | Samsung Techwin Co., Ltd. | Digital photographing apparatus and method for controlling the same |
US8368764B2 (en) * | 2007-01-17 | 2013-02-05 | Samsung Electronics Co., Ltd. | Digital photographing apparatus and method for controlling the same |
US9467647B2 (en) * | 2007-07-17 | 2016-10-11 | Carnegie Mellon University | Multiple resolution video network with context based control |
US20100231734A1 (en) * | 2007-07-17 | 2010-09-16 | Yang Cai | Multiple resolution video network with context based control |
US20100283843A1 (en) * | 2007-07-17 | 2010-11-11 | Yang Cai | Multiple resolution video network with eye tracking based control |
US8565190B2 (en) | 2007-07-27 | 2013-10-22 | Sony Computer Entertainment Inc. | NAT traversal for mobile network devices |
US20110200009A1 (en) * | 2007-07-27 | 2011-08-18 | Sony Computer Entertainment Inc. | Nat traversal for mobile network devices |
USRE47566E1 (en) | 2007-07-27 | 2019-08-06 | Sony Interactive Entertainment Inc. | NAT traversal for mobile network devices |
US20090041368A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US7859572B2 (en) * | 2007-08-06 | 2010-12-28 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090040323A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US8063941B2 (en) * | 2007-08-06 | 2011-11-22 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090040322A1 (en) * | 2007-08-06 | 2009-02-12 | Microsoft Corporation | Enhancing digital images using secondary optical systems |
US20090129693A1 (en) * | 2007-11-15 | 2009-05-21 | Bloebaum L Scott | System and method for generating a photograph with variable image quality |
WO2010008705A3 (en) * | 2008-06-24 | 2010-03-11 | Precoad, Inc. Et Al. | Providing and displaying video at multiple resolution and quality levels |
WO2010008705A2 (en) * | 2008-06-24 | 2010-01-21 | Precoad, Inc. Et Al. | Providing and displaying video at multiple resolution and quality levels |
US20090320081A1 (en) * | 2008-06-24 | 2009-12-24 | Chui Charles K | Providing and Displaying Video at Multiple Resolution and Quality Levels |
FR2953088A1 (en) * | 2009-11-26 | 2011-05-27 | Defiboat Technology | Method for transmitting image data e.g. video stream, involves transmitting image data to control room based on received request, and displaying image data transmitted by cameras simultaneously in distinct display windows |
US9619861B2 (en) * | 2012-04-04 | 2017-04-11 | Samsung Electronics Co., Ltd | Apparatus and method for improving quality of enlarged image |
US20130265311A1 (en) * | 2012-04-04 | 2013-10-10 | Samsung Electronics Co., Ltd. | Apparatus and method for improving quality of enlarged image |
US10255889B2 (en) | 2014-12-29 | 2019-04-09 | Beijing Zhigu Rui Tuo Tech Co., Ltd. | Light field display control methods and apparatuses, light field display devices |
WO2016107329A1 (en) * | 2014-12-29 | 2016-07-07 | Beijing Zhigu Rui Tuo Tech Co., Ltd. | Light field display control methods and apparatuses, light field display devices |
EP3107279A1 (en) | 2015-06-15 | 2016-12-21 | Coherent Synchro, S.L. | Method, device and installation for composing a video signal |
WO2016203081A1 (en) | 2015-06-15 | 2016-12-22 | Coherent Synchro, S.L. | Method, device and installation for composing a video signal |
KR102501127B1 (en) | 2015-06-15 | 2023-02-16 | 코히런트 싱크로 에스엘 | Method, device and installation for composing a video signal |
KR20180018727A (en) * | 2015-06-15 | 2018-02-21 | 코히런트 싱크로 에스엘 | Method, device and installation for composing a video signal |
CN107750453A (en) * | 2015-06-15 | 2018-03-02 | 相干同步公司 | For forming the method, equipment and erecting device of vision signal |
US10567676B2 (en) * | 2015-06-15 | 2020-02-18 | Coherent Synchro, S.L. | Method, device and installation for composing a video signal |
US20170294033A1 (en) * | 2016-04-06 | 2017-10-12 | Varex Imaging Corporation | Dose efficient x-ray detector and method |
US9936129B2 (en) * | 2016-06-15 | 2018-04-03 | Obsidian Sensors, Inc. | Generating high resolution images |
US20180098026A1 (en) * | 2016-10-04 | 2018-04-05 | Avaya Inc. | System and Method for Processing Digital Images During Videoconference |
US9936162B1 (en) * | 2016-10-04 | 2018-04-03 | Avaya Inc. | System and method for processing digital images during videoconference |
US9774823B1 (en) | 2016-10-04 | 2017-09-26 | Avaya Inc. | System and method for processing digital images during videoconference |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20070076099A1 (en) | Device and method for hybrid resolution video frames | |
US9398214B2 (en) | Multiple view and multiple object processing in wide-angle video camera | |
US9602700B2 (en) | Method and system of simultaneously displaying multiple views for video surveillance | |
US7834907B2 (en) | Image-taking apparatus and image processing method | |
US7161615B2 (en) | System and method for tracking objects and obscuring fields of view under video surveillance | |
US20050007478A1 (en) | Multiple-view processing in wide-angle video camera | |
US8994785B2 (en) | Method for generating video data and image photographing device thereof | |
GB2411310A (en) | Image stabilisation using field of view and image analysis. | |
CN107113376A (en) | A kind of image processing method, device and video camera | |
JP2006352851A (en) | Method and device for acquiring image of scene using composite camera | |
JP2006262030A (en) | Angle of view adjusting apparatus, camera system, and angle of view adjusting method | |
SG191198A1 (en) | Imaging system for immersive surveillance | |
CN108717704B (en) | Target tracking method based on fisheye image, computer device and computer readable storage medium | |
KR101916419B1 (en) | Apparatus and method for generating multi-view image from wide angle camera | |
KR101778744B1 (en) | Monitoring system through synthesis of multiple camera inputs | |
US20150116370A1 (en) | Method for successively displaying sections of screen and computer-readable medium | |
KR20180017591A (en) | Camera apparatus, display apparatus and method of correcting a movement therein | |
KR20140055541A (en) | Apparatus for monitoring image | |
US8169495B2 (en) | Method and apparatus for dynamic panoramic capturing | |
JP2000341574A (en) | Camera device and camera control system | |
KR102009988B1 (en) | Method for compensating image camera system for compensating distortion of lens using super wide angle camera and Transport Video Interface Apparatus used in it | |
GB2347575A (en) | Synthesizing a zoomed out image using several overlapping images | |
US20100045813A1 (en) | Digital image capture device and video capturing method thereof | |
US20200252548A1 (en) | Method of generating a digital video image using a wide-angle field of view lens | |
JPH08305841A (en) | Distorted image correcting display device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DVTEL, INC., NEW JERSEY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ESHED, EYAL;KIDRON, BEN;THOMPSON, EDWIN;REEL/FRAME:017948/0614;SIGNING DATES FROM 20060516 TO 20060517 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: SQUARE 1 BANK, NORTH CAROLINA Free format text: SECURITY AGREEMENT;ASSIGNOR:DVTEL, INC.;REEL/FRAME:030661/0033 Effective date: 20130430 |
|
AS | Assignment |
Owner name: DVTEL, INC., NEW JERSEY Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:PACIFIC WESTERN BANK, AS SUCCESSOR IN INTEREST BY MERGER TO SQUARE 1 BANK;REEL/FRAME:037377/0892 Effective date: 20151201 Owner name: DVTEL, LLC, NEW JERSEY Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:PACIFIC WESTERN BANK, AS SUCCESSOR IN INTEREST BY MERGER TO SQUARE 1 BANK;REEL/FRAME:037377/0892 Effective date: 20151201 |