WO2000031981A1 - Extraction of foreground information for stereoscopic video coding - Google Patents
Extraction of foreground information for stereoscopic video coding Download PDFInfo
- Publication number
- WO2000031981A1 WO2000031981A1 PCT/EP1999/008243 EP9908243W WO0031981A1 WO 2000031981 A1 WO2000031981 A1 WO 2000031981A1 EP 9908243 W EP9908243 W EP 9908243W WO 0031981 A1 WO0031981 A1 WO 0031981A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- foreground
- images
- information
- stereo pair
- pixel information
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/174—Segmentation; Edge detection involving the use of two or more images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/194—Transmission of image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/286—Image signal generators having separate monoscopic and stereoscopic modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0081—Depth or disparity estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0092—Image segmentation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0096—Synchronisation or controlling aspects
Definitions
- the invention relates in general to image processing and in particular to the extraction and variable bit rate encoding of foreground and background information from a stereo pair of images for video conferencing applications.
- the bandwidth of communication between the participants is typically limited, about 64 kilo bits per second for a telephone line connection.
- Better compression standards have been developed over the years for efficiently compressing low-bitrate audio and video data, for example H.263 and MPEG-4.
- a majority of the picture data in any given scene consists of irrelevant information, for example objects in the background. Compression algorithms cannot distinguish between relevant and irrelevant objects and if all of this information is transmitted on a low bandwidth channel, the result is a delayed jumpy looking video of a video conference participant.
- the problems with such systems is that the background looks artificial since it lacks all motion and the contour of the video conference participant must be defined with a certain degree of accuracy.
- the encoder which is typically optimized for a rectangular image such as an 8 x 8 block of DCT coefficients must encode an oddly shaped image which follows the contour of the video conference participant. This "oddly" shaped information must also be transmitted separately which is a load on both bandwidth and computational resources at both the encoder and decoder sides.
- This object is achieved by the use of a pair of cameras arranged such that each camera has a slightly different view of the scene. Two images are produced and the difference in location of corresponding matching pixels in each image is computed and the disparity in location of these pixels is determined. A small disparity between the location of two identical pixels indicates the pixels constitute background information. A large disparity indicates the pixels constitute foreground information. The foreground pixels are then transmitted at the higher bit rate while the background pixels are transmitted at the lower bit rate.
- This object is achieved by using the 8 x 8 DCT blocks of coefficients to define the contour. Any block that includes a predefined number of foreground pixels is encoded at the higher bit rate, while those blocks that fall below this predefined number are encoded at the lower bit rate.
- Figure 1 shows a video conference scheme which uses a stereo pair of cameras
- Figures 2 A and 2B show the images that result from the cameras in Fig. 1.
- Figure 3A shows the identification of the foreground information
- Figure 3B shows the DCT blocks which are transmitted at the higher bit rate
- Figure 4 shows a block diagram of a video conference device in accordance with the invention
- Figure 5 shows a PC configured for operating the instant invention
- Figure 6 shows the internal structure of the PC in Figure 5.
- Fig. 1 shows a video conference set up in accordance with the invention.
- a video conference participant 30 sits at a desk 32 in front of two cameras 10 and 20 slightly spaced from one another.
- a computer 40 In the background there is a computer 40, a door 50 with people walking m and out, and a clock 60.
- the view of camera 10 is shown m Fig. 2A as follows: the video conference participant 30 is positioned to the ⁇ ght of the lens of camera 10, the computer 40 since it is a distance from the cameras it remains basically in the center of the image.
- the door 50 is in the ⁇ ght hand portion of the image.
- the clock 60 is in the left hand corner of the image.
- the view of camera 20 is shown in Fig. 2B as follows:
- the video conference participant 30 is off to the left in the image.
- the clock 60 is to the left of the video conference participant 30.
- the computer 40 is to the ⁇ ght of the video conference participant 30 but still remains basically in the center of the image.
- the door 50 is in the upper ⁇ ght hand comer of the image
- the images received from the two cameras are compared to locate pixels of foreground information.
- the image from the left camera 10 (image A) is compared to the image from the ⁇ ght camera 20 (image B).
- the scan lines are lined up, e.g. scan line 19 of image A matches scan line 19 of image B.
- a pixel on scan line 19 of image A is then matched to its corresponding pixel in scan line 19 of image B.
- a dispa ⁇ ty threshold is then chosen, e g. 7, and any dispa ⁇ ty above the threshold 7 indicates the pixel is foreground information while any dispa ⁇ ty below 7 indicates the pixel is background information.
- image B and another block of data which is of the same size as the image data and indicates which pixels are foreground pixels, e.g. T, and which are background pixels, e.g. '0'.
- DCT block classifier 52 which creates 8 x 8 DCT blocks of the image and also binary blocks which indicate which DCT blocks of the image are foreground and which are background information.
- the block will either be identified to the encoder 56 as a foreground block (triggering high bit-rate encoding 56A) or a background block (triggering low bit-rate encoding 56B).
- Fig. 3A shows image B with the dashed lines representing the information that is encoded as foreground information in accordance with the invention. Assume each square represents an 8 x 8 DCT block. A foreground threshold is set such that if any pixel within an 8 x 8 block is foreground information then the entire block must be encoded as foreground information.
- the dashed lines in Fig. 3A indicate the DCT blocks identified as foreground information, these blocks will be encoded with a finer quantization level.
- Fig. 3B shows a binary DCT disparity block which is the output of DCT block classifier 52.
- Encoder 56 receives both the image B and the binary DCT disparity blocks. Any DCT block which corresponds to a logic '1' DCT disparity block is encoded finely. Any DCT block which corresponds to a logic '0' DCT disparity block is encoded coarsely. The result is most of the bandwidth of the channel is dedicated to the foreground information and only a small portion allocated to background information.
- a decoder 58 (shown in Fig.4) receives the bitstream and decodes it according to the quantization levels provided in the bitstream.
- This invention has applications wherever there is a transmission of moving images over a network such as the Internet, telephone lines, videomail, video phones, digital television receivers etc.
- the invention is implemented on a digital television platform using a Trimedia processor for processing and the television monitor for display.
- the invention can also be implemented similarly on a personal computer.
- Figure 5 shows a representative embodiment of a computer system 7 on which the present invention may be implemented.
- personal computer (“PC") 8 includes network connection 11 for interfacing to a network, such as a variable-bandwidth network or the Internet, and fax/modem connection 12 for interfacing with other remote sources such as a video camera (not shown).
- PC 8 also includes display screen 14 for displaying information (including video data) to a user, keyboard 15 for inputting text and user commands, mouse 13 for positioning a cursor on display screen 14 and for inputting user commands, disk d ⁇ ve 16 for reading from and w ⁇ ting to floppy disks installed therein, and CD-ROM d ⁇ ve 17 for accessing information stored on CD-ROM PC 8 may also have one or more pe ⁇ pheral devices attached thereto, such as a pair of video conference cameras for inputting images, or the like, and p ⁇ nter 19 for outputting images, text, or the like.
- pe ⁇ pheral devices attached thereto, such as a pair of video conference cameras for inputting images, or the like, and p ⁇ nter 19 for outputting images, text, or the like.
- FIG 6 shows the internal structure of PC 8.
- PC 8 includes memory 25, which comp ⁇ ses a computer-readable medium such as a computer hard disk.
- Memory 25 stores data 23, applications 25, p ⁇ nt d ⁇ ver 24, and operating system 26.
- operating system 26 is a windowing operating system, such as Microsoft® Wmdows95; although the invention may be used with other operating systems as well.
- applications stored in applications area 51 of the memory 25 are foreground information detector/DCT block classifier/video coder 21 ('video coder 21') and video decoder 22.
- Video coder 21 performs video data encoding in the manner set forth in detail above, and video decoder 22 decodes video data which has been coded in the manner presc ⁇ bed by video coder 21. The operation of these applications has been desc ⁇ bed in detail above.
- PC 8 Also included in PC 8 are display interface 29, keyboard interface 41, mouse interface 31, disk d ⁇ ve interface 42, CD-ROM d ⁇ ve interface 34, computer bus 36, RAM 37, processor 38, and p ⁇ nter interface 43
- Processor 38 preferably comp ⁇ ses a microprocessor or the like for executing applications, such those noted above, out of RAM 37.
- Such applications may be stored memory 25 (as noted above) or, alternatively, on a floppy disk in disk d ⁇ ve 16 or a CD-ROM in CD-ROM d ⁇ ve 17
- Processor 38 accesses applications (or other data) stored on a floppy disk via disk d ⁇ ve interface 32 and accesses applications (or other data) stored on a CD-ROM via CD-ROM d ⁇ ve interface 34.
- Application execution and other tasks of PC 8 may be initiated using keyboard 15 or mouse 13, commands from which are transmitted to processor 38 via keyboard interface 41 and mouse interface 31, respectively
- Output results from applications running on PC 8 may be processed by display interface 29 and then displayed to a user on display 14 or, alternatively, output via network connection 11
- input video data which has been coded by video coder 21 is typically output via network connection 11
- coded video data which has been received from, e.g , a vanable bandwidth-network is decoded by video decoder 22 and then displayed on display 14
- display interface 29 preferably comprises a display processor for forming video images based on decoded video data provided by processor 38 over computer bus 36, and for outputting those images to display 14.
- Output results from other applications, such as word processing programs, running on PC 8 may be provided to printer 19 via printer interface 43.
- Processor 38 executes print driver 24 so as to perform appropriate formatting of such print jobs prior to their transmission to printer 19.
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020007007936A KR100669837B1 (en) | 1998-11-20 | 1999-10-27 | Extraction of foreground information for stereoscopic video coding |
JP2000584695A JP2002531020A (en) | 1998-11-20 | 1999-10-27 | Foreground information extraction method in stereoscopic image coding |
EP99972820A EP1050169A1 (en) | 1998-11-20 | 1999-10-27 | Extraction of foreground information for stereoscopic video coding |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/196,574 US20020051491A1 (en) | 1998-11-20 | 1998-11-20 | Extraction of foreground information for video conference |
US09/196,574 | 1998-11-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2000031981A1 true WO2000031981A1 (en) | 2000-06-02 |
Family
ID=22725937
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP1999/008243 WO2000031981A1 (en) | 1998-11-20 | 1999-10-27 | Extraction of foreground information for stereoscopic video coding |
Country Status (5)
Country | Link |
---|---|
US (1) | US20020051491A1 (en) |
EP (1) | EP1050169A1 (en) |
JP (1) | JP2002531020A (en) |
KR (1) | KR100669837B1 (en) |
WO (1) | WO2000031981A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2999221A4 (en) * | 2013-08-19 | 2016-06-22 | Huawei Tech Co Ltd | Image processing method and device |
WO2019219065A1 (en) * | 2018-05-17 | 2019-11-21 | 杭州海康威视数字技术股份有限公司 | Video analysis method and device |
EP3605471A4 (en) * | 2017-12-14 | 2021-01-06 | Canon Kabushiki Kaisha | System, method, and program for generating virtual viewpoint image |
Families Citing this family (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4670303B2 (en) * | 2004-10-06 | 2011-04-13 | ソニー株式会社 | Image processing method and image processing apparatus |
JP4251650B2 (en) * | 2005-03-28 | 2009-04-08 | 株式会社カシオ日立モバイルコミュニケーションズ | Image processing apparatus and program |
US9158975B2 (en) * | 2005-05-31 | 2015-10-13 | Avigilon Fortress Corporation | Video analytics for retail business process monitoring |
US20090316777A1 (en) * | 2008-06-20 | 2009-12-24 | Xin Feng | Method and Apparatus for Improved Broadcast Bandwidth Efficiency During Transmission of a Static Code Page of an Advertisement |
US9729899B2 (en) | 2009-04-20 | 2017-08-08 | Dolby Laboratories Licensing Corporation | Directed interpolation and data post-processing |
US9628722B2 (en) | 2010-03-30 | 2017-04-18 | Personify, Inc. | Systems and methods for embedding a foreground video into a background feed based on a control input |
US8649592B2 (en) | 2010-08-30 | 2014-02-11 | University Of Illinois At Urbana-Champaign | System for background subtraction with 3D camera |
US9171075B2 (en) | 2010-12-30 | 2015-10-27 | Pelco, Inc. | Searching recorded video |
US9049447B2 (en) * | 2010-12-30 | 2015-06-02 | Pelco, Inc. | Video coding |
US9681125B2 (en) * | 2011-12-29 | 2017-06-13 | Pelco, Inc | Method and system for video coding with noise filtering |
US9414016B2 (en) | 2013-12-31 | 2016-08-09 | Personify, Inc. | System and methods for persona identification using combined probability maps |
US9485433B2 (en) | 2013-12-31 | 2016-11-01 | Personify, Inc. | Systems and methods for iterative adjustment of video-capture settings based on identified persona |
US9916668B2 (en) | 2015-05-19 | 2018-03-13 | Personify, Inc. | Methods and systems for identifying background in video data using geometric primitives |
US9563962B2 (en) | 2015-05-19 | 2017-02-07 | Personify, Inc. | Methods and systems for assigning pixels distance-cost values using a flood fill technique |
US9607397B2 (en) | 2015-09-01 | 2017-03-28 | Personify, Inc. | Methods and systems for generating a user-hair-color model |
US9883155B2 (en) | 2016-06-14 | 2018-01-30 | Personify, Inc. | Methods and systems for combining foreground video and background video using chromatic matching |
CN107662872B (en) * | 2016-07-29 | 2021-03-12 | 奥的斯电梯公司 | Monitoring system and monitoring method for passenger conveyor |
US9881207B1 (en) | 2016-10-25 | 2018-01-30 | Personify, Inc. | Methods and systems for real-time user extraction using deep learning networks |
KR20190004010A (en) * | 2017-07-03 | 2019-01-11 | 삼성에스디에스 주식회사 | Method and Apparatus for extracting foreground |
GB201717011D0 (en) * | 2017-10-17 | 2017-11-29 | Nokia Technologies Oy | An apparatus a method and a computer program for volumetric video |
GB2595679A (en) * | 2020-06-02 | 2021-12-08 | Athlone Institute Of Tech | Video storage system |
US11800056B2 (en) | 2021-02-11 | 2023-10-24 | Logitech Europe S.A. | Smart webcam system |
US11800048B2 (en) | 2021-02-24 | 2023-10-24 | Logitech Europe S.A. | Image generating system with background replacement or modification capabilities |
US11831696B2 (en) | 2022-02-02 | 2023-11-28 | Microsoft Technology Licensing, Llc | Optimizing richness in a remote meeting |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0731608A2 (en) * | 1995-03-10 | 1996-09-11 | Sharp Kabushiki Kaisha | Image encoder and decoder with area selection |
WO1997024000A1 (en) * | 1995-12-22 | 1997-07-03 | Xenotech Research Pty. Ltd. | Image conversion and encoding techniques |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0330455A3 (en) * | 1988-02-22 | 1990-07-04 | Kabushiki Kaisha Toshiba | Image encoding apparatus |
DE4118571A1 (en) * | 1991-06-06 | 1992-12-10 | Philips Patentverwaltung | DEVICE FOR CONTROLLING THE QUANTIZER OF A HYBRID ENCODER |
JP3258840B2 (en) * | 1994-12-27 | 2002-02-18 | シャープ株式会社 | Video encoding device and region extraction device |
US5710829A (en) * | 1995-04-27 | 1998-01-20 | Lucent Technologies Inc. | System and method for focused-based image segmentation for video signals |
US5832115A (en) * | 1997-01-02 | 1998-11-03 | Lucent Technologies Inc. | Ternary image templates for improved semantic compression |
-
1998
- 1998-11-20 US US09/196,574 patent/US20020051491A1/en not_active Abandoned
-
1999
- 1999-10-27 EP EP99972820A patent/EP1050169A1/en not_active Withdrawn
- 1999-10-27 JP JP2000584695A patent/JP2002531020A/en not_active Withdrawn
- 1999-10-27 WO PCT/EP1999/008243 patent/WO2000031981A1/en not_active Application Discontinuation
- 1999-10-27 KR KR1020007007936A patent/KR100669837B1/en not_active IP Right Cessation
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0731608A2 (en) * | 1995-03-10 | 1996-09-11 | Sharp Kabushiki Kaisha | Image encoder and decoder with area selection |
WO1997024000A1 (en) * | 1995-12-22 | 1997-07-03 | Xenotech Research Pty. Ltd. | Image conversion and encoding techniques |
Non-Patent Citations (1)
Title |
---|
TZOVARAS D ET AL: "THREE-DIMENSIONAL CAMERA MOTION ESTIMATION AND FOREGROUND/BACKGROUND SEPARATION FOR STEREOSCOPIC IMAGE SEQUENCES", OPTICAL ENGINEERING,US,SOC. OF PHOTO-OPTICAL INSTRUMENTATION ENGINEERS. BELLINGHAM, vol. 36, no. 2, 1 February 1997 (1997-02-01), pages 574 - 579, XP000686883, ISSN: 0091-3286 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2999221A4 (en) * | 2013-08-19 | 2016-06-22 | Huawei Tech Co Ltd | Image processing method and device |
US9392218B2 (en) | 2013-08-19 | 2016-07-12 | Huawei Technologies Co., Ltd. | Image processing method and device |
EP3605471A4 (en) * | 2017-12-14 | 2021-01-06 | Canon Kabushiki Kaisha | System, method, and program for generating virtual viewpoint image |
CN112489182A (en) * | 2017-12-14 | 2021-03-12 | 佳能株式会社 | System, method, and storage medium for generating an image |
WO2019219065A1 (en) * | 2018-05-17 | 2019-11-21 | 杭州海康威视数字技术股份有限公司 | Video analysis method and device |
Also Published As
Publication number | Publication date |
---|---|
US20020051491A1 (en) | 2002-05-02 |
JP2002531020A (en) | 2002-09-17 |
KR100669837B1 (en) | 2007-01-18 |
EP1050169A1 (en) | 2000-11-08 |
KR20010034256A (en) | 2001-04-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100669837B1 (en) | Extraction of foreground information for stereoscopic video coding | |
US8295350B2 (en) | Image coding apparatus with segment classification and segmentation-type motion prediction circuit | |
JP3197420B2 (en) | Image coding device | |
EP0909096B1 (en) | Picture encoder and picture decoder | |
US20030058939A1 (en) | Video telecommunication system | |
CA2177866A1 (en) | Automatic face and facial feature location detection for low bit rate model-assisted h.261 compatible coding of video | |
CN112954398B (en) | Encoding method, decoding method, device, storage medium and electronic equipment | |
CN103716643A (en) | System and method for improving video encoding using content information | |
US7489728B2 (en) | Apparatus and method for coding moving image | |
JP2002125233A (en) | Image compression system for weighting video contents | |
EP1747674A1 (en) | Image compression for transmission over mobile networks | |
KR20190023546A (en) | Video encoding apparatus and video encoding system | |
US9986257B2 (en) | Method of lookup table size reduction for depth modelling mode in depth coding | |
KR100575733B1 (en) | Method for segmenting motion object of compressed motion pictures | |
CN114387440A (en) | Video clipping method and device and storage medium | |
NO20200708A1 (en) | Method, computer program and system for detecting changes and moving objects in a video view | |
Strutz | Improved probability modelling for exception handling in lossless screen content coding | |
JPH0998416A (en) | Encoder for image signal and recognition device for image | |
JP2828977B2 (en) | Video encoding device | |
KR102320315B1 (en) | Method and Apparatus of Encoding Tile based on Region of Interest for Tiled Streaming | |
CN110784716B (en) | Media data processing method, device and medium | |
US20020009140A1 (en) | Method of encoding video signals | |
Strat | Object-based encoding: next-generation video compression | |
CN114422794A (en) | Dynamic video definition processing method based on front camera | |
JPH0767107A (en) | Image encoder |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): JP KR |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020007007936 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1999972820 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWP | Wipo information: published in national office |
Ref document number: 1999972820 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1020007007936 Country of ref document: KR |
|
WWG | Wipo information: grant in national office |
Ref document number: 1020007007936 Country of ref document: KR |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 1999972820 Country of ref document: EP |