US20090263106A1 - Text subtitle decoder and method for decoding text subtitle streams - Google Patents
Text subtitle decoder and method for decoding text subtitle streams Download PDFInfo
- Publication number
- US20090263106A1 US20090263106A1 US12/487,992 US48799209A US2009263106A1 US 20090263106 A1 US20090263106 A1 US 20090263106A1 US 48799209 A US48799209 A US 48799209A US 2009263106 A1 US2009263106 A1 US 2009263106A1
- Authority
- US
- United States
- Prior art keywords
- text
- region
- style
- information
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000009877 rendering Methods 0.000 claims abstract description 43
- 239000000872 buffer Substances 0.000 claims description 28
- 230000036316 preload Effects 0.000 claims description 6
- 239000012160 loading buffer Substances 0.000 abstract description 4
- 230000003287 optical effect Effects 0.000 description 25
- 230000008901 benefit Effects 0.000 description 5
- 230000002452 interceptive effect Effects 0.000 description 3
- 230000001360 synchronised effect Effects 0.000 description 3
- 102000009913 Peroxisomal Targeting Signal 2 Receptor Human genes 0.000 description 2
- 108010077056 Peroxisomal Targeting Signal 2 Receptor Proteins 0.000 description 2
- 101100465559 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) PRE7 gene Proteins 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 101150076896 pts1 gene Proteins 0.000 description 2
- GUGNSJAORJLKGP-UHFFFAOYSA-K sodium 8-methoxypyrene-1,3,6-trisulfonate Chemical compound [Na+].[Na+].[Na+].C1=C2C(OC)=CC(S([O-])(=O)=O)=C(C=C3)C2=C2C3=C(S([O-])(=O)=O)C=C(S([O-])(=O)=O)C2=C1 GUGNSJAORJLKGP-UHFFFAOYSA-K 0.000 description 2
- 238000000605 extraction Methods 0.000 description 1
- 238000001824 photoionisation detection Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/10527—Audio or video recording; Data buffering arrangements
- G11B2020/1062—Data buffering arrangements, e.g. recording or playback buffers
- G11B2020/1075—Data buffering arrangements, e.g. recording or playback buffers the usage of the buffer being restricted to a specific kind of data
- G11B2020/10759—Data buffering arrangements, e.g. recording or playback buffers the usage of the buffer being restricted to a specific kind of data content data
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2541—Blu-ray discs; Blue laser DVR discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/806—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
- H04N9/8063—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8233—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a character code signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8233—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a character code signal
- H04N9/8244—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a character code signal involving the use of subcodes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/87—Regeneration of colour television signals
- H04N9/8715—Regeneration of colour television signals involving the mixing of the reproduced video signal with a non-recorded signal, e.g. a text signal
Definitions
- the present invention relates to a text subtitle decoder and a method for decoding text subtitle streams recorded on a recording medium, an example of which is a Blu-ray disc (BD).
- BD Blu-ray disc
- Optical discs are widely used as an optical recording medium for recording mass data.
- a new high-density digital video disc hereinafter referred to as “HD-DVD”
- HD-DVD high-density digital video disc
- BD Blu-ray Disc
- BD global standard technical specifications of the Blu-ray Disc (BD)
- BD next generation HD-DVD technology
- optical reproducing apparatuses having the Blu-ray Disc (BD) standards applied thereto are also being developed.
- the Blu-ray Disc (BD) standards are yet to be completed, there have been many difficulties in developing a complete optical reproducing apparatus.
- the main AV data as well as various data required for a user's convenience, such as subtitle information as the supplementary data related to the main AV data, be provided, but also managing information for reproducing the main data and the subtitle data recorded in the optical disc should be systemized and provided.
- the present invention is directed to a text subtitle decoder and a method for decoding text subtitle streams recorded on a recording medium that substantially obviates one or more problems due to limitations and disadvantages of the related art.
- An object of the present invention is to provide a method and a text subtitle decoder for decoding a text subtitle stream recorded on a recording medium, which includes text strings for each dialog region and composition and rendering information required for decoding the text strings.
- a method for decoding a text subtitle stream recorded on a recording medium is provided. Initially, a text subtitle stream recorded on the recording medium is loaded into a subtitle loading buffer, where the text subtitle stream includes a dialog style segment and a dialog presentation segment.
- the dialog style segment defines a group of region styles, and the dialog presentation segment includes dialog presentation information and dialog text data for dialog having at least one region. Thereafter, the dialog presentation segment is parsed into composition information, rendering information, and the dialog text data for each region.
- the parsed composition information and rendering information are then stored in a first buffer, and the parsed dialog text data are stored in a second buffer, where the dialog text data stored in the second buffer includes one or more text strings for each region.
- the text strings stored in the second buffer are rendered into a bitmap object for each region according to the rendering information, and the rendered bitmap object is stored into a third buffer.
- the stored bitmap object is composed in a graphics plane for each region according to the composition information.
- a text subtitle decoder for decoding a text subtitle stream recorded on a recording medium includes a subtitle loading buffer, a text subtitle processor, a dialog composition buffer, a dialog buffer, a text renderer, a bitmap object buffer, and a graphics plane.
- the subtitle loading buffer initially loads the text subtitle stream, which includes a dialog style segment defining a group of region styles and a dialog presentation segment including dialog presentation information and dialog text data for a dialog having at least one region.
- the text subtitle processor parses the dialog presentation segment into composition information, rendering information, and the dialog text data for each region.
- the dialog composition buffer stores the composition and rendering information parsed from the text subtitle processor
- the dialog buffer stores the dialog text data, which includes one or more text strings for each region.
- each bitmap object stored in the bitmap object buffer is composed in the graphics plane according to the composition information.
- an optical disc player for reproducing text subtitle streams recorded on an optical disc includes an audio decoder configured to decode audio streams recorded on the optical disc into audio data, a video decoder configured to decode video streams recorded on the optical disc into video image data, a text subtitle decoder configured to decode a text subtitle stream recorded on the optical disc into text subtitle image data, and an image superimposition unit configured to superimpose the decoded text subtitle image data with the decoded video image data.
- the text subtitle decoder includes a text subtitle processor, a text renderer, and a graphics plane.
- the text subtitle processor initially parses the text subtitle stream into composition information, rendering information, and dialog text data for a dialog having at least one region, where the dialog text data include one or more text strings for each region.
- the text renderer renders the text strings into graphic data for each region according to the rendering information, and the graphics plane composes the rendered graphic data according to the composition information.
- FIG. 1 illustrates a structure of the data files recorded in an optical disc according to the present invention
- FIG. 2 illustrates data storage areas of an optical disc according to the present invention
- FIG. 3 illustrates a text subtitle and a main image presented on a display screen according to the present invention
- FIG. 4 illustrates a schematic diagram illustrating reproduction control of a main AV clip and text subtitle clips according to the present invention
- FIG. 5A illustrates a dialog presented on a display screen according to the present invention
- FIG. 5B illustrates regions of a dialog presented on a display screen according to the present invention
- FIG. 5C illustrates style information for regions of a dialog according to the present invention
- FIG. 6A illustrates presentations of text subtitle dialogs on a display screen in presentation time stamps
- FIG. 6B illustrates continuities between text subtitle dialogs being presented on a display screen
- FIG. 7A illustrates the structure of a text subtitle stream file according to the present invention
- FIG. 7B illustrates specific information contained in a DPU and a DSU included in a text subtitle stream according to the present invention
- FIG. 8 illustrates an apparatus for decoding main AV streams and text subtitle streams according to the present invention
- FIG. 9 illustrates a text subtitle decoder according to the present invention.
- FIG. 10 illustrates a graphical representation illustrating parsing, rendering, and composition processes performing by a text subtitle decoder according to the present invention.
- FIG. 11 illustrates operation times of various components included in a text subtitle decoder according to the present invention.
- main data represent audio/video (AV) data that belong to a title (e.g., a movie title) recorded in an optical disc by an author.
- the AV data are recorded in MPEG2 format and are often called AV streams or main AV streams.
- supplementary data represent all other data required for reproducing the main data, examples of which are text subtitle streams, interactive graphic streams, presentation graphic streams, and supplementary audio streams (e.g., for a browsable slideshow).
- Theses supplementary data streams may be recorded in MPEG2 format or in any other data format. They could be multiplexed with the AV streams or could exist as independent data files within the optical disc.
- a subtitle represents caption information corresponding to video (image) data being reproduced, and it may be represented in a predetermined language. For example, when a user selects an option for viewing one of a plurality of subtitles represented in various languages while viewing images on a display screen, the caption information corresponding to the selected subtitle is displayed on a predetermined portion of the display screen. If the displayed caption information is text data (e.g., characters), the selected subtitle is often called a text subtitle.
- a plurality of text subtitle streams in MPEG2 format may be recorded in an optical disc, and they may exist as a plurality of independent stream files. Each text subtitle stream file includes text data for a text subtitle and reproduction control data required for reproduction of the text data.
- only a single text subtitle stream in MPEG2 format may be recorded in an optical disc.
- FIG. 1 illustrates a file structure of the data files recorded in a Blu-ray disc (hereinafter “BD”) according to the present invention.
- BD Blu-ray disc
- FIG. 1 illustrates a file structure of the data files recorded in a Blu-ray disc (hereinafter “BD”) according to the present invention.
- BDMV Blu-ray disc
- index.bdmv an index file
- object file An object file
- the index file may contain data representing an index table having a plurality of selectable menus and movie titles.
- Each BD directory further includes four file directories that include audio/video (AV) data to be reproduced and various data required for reproduction of the AV data.
- AV audio/video
- the file directories included in each BD directory are a stream directory (STREAM), a clip information directory (CLIPINF), a playlist directory (PLAYLIST), and an auxiliary data directory (AUX DATA).
- the stream directory (STREAM) includes audio/video (AV) stream files having a particular data format.
- the AV stream files may be in the form of MPEG2 transport packets and be named as “*.m2ts”, as shown in FIG. 1 .
- the stream directory may further include one or more text subtitle stream files, where each text subtitle stream file includes text (e.g., characters) data for a text subtitle represented in a particular language and reproduction control information of the text data.
- the text subtitle stream files exist as independent stream files within the stream directory and may be named as “*.m2ts” or “*.txtst”, as shown in FIG. 1 .
- An AV stream file or text subtitle stream file included in the stream directory is often called a clip stream file.
- the clip information directory includes clip information files that correspond to the stream files (AV or text subtitle) included in the stream directory, respectively.
- Each clip information file contains property and reproduction timing information of a corresponding stream file.
- a clip information file may includes mapping information, in which presentation time stamps (PTS) and source packet numbers (SPN) are one-to-one mapped by an entry point map (EPM).
- PTS presentation time stamps
- SPN source packet numbers
- EPM entry point map
- a particular location of a stream file may be determined from timing information (In-Time and Out-Time) provided by a PlayItem or SubPlayItem, which will be discussed later in more details.
- each pair of a stream file and its corresponding clip information file is designated as a clip.
- 01000.clpi included in CLIPINF includes property and reproduction timing information of 01000.m2ts included in STREAM, and 01000.clpi and 01000.m2ts form a clip.
- the playlist directory includes one or more PlayList files (*.mpls), where each PlayList file includes at least one PlayItem which designates at least one main AV clip and the reproduction time of the main AV clip. More specifically, a PlayItem contains information designating In-Time and Out-Time, which represent reproduction begin and end times for a main AV clip designated by Clip_Information_File_Name within the PlayItem. Therefore, a PlayList file represents the basic reproduction control information for one or more main AV clips. In addition, the PlayList file may further include a SubPlayItem, which represents the basic reproduction control information for a text subtitle stream file.
- the main function of a SubPlayItem is to control reproduction of one or more text subtitle stream files.
- auxiliary data directory may include supplementary data stream files, examples of which are font files (e.g., aaaaa.font), pop-up menu files (not illustrated), and sound files (e.g., Sound.bdmv) for generating click sound.
- font files e.g., aaaaa.font
- pop-up menu files not illustrated
- sound files e.g., Sound.bdmv
- the text subtitle stream files mentioned earlier may be included in the auxiliary data directory instead of the stream directory.
- FIG. 2 illustrates data storage areas of an optical disc according to the present invention.
- the optical disc includes a file system information area occupying the inmost portion of the disc volume, a stream area occupying the outmost portion of the disc volume, and a database area occupied between the file system information area and the stream area.
- system information for managing the entire data files shown in FIG. 1 is stored.
- AV streams and one or more text subtitle streams are stored in the stream area.
- the general files, PlayList files, and clip information files shown in FIG. 1 are stored in the database area of the disc volume.
- the general files include an index file and an object file
- the PlayList files and clip information files include information required to reproduce the AV streams and the text subtitle streams stored in the stream area.
- FIG. 3 illustrates a text subtitle and a main image presented on a display screen according to the present invention.
- the main image and the text subtitle are simultaneously displayed on the display screen when a main AV stream and a corresponding text subtitle stream are reproduced in synchronization.
- FIG. 4 is a schematic diagram illustrating reproduction control of a main AV clip and text subtitle clips according to the present invention.
- a PlayList file includes at least one PlayItem controlling reproduction of at least one main AV clip and a SubPlayItem controlling reproduction of a plurality of text subtitle clips.
- One of text subtitle clip 1 and text subtitle clip 2 shown in FIG. 4 for English and Korean text subtitles may be synchronized with the main AV clip such that a main image and a corresponding text subtitle are displayed on a display screen simultaneously at a particular presentation time.
- display control information e.g., position and size information
- presentation time information examples of which are illustrated in FIG. 5A to FIG. 5C , are required.
- FIG. 5A illustrates a dialog presented on a display screen according to the present invention.
- a dialog represents entire text subtitle data displayed on a display screen during a given presentation time.
- presentation times of the dialog may be represented in presentation time stamps (PTS).
- PTS presentation time stamps
- presentation of the dialog shown in FIG. 5A starts at PTS (k) and ends at PTS (k+1). Therefore, the dialog shown in FIG. 5A represents an entire unit of text subtitle data which are displayed on the display screen between PTS (k) and PTS (k+1).
- a dialog includes at least one line of subtitle text (characters). When there are two or more lines of subtitle text in a dialog, entire text data may be displayed according to a style defined for the dialog.
- FIG. 5B illustrates regions of a dialog according to the present invention.
- a region represents a divided portion of text subtitle data (dialog) displayed on a display screen during a given presentation time.
- a dialog includes at least one region, and each region may include at least one line of subtitle text.
- the entire text subtitle data representing a region may be displayed on the display screen according to a region style (global style) assigned to the region.
- the maximum number of regions included in a dialog should be determined based on a desired decoding rate of the subtitle data because the greater number of regions generally results a lower decoding ratio. For example, the maximum number of regions for a dialog may be limited to two in order to achieve a reasonably high decoding rate. However, the maximum number could be greater than two for other purposes.
- FIG. 5C illustrates style information for regions of a dialog according to the present invention.
- Style information represents information defining properties required for displaying at least a portion of a region included in a dialog. Some of the examples of the style information are position, region size, background color, text alignment, text flow information, and many others.
- the style information may be classified into region style information (global style information) and inline style information (local style information).
- Region style information defines a region style (global style) which is applied to an entire region of a dialog.
- the region style information may contain at least one of a region position, region size, font color, background color, text flow, text alignment, line space, font name, font style, and font size of the region.
- region style information may contain at least one of a region position, region size, font color, background color, text flow, text alignment, line space, font name, font style, and font size of the region.
- two different region styles are applied to region 1 and region 2 , as shown in FIG. 5C .
- a region style with position 1 , size 1 , and blue background color is applied to Region 1
- a different region style with position 2 , size 2 , and red background color is applied to Region 2 .
- inline style information defines an inline style (local style) which is applied to a particular portion of text strings included in a region.
- the inline style information may contain at least one of a font type, font size, font style, and font color.
- the particular portion of text strings may be an entire text line within a region or a particular portion of the text line.
- a particular inline style is applied to the text portion “mountain” included in Region 1 .
- at least one of the font type, font size, font style, and font color of the particular portion of text strings is different from the remaining portion of the text strings within Region 1 .
- FIG. 6A illustrates presentations of text subtitle dialogs on a display screen in presentation time stamp (PTS) intervals.
- PTS presentation time stamp
- Information defining a dialog includes dialog presentation time information and dialog text data including style information and text strings to be displayed within each region of the dialog.
- An example of the presentation time information is a set of start PTS start and PTS end, and the style information includes region (global) style information and inline (local) style information described above. It is shown in FIG. 6A that different style information sets may be applied to the dialogs.
- FIG. 6B illustrates continuities between text subtitle dialogs being presented on a display screen in PTS intervals.
- a continuity exists between Dialog # 1 and Dialog # 2 .
- Display of Text # 1 in a region of Dialog # 1 is continuous with display of Text # 1 in Region 1 of Dialog # 2 .
- PTS intervals of both dialogs are continuous and same style information (region and inline) is used when presenting Text # 1 in both regions.
- another continuity exists between Dialog # 2 and Dialog # 3 because display of Text # 2 in Region 2 of Dialog # 2 is continuous with display of Text # 2 in a region of Dialog # 3 .
- presentation times (PTS intervals) of the dialogs must be continuous.
- same region and inline style information must be used when presenting the same text in the regions, respectively.
- PTS intervals are not continuous.
- An indicator e.g., continuous_presentation_flag
- continuous_presentation_flag may be included in presentation information of a current dialog to indicate whether the dialog is continuous with a previous dialog.
- TP transport packets
- PID packet identifier
- a disc player receives many input streams including a particular text subtitle stream, it finds all the transport packets that belong to the text subtitle stream using their PIDs.
- PES packet elementary stream
- One of the PES packets shown in FIG. 7A corresponds to a dialog style unit (DSU) defining a group of region styles.
- DSU dialog style unit
- a DSU is also often referred as a dialog style segment (DSS). All the remaining PES packets correspond to dialog presentation units (DPUs), each of which includes presentation information for a dialog having at least one region, and dialog text data which includes an region style indicator, inline style information, and text strings for each region.
- DPS dialog style segment
- FIG. 7B illustrates specific information contained within a DPU and a DSU included in a text subtitle stream according to the present invention.
- a DSU contains information sets defining a group of region styles, each of which is applied to a corresponding region of a dialog.
- a DPU contains dialog text data and dialog presentation information for a dialog.
- the dialog text data includes text strings to be included in each region of the dialog, inline style information to be applied to a particular portion of the text strings, and a region style identifier indicating a region style to be applied to each dialog region.
- the region style identifier identifies one of the group of region styles defined in the DSU.
- the dialog presentation information includes presentation time information and palette (color) update information for a dialog.
- the presentation time information may include presentation start time (e.g, PTS_start) and presentation end time (e.g., PTS_end) for presenting the dialog on a display screen
- the palette update information may include an indicator (e.g., palette_update_flag) indicating whether to update display colors of the dialog and palette information (e.g., Palette for update) to be applied when updating the display colors.
- All the data included in a text subtitle stream may be classified into three types of data based on their basic functions.
- the data could be classified into dialog text data, composition information, and rendering information, as shown in FIG. 7B .
- the dialog text data include text string(s), inline style information, and a region style identifier for each region of a dialog.
- the composition information includes presentation time information, examples of which are presentation start and end times, position information for a dialog region, and palette update information for a dialog.
- the rendering information includes information required for rendering the text strings to graphic data for presentation. Referring to FIG.
- the horizontal and vertical positions of each region included in the DSU is a part of the composition information, and the region width, region height, font color, background color, text flow, text alignment, line space, font name, font style, and font size included in the DSU represent the rendering information.
- the apparatus includes a packet identifier (PID) filter 5 for separating input streams into video streams, audio streams, graphic streams, and text subtitle streams based on their packet identifiers, a video decoding part 20 for decoding the video streams, an audio decoding part 10 for decoding the audio streams, a graphic decoding part 30 for decoding the graphic streams, and a text subtitle decoding part 40 for decoding the text subtitle streams.
- PID packet identifier
- the text subtitle streams may be extracted from an optical disc or from an additional external source, as shown in FIG. 8 .
- the apparatus additionally includes a switch 6 which selects an input data source. Therefore, if the text subtitle streams are extracted from the optical disc, the switch 6 selects data line A connected to the PID filter 5 . On the other hand, if they are inputted from the external source, the switch 6 selects line B connected to the external source.
- the audio decoding part 10 , video decoding part 20 , and graphic decoding part 30 include transport buffers 11 , 21 , and 31 , respectively, for storing stream data to be decoded.
- a video plane (VP) 23 and a graphic plane 33 are included in the video decoding part 20 and the graphic decoding part 30 , respectively, for converting decoded signals into displayable video and graphic images.
- the graphic decoding part 30 includes a color look up table (CLUT) 34 for controlling color and transparency levels of the displayable graphic images.
- CLUT color look up table
- the text subtitle decoding part 40 When the text subtitle decoding part 40 receives a text subtitle stream supporting a single language from the switch 6 , an entire portion of the text subtitle stream may be preloaded into a subtitle preloading buffer (SPB) 41 at once. Alternatively, when there are more than one text subtitle streams for supporting multi-languages, all the text subtitle streams may be preloaded into the SPB 41 at once. Therefore, the size of the SPB 41 should be determined based on a total number of text subtitle stream files received from the switch 6 . For example, the size of the SPB 41 should be greater than or equal to 0.5 megabytes for preloading a 0.5 megabyte text subtitle stream file.
- the size of the SPB 41 should be greater than or equal to 1 megabytes.
- the size of the SPB 42 should be large enough to preload all the required text subtitle stream files at once.
- the text subtitle decoding part 40 shown in FIG. 8 further includes a font preloading buffer (FPB) 410 for storing all the associated font files which may be included in the auxiliary data directory shown in FIG. 1 .
- FPB font preloading buffer
- the size of the FPB 410 should be large enough to preload all the required font files at once in order to ensure seamless presentation of a text subtitle supporting one or more languages. Since all the available text subtitle stream files and related font files are preloaded, extraction and use of the preloaded data can be done in a simple manner. Also the control of the SPB 41 and the FPB 410 could be quite simple due to the this reason.
- the text subtitle decoding part 40 further includes a text subtitle decoder 42 which decodes each text subtitle stream stored in the SPB 41 , a graphic plane 43 in which the decoded subtitle data are composed as displayable subtitle images, and a color look up table (CLUT) 44 controlling at least one of color and transparency levels of the converted subtitle images.
- a text subtitle decoder 42 which decodes each text subtitle stream stored in the SPB 41
- a graphic plane 43 in which the decoded subtitle data are composed as displayable subtitle images
- CLUT color look up table
- the apparatus shown in FIG. 8 further includes an image superimposition part 50 which superimposes the images outputted from the video decoding part 20 , the graphic decoding part 30 , and the text subtitle decoding part 40 . These combined images are displayed on a display screen, as shown in FIG. 3 .
- the video images outputted from the VP 23 of the video decoding part 20 may be displayed as a background of the display screen, and the images outputted from the graphic decoding part 30 and/or text subtitle decoding part 40 may be superimposed over the video images in a predetermined order.
- the output images of the graphic decoding part 30 are presentation graphic images
- these images may be initially superimposed over the video images by a first adder 52
- the text subtitle images from the text subtitle decoding part 40 may be superimposed over the video images by a second adder 53 .
- the output images of the graphic decoding part 30 are interactive graphic images
- the text subtitle images from the text subtitle decoding part 40 may be initially superimposed over the video images by the first adder 52 .
- the interactive graphic images may be further superimposed over the subtitle-superimposed images by the second adder 53 .
- the apparatus shown in FIG. 8 further includes a system decoder 4 for decoding input transport streams (e.g., MPEG transport streams), and a microprocessor 3 for controlling operations of all the components of the apparatus mentioned above.
- a system decoder 4 for decoding input transport streams (e.g., MPEG transport streams)
- a microprocessor 3 for controlling operations of all the components of the apparatus mentioned above.
- At least one of text subtitle stream files for Text Subtitle Clip 1 (English) and Text Subtitle Clip 2 (Korean) may be initially preloaded into the SPB 41 .
- all the font files related to the text subtitle stream files, which may be designated in clip information files of the text subtitle clips, are preloaded into the FPB 410 .
- playback of the PlayList is started.
- AV streams included in the main AV clip are decoded by the audio decoding part 10 and the video decoding part 20 , and one of the preloaded text subtitle streams is decoded by the text subtitle decoder 42 .
- the decoded text subtitle images are superimposed over the decoded main video images, and the entire images are displayed on a display screen.
- FIG. 9 illustrates the text subtitle decoder 42 included in the decoding apparatus shown in FIG. 8 in more details.
- the decoding process performed by the text subtitle decoder 42 includes parsing, rendering, and composition steps.
- a text subtitle stream stored in the SPB 41 is parsed into composition information, rendering information, and dialog text data.
- the rendering step text strings included in the dialog text data are rendered into graphic data (bitmap data) for each region using the rendering information, which represents information required for rendering the text strings into the graphic data.
- the rendered text strings are added into the GP 43 according to the composition information, which represents dialog presentation information, palette update information, and position information for a dialog.
- a text subtitle processor 421 included in the text subtitle decoder 42 parses a text subtitle stream preloaded in the SPB 41 into composition information, rendering information, and dialog text data. More particularly, the text subtitle processor 421 initially transfers a dialog style unit (DSU) included in the preloaded subtitle stream to a dialog composition buffer (DCB) 425 , and it parses a dialog presentation unit (DPU) further included in the preloaded text subtitle stream into composition information, rendering information, and dialog text data. The composition and rendering information are then stored in the DCB 425 , and the dialog text data are stored in a dialog buffer (DB) 422 .
- the dialog text data stored in the DB 422 include a region style identifier, text strings, and inline style information for each dialog region.
- a text renderer 423 renders the text strings stored in the DB 422 into a bitmap object (graphic data) for each dialog region under the control of a dialog presentation controller 426 .
- the text renderer 423 renders the text strings stored in the DB 422 into a bitmap object for each dialog region using the region style identifier and inline style information stored in the DB 422 , the rendering information provided from the dialog presentation controller 426 , and related font data provided from the FPB 410 .
- the bitmap object rendered by the text renderer 423 for each dialog region is then stored in a bitmap object buffer (BOB) 424 .
- BOB bitmap object buffer
- each bitmap object stored in the BOB 424 is composed within (added to) the GP 43 according to the composition information provided by the dialog presentation controller 426 .
- the CLUT 44 uses palette update information included in the composition information to adjust color and/or transparency levels of an output of the GP 43 .
- particular style information selected by a user may be applied.
- the dialog presentation controller 426 may receive such user-selected style information and provide this information to the text renderer 423 and/or the GP 43 . Examples of the user-selectable style information are a region position and a font size.
- FIG. 10 is a simple graphical representation illustrating parsing, rendering, and composition processes performed by the text subtitle decoder 42 shown in FIG. 9 .
- a DSU is initially transferred to the DCB 425 and a DPU is parsed into composition information, dialog text data, and rendering information.
- Text strings for each dialog region included in the dialog text data are rendered into a bitmap object using the rendering information, and the bitmap object rendered for each region is composed within the GP 43 according to the composition information, which may include presentation information, palette information, and region position information.
- FIG. 11 illustrates operation times of the GP 43 , BOB 424 , DCB 425 , and DB 422 in a global time axis perceived by a user.
- a decoding start time represents a time at which the text subtitle decoder 42 starts to decode a text subtitle stream preloaded in the SPB 41 . More particularly, a DST represents a time at which the DB 422 and the DCB 425 start storing dialog text data and composition and rendering information parsed by the text subtitle processor 421 .
- a bitmap object ready time (BORT) represents a time at which the BOB 424 stores all the bitmap objects rendered by the text renderer 423 and is ready to output the bitmap objects.
- a presentation time stamp represents a time at which all the bitmap objects are composed within the GP 43 .
- the text subtitle processor 421 starts parsing a DSU and DPU # 1 into dialog text data, composition information, and rendering information
- the DB 422 starts storing the dialog text data at DST 1 .
- DCB 425 starts storing the composition and rendering information.
- the text renderer 423 renders text strings included in the dialog text data into an bitmap object for each dialog region and the BOB 424 stores all the bitmap objects and is ready to output the stored objects at BORT 1 .
- all the bitmap objects are composed within the GP 43 between PTS start 1 and PTS end 1 .
- the text subtitle processor 421 Prior to PTS end 1 , the text subtitle processor 421 start parsing a DSU and DPU # 2 into dialog text data, composition information, and rendering information, and all the steps described above for decoding DPU # 1 are repeated again for decoding DPU # 2 , as shown in FIG. 11 .
- the dialog presentation period for a text subtitle dialog may be limited (e.g., greater than or equal to one second) so as to avoid frequent changes of dialogs within a display screen.
- the bitmap objects stored in the BOB 424 may be deleted when all the bitmap objects are composed within the GP 43 . However, when two consecutive DPUs are continuous as shown in FIG. 11 , at least a portion of the bitmap objects stored in the BOB 424 for a previous DPU may be used when generating bitmap objects for a current DPU. However, the bitmap objects may be deleted from the BOB 424 when a discontinuity exists between two DPUs.
- the data structure of the data included in a text subtitle stream recorded on a recording medium is defined such that the text subtitle stream could be reproduced with main AV streams in a very efficient and standardized manner. Also, seamless presentation of a text subtitle supporting multi-languages is ensured by preloading all the necessary text subtitle streams and related font files in buffers, respectively.
Abstract
A method and a text subtitle decoder for decoding a text subtitle stream recorded on a recording medium are disclosed. A subtitle loading buffer initially loads a text subtitle stream including a dialog style segment defining a group of region styles and a dialog presentation segment having dialog presentation information and dialog text data for a dialog having at least one region. A text subtitle processor parses the dialog presentation segment into composition information, rendering information, and the dialog text data for each region. Next, a text renderer renders the text strings into a bitmap object for each region according to the rendering information, and the rendered bitmap object is composed in a graphics plane.
Description
- This application is a continuation of and claims the benefit under 35 U.S.C. § 120 to U.S. application Ser. No. 11/013,379, filed on Dec. 17, 2004, which claims the benefit under 35 U.S.C. § 119 to Korean Patent Application No. 10-2004-0016564, filed on Mar. 11, 2004, and Provisional Applications 60/543,328 filed Feb. 11, 2004; 60/548,852 filed Feb. 10, 2004; and 60/548,850 filed Feb. 10, 2004; all of which are hereby incorporated by reference.
- 1. Field
- The present invention relates to a text subtitle decoder and a method for decoding text subtitle streams recorded on a recording medium, an example of which is a Blu-ray disc (BD).
- 2. Discussion of the Related Art
- Optical discs are widely used as an optical recording medium for recording mass data. Presently, among a wide range of optical discs, a new high-density digital video disc (hereinafter referred to as “HD-DVD”), such as a Blu-ray Disc (hereafter referred to as “BD”), is under development for writing and storing high definition video and audio data. Currently, global standard technical specifications of the Blu-ray Disc (BD), which is known to be the next generation HD-DVD technology, are under establishment as a next generation optical recording solution that is able to have a data significantly surpassing the conventional DVD, along with many other digital apparatuses.
- Accordingly, optical reproducing apparatuses having the Blu-ray Disc (BD) standards applied thereto are also being developed. However, since the Blu-ray Disc (BD) standards are yet to be completed, there have been many difficulties in developing a complete optical reproducing apparatus. Particularly, in order to effectively reproduce the data from the Blu-ray Disc (BD), not only should the main AV data as well as various data required for a user's convenience, such as subtitle information as the supplementary data related to the main AV data, be provided, but also managing information for reproducing the main data and the subtitle data recorded in the optical disc should be systemized and provided.
- However, in the present Blu-ray Disc (BD) standards, since the standards of the supplementary data, particularly the subtitle information, are not completely consolidated, there are many restrictions in the full-scale development of a Blu-ray Disc (BD) basis optical reproducing apparatus. And, such restrictions cause problems in providing the supplementary data such as subtitles to the user.
- Accordingly, the present invention is directed to a text subtitle decoder and a method for decoding text subtitle streams recorded on a recording medium that substantially obviates one or more problems due to limitations and disadvantages of the related art.
- An object of the present invention is to provide a method and a text subtitle decoder for decoding a text subtitle stream recorded on a recording medium, which includes text strings for each dialog region and composition and rendering information required for decoding the text strings.
- Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objectives and other advantages of the invention may be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
- To achieve these objects and other advantages and in accordance with the purpose of the invention, as embodied and broadly described herein, a method for decoding a text subtitle stream recorded on a recording medium is provided. Initially, a text subtitle stream recorded on the recording medium is loaded into a subtitle loading buffer, where the text subtitle stream includes a dialog style segment and a dialog presentation segment. The dialog style segment defines a group of region styles, and the dialog presentation segment includes dialog presentation information and dialog text data for dialog having at least one region. Thereafter, the dialog presentation segment is parsed into composition information, rendering information, and the dialog text data for each region.
- The parsed composition information and rendering information are then stored in a first buffer, and the parsed dialog text data are stored in a second buffer, where the dialog text data stored in the second buffer includes one or more text strings for each region. The text strings stored in the second buffer are rendered into a bitmap object for each region according to the rendering information, and the rendered bitmap object is stored into a third buffer. Finally, the stored bitmap object is composed in a graphics plane for each region according to the composition information.
- In another aspect of the present invention, a text subtitle decoder for decoding a text subtitle stream recorded on a recording medium includes a subtitle loading buffer, a text subtitle processor, a dialog composition buffer, a dialog buffer, a text renderer, a bitmap object buffer, and a graphics plane. The subtitle loading buffer initially loads the text subtitle stream, which includes a dialog style segment defining a group of region styles and a dialog presentation segment including dialog presentation information and dialog text data for a dialog having at least one region. The text subtitle processor parses the dialog presentation segment into composition information, rendering information, and the dialog text data for each region. Next, the dialog composition buffer stores the composition and rendering information parsed from the text subtitle processor, and the dialog buffer stores the dialog text data, which includes one or more text strings for each region.
- Thereafter, the text renderer included in the text subtitle decoder renders the text strings stored in the dialog buffer into a bitmap object for each region according to the rendering information, and the bitmap object buffer stores the rendered bitmap object. Finally, each bitmap object stored in the bitmap object buffer is composed in the graphics plane according to the composition information.
- In further aspect of the present invention, an optical disc player for reproducing text subtitle streams recorded on an optical disc includes an audio decoder configured to decode audio streams recorded on the optical disc into audio data, a video decoder configured to decode video streams recorded on the optical disc into video image data, a text subtitle decoder configured to decode a text subtitle stream recorded on the optical disc into text subtitle image data, and an image superimposition unit configured to superimpose the decoded text subtitle image data with the decoded video image data. The text subtitle decoder includes a text subtitle processor, a text renderer, and a graphics plane. The text subtitle processor initially parses the text subtitle stream into composition information, rendering information, and dialog text data for a dialog having at least one region, where the dialog text data include one or more text strings for each region. The text renderer renders the text strings into graphic data for each region according to the rendering information, and the graphics plane composes the rendered graphic data according to the composition information.
- It is to be understood that both the foregoing general description and the following detailed description of the present invention are exemplary and explanatory and are intended to provide further explanation of the invention as claimed.
- The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the principle of the invention. In the drawings;
-
FIG. 1 illustrates a structure of the data files recorded in an optical disc according to the present invention; -
FIG. 2 illustrates data storage areas of an optical disc according to the present invention; -
FIG. 3 illustrates a text subtitle and a main image presented on a display screen according to the present invention; -
FIG. 4 illustrates a schematic diagram illustrating reproduction control of a main AV clip and text subtitle clips according to the present invention; -
FIG. 5A illustrates a dialog presented on a display screen according to the present invention; -
FIG. 5B illustrates regions of a dialog presented on a display screen according to the present invention; -
FIG. 5C illustrates style information for regions of a dialog according to the present invention; -
FIG. 6A illustrates presentations of text subtitle dialogs on a display screen in presentation time stamps; -
FIG. 6B illustrates continuities between text subtitle dialogs being presented on a display screen; -
FIG. 7A illustrates the structure of a text subtitle stream file according to the present invention; -
FIG. 7B illustrates specific information contained in a DPU and a DSU included in a text subtitle stream according to the present invention; -
FIG. 8 illustrates an apparatus for decoding main AV streams and text subtitle streams according to the present invention; -
FIG. 9 illustrates a text subtitle decoder according to the present invention; -
FIG. 10 illustrates a graphical representation illustrating parsing, rendering, and composition processes performing by a text subtitle decoder according to the present invention; and -
FIG. 11 illustrates operation times of various components included in a text subtitle decoder according to the present invention. - Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.
- In this detailed description, main data represent audio/video (AV) data that belong to a title (e.g., a movie title) recorded in an optical disc by an author. In general, the AV data are recorded in MPEG2 format and are often called AV streams or main AV streams. In addition, supplementary data represent all other data required for reproducing the main data, examples of which are text subtitle streams, interactive graphic streams, presentation graphic streams, and supplementary audio streams (e.g., for a browsable slideshow). Theses supplementary data streams may be recorded in MPEG2 format or in any other data format. They could be multiplexed with the AV streams or could exist as independent data files within the optical disc.
- A subtitle represents caption information corresponding to video (image) data being reproduced, and it may be represented in a predetermined language. For example, when a user selects an option for viewing one of a plurality of subtitles represented in various languages while viewing images on a display screen, the caption information corresponding to the selected subtitle is displayed on a predetermined portion of the display screen. If the displayed caption information is text data (e.g., characters), the selected subtitle is often called a text subtitle. According to one aspect of the present invention, a plurality of text subtitle streams in MPEG2 format may be recorded in an optical disc, and they may exist as a plurality of independent stream files. Each text subtitle stream file includes text data for a text subtitle and reproduction control data required for reproduction of the text data. According to another aspect of the present invention, only a single text subtitle stream in MPEG2 format may be recorded in an optical disc.
-
FIG. 1 illustrates a file structure of the data files recorded in a Blu-ray disc (hereinafter “BD”) according to the present invention. Referring toFIG. 1 , at least one BD directory (BDMV) is included in a root directory (root). Each BD directory includes an index file (index.bdmv) and an object file (MovieObject.bdmv), which are used for interacting with one or more users. For example, the index file may contain data representing an index table having a plurality of selectable menus and movie titles. Each BD directory further includes four file directories that include audio/video (AV) data to be reproduced and various data required for reproduction of the AV data. - The file directories included in each BD directory are a stream directory (STREAM), a clip information directory (CLIPINF), a playlist directory (PLAYLIST), and an auxiliary data directory (AUX DATA). First of all, the stream directory (STREAM) includes audio/video (AV) stream files having a particular data format. For example, the AV stream files may be in the form of MPEG2 transport packets and be named as “*.m2ts”, as shown in
FIG. 1 . The stream directory may further include one or more text subtitle stream files, where each text subtitle stream file includes text (e.g., characters) data for a text subtitle represented in a particular language and reproduction control information of the text data. The text subtitle stream files exist as independent stream files within the stream directory and may be named as “*.m2ts” or “*.txtst”, as shown inFIG. 1 . An AV stream file or text subtitle stream file included in the stream directory is often called a clip stream file. - Next, the clip information directory (CLIPINF) includes clip information files that correspond to the stream files (AV or text subtitle) included in the stream directory, respectively. Each clip information file contains property and reproduction timing information of a corresponding stream file. For example, A clip information file may includes mapping information, in which presentation time stamps (PTS) and source packet numbers (SPN) are one-to-one mapped by an entry point map (EPM). Using the mapping information, a particular location of a stream file may be determined from timing information (In-Time and Out-Time) provided by a PlayItem or SubPlayItem, which will be discussed later in more details. In the industry standard, each pair of a stream file and its corresponding clip information file is designated as a clip. For example, 01000.clpi included in CLIPINF includes property and reproduction timing information of 01000.m2ts included in STREAM, and 01000.clpi and 01000.m2ts form a clip.
- Referring back to
FIG. 1 , the playlist directory (PLAYLIST) includes one or more PlayList files (*.mpls), where each PlayList file includes at least one PlayItem which designates at least one main AV clip and the reproduction time of the main AV clip. More specifically, a PlayItem contains information designating In-Time and Out-Time, which represent reproduction begin and end times for a main AV clip designated by Clip_Information_File_Name within the PlayItem. Therefore, a PlayList file represents the basic reproduction control information for one or more main AV clips. In addition, the PlayList file may further include a SubPlayItem, which represents the basic reproduction control information for a text subtitle stream file. When a SubPlayItem is included in a PlayList file to reproduce one or more text subtitle stream files, the SubPlayItem is synchronized with the PlayItem(s). On the other hand, when the SubPlayItem is used to reproduce a browsable slideshow, it may not be synchronized with the PlayItem(s). According to the present invention, the main function of a SubPlayItem is to control reproduction of one or more text subtitle stream files. - Lastly, the auxiliary data directory (AUX DATA) may include supplementary data stream files, examples of which are font files (e.g., aaaaa.font), pop-up menu files (not illustrated), and sound files (e.g., Sound.bdmv) for generating click sound. The text subtitle stream files mentioned earlier may be included in the auxiliary data directory instead of the stream directory.
-
FIG. 2 illustrates data storage areas of an optical disc according to the present invention. Referring toFIG. 2 , the optical disc includes a file system information area occupying the inmost portion of the disc volume, a stream area occupying the outmost portion of the disc volume, and a database area occupied between the file system information area and the stream area. In the file system information area, system information for managing the entire data files shown inFIG. 1 is stored. Next, AV streams and one or more text subtitle streams are stored in the stream area. The general files, PlayList files, and clip information files shown inFIG. 1 are stored in the database area of the disc volume. As discussed above, the general files include an index file and an object file, and the PlayList files and clip information files include information required to reproduce the AV streams and the text subtitle streams stored in the stream area. Using the information stored in the database area and/or stream area, a user is able to select a specific playback mode and to reproduce the main AV and text subtitle streams in the selected playback mode. -
FIG. 3 illustrates a text subtitle and a main image presented on a display screen according to the present invention. The main image and the text subtitle are simultaneously displayed on the display screen when a main AV stream and a corresponding text subtitle stream are reproduced in synchronization. -
FIG. 4 is a schematic diagram illustrating reproduction control of a main AV clip and text subtitle clips according to the present invention. Referring toFIG. 4 , a PlayList file includes at least one PlayItem controlling reproduction of at least one main AV clip and a SubPlayItem controlling reproduction of a plurality of text subtitle clips. One oftext subtitle clip 1 andtext subtitle clip 2 shown inFIG. 4 for English and Korean text subtitles may be synchronized with the main AV clip such that a main image and a corresponding text subtitle are displayed on a display screen simultaneously at a particular presentation time. In order to display the text subtitle on the display screen, display control information (e.g., position and size information) and presentation time information, examples of which are illustrated inFIG. 5A toFIG. 5C , are required. -
FIG. 5A illustrates a dialog presented on a display screen according to the present invention. A dialog represents entire text subtitle data displayed on a display screen during a given presentation time. In general, presentation times of the dialog may be represented in presentation time stamps (PTS). For example, presentation of the dialog shown inFIG. 5A starts at PTS (k) and ends at PTS (k+1). Therefore, the dialog shown inFIG. 5A represents an entire unit of text subtitle data which are displayed on the display screen between PTS (k) and PTS (k+1). A dialog includes at least one line of subtitle text (characters). When there are two or more lines of subtitle text in a dialog, entire text data may be displayed according to a style defined for the dialog. - In addition,
FIG. 5B illustrates regions of a dialog according to the present invention. A region represents a divided portion of text subtitle data (dialog) displayed on a display screen during a given presentation time. In other words, a dialog includes at least one region, and each region may include at least one line of subtitle text. The entire text subtitle data representing a region may be displayed on the display screen according to a region style (global style) assigned to the region. The maximum number of regions included in a dialog should be determined based on a desired decoding rate of the subtitle data because the greater number of regions generally results a lower decoding ratio. For example, the maximum number of regions for a dialog may be limited to two in order to achieve a reasonably high decoding rate. However, the maximum number could be greater than two for other purposes. -
FIG. 5C illustrates style information for regions of a dialog according to the present invention. Style information represents information defining properties required for displaying at least a portion of a region included in a dialog. Some of the examples of the style information are position, region size, background color, text alignment, text flow information, and many others. The style information may be classified into region style information (global style information) and inline style information (local style information). - Region style information defines a region style (global style) which is applied to an entire region of a dialog. For example, the region style information may contain at least one of a region position, region size, font color, background color, text flow, text alignment, line space, font name, font style, and font size of the region. For example, two different region styles are applied to
region 1 andregion 2, as shown inFIG. 5C . A region style withposition 1,size 1, and blue background color is applied toRegion 1, and a different region style withposition 2,size 2, and red background color is applied toRegion 2. - On the other hand, inline style information defines an inline style (local style) which is applied to a particular portion of text strings included in a region. For example, the inline style information may contain at least one of a font type, font size, font style, and font color. The particular portion of text strings may be an entire text line within a region or a particular portion of the text line. Referring to
FIG. 5C , a particular inline style is applied to the text portion “mountain” included inRegion 1. In other words, at least one of the font type, font size, font style, and font color of the particular portion of text strings is different from the remaining portion of the text strings withinRegion 1. -
FIG. 6A illustrates presentations of text subtitle dialogs on a display screen in presentation time stamp (PTS) intervals. There are four dialogs to be displayed between PTS1 to PTS6. More specifically,Dialog # 1 has only one region andText # 1 is displayed within this region between PTS1 to PTS2. Next,Dialog # 2 hasRegion 1 andRegion 2 andText # 1 andText # 2 are displayed withinRegion 1 andRegion 2, respectively, between PTS2 to PTS3. Thereafter,Dialog # 3 also has only one region andText # 2 is displayed within this region between PTS3 and PTS4. There is no dialog to be presented between PTS4 to PTS5, andText # 3 is displayed within a region ofDialog # 4 between PTS5 to PTS6. Information defining a dialog includes dialog presentation time information and dialog text data including style information and text strings to be displayed within each region of the dialog. An example of the presentation time information is a set of start PTS start and PTS end, and the style information includes region (global) style information and inline (local) style information described above. It is shown inFIG. 6A that different style information sets may be applied to the dialogs. -
FIG. 6B illustrates continuities between text subtitle dialogs being presented on a display screen in PTS intervals. Referring toFIG. 6B , a continuity exists betweenDialog # 1 andDialog # 2. Display ofText # 1 in a region ofDialog # 1 is continuous with display ofText # 1 inRegion 1 ofDialog # 2. In other words, PTS intervals of both dialogs are continuous and same style information (region and inline) is used when presentingText # 1 in both regions. Similarly, another continuity exists betweenDialog # 2 andDialog # 3 because display ofText # 2 inRegion 2 ofDialog # 2 is continuous with display ofText # 2 in a region ofDialog # 3. In order to ensure a continuity between two consecutive dialogs displaying same subtitle text, presentation times (PTS intervals) of the dialogs must be continuous. In addition, same region and inline style information must be used when presenting the same text in the regions, respectively. Referring back toFIG. 6B , there is no continuity betweenDialog # 3 andDialog # 4 because their PTS intervals are not continuous. An indicator (e.g., continuous_presentation_flag) may be included in presentation information of a current dialog to indicate whether the dialog is continuous with a previous dialog. -
FIG. 7A illustrates a text subtitle stream file (e.g., 10001.m2ts shown inFIG. 1 ) according to the present invention. It may be formed of an MPEG2 transport stream including a plurality of transport packets (TP), all of which have a same packet identifier (e.g., PID=0x18xx). When a disc player receives many input streams including a particular text subtitle stream, it finds all the transport packets that belong to the text subtitle stream using their PIDs. Referring toFIG. 7A , each sub-set of transport packets form a packet elementary stream (PES) packet. One of the PES packets shown inFIG. 7A corresponds to a dialog style unit (DSU) defining a group of region styles. A DSU is also often referred as a dialog style segment (DSS). All the remaining PES packets correspond to dialog presentation units (DPUs), each of which includes presentation information for a dialog having at least one region, and dialog text data which includes an region style indicator, inline style information, and text strings for each region. -
FIG. 7B illustrates specific information contained within a DPU and a DSU included in a text subtitle stream according to the present invention. A DSU contains information sets defining a group of region styles, each of which is applied to a corresponding region of a dialog. In addition, a DPU contains dialog text data and dialog presentation information for a dialog. The dialog text data includes text strings to be included in each region of the dialog, inline style information to be applied to a particular portion of the text strings, and a region style identifier indicating a region style to be applied to each dialog region. The region style identifier identifies one of the group of region styles defined in the DSU. On the other hand, the dialog presentation information includes presentation time information and palette (color) update information for a dialog. The presentation time information may include presentation start time (e.g, PTS_start) and presentation end time (e.g., PTS_end) for presenting the dialog on a display screen, and the palette update information may include an indicator (e.g., palette_update_flag) indicating whether to update display colors of the dialog and palette information (e.g., Palette for update) to be applied when updating the display colors. - All the data included in a text subtitle stream may be classified into three types of data based on their basic functions. For example, the data could be classified into dialog text data, composition information, and rendering information, as shown in
FIG. 7B . The dialog text data include text string(s), inline style information, and a region style identifier for each region of a dialog. The composition information includes presentation time information, examples of which are presentation start and end times, position information for a dialog region, and palette update information for a dialog. Lastly, the rendering information includes information required for rendering the text strings to graphic data for presentation. Referring toFIG. 7B , the horizontal and vertical positions of each region included in the DSU is a part of the composition information, and the region width, region height, font color, background color, text flow, text alignment, line space, font name, font style, and font size included in the DSU represent the rendering information. - Reference will now be made in detail to an apparatus for decoding man AV streams and text subtitle streams according to the present invention, an example of which is illustrated in
FIG. 8 . The apparatus includes a packet identifier (PID)filter 5 for separating input streams into video streams, audio streams, graphic streams, and text subtitle streams based on their packet identifiers, avideo decoding part 20 for decoding the video streams, anaudio decoding part 10 for decoding the audio streams, agraphic decoding part 30 for decoding the graphic streams, and a textsubtitle decoding part 40 for decoding the text subtitle streams. - The text subtitle streams may be extracted from an optical disc or from an additional external source, as shown in
FIG. 8 . For this reason, the apparatus additionally includes aswitch 6 which selects an input data source. Therefore, if the text subtitle streams are extracted from the optical disc, theswitch 6 selects data line A connected to thePID filter 5. On the other hand, if they are inputted from the external source, theswitch 6 selects line B connected to the external source. - Referring back to
FIG. 8 , theaudio decoding part 10,video decoding part 20, andgraphic decoding part 30 includetransport buffers graphic plane 33 are included in thevideo decoding part 20 and thegraphic decoding part 30, respectively, for converting decoded signals into displayable video and graphic images. Thegraphic decoding part 30 includes a color look up table (CLUT) 34 for controlling color and transparency levels of the displayable graphic images. - When the text
subtitle decoding part 40 receives a text subtitle stream supporting a single language from theswitch 6, an entire portion of the text subtitle stream may be preloaded into a subtitle preloading buffer (SPB) 41 at once. Alternatively, when there are more than one text subtitle streams for supporting multi-languages, all the text subtitle streams may be preloaded into theSPB 41 at once. Therefore, the size of theSPB 41 should be determined based on a total number of text subtitle stream files received from theswitch 6. For example, the size of theSPB 41 should be greater than or equal to 0.5 megabytes for preloading a 0.5 megabyte text subtitle stream file. In addition, in order to ensure seamless presentation of a text subtitle when a user switches among two 0.5 megabyte text subtitle stream files, the size of theSPB 41 should be greater than or equal to 1 megabytes. The size of theSPB 42 should be large enough to preload all the required text subtitle stream files at once. - The text
subtitle decoding part 40 shown inFIG. 8 further includes a font preloading buffer (FPB) 410 for storing all the associated font files which may be included in the auxiliary data directory shown inFIG. 1 . Similarly, the size of theFPB 410 should be large enough to preload all the required font files at once in order to ensure seamless presentation of a text subtitle supporting one or more languages. Since all the available text subtitle stream files and related font files are preloaded, extraction and use of the preloaded data can be done in a simple manner. Also the control of theSPB 41 and theFPB 410 could be quite simple due to the this reason. The textsubtitle decoding part 40 further includes atext subtitle decoder 42 which decodes each text subtitle stream stored in theSPB 41, agraphic plane 43 in which the decoded subtitle data are composed as displayable subtitle images, and a color look up table (CLUT) 44 controlling at least one of color and transparency levels of the converted subtitle images. - The apparatus shown in
FIG. 8 further includes animage superimposition part 50 which superimposes the images outputted from thevideo decoding part 20, thegraphic decoding part 30, and the textsubtitle decoding part 40. These combined images are displayed on a display screen, as shown inFIG. 3 . In general, the video images outputted from theVP 23 of thevideo decoding part 20 may be displayed as a background of the display screen, and the images outputted from thegraphic decoding part 30 and/or textsubtitle decoding part 40 may be superimposed over the video images in a predetermined order. For example, if the output images of thegraphic decoding part 30 are presentation graphic images, these images may be initially superimposed over the video images by afirst adder 52, and subsequently, the text subtitle images from the textsubtitle decoding part 40 may be superimposed over the video images by asecond adder 53. However, if the output images of thegraphic decoding part 30 are interactive graphic images, the text subtitle images from the textsubtitle decoding part 40 may be initially superimposed over the video images by thefirst adder 52. Thereafter, the interactive graphic images may be further superimposed over the subtitle-superimposed images by thesecond adder 53. - Lastly, the apparatus shown in
FIG. 8 further includes asystem decoder 4 for decoding input transport streams (e.g., MPEG transport streams), and amicroprocessor 3 for controlling operations of all the components of the apparatus mentioned above. - Reference will now be made in detail to a method and a text subtitle decoder for reproducing text subtitle streams according to the present invention. When an optical disc is preloaded by an optical disc player, an example of which is illustrated in
FIG. 8 , information required for reproducing main data recorded on the disc is initially extracted from the disc and is stored in a storage (not illustrated). When a particular title that associates with a PlayList is selected by user, at least one text subtitle stream file designated in the PlayList file is initially preloaded in theSPB 41 and font files related to the subtitle stream file are preloaded into theFPB 410. - For example, when a title that associates with the PlayList shown in
FIG. 4 is selected by a user, at least one of text subtitle stream files for Text Subtitle Clip 1 (English) and Text Subtitle Clip 2 (Korean) may be initially preloaded into theSPB 41. Also, all the font files related to the text subtitle stream files, which may be designated in clip information files of the text subtitle clips, are preloaded into theFPB 410. Thereafter, playback of the PlayList is started. During the playback of the PlayList, AV streams included in the main AV clip are decoded by theaudio decoding part 10 and thevideo decoding part 20, and one of the preloaded text subtitle streams is decoded by thetext subtitle decoder 42. Then the decoded text subtitle images are superimposed over the decoded main video images, and the entire images are displayed on a display screen. -
FIG. 9 illustrates thetext subtitle decoder 42 included in the decoding apparatus shown inFIG. 8 in more details. The decoding process performed by thetext subtitle decoder 42 includes parsing, rendering, and composition steps. In the parsing step, a text subtitle stream stored in theSPB 41 is parsed into composition information, rendering information, and dialog text data. In the rendering step, text strings included in the dialog text data are rendered into graphic data (bitmap data) for each region using the rendering information, which represents information required for rendering the text strings into the graphic data. Finally, in the composition step, the rendered text strings are added into theGP 43 according to the composition information, which represents dialog presentation information, palette update information, and position information for a dialog. - After one or more text subtitle streams and the related font files are preloaded into the
SPB 41 and theFPB 410, respectively, atext subtitle processor 421 included in thetext subtitle decoder 42 parses a text subtitle stream preloaded in theSPB 41 into composition information, rendering information, and dialog text data. More particularly, thetext subtitle processor 421 initially transfers a dialog style unit (DSU) included in the preloaded subtitle stream to a dialog composition buffer (DCB) 425, and it parses a dialog presentation unit (DPU) further included in the preloaded text subtitle stream into composition information, rendering information, and dialog text data. The composition and rendering information are then stored in theDCB 425, and the dialog text data are stored in a dialog buffer (DB) 422. The dialog text data stored in theDB 422 include a region style identifier, text strings, and inline style information for each dialog region. - Next, a
text renderer 423 renders the text strings stored in theDB 422 into a bitmap object (graphic data) for each dialog region under the control of adialog presentation controller 426. In other words, thetext renderer 423 renders the text strings stored in theDB 422 into a bitmap object for each dialog region using the region style identifier and inline style information stored in theDB 422, the rendering information provided from thedialog presentation controller 426, and related font data provided from theFPB 410. The bitmap object rendered by thetext renderer 423 for each dialog region is then stored in a bitmap object buffer (BOB) 424. - Finally, each bitmap object stored in the
BOB 424 is composed within (added to) theGP 43 according to the composition information provided by thedialog presentation controller 426. TheCLUT 44 uses palette update information included in the composition information to adjust color and/or transparency levels of an output of theGP 43. During the rendering and composition processes performed by thetext renderer 423 and theGP 43, particular style information selected by a user may be applied. Thedialog presentation controller 426 may receive such user-selected style information and provide this information to thetext renderer 423 and/or theGP 43. Examples of the user-selectable style information are a region position and a font size. -
FIG. 10 is a simple graphical representation illustrating parsing, rendering, and composition processes performed by thetext subtitle decoder 42 shown inFIG. 9 . In the parsing process, a DSU is initially transferred to theDCB 425 and a DPU is parsed into composition information, dialog text data, and rendering information. Text strings for each dialog region included in the dialog text data are rendered into a bitmap object using the rendering information, and the bitmap object rendered for each region is composed within theGP 43 according to the composition information, which may include presentation information, palette information, and region position information. -
FIG. 11 illustrates operation times of theGP 43,BOB 424,DCB 425, andDB 422 in a global time axis perceived by a user. Referring toFIG. 11 , a decoding start time (DST) represents a time at which thetext subtitle decoder 42 starts to decode a text subtitle stream preloaded in theSPB 41. More particularly, a DST represents a time at which theDB 422 and theDCB 425 start storing dialog text data and composition and rendering information parsed by thetext subtitle processor 421. A bitmap object ready time (BORT) represents a time at which theBOB 424 stores all the bitmap objects rendered by thetext renderer 423 and is ready to output the bitmap objects. A presentation time stamp (PTS) represents a time at which all the bitmap objects are composed within theGP 43. - For example, when the
text subtitle processor 421 starts parsing a DSU andDPU # 1 into dialog text data, composition information, and rendering information, theDB 422 starts storing the dialog text data at DST1. At the same time,DCB 425 starts storing the composition and rendering information. Thereafter, thetext renderer 423 renders text strings included in the dialog text data into an bitmap object for each dialog region and theBOB 424 stores all the bitmap objects and is ready to output the stored objects at BORT1. Next, all the bitmap objects are composed within theGP 43 between PTSstart 1 and PTSend 1. Prior to PTSend 1, thetext subtitle processor 421 start parsing a DSU andDPU # 2 into dialog text data, composition information, and rendering information, and all the steps described above fordecoding DPU # 1 are repeated again fordecoding DPU # 2, as shown inFIG. 11 . - The dialog presentation period for a text subtitle dialog (e.g., between PTSstart 1 and PTSend 1) may be limited (e.g., greater than or equal to one second) so as to avoid frequent changes of dialogs within a display screen. In addition, the bitmap objects stored in the
BOB 424 may be deleted when all the bitmap objects are composed within theGP 43. However, when two consecutive DPUs are continuous as shown inFIG. 11 , at least a portion of the bitmap objects stored in theBOB 424 for a previous DPU may be used when generating bitmap objects for a current DPU. However, the bitmap objects may be deleted from theBOB 424 when a discontinuity exists between two DPUs. - It will be apparent to those skilled in the art that various modifications and variations can be made in the present invention without departing from the spirit or scope of the inventions. Thus, it is intended that the present invention covers the modifications and variations of this invention provided they come within the scope of the appended claims and their equivalents.
- According to the present invention, the data structure of the data included in a text subtitle stream recorded on a recording medium is defined such that the text subtitle stream could be reproduced with main AV streams in a very efficient and standardized manner. Also, seamless presentation of a text subtitle supporting multi-languages is ensured by preloading all the necessary text subtitle streams and related font files in buffers, respectively.
Claims (21)
1. An apparatus for reproducing a text subtitle stream recorded on a recording medium, the apparatus comprising:
a reproducing unit configured to reproduce the text subtitle stream; and
a decoder configured to,
parse the text subtitle stream into composition information, rendering information, and text data for at least one region, the text data including text strings for the region,
render the text strings into graphic data for the region according to the rendering information, and
compose the rendered graphic data according to the composition information.
2. The apparatus of claim 1 , further comprising:
a subtitle preloading buffer configured to preload the text subtitle stream.
3. The apparatus of claim 2 , wherein seamless presentation of AV data can be guaranteed if a size of the text subtitle stream is smaller than or equal to a size of the subtitle preloading buffer.
4. The apparatus of claim 2 , wherein the subtitle preloading buffer is configured to preload the text subtitle stream at once before starting playback of a video data to be presented with the text data.
5. The apparatus of claim 1 , wherein the rendering information includes at least one of a region size, font color, background color, text flow, text alignment, line space, font name, font style, and font size.
6. The apparatus of claim 5 , wherein the composition information includes at least one of presentation time information, palette update information, and a region position.
7. The apparatus of claim 1 , wherein the text subtitle stream includes,
a style segment defining at least one region style, and
a presentation segment including at least one region and a region style identifier identifying one of the region styles for the at least one region, and wherein the decoder is configured to apply the identified region style defined in the style segment to the at least one region included in the presentation segment.
8. The apparatus of claim 7 , wherein the presentation segment includes at most two regions, and wherein the decoder is configured to apply a different region style to each of the regions.
9. The apparatus of claim 7 , wherein the presentation segment further includes at least one inline style for the region, and wherein the inline style overrides the region style.
10. The apparatus of claim 1 , further comprising:
a font preloading buffer configured to preload related font data, wherein the decoder is configured to render the text strings into the graphic data according to the rendering information and the preloaded font data.
11. A method for reproducing a text subtitle stream recorded on a recording medium, the method comprising:
parsing the text subtitle stream into composition information, rendering information, and text data for at least one region, the text data including text strings for the region;
rendering the text strings into graphic data for the region according to the rendering information; and
composing the rendered graphic data according to the composition information.
12. The method of claim 11 , further comprising:
preloading the text subtitle stream into a subtitle preloading buffer.
13. The method of claim 12 , wherein the preloading step is configured to preload the text subtitle stream into the subtitle preloading buffer at once before starting playback of a video data to be presented with the text data.
14. The method of claim 12 , wherein seamless presentation of AV data can be guaranteed if a size of the text subtitle stream is smaller than or equal to a size of the subtitle preloading buffer.
15. The method of claim 11 , wherein the rendering information includes at least one of a region size, font color, background color, text flow, text alignment, line space, font name, font style, and font size.
16. The method of claim 15 , wherein the composition information includes at least one of presentation time information, palette update information, and a region position.
17. The method of claim 11 , wherein the text subtitle stream includes,
a style segment defining at least one region style, and
a presentation segment including at least one region and a region style identifier identifying one of the region styles for the at least one region, and wherein the decoder is configured to apply the identified region style defined in the style segment to the at least one region included in the presentation segment.
18. The method of claim 17 , further comprising:
applying the identified region style defined in the style segment to the at least one region included in the presentation segment.
19. The method of claim 18 , wherein the presentation segment includes at most two regions and each of the regions is applied with a different region style.
20. The method of claim 17 , wherein the presentation segment further includes at least one inline style for the region, and the inline style overrides the region style.
21. The method of claim 11 , further comprising:
preloading related font data into a font preloading buffer; and
rendering the text strings into the graphic data according to the rendering information and the preloaded font data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/487,992 US20090263106A1 (en) | 2004-02-10 | 2009-06-19 | Text subtitle decoder and method for decoding text subtitle streams |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US54285204P | 2004-02-10 | 2004-02-10 | |
US54285004P | 2004-02-10 | 2004-02-10 | |
US54332804P | 2004-02-11 | 2004-02-11 | |
KR10-2004-0016564 | 2004-03-11 | ||
KR1020040016564A KR20050091228A (en) | 2004-03-11 | 2004-03-11 | Apparatus and method for reproducing a text subtitle stream of high density optical disc |
US11/013,379 US7561780B2 (en) | 2004-02-10 | 2004-12-17 | Text subtitle decoder and method for decoding text subtitle streams |
US12/487,992 US20090263106A1 (en) | 2004-02-10 | 2009-06-19 | Text subtitle decoder and method for decoding text subtitle streams |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/013,379 Continuation US7561780B2 (en) | 2004-02-10 | 2004-12-17 | Text subtitle decoder and method for decoding text subtitle streams |
Publications (1)
Publication Number | Publication Date |
---|---|
US20090263106A1 true US20090263106A1 (en) | 2009-10-22 |
Family
ID=34841854
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/013,379 Expired - Fee Related US7561780B2 (en) | 2004-02-10 | 2004-12-17 | Text subtitle decoder and method for decoding text subtitle streams |
US12/487,992 Abandoned US20090263106A1 (en) | 2004-02-10 | 2009-06-19 | Text subtitle decoder and method for decoding text subtitle streams |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/013,379 Expired - Fee Related US7561780B2 (en) | 2004-02-10 | 2004-12-17 | Text subtitle decoder and method for decoding text subtitle streams |
Country Status (4)
Country | Link |
---|---|
US (2) | US7561780B2 (en) |
EP (1) | EP1716701A1 (en) |
KR (1) | KR20070028325A (en) |
WO (1) | WO2005076601A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070025698A1 (en) * | 2005-07-29 | 2007-02-01 | Kabushiki Kaisha Toshiba | Information playback method using information recording medium |
US20080298775A1 (en) * | 2004-02-21 | 2008-12-04 | Samsung Electronics Co., Ltd. | Storage medium for storing text-based subtitle data including style information, and reproducing apparatus and method for reproducing text-based subtitle data including style information |
US20090185075A1 (en) * | 2004-02-28 | 2009-07-23 | Samsung Electronics Co., Ltd | Storage medium recording text-based subtitle stream, reproducing apparatus and reproducing method for reproducing text-based subtitle stream recorded on the storage medium |
US20110227911A1 (en) * | 2010-03-22 | 2011-09-22 | Lg Electronics Inc. | Image display device and method for operating the same |
Families Citing this family (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ATE517413T1 (en) * | 2003-04-09 | 2011-08-15 | Lg Electronics Inc | RECORDING MEDIUM HAVING A DATA STRUCTURE FOR MANAGING THE PLAYBACK OF TEXT CAPTION DATA AND METHOD AND APPARATUS FOR RECORDING AND REPLAYING |
US8472792B2 (en) | 2003-12-08 | 2013-06-25 | Divx, Llc | Multimedia distribution system |
US7519274B2 (en) | 2003-12-08 | 2009-04-14 | Divx, Inc. | File format for multiple track digital data |
KR20050078907A (en) * | 2004-02-03 | 2005-08-08 | 엘지전자 주식회사 | Method for managing and reproducing a subtitle of high density optical disc |
KR101067778B1 (en) | 2004-03-26 | 2011-09-28 | 엘지전자 주식회사 | Recording medium and method and apparatus for reproducing text subtitle stream recorded on the recording medium |
BRPI0509231A (en) * | 2004-03-26 | 2007-09-04 | Lg Electronics Inc | recording medium, method and apparatus for reproducing text subtitle streams |
EP1733385B1 (en) * | 2004-03-26 | 2009-12-09 | LG Electronics Inc. | Recording medium and method and apparatus for reproducing and recording text subtitle streams |
WO2007106844A2 (en) | 2006-03-14 | 2007-09-20 | Divx, Inc. | Federated digital rights management scheme including trusted systems |
TWM304101U (en) * | 2006-06-14 | 2007-01-01 | Wei-Jing Yang | DVD player capable of showing multi-national captions |
CN101179669B (en) * | 2006-11-08 | 2011-03-16 | 中兴通讯股份有限公司 | Session television terminal subtitling generating and stacking method |
EP2223232A4 (en) | 2007-11-16 | 2015-02-25 | Sonic Ip Inc | Hierarchical and reduced index structures for multimedia files |
WO2009149440A1 (en) * | 2008-06-06 | 2009-12-10 | Divx, Inc. | Multimedia distribution and playback systems and methods using enhanced metadata structures |
KR101635876B1 (en) | 2009-01-07 | 2016-07-04 | 쏘닉 아이피, 아이엔씨. | Singular, collective and automated creation of a media guide for online content |
KR20110018261A (en) * | 2009-08-17 | 2011-02-23 | 삼성전자주식회사 | Method and apparatus for processing text subtitle data |
CA2782825C (en) | 2009-12-04 | 2016-04-26 | Divx, Llc | Elementary bitstream cryptographic material transport systems and methods |
US8914534B2 (en) | 2011-01-05 | 2014-12-16 | Sonic Ip, Inc. | Systems and methods for adaptive bitrate streaming of media stored in matroska container files using hypertext transfer protocol |
US9467708B2 (en) | 2011-08-30 | 2016-10-11 | Sonic Ip, Inc. | Selection of resolutions for seamless resolution switching of multimedia content |
US9955195B2 (en) | 2011-08-30 | 2018-04-24 | Divx, Llc | Systems and methods for encoding and streaming video encoded using a plurality of maximum bitrate levels |
US8818171B2 (en) | 2011-08-30 | 2014-08-26 | Kourosh Soroushian | Systems and methods for encoding alternative streams of video for playback on playback devices having predetermined display aspect ratios and network connection maximum data rates |
US8964977B2 (en) | 2011-09-01 | 2015-02-24 | Sonic Ip, Inc. | Systems and methods for saving encoded media streamed using adaptive bitrate streaming |
US8909922B2 (en) | 2011-09-01 | 2014-12-09 | Sonic Ip, Inc. | Systems and methods for playing back alternative streams of protected content protected using common cryptographic information |
US10452715B2 (en) | 2012-06-30 | 2019-10-22 | Divx, Llc | Systems and methods for compressing geotagged video |
ITTO20120966A1 (en) * | 2012-11-06 | 2014-05-07 | Inst Rundfunktechnik Gmbh | MEHRSPRACHIGE GRAFIKANSTEUERUNG IN FERNSEHSENDUNGEN |
US9191457B2 (en) | 2012-12-31 | 2015-11-17 | Sonic Ip, Inc. | Systems, methods, and media for controlling delivery of content |
US9313510B2 (en) | 2012-12-31 | 2016-04-12 | Sonic Ip, Inc. | Use of objective quality measures of streamed content to reduce streaming bandwidth |
US10397292B2 (en) | 2013-03-15 | 2019-08-27 | Divx, Llc | Systems, methods, and media for delivery of content |
US9906785B2 (en) | 2013-03-15 | 2018-02-27 | Sonic Ip, Inc. | Systems, methods, and media for transcoding video data according to encoding parameters indicated by received metadata |
US9094737B2 (en) | 2013-05-30 | 2015-07-28 | Sonic Ip, Inc. | Network video streaming with trick play based on separate trick play files |
US9967305B2 (en) | 2013-06-28 | 2018-05-08 | Divx, Llc | Systems, methods, and media for streaming media content |
US9866608B2 (en) * | 2014-03-24 | 2018-01-09 | Qualcomm Incorporated | Processing continuous multi-period content |
US9866878B2 (en) | 2014-04-05 | 2018-01-09 | Sonic Ip, Inc. | Systems and methods for encoding and playing back video at different frame rates using enhancement layers |
US10511882B2 (en) * | 2016-01-26 | 2019-12-17 | Sony Corporation | Reception apparatus, reception method, and transmission apparatus |
US10148989B2 (en) | 2016-06-15 | 2018-12-04 | Divx, Llc | Systems and methods for encoding video content |
EP3364660B1 (en) * | 2016-09-14 | 2020-12-23 | Sony Corporation | Transmitting device, transmitting method, receiving device and receiving method |
US10498795B2 (en) | 2017-02-17 | 2019-12-03 | Divx, Llc | Systems and methods for adaptive switching between multiple content delivery networks during adaptive bitrate streaming |
Citations (95)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3128434A (en) * | 1960-04-28 | 1964-04-07 | Bendix Corp | Transfluxor with amplitude modulated driving pulse input converted to alternating sine wave output |
GB1556366A (en) * | 1976-08-12 | 1979-11-21 | English Electric Co Ltd | Receiving apparatus for combined television/data display |
US4218698A (en) * | 1978-03-13 | 1980-08-19 | Rca Corporation | TV Graphics and mixing control |
US5253530A (en) * | 1991-08-12 | 1993-10-19 | Letcher Iii John H | Method and apparatus for reflective ultrasonic imaging |
US5467142A (en) * | 1992-04-24 | 1995-11-14 | Victor Company Of Japan, Ltd. | Television receiver for reproducing video images having different aspect ratios and characters transmitted with video images |
US5519443A (en) * | 1991-12-24 | 1996-05-21 | National Captioning Institute, Inc. | Method and apparatus for providing dual language captioning of a television program |
US5537151A (en) * | 1994-02-16 | 1996-07-16 | Ati Technologies Inc. | Close caption support with timewarp |
US5731847A (en) * | 1995-03-20 | 1998-03-24 | Sony Corporation | Subtitle encoding/decoding method and apparatus |
US5758007A (en) * | 1995-02-03 | 1998-05-26 | Kabushiki Kaisha Toshiba | Image information encoding/decoding system |
US5778142A (en) * | 1994-11-24 | 1998-07-07 | Kabushiki Kaisha Toshiba | Large capacity recording medium, method and apparatus for reproducing data from a large-capacity recording medium, and method and apparatus for recording data on a large-capacity recording medium |
US5781687A (en) * | 1993-05-27 | 1998-07-14 | Studio Nemo, Inc. | Script-based, real-time, video editor |
US5832530A (en) * | 1994-09-12 | 1998-11-03 | Adobe Systems Incorporated | Method and apparatus for identifying words described in a portable electronic document |
US5847770A (en) * | 1995-09-25 | 1998-12-08 | Sony Corporation | Apparatus and method for encoding and decoding a subtitle signal |
US5848217A (en) * | 1995-08-02 | 1998-12-08 | Sony Corporation | Subtitle encoding/decoding method and apparatus |
US5987214A (en) * | 1995-06-30 | 1999-11-16 | Sony Corporation | Apparatus and method for decoding an information page having header information and page data |
US6009234A (en) * | 1995-04-14 | 1999-12-28 | Kabushiki Kaisha Toshiba | Method of reproducing information |
US6115077A (en) * | 1995-08-04 | 2000-09-05 | Sony Corporation | Apparatus and method for encoding and decoding digital video data operable to remove noise from subtitle date included therewith |
US6128434A (en) * | 1993-10-29 | 2000-10-03 | Kabushiki Kaisha Toshiba | Multilingual recording medium and reproduction apparatus |
US6148140A (en) * | 1997-09-17 | 2000-11-14 | Matsushita Electric Industrial Co., Ltd. | Video data editing apparatus, optical disc for use as a recording medium of a video data editing apparatus, and computer readable recording medium storing an editing program |
US6173113B1 (en) * | 1995-09-29 | 2001-01-09 | Matsushita Electric Industrial Co., Ltd. | Machine readable information recording medium having audio gap information stored therein for indicating a start time and duration of an audio presentation discontinuous period |
US6204883B1 (en) * | 1993-12-21 | 2001-03-20 | Sony Corporation | Video subtitle processing system |
US6219043B1 (en) * | 1995-07-13 | 2001-04-17 | Kabushiki Kaisha Toshiba | Method and system to replace sections of an encoded video bitstream |
US6222532B1 (en) * | 1997-02-03 | 2001-04-24 | U.S. Philips Corporation | Method and device for navigating through video matter by means of displaying a plurality of key-frames in parallel |
US6230295B1 (en) * | 1997-04-10 | 2001-05-08 | Lsi Logic Corporation | Bitstream assembler for comprehensive verification of circuits, devices, and systems |
US6253221B1 (en) * | 1996-06-21 | 2001-06-26 | Lg Electronics Inc. | Character display apparatus and method for a digital versatile disc |
US6259858B1 (en) * | 1998-12-16 | 2001-07-10 | Kabushiki Kaisha Toshiba | Optical disc for storing moving pictures with text information and apparatus using the disc |
US6262775B1 (en) * | 1997-06-17 | 2001-07-17 | Samsung Electronics Co., Ltd. | Caption data processing circuit and method therefor |
US6297797B1 (en) * | 1997-10-30 | 2001-10-02 | Kabushiki Kaisha Toshiba | Computer system and closed caption display method |
US6320621B1 (en) * | 1999-03-27 | 2001-11-20 | Sharp Laboratories Of America, Inc. | Method of selecting a digital closed captioning service |
US20010044809A1 (en) * | 2000-03-29 | 2001-11-22 | Parasnis Shashank Mohan | Process of localizing objects in markup language documents |
US20020004755A1 (en) * | 2000-06-29 | 2002-01-10 | Neil Balthaser | Methods, systems, and processes for the design and creation of rich-media applications via the internet |
US20020010924A1 (en) * | 2000-05-03 | 2002-01-24 | Morteza Kalhour | Push method and system |
US6356567B2 (en) * | 1997-09-26 | 2002-03-12 | International Business Machines Corporation | Embedded clock recovery and difference filtering for an MPEG-2 compliant transport stream |
US6393196B1 (en) * | 1996-09-27 | 2002-05-21 | Matsushita Electric Industrial Co., Ltd. | Multimedia stream generating method enabling alternative reproduction of video data, and a multimedia optical disk authoring system |
US20020067744A1 (en) * | 1995-03-29 | 2002-06-06 | Yukio Fujii | Decoder for compressed and multiplexed video and audio data |
US20020087999A1 (en) * | 2000-04-26 | 2002-07-04 | Sony Corporation | Scalable filtering table |
US20020087569A1 (en) * | 2000-12-07 | 2002-07-04 | International Business Machines Corporation | Method and system for the automatic generation of multi-lingual synchronized sub-titles for audiovisual data |
US20020106193A1 (en) * | 2001-02-05 | 2002-08-08 | Park Sung-Wook | Data storage medium in which multiple bitstreams are recorded, apparatus and method for reproducing the multiple bitstreams, and apparatus and method for reproducing the multiple bitstreams |
US20020135607A1 (en) * | 2000-04-21 | 2002-09-26 | Motoki Kato | Information processing apparatus and method, program, and recorded medium |
US20020135608A1 (en) * | 2000-04-21 | 2002-09-26 | Toshiya Hamada | Recording apparatus and method, reproducing apparatus and method, recorded medium, and program |
US20020151992A1 (en) * | 1999-02-01 | 2002-10-17 | Hoffberg Steven M. | Media recording device with packet data interface |
US20020159757A1 (en) * | 1998-12-16 | 2002-10-31 | Hideo Ando | Optical disc for storing moving pictures with text information and apparatus using the disc |
US20020194618A1 (en) * | 2001-04-02 | 2002-12-19 | Matsushita Electric Industrial Co., Ltd. | Video reproduction apparatus, video reproduction method, video reproduction program, and package media for digital video content |
US20030007096A1 (en) * | 2001-07-03 | 2003-01-09 | Takahiro Katayama | Television receiver |
US20030039472A1 (en) * | 2001-08-25 | 2003-02-27 | Kim Doo-Nam | Method of and apparatus for selecting subtitles from an optical recording medium |
US20030078858A1 (en) * | 2001-10-19 | 2003-04-24 | Angelopoulos Tom A. | System and methods for peer-to-peer electronic commerce |
US20030085997A1 (en) * | 2000-04-10 | 2003-05-08 | Satoshi Takagi | Asset management system and asset management method |
US20030086690A1 (en) * | 2001-06-16 | 2003-05-08 | Samsung Electronics Co., Ltd. | Storage medium having preloaded font information, and apparatus for and method of reproducing data from storage medium |
US20030099464A1 (en) * | 2001-11-29 | 2003-05-29 | Oh Yeong-Heon | Optical recording medium and apparatus and method to play the optical recording medium |
US20030103604A1 (en) * | 2000-04-21 | 2003-06-05 | Motoki Kato | Information processing apparatus and method, program and recorded medium |
US6580437B1 (en) * | 2000-06-26 | 2003-06-17 | Siemens Corporate Research, Inc. | System for organizing videos based on closed-caption information |
US20030123845A1 (en) * | 2001-12-28 | 2003-07-03 | Pioneer Corporation | Information recording medium, information recording and/or reproducing apparatus and method, program storage device and computer data signal embodied in carrier wave for controlling record or reproduction and data structure including control signal |
US6597861B1 (en) * | 1996-03-15 | 2003-07-22 | Pioneer Electronic Corporation | Information record medium, apparatus for recording the same and apparatus for reproducing the same |
US20030147629A1 (en) * | 1998-01-21 | 2003-08-07 | Shinichi Kikuchi | Digital information recording/playback system and digital information recording medium |
US20030188312A1 (en) * | 2002-02-28 | 2003-10-02 | Bae Chang Seok | Apparatus and method of reproducing subtitle recorded in digital versatile disk player |
US20030189669A1 (en) * | 2002-04-05 | 2003-10-09 | Bowser Todd S. | Method for off-image data display |
US20030190147A1 (en) * | 2002-03-20 | 2003-10-09 | Lg Electronics Inc. | Method for reproducing sub-picture data in optical disc device, and method for displaying multi-text in optical disc device |
US20030189571A1 (en) * | 1999-11-09 | 2003-10-09 | Macinnis Alexander G. | Video and graphics system with parallel processing of graphics windows |
US20030194211A1 (en) * | 1998-11-12 | 2003-10-16 | Max Abecassis | Intermittently playing a video |
US20030202431A1 (en) * | 2002-04-24 | 2003-10-30 | Kim Mi Hyun | Method for managing summary information of play lists |
US20030206553A1 (en) * | 2001-12-13 | 2003-11-06 | Andre Surcouf | Routing and processing data |
US20030216922A1 (en) * | 2002-05-20 | 2003-11-20 | International Business Machines Corporation | Method and apparatus for performing real-time subtitles translation |
US6661467B1 (en) * | 1994-12-14 | 2003-12-09 | Koninklijke Philips Electronics N.V. | Subtitling transmission system |
US20030235402A1 (en) * | 2002-06-21 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure for managing reproduction of video data recorded thereon |
US20030235404A1 (en) * | 2002-06-24 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure for managing reproduction of multiple reproduction path video data for at least a segment of a title recorded thereon and recording and reproducing methods and apparatuses |
US20030235406A1 (en) * | 2002-06-24 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure including navigation control information for managing reproduction of video data recorded thereon and recording and reproducing methods and apparatuses |
US20040003347A1 (en) * | 2002-06-28 | 2004-01-01 | Ubs Painewebber Inc. | System and method for providing on-line services for multiple entities |
US20040001699A1 (en) * | 2002-06-28 | 2004-01-01 | Seo Kang Soo | Recording medium having data structure for managing reproduction of multiple playback path video data recorded thereon and recording and reproducing methods and apparatuses |
US20040027369A1 (en) * | 2000-12-22 | 2004-02-12 | Peter Rowan Kellock | System and method for media production |
US20040047605A1 (en) * | 2002-09-05 | 2004-03-11 | Seo Kang Soo | Recording medium having data structure for managing reproduction of slideshows recorded thereon and recording and reproducing methods and apparatuses |
US20040054771A1 (en) * | 2002-08-12 | 2004-03-18 | Roe Glen E. | Method and apparatus for the remote retrieval and viewing of diagnostic information from a set-top box |
US20040081434A1 (en) * | 2002-10-15 | 2004-04-29 | Samsung Electronics Co., Ltd. | Information storage medium containing subtitle data for multiple languages using text data and downloadable fonts and apparatus therefor |
US6744998B2 (en) * | 2002-09-23 | 2004-06-01 | Hewlett-Packard Development Company, L.P. | Printer with video playback user interface |
US6747920B2 (en) * | 2001-06-01 | 2004-06-08 | Pioneer Corporation | Information reproduction apparatus and information reproduction |
US20040151472A1 (en) * | 2003-01-20 | 2004-08-05 | Seo Kang Soo | Recording medium having data structure for managing reproduction of still pictures recorded thereon and recording and reproducing methods and apparatuses |
US6792577B1 (en) * | 1999-06-21 | 2004-09-14 | Sony Corporation | Data distribution method and apparatus, and data receiving method and apparatus |
US20040184785A1 (en) * | 2003-01-31 | 2004-09-23 | Jean-Marie Steyer | Device and process for the read-synchronization of video data and of ancillary data and associated products |
US20040202454A1 (en) * | 2003-04-09 | 2004-10-14 | Kim Hyung Sun | Recording medium having a data structure for managing reproduction of text subtitle data and methods and apparatuses of recording and reproducing |
US20040252234A1 (en) * | 2003-06-12 | 2004-12-16 | Park Tae Jin | Management method of option for caption display |
US20050013207A1 (en) * | 2003-05-13 | 2005-01-20 | Yasufumi Tsumagari | Information storage medium, information reproduction device, information reproduction method |
US20050105888A1 (en) * | 2002-11-28 | 2005-05-19 | Toshiya Hamada | Reproducing device, reproduction method, reproduction program, and recording medium |
US20050147387A1 (en) * | 2004-01-06 | 2005-07-07 | Seo Kang S. | Recording medium and method and apparatus for reproducing and recording text subtitle streams |
US20060013563A1 (en) * | 2002-11-15 | 2006-01-19 | Dirk Adolph | Method and apparatus for composition of subtitles |
US20060098936A1 (en) * | 2002-09-25 | 2006-05-11 | Wataru Ikeda | Reproduction device, optical disc, recording medium, program, and reproduction method |
US20060156358A1 (en) * | 2002-10-11 | 2006-07-13 | Dirk Adolph | Method and apparatus for synchronizing data streams containing audio, video and/or other data |
US7134074B2 (en) * | 1998-12-25 | 2006-11-07 | Matsushita Electric Industrial Co., Ltd. | Data processing method and storage medium, and program for causing computer to execute the data processing method |
US20060259941A1 (en) * | 2000-08-23 | 2006-11-16 | Jason Goldberg | Distributed publishing network |
US7151617B2 (en) * | 2001-01-19 | 2006-12-19 | Fuji Photo Film Co., Ltd. | Image synthesizing apparatus |
US7174560B1 (en) * | 1999-02-25 | 2007-02-06 | Sharp Laboratories Of America, Inc. | Method of synchronizing events with a digital television audio-visual program |
US7370274B1 (en) * | 2003-09-18 | 2008-05-06 | Microsoft Corporation | System and method for formatting objects on a page of an electronic document by reference |
US7488353B2 (en) * | 2004-02-28 | 2009-02-10 | Wella Ag | Cationic heteroarylazine dyes and colorants containing these compounds |
US7502549B2 (en) * | 2002-12-26 | 2009-03-10 | Canon Kabushiki Kaisha | Reproducing apparatus |
US7526718B2 (en) * | 2003-04-30 | 2009-04-28 | Hewlett-Packard Development Company, L.P. | Apparatus and method for recording “path-enhanced” multimedia |
US7536705B1 (en) * | 1999-02-22 | 2009-05-19 | Tvworks, Llc | System and method for interactive distribution of selectable presentations |
US7587405B2 (en) * | 2004-02-10 | 2009-09-08 | Lg Electronics Inc. | Recording medium and method and apparatus for decoding text subtitle streams |
Family Cites Families (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB1586431A (en) * | 1978-04-11 | 1981-03-18 | Philips Electronic Associated | Data transmission |
US5497241A (en) | 1993-10-29 | 1996-03-05 | Time Warner Entertainment Co., L.P. | System and method for controlling display of motion picture subtitles in a selected language during play of a software carrier |
RU2229174C2 (en) | 1994-06-03 | 2004-05-20 | Конинклийке Филипс Электроникс Н.В. | Mpeg format data signal recording on data medium and data reproduction |
JPH08275205A (en) | 1995-04-03 | 1996-10-18 | Sony Corp | Method and device for data coding/decoding and coded data recording medium |
JP3577794B2 (en) | 1995-07-18 | 2004-10-13 | ソニー株式会社 | Data decryption device |
JPH09102940A (en) | 1995-08-02 | 1997-04-15 | Sony Corp | Encoding method, encoder, decoder, recording medium and transmitting method for moving image signal |
WO1997019451A1 (en) | 1995-11-24 | 1997-05-29 | Kabushiki Kaisha Toshiba | Multi-language recording medium and reproducing device for the same |
JPH11252518A (en) | 1997-10-29 | 1999-09-17 | Matsushita Electric Ind Co Ltd | Sub-video unit title preparing device and storing medium |
KR100297206B1 (en) | 1999-01-08 | 2001-09-26 | 노영훈 | Caption MP3 data format and a player for reproducing the same |
US7188353B1 (en) * | 1999-04-06 | 2007-03-06 | Sharp Laboratories Of America, Inc. | System for presenting synchronized HTML documents in digital television receivers |
JP4140745B2 (en) | 1999-05-14 | 2008-08-27 | 独立行政法人情報通信研究機構 | How to add timing information to subtitles |
KR20010001725A (en) * | 1999-06-08 | 2001-01-05 | 윤종용 | Method for controlling display of a caption graphic signal |
EP1178691A1 (en) | 2000-07-17 | 2002-02-06 | Deutsche Thomson-Brandt Gmbh | Method and device for recording digital supplementary data |
KR100363170B1 (en) | 2000-12-04 | 2002-12-05 | 삼성전자 주식회사 | Recording medium, reproducing apparatus, and text displaying method thereof |
JP2002290895A (en) | 2001-03-27 | 2002-10-04 | Denon Ltd | Optical disk reproducer |
JP2003061098A (en) | 2001-08-21 | 2003-02-28 | Canon Inc | Image processor, image processing method, recording medium and program |
KR20030030554A (en) * | 2001-10-11 | 2003-04-18 | 삼성전자주식회사 | Caption data transport system and method capable of editting caption data |
JP4078581B2 (en) | 2002-02-04 | 2008-04-23 | ソニー株式会社 | Image processing apparatus and method, recording medium, and program |
EP1571667A3 (en) | 2002-06-05 | 2009-12-09 | Lg Electronics Inc. | Recording medium with a linking area including a synch pattern thereon and apparatus and method for forming, recording, and reproducing the recording medium |
WO2004036574A1 (en) | 2002-10-15 | 2004-04-29 | Samsung Electronics Co., Ltd. | Information storage medium containing subtitle data for multiple languages using text data and downloadable fonts and apparatus therefor |
JP3718498B2 (en) | 2002-11-28 | 2005-11-24 | シャープ株式会社 | Moving image recording / playback method |
TW578068B (en) | 2002-12-30 | 2004-03-01 | Inst Information Industry | Method to set up film search database and recording medium thereof |
KR100739682B1 (en) | 2003-10-04 | 2007-07-13 | 삼성전자주식회사 | Information storage medium storing text based sub-title, processing apparatus and method thereof |
CN1864220B (en) | 2003-10-04 | 2012-08-22 | 三星电子株式会社 | Apparatus for processing text-based subtitle |
KR100788655B1 (en) | 2003-11-10 | 2007-12-26 | 삼성전자주식회사 | Storage medium recorded text-based subtitle data including style information thereon, display playback device and display playback method thereof |
KR100619053B1 (en) * | 2003-11-10 | 2006-08-31 | 삼성전자주식회사 | Information storage medium for subtitle and apparatus therefor |
EP2141704A3 (en) | 2004-02-10 | 2011-03-16 | LG Electronics, Inc. | Recording medium having a data structure for managing data and recording and reproducing method and apparatus |
KR100739680B1 (en) | 2004-02-21 | 2007-07-13 | 삼성전자주식회사 | Storage medium for recording text-based subtitle data including style information, reproducing apparatus, and method therefor |
US7529467B2 (en) | 2004-02-28 | 2009-05-05 | Samsung Electronics Co., Ltd. | Storage medium recording text-based subtitle stream, reproducing apparatus and reproducing method for reproducing text-based subtitle stream recorded on the storage medium |
-
2004
- 2004-12-08 EP EP04808336A patent/EP1716701A1/en not_active Ceased
- 2004-12-08 KR KR1020067018156A patent/KR20070028325A/en not_active Application Discontinuation
- 2004-12-08 WO PCT/KR2004/003205 patent/WO2005076601A1/en active Application Filing
- 2004-12-17 US US11/013,379 patent/US7561780B2/en not_active Expired - Fee Related
-
2009
- 2009-06-19 US US12/487,992 patent/US20090263106A1/en not_active Abandoned
Patent Citations (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3128434A (en) * | 1960-04-28 | 1964-04-07 | Bendix Corp | Transfluxor with amplitude modulated driving pulse input converted to alternating sine wave output |
GB1556366A (en) * | 1976-08-12 | 1979-11-21 | English Electric Co Ltd | Receiving apparatus for combined television/data display |
US4218698A (en) * | 1978-03-13 | 1980-08-19 | Rca Corporation | TV Graphics and mixing control |
US5253530A (en) * | 1991-08-12 | 1993-10-19 | Letcher Iii John H | Method and apparatus for reflective ultrasonic imaging |
US5519443A (en) * | 1991-12-24 | 1996-05-21 | National Captioning Institute, Inc. | Method and apparatus for providing dual language captioning of a television program |
US5467142A (en) * | 1992-04-24 | 1995-11-14 | Victor Company Of Japan, Ltd. | Television receiver for reproducing video images having different aspect ratios and characters transmitted with video images |
US5781687A (en) * | 1993-05-27 | 1998-07-14 | Studio Nemo, Inc. | Script-based, real-time, video editor |
US6128434A (en) * | 1993-10-29 | 2000-10-03 | Kabushiki Kaisha Toshiba | Multilingual recording medium and reproduction apparatus |
US6204883B1 (en) * | 1993-12-21 | 2001-03-20 | Sony Corporation | Video subtitle processing system |
US5537151A (en) * | 1994-02-16 | 1996-07-16 | Ati Technologies Inc. | Close caption support with timewarp |
US5832530A (en) * | 1994-09-12 | 1998-11-03 | Adobe Systems Incorporated | Method and apparatus for identifying words described in a portable electronic document |
US5778142A (en) * | 1994-11-24 | 1998-07-07 | Kabushiki Kaisha Toshiba | Large capacity recording medium, method and apparatus for reproducing data from a large-capacity recording medium, and method and apparatus for recording data on a large-capacity recording medium |
US6661467B1 (en) * | 1994-12-14 | 2003-12-09 | Koninklijke Philips Electronics N.V. | Subtitling transmission system |
US5758007A (en) * | 1995-02-03 | 1998-05-26 | Kabushiki Kaisha Toshiba | Image information encoding/decoding system |
US5731847A (en) * | 1995-03-20 | 1998-03-24 | Sony Corporation | Subtitle encoding/decoding method and apparatus |
US20020067744A1 (en) * | 1995-03-29 | 2002-06-06 | Yukio Fujii | Decoder for compressed and multiplexed video and audio data |
US6009234A (en) * | 1995-04-14 | 1999-12-28 | Kabushiki Kaisha Toshiba | Method of reproducing information |
US5987214A (en) * | 1995-06-30 | 1999-11-16 | Sony Corporation | Apparatus and method for decoding an information page having header information and page data |
US6219043B1 (en) * | 1995-07-13 | 2001-04-17 | Kabushiki Kaisha Toshiba | Method and system to replace sections of an encoded video bitstream |
US5848217A (en) * | 1995-08-02 | 1998-12-08 | Sony Corporation | Subtitle encoding/decoding method and apparatus |
US6115077A (en) * | 1995-08-04 | 2000-09-05 | Sony Corporation | Apparatus and method for encoding and decoding digital video data operable to remove noise from subtitle date included therewith |
US5847770A (en) * | 1995-09-25 | 1998-12-08 | Sony Corporation | Apparatus and method for encoding and decoding a subtitle signal |
US6173113B1 (en) * | 1995-09-29 | 2001-01-09 | Matsushita Electric Industrial Co., Ltd. | Machine readable information recording medium having audio gap information stored therein for indicating a start time and duration of an audio presentation discontinuous period |
US7330643B2 (en) * | 1996-03-15 | 2008-02-12 | Pioneer Electronic Corporation | Information record medium, apparatus for recording the same and apparatus for reproducing the same |
US20030206727A1 (en) * | 1996-03-15 | 2003-11-06 | Pioneer Electronic Corporation | Information record medium, apparatus for recording the same and apparatus for reproducing the same |
US6597861B1 (en) * | 1996-03-15 | 2003-07-22 | Pioneer Electronic Corporation | Information record medium, apparatus for recording the same and apparatus for reproducing the same |
US6253221B1 (en) * | 1996-06-21 | 2001-06-26 | Lg Electronics Inc. | Character display apparatus and method for a digital versatile disc |
US6393196B1 (en) * | 1996-09-27 | 2002-05-21 | Matsushita Electric Industrial Co., Ltd. | Multimedia stream generating method enabling alternative reproduction of video data, and a multimedia optical disk authoring system |
US6222532B1 (en) * | 1997-02-03 | 2001-04-24 | U.S. Philips Corporation | Method and device for navigating through video matter by means of displaying a plurality of key-frames in parallel |
US6230295B1 (en) * | 1997-04-10 | 2001-05-08 | Lsi Logic Corporation | Bitstream assembler for comprehensive verification of circuits, devices, and systems |
US6262775B1 (en) * | 1997-06-17 | 2001-07-17 | Samsung Electronics Co., Ltd. | Caption data processing circuit and method therefor |
US6148140A (en) * | 1997-09-17 | 2000-11-14 | Matsushita Electric Industrial Co., Ltd. | Video data editing apparatus, optical disc for use as a recording medium of a video data editing apparatus, and computer readable recording medium storing an editing program |
US6356567B2 (en) * | 1997-09-26 | 2002-03-12 | International Business Machines Corporation | Embedded clock recovery and difference filtering for an MPEG-2 compliant transport stream |
US6297797B1 (en) * | 1997-10-30 | 2001-10-02 | Kabushiki Kaisha Toshiba | Computer system and closed caption display method |
US20030147629A1 (en) * | 1998-01-21 | 2003-08-07 | Shinichi Kikuchi | Digital information recording/playback system and digital information recording medium |
US20030194211A1 (en) * | 1998-11-12 | 2003-10-16 | Max Abecassis | Intermittently playing a video |
US6259858B1 (en) * | 1998-12-16 | 2001-07-10 | Kabushiki Kaisha Toshiba | Optical disc for storing moving pictures with text information and apparatus using the disc |
US20020159757A1 (en) * | 1998-12-16 | 2002-10-31 | Hideo Ando | Optical disc for storing moving pictures with text information and apparatus using the disc |
US7134074B2 (en) * | 1998-12-25 | 2006-11-07 | Matsushita Electric Industrial Co., Ltd. | Data processing method and storage medium, and program for causing computer to execute the data processing method |
US20020151992A1 (en) * | 1999-02-01 | 2002-10-17 | Hoffberg Steven M. | Media recording device with packet data interface |
US7536705B1 (en) * | 1999-02-22 | 2009-05-19 | Tvworks, Llc | System and method for interactive distribution of selectable presentations |
US7174560B1 (en) * | 1999-02-25 | 2007-02-06 | Sharp Laboratories Of America, Inc. | Method of synchronizing events with a digital television audio-visual program |
US6320621B1 (en) * | 1999-03-27 | 2001-11-20 | Sharp Laboratories Of America, Inc. | Method of selecting a digital closed captioning service |
US6792577B1 (en) * | 1999-06-21 | 2004-09-14 | Sony Corporation | Data distribution method and apparatus, and data receiving method and apparatus |
US20030189571A1 (en) * | 1999-11-09 | 2003-10-09 | Macinnis Alexander G. | Video and graphics system with parallel processing of graphics windows |
US20010044809A1 (en) * | 2000-03-29 | 2001-11-22 | Parasnis Shashank Mohan | Process of localizing objects in markup language documents |
US20030085997A1 (en) * | 2000-04-10 | 2003-05-08 | Satoshi Takagi | Asset management system and asset management method |
US20020135608A1 (en) * | 2000-04-21 | 2002-09-26 | Toshiya Hamada | Recording apparatus and method, reproducing apparatus and method, recorded medium, and program |
US20020135607A1 (en) * | 2000-04-21 | 2002-09-26 | Motoki Kato | Information processing apparatus and method, program, and recorded medium |
US20030103604A1 (en) * | 2000-04-21 | 2003-06-05 | Motoki Kato | Information processing apparatus and method, program and recorded medium |
US20020087999A1 (en) * | 2000-04-26 | 2002-07-04 | Sony Corporation | Scalable filtering table |
US20020010924A1 (en) * | 2000-05-03 | 2002-01-24 | Morteza Kalhour | Push method and system |
US6580437B1 (en) * | 2000-06-26 | 2003-06-17 | Siemens Corporate Research, Inc. | System for organizing videos based on closed-caption information |
US20020004755A1 (en) * | 2000-06-29 | 2002-01-10 | Neil Balthaser | Methods, systems, and processes for the design and creation of rich-media applications via the internet |
US20060259941A1 (en) * | 2000-08-23 | 2006-11-16 | Jason Goldberg | Distributed publishing network |
US20020087569A1 (en) * | 2000-12-07 | 2002-07-04 | International Business Machines Corporation | Method and system for the automatic generation of multi-lingual synchronized sub-titles for audiovisual data |
US20040027369A1 (en) * | 2000-12-22 | 2004-02-12 | Peter Rowan Kellock | System and method for media production |
US7151617B2 (en) * | 2001-01-19 | 2006-12-19 | Fuji Photo Film Co., Ltd. | Image synthesizing apparatus |
US20020106193A1 (en) * | 2001-02-05 | 2002-08-08 | Park Sung-Wook | Data storage medium in which multiple bitstreams are recorded, apparatus and method for reproducing the multiple bitstreams, and apparatus and method for reproducing the multiple bitstreams |
US20020194618A1 (en) * | 2001-04-02 | 2002-12-19 | Matsushita Electric Industrial Co., Ltd. | Video reproduction apparatus, video reproduction method, video reproduction program, and package media for digital video content |
US6747920B2 (en) * | 2001-06-01 | 2004-06-08 | Pioneer Corporation | Information reproduction apparatus and information reproduction |
US20030086690A1 (en) * | 2001-06-16 | 2003-05-08 | Samsung Electronics Co., Ltd. | Storage medium having preloaded font information, and apparatus for and method of reproducing data from storage medium |
US20030007096A1 (en) * | 2001-07-03 | 2003-01-09 | Takahiro Katayama | Television receiver |
US20030039472A1 (en) * | 2001-08-25 | 2003-02-27 | Kim Doo-Nam | Method of and apparatus for selecting subtitles from an optical recording medium |
US20030078858A1 (en) * | 2001-10-19 | 2003-04-24 | Angelopoulos Tom A. | System and methods for peer-to-peer electronic commerce |
US20030099464A1 (en) * | 2001-11-29 | 2003-05-29 | Oh Yeong-Heon | Optical recording medium and apparatus and method to play the optical recording medium |
US20030206553A1 (en) * | 2001-12-13 | 2003-11-06 | Andre Surcouf | Routing and processing data |
US20030123845A1 (en) * | 2001-12-28 | 2003-07-03 | Pioneer Corporation | Information recording medium, information recording and/or reproducing apparatus and method, program storage device and computer data signal embodied in carrier wave for controlling record or reproduction and data structure including control signal |
US20030188312A1 (en) * | 2002-02-28 | 2003-10-02 | Bae Chang Seok | Apparatus and method of reproducing subtitle recorded in digital versatile disk player |
US20030190147A1 (en) * | 2002-03-20 | 2003-10-09 | Lg Electronics Inc. | Method for reproducing sub-picture data in optical disc device, and method for displaying multi-text in optical disc device |
US20030189669A1 (en) * | 2002-04-05 | 2003-10-09 | Bowser Todd S. | Method for off-image data display |
US20030202431A1 (en) * | 2002-04-24 | 2003-10-30 | Kim Mi Hyun | Method for managing summary information of play lists |
US20030216922A1 (en) * | 2002-05-20 | 2003-11-20 | International Business Machines Corporation | Method and apparatus for performing real-time subtitles translation |
US20030235402A1 (en) * | 2002-06-21 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure for managing reproduction of video data recorded thereon |
US20030235404A1 (en) * | 2002-06-24 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure for managing reproduction of multiple reproduction path video data for at least a segment of a title recorded thereon and recording and reproducing methods and apparatuses |
US20030235406A1 (en) * | 2002-06-24 | 2003-12-25 | Seo Kang Soo | Recording medium having data structure including navigation control information for managing reproduction of video data recorded thereon and recording and reproducing methods and apparatuses |
US20040001699A1 (en) * | 2002-06-28 | 2004-01-01 | Seo Kang Soo | Recording medium having data structure for managing reproduction of multiple playback path video data recorded thereon and recording and reproducing methods and apparatuses |
US20040003347A1 (en) * | 2002-06-28 | 2004-01-01 | Ubs Painewebber Inc. | System and method for providing on-line services for multiple entities |
US20040054771A1 (en) * | 2002-08-12 | 2004-03-18 | Roe Glen E. | Method and apparatus for the remote retrieval and viewing of diagnostic information from a set-top box |
US20040047592A1 (en) * | 2002-09-05 | 2004-03-11 | Seo Kang Soo | Recording medium having data structure of playlist marks for managing reproduction of still images recorded thereon and recording and reproducing methods and apparatuses |
US20040047605A1 (en) * | 2002-09-05 | 2004-03-11 | Seo Kang Soo | Recording medium having data structure for managing reproduction of slideshows recorded thereon and recording and reproducing methods and apparatuses |
US20040047591A1 (en) * | 2002-09-05 | 2004-03-11 | Seo Kang Soo | Recording medium having data structure for managing reproduction of still images recorded thereon and recording and reproducing methods and apparatuses |
US6744998B2 (en) * | 2002-09-23 | 2004-06-01 | Hewlett-Packard Development Company, L.P. | Printer with video playback user interface |
US20060098936A1 (en) * | 2002-09-25 | 2006-05-11 | Wataru Ikeda | Reproduction device, optical disc, recording medium, program, and reproduction method |
US20060156358A1 (en) * | 2002-10-11 | 2006-07-13 | Dirk Adolph | Method and apparatus for synchronizing data streams containing audio, video and/or other data |
US20040081434A1 (en) * | 2002-10-15 | 2004-04-29 | Samsung Electronics Co., Ltd. | Information storage medium containing subtitle data for multiple languages using text data and downloadable fonts and apparatus therefor |
US20060013563A1 (en) * | 2002-11-15 | 2006-01-19 | Dirk Adolph | Method and apparatus for composition of subtitles |
US20050105888A1 (en) * | 2002-11-28 | 2005-05-19 | Toshiya Hamada | Reproducing device, reproduction method, reproduction program, and recording medium |
US7502549B2 (en) * | 2002-12-26 | 2009-03-10 | Canon Kabushiki Kaisha | Reproducing apparatus |
US20040151472A1 (en) * | 2003-01-20 | 2004-08-05 | Seo Kang Soo | Recording medium having data structure for managing reproduction of still pictures recorded thereon and recording and reproducing methods and apparatuses |
US20040184785A1 (en) * | 2003-01-31 | 2004-09-23 | Jean-Marie Steyer | Device and process for the read-synchronization of video data and of ancillary data and associated products |
US20040202454A1 (en) * | 2003-04-09 | 2004-10-14 | Kim Hyung Sun | Recording medium having a data structure for managing reproduction of text subtitle data and methods and apparatuses of recording and reproducing |
US7526718B2 (en) * | 2003-04-30 | 2009-04-28 | Hewlett-Packard Development Company, L.P. | Apparatus and method for recording “path-enhanced” multimedia |
US20050013207A1 (en) * | 2003-05-13 | 2005-01-20 | Yasufumi Tsumagari | Information storage medium, information reproduction device, information reproduction method |
US20040252234A1 (en) * | 2003-06-12 | 2004-12-16 | Park Tae Jin | Management method of option for caption display |
US7370274B1 (en) * | 2003-09-18 | 2008-05-06 | Microsoft Corporation | System and method for formatting objects on a page of an electronic document by reference |
US20050147387A1 (en) * | 2004-01-06 | 2005-07-07 | Seo Kang S. | Recording medium and method and apparatus for reproducing and recording text subtitle streams |
US7587405B2 (en) * | 2004-02-10 | 2009-09-08 | Lg Electronics Inc. | Recording medium and method and apparatus for decoding text subtitle streams |
US7488353B2 (en) * | 2004-02-28 | 2009-02-10 | Wella Ag | Cationic heteroarylazine dyes and colorants containing these compounds |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080298775A1 (en) * | 2004-02-21 | 2008-12-04 | Samsung Electronics Co., Ltd. | Storage medium for storing text-based subtitle data including style information, and reproducing apparatus and method for reproducing text-based subtitle data including style information |
US8873931B2 (en) | 2004-02-21 | 2014-10-28 | Samsung Electronics Co., Ltd. | Storage medium for storing text-based subtitle data including style information, and reproducing apparatus and method for reproducing text-based subtitle data including style information |
US20090185075A1 (en) * | 2004-02-28 | 2009-07-23 | Samsung Electronics Co., Ltd | Storage medium recording text-based subtitle stream, reproducing apparatus and reproducing method for reproducing text-based subtitle stream recorded on the storage medium |
US8437612B2 (en) | 2004-02-28 | 2013-05-07 | Samsung Electronics Co., Ltd. | Storage medium recording text-based subtitle stream, reproducing apparatus and reproducing method for reproducing text-based subtitle stream recorded on the storage medium |
US20070025698A1 (en) * | 2005-07-29 | 2007-02-01 | Kabushiki Kaisha Toshiba | Information playback method using information recording medium |
US20110227911A1 (en) * | 2010-03-22 | 2011-09-22 | Lg Electronics Inc. | Image display device and method for operating the same |
Also Published As
Publication number | Publication date |
---|---|
KR20070028325A (en) | 2007-03-12 |
US7561780B2 (en) | 2009-07-14 |
US20050196147A1 (en) | 2005-09-08 |
WO2005076601A1 (en) | 2005-08-18 |
EP1716701A1 (en) | 2006-11-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7561780B2 (en) | Text subtitle decoder and method for decoding text subtitle streams | |
US7587405B2 (en) | Recording medium and method and apparatus for decoding text subtitle streams | |
US7643732B2 (en) | Recording medium and method and apparatus for decoding text subtitle streams | |
US7982802B2 (en) | Text subtitle decoder and method for decoding text subtitle streams | |
US7848617B2 (en) | Recording medium, method, and apparatus for reproducing text subtitle streams | |
US7756398B2 (en) | Recording medium and method and apparatus for reproducing text subtitle stream for updating palette information | |
US20050196146A1 (en) | Method for reproducing text subtitle and text subtitle decoding system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: LG ELECTRONICS INC., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SEO, KANG SOO;KIM, BYUNG JIN;YOO, JEA YONG;REEL/FRAME:022883/0948 Effective date: 20041213 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |