US20090254562A1 - Automatic Metadata Extraction and Metadata Controlled Production Process - Google Patents
Automatic Metadata Extraction and Metadata Controlled Production Process Download PDFInfo
- Publication number
- US20090254562A1 US20090254562A1 US11/990,804 US99080406A US2009254562A1 US 20090254562 A1 US20090254562 A1 US 20090254562A1 US 99080406 A US99080406 A US 99080406A US 2009254562 A1 US2009254562 A1 US 2009254562A1
- Authority
- US
- United States
- Prior art keywords
- metadata
- field device
- server
- audio
- edited
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/16—Analogue secrecy systems; Analogue subscription systems
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/11—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
Definitions
- This invention relates to a technique for capturing and processing metadata during production of an audio visual program such as a television news story.
- Metadata comprises data about the captured audio and/or video information.
- metadata can include simple information, such as the time and date of the capture of the associated audio and/or video information.
- More complex metadata can include identification of the content of the audio and/or video information, as well as data associated with authoring and editing of that content.
- a technique for associating metadata with at least one of audio and video information commences by transmitting metadata to a field device associated with the capture of the at least one of the audio and video information so that the metadata can be used by an operator of that device.
- the metadata is also transmitted to a storage mechanism, such as a server or the like, destined to receive the at least one of the audio and video information along with edited metadata, such as edited metadata received from the field device.
- the metadata received at the server undergoes updating in accordance with the edited metadata. In this way metadata created at the outset of production can undergo editing by field personnel in association with the capture of the audio and/or video information
- the edited metadata from the field can serve to update the original metadata at a server that receives the audio and/or video information captured in the field by a capture device, such as a camcorder to the like.
- a capture device such as a camcorder to the like.
- the updated metadata stored at the server will provide information useful for association with essence objects captured by the capture device to enable secure identification of such objects.
- FIG. 1 depicts a block schematic diagram of a system for audio visual production illustrating metadata flow in accordance with the present principles.
- FIG. 1 depicts a block schematic diagram of a system 10 , in accordance with the present principles for capturing useful metadata in connection with the capture of audio and/or video information.
- the system 10 takes the form of a system for the production of television news although the metadata capture technique of the present principles has application to other systems.
- the solid lines within FIG. 1 indicate metadata flow whereas the dashed blocks indicate the type of metadata added and/or transmitted.
- the system of FIG. 10 comprises a programmed processor 12 , typically in the form of a news room computer system (NRCS) capable of generating assignment information for use by personnel who gather news stories.
- NRCS news room computer system
- the assignment information generated by the processor 12 can include metadata associated with a news story to be pursued by news gathering personnel.
- metadata can originate from a variety of sources.
- the processor 12 can receive metadata from one or more news wires services 14 (only one of which is shown), such as Reuters, Associated Press, United Press International, for example. Metadata from the newswire service 14 can include names, story location(s) any other information, as indicated by the legend in block 16 , for attachment to a story assignment generated by the processor 12 .
- the processor 12 can also receive metadata from an assignment editor 18 , an individual who creates and edits assignment information for the news gathering personnel.
- the assignment editor 18 will typically add metadata from a file, sometimes referred to as a “tickler” file, that can include the names of individuals for interview and story locations, as indicated by the legend in block 20 .
- the processor 12 can receive status metadata, as indicated by the legend in block 21 .
- the processor 12 provides assignment information, in the form of an assignment grid 22 that designates which news gathering personnel handle which stories.
- the processor 12 communicates the assignment information and metadata to a server 24 which also stores information.
- the server 24 can also store metadata as received through an interface 25 , designated by the legend “mailer”.
- Such status information can include slug data (i.e., data related to slugs which constitute blank space reserved for future stories), story locations, names, set-up times, show times as indicated by the legend in block 26 .
- a communications network 28 typically a wireless communications network transmits metadata from the server 24 to a field device 30 , such as a laptop or personal computer for example, associated with a reporter 32 and videographer (cameraman) 34 assigned to a particular news story.
- the server 24 makes use of the assignment data contained in the assignment grid 22 to transmit appropriate metadata to the corresponding field device 30 so that the reporter 32 and videographer receive information related stories to which they have been assigned.
- the server 24 provides the reporter 32 and videographer 34 with instructions where to go, whom to interview, background wire stories from the wire service 14 , notes from previous coverage and other information relevant to the assigned story or stories.
- the metadata from the server 24 also includes the identity of the specific story assigned to a particular news team (e.g., the reporter 32 and videographer 34 ). This identity is usually referred to by the term Media Object Server Identification or MOS-ID.
- the reporter 32 can add and/or edit the metadata transmitted from the server 24 .
- the reporter 32 can add information related the identity (e.g., the name) of individuals interviewed as well as notes created by the reporter, as indicated by the legend in block 36 .
- the videographer 34 can add metadata indicative of preferred settings and/or other information related to the captured image, as indicated by the legend in block 38 .
- FIG. 1 depicts both the reporter 32 and videographer 34 , a single individual could easily serve both functions.
- the field device 30 communicates the metadata it has received to a capture device 40 that captures at least one of audio and/or video information.
- the capture device 40 takes the form of a camcorder, such as the Infinity series camcorder available from Thomson Grass Valley, although the capture device could take comprise the combination of a video camera and a recording device, such as a videotape recorder, a video disc recorder or a server.
- the metadata received by the capture device 40 from the field device 30 can also include one of more of the following: Global Positioning Satellite information, compass bearings, lens settings, aspect ratio data, as well as any other data generated by the field device 30 and/or entered manually by one or more of the reporter 32 and videographer 34 .
- Such metadata is indicated by the legend in block 41 .
- the metadata can also include information entered directly to the capture device 40 by the videographer 34 through entry of one or more keys (not shown) on the capture device. Note that while the field device 30 and the capture device 40 are depicted as separate elements, it is possible that a single unit could serve the functions of the capture device and the field device.
- the metadata from the field device 30 sent to the capture device 40 gets entered into the file structure of the captured audio and/or video information.
- the file structure comprises the Media Exchange File or MXF structure but other file structures could be used. Entering such metadata into the file structure of the captured audio and/or visual information in the manner described resolves the long-standing metadata paradox, namely, how to create metadata at the outset of creating an audio-visual program, such as a news program.
- the metadata incorporated in the file structure of the captured audio and/or video information captured by the capture device 40 includes the metadata already previously created by the assignment editor 18 and the wire service 14 .
- the field device 30 simply “harvests” metadata already existing on the server 24 .
- the audio and/or video information captured by the capture device, and the metadata entered into the file structure of such information undergoes downloading from capture device 40 to a storage mechanism 42 .
- the storage mechanism 42 bears the designation “Ingest Station (IS)” because it serves to ingest (receive) the audio and/or video information and associated metadata from the capture device 40 .
- the storage mechanism 42 In addition to receiving metadata from the field device 30 via the audio and/or video information downloaded from the capture device 40 , the storage mechanism 42 also receives the same metadata originally sent to the field device 30 by the server 24 , including the MOS-ID identifying the server.
- Additional metadata such as metadata related to stories, slugs schedules and estimated time of arrival (ETA), as indicated by the legend in block 44 , can be added to the metadata from the processor 12 .
- the storage mechanism 42 can also metadata from an external source (not shown) related to show schedules as indicated by the legend in box 45 .
- the storage mechanism 42 has the ability to match the metadata received from the capture device 40 with the metadata received from the server 24 by matching the MOS-ID of the metadata received from the various sources.
- the storage mechanism 42 will look for the MOS-ID comprising part of the metadata in the file structure of the audio and/or video information downloaded by the capture device 40 to match it with the MOS-ID in the metadata received from the server 24 .
- the server 42 can update the status information associated with audio and/or video information (e.g., news stories) created via the system 10 based on the updating of the metadata created by the processor 12 with edited metadata from the capture device 40 .
- an editor using a news edit block comprised of editing software running on a processor with an associated terminal, can edit audio and/or video information downloaded to the storage mechanism 42 using the metadata information to determine not only the identity of the reporter and videographer assigned to the story, but also the status of the story, including, but not limited to the estimated time of arrival (ETA) for material not yet downloaded.
- ETA estimated time of arrival
- Extraction and use of the metadata can occur elsewhere in production process.
- a reporter 32 and videographer 34 have the task of interviewing a certain individual, say John Doe.
- the cameraman 32 and/or videographer 34 can check John Doe's name against the list of names in the assignment information comprising part of the metadata transmitted to the field device 30 . If the name matches, then the cameraman 32 or videographer can add the name to one or more of the video frames captured by the capture device 40 .
- the non-linear editor (NLE) program running on the news edit block 48 or elsewhere in the system 10 of FIG. 1 can enter the name into a character generator (CG) template with one more commands, thus avoiding spelling or transcription errors. This will also reduce the likelihood of misidentification of an interview subject.
- CG character generator
- the server 42 serves as the source of metadata for the news edit block 48 . Additionally, the server 42 as serves as the metadata source for a simple data base (SDB), not shown that stories a play list of stories edited via the news edit block 48 . Additionally the server 42 can also provide metadata to a news room computer system, an example of which is the Electronic News Room Product System (ENPS) available from Associated Press.
- ENPS Electronic News Room Product System
- the news edit block 48 not only receives the both audio and/or video information and associated metadata from the server 42 but also receives metadata associated with stories, slugs running order computer graphics production techniques and the like, as indicated by the legend in block 51 .
- Such metadata can originate from one of several sources (not shown) such as an news room computer system (not shown) and or an integrated television production system (IPS) (not shown).
- the new edit terminal can provide edited news story clips (e.g., edited audio and/or video information), to a playback unit 50 .
- the new edit terminal 48 can supply edited news clips (e.g., edited audio and/or video information) and accompanying metadata to the IPS.
- the edited news clips (and accompany metadata) provided by the server 42 to the Integrated production System can include information, including metadata, from a character generator (not shown) as indicated by the legend in block 53 .
- the new edit block 48 can also supply the edited news story clips to an asset management system (not shown) for other distribution.
- asset management system can comprise the “NewsBrowse” system available from Thomson Grass Valley.
- the playback unit 50 not only receives the news clips from the new block 48 but also receives audio and/or video information from the server 42 and from the IPS.
- the information from the IPS can also include metadata representing status information, as indicated by the legend in block 54 .
- the playback unit 50 will feed edited news clips from the new edit terminal 48 , as well as audio and/or video information from the server 42 and the IPS, to one or more of the news edit block 48 , the news room computer system, the asset management system (e.g., NewsBrowse), an editor (not shown) and a switcher/production system (PS) 57 .
- the clips provided by the playback system 50 to the news edit block 48 can include metadata associated with the slugs and their running order as indicated by the legend in block 56 .
- the clips provided by the playback system 50 to the news room computer system, the editor and/or asset management system can include status metadata, as indicated by the legend in block 58 .
- Metadata typically in the form of status information, as indicated by the legend in block 59 can accompany the audio and/or video information received by the switcher/production system 57 from the payback unit 52 .
- Metadata, typically containing formation related to stories, slugs, running order on-air talent graphics special effects, character generator data, production techniques, camera data and the like, as indicated by the legend in block 60 can accompany assignment information and other data received by the switcher/production system 57 from the server 24 .
- Metadata, typically in the form of character generator proxy and insertion information, as indicated by the legend in block 62 can accompany the audio and/or video information, and other metadata, received by the switcher/production system 56 from the news edit terminal 48 .
- the switcher/production system 57 can supply audio and/or video information (and accompanying metadata) to the playback unit 50 and to other system devices (not shown).
- the audio and/or video information supplied by the switcher/production system 57 to the playback unit 50 can include metadata containing status information; such information associated with released audio and/or video information, as indicated by the legend in block 64 .
- the audio and/or video information and accompanying metadata supplied from the switcher/production system 57 to the other system devices can include metadata related to Global Positioning Satellite data and lens information for the capture device 40 or other such devices (not shown), graphics and character generator information, aspect ratio to the switcher/production system, for example, as indicated by the legend in block 66 .
- the foregoing describes a technique capturing and processing metadata during production of an audio visual program.
Abstract
Metadata generated at the outset of an audio visual program, such as a television undergoes transmission to a field device associated with a capture device operated by production personnel, such as a news reporter and/or a videographer to capture one of audio and/or video information. The production personnel will typically edit the metadata for incorporation into the file structure of audio and/or visual information captured by the capture device. A server 42 receives and updates the original metadata using the metadata in the file structure of the capture audio and/or video information, thus effectively harvesting the original metadata.
Description
- This application claims priority under 35 U.S.C. 119(e) to U.S. Provisional Patent Application Ser. No. 60/713,848, filed Sep. 2, 2005, the teachings of which are incorporated herein.
- This invention relates to a technique for capturing and processing metadata during production of an audio visual program such as a television news story.
- Advances in the development of television production equipment allow recording devices like video recorders, servers and camcorders, for example, to record not only audio and/or video information, but metadata as well. Such metadata comprises data about the captured audio and/or video information. Such metadata can include simple information, such as the time and date of the capture of the associated audio and/or video information. More complex metadata can include identification of the content of the audio and/or video information, as well as data associated with authoring and editing of that content.
- Techniques for capturing metadata in conjunction with the capture of audio and/or video information, as well as techniques for associating such metadata have become well known. What has proven elusive is the ability to make good use of the metadata. In other words, the problem facing production personnel is the generation of “useful” metadata that can assist production personnel, rather than the creation of metadata that simply gets stored without aiding in the production process.
- Thus, a need exists for a technique for generating and associating useful metadata in conjunction with the production of audio-visual content, and particularly, event driven audio visual content, such a television news material.
- Briefly, in accordance with an illustrative embodiment of the present principles, there is provided a technique for associating metadata with at least one of audio and video information. The method commences by transmitting metadata to a field device associated with the capture of the at least one of the audio and video information so that the metadata can be used by an operator of that device. The metadata is also transmitted to a storage mechanism, such as a server or the like, destined to receive the at least one of the audio and video information along with edited metadata, such as edited metadata received from the field device. The metadata received at the server undergoes updating in accordance with the edited metadata. In this way metadata created at the outset of production can undergo editing by field personnel in association with the capture of the audio and/or video information
- In practice, the edited metadata from the field can serve to update the original metadata at a server that receives the audio and/or video information captured in the field by a capture device, such as a camcorder to the like. Thus, the updated metadata stored at the server will provide information useful for association with essence objects captured by the capture device to enable secure identification of such objects.
-
FIG. 1 depicts a block schematic diagram of a system for audio visual production illustrating metadata flow in accordance with the present principles. -
FIG. 1 depicts a block schematic diagram of a system 10, in accordance with the present principles for capturing useful metadata in connection with the capture of audio and/or video information. In the illustrated embodiment, the system 10 takes the form of a system for the production of television news although the metadata capture technique of the present principles has application to other systems. The solid lines withinFIG. 1 indicate metadata flow whereas the dashed blocks indicate the type of metadata added and/or transmitted. - The system of
FIG. 10 comprises a programmedprocessor 12, typically in the form of a news room computer system (NRCS) capable of generating assignment information for use by personnel who gather news stories. Newsroom computer systems capable of performing this function are available from manufacturers of automated news systems. The assignment information generated by theprocessor 12 can include metadata associated with a news story to be pursued by news gathering personnel. Such metadata can originate from a variety of sources. For example, theprocessor 12 can receive metadata from one or more news wires services 14 (only one of which is shown), such as Reuters, Associated Press, United Press International, for example. Metadata from thenewswire service 14 can include names, story location(s) any other information, as indicated by the legend inblock 16, for attachment to a story assignment generated by theprocessor 12. In practice, theprocessor 12 can also receive metadata from anassignment editor 18, an individual who creates and edits assignment information for the news gathering personnel. Theassignment editor 18 will typically add metadata from a file, sometimes referred to as a “tickler” file, that can include the names of individuals for interview and story locations, as indicated by the legend inblock 20. In addition, theprocessor 12 can receive status metadata, as indicated by the legend inblock 21. - In the process of generating assignment information, the
processor 12 provides assignment information, in the form of anassignment grid 22 that designates which news gathering personnel handle which stories. Theprocessor 12 communicates the assignment information and metadata to aserver 24 which also stores information. Theserver 24 can also store metadata as received through an interface 25, designated by the legend “mailer”. Such status information can include slug data (i.e., data related to slugs which constitute blank space reserved for future stories), story locations, names, set-up times, show times as indicated by the legend inblock 26. - A
communications network 28, typically a wireless communications network transmits metadata from theserver 24 to afield device 30, such as a laptop or personal computer for example, associated with areporter 32 and videographer (cameraman) 34 assigned to a particular news story. Theserver 24 makes use of the assignment data contained in theassignment grid 22 to transmit appropriate metadata to thecorresponding field device 30 so that thereporter 32 and videographer receive information related stories to which they have been assigned. In other words, theserver 24 provides thereporter 32 andvideographer 34 with instructions where to go, whom to interview, background wire stories from thewire service 14, notes from previous coverage and other information relevant to the assigned story or stories. Additionally, the metadata from theserver 24 also includes the identity of the specific story assigned to a particular news team (e.g., thereporter 32 and videographer 34). This identity is usually referred to by the term Media Object Server Identification or MOS-ID. Using thefield device 30, thereporter 32 can add and/or edit the metadata transmitted from theserver 24. For example, thereporter 32 can add information related the identity (e.g., the name) of individuals interviewed as well as notes created by the reporter, as indicated by the legend inblock 36. Thevideographer 34 can add metadata indicative of preferred settings and/or other information related to the captured image, as indicated by the legend inblock 38. AlthoughFIG. 1 depicts both thereporter 32 andvideographer 34, a single individual could easily serve both functions. - The
field device 30 communicates the metadata it has received to acapture device 40 that captures at least one of audio and/or video information. In practice, thecapture device 40 takes the form of a camcorder, such as the Infinity series camcorder available from Thomson Grass Valley, although the capture device could take comprise the combination of a video camera and a recording device, such as a videotape recorder, a video disc recorder or a server. The metadata received by thecapture device 40 from thefield device 30 can also include one of more of the following: Global Positioning Satellite information, compass bearings, lens settings, aspect ratio data, as well as any other data generated by thefield device 30 and/or entered manually by one or more of thereporter 32 andvideographer 34. Such metadata is indicated by the legend inblock 41. The metadata can also include information entered directly to thecapture device 40 by thevideographer 34 through entry of one or more keys (not shown) on the capture device. Note that while thefield device 30 and thecapture device 40 are depicted as separate elements, it is possible that a single unit could serve the functions of the capture device and the field device. - The metadata from the
field device 30 sent to thecapture device 40, along with metadata entered to the capture device by thevideographer 34, gets entered into the file structure of the captured audio and/or video information. Typically the file structure comprises the Media Exchange File or MXF structure but other file structures could be used. Entering such metadata into the file structure of the captured audio and/or visual information in the manner described resolves the long-standing metadata paradox, namely, how to create metadata at the outset of creating an audio-visual program, such as a news program. As discussed, the metadata incorporated in the file structure of the captured audio and/or video information captured by thecapture device 40 includes the metadata already previously created by theassignment editor 18 and thewire service 14. Thus, thefield device 30 simply “harvests” metadata already existing on theserver 24. - The audio and/or video information captured by the capture device, and the metadata entered into the file structure of such information undergoes downloading from
capture device 40 to astorage mechanism 42. In the illustrative embodiment ofFIG. 1 , thestorage mechanism 42 bears the designation “Ingest Station (IS)” because it serves to ingest (receive) the audio and/or video information and associated metadata from thecapture device 40. In addition to receiving metadata from thefield device 30 via the audio and/or video information downloaded from thecapture device 40, thestorage mechanism 42 also receives the same metadata originally sent to thefield device 30 by theserver 24, including the MOS-ID identifying the server. Additional metadata, such as metadata related to stories, slugs schedules and estimated time of arrival (ETA), as indicated by the legend inblock 44, can be added to the metadata from theprocessor 12, Further, thestorage mechanism 42 can also metadata from an external source (not shown) related to show schedules as indicated by the legend inbox 45. - The
storage mechanism 42 has the ability to match the metadata received from thecapture device 40 with the metadata received from theserver 24 by matching the MOS-ID of the metadata received from the various sources. In particular, thestorage mechanism 42 will look for the MOS-ID comprising part of the metadata in the file structure of the audio and/or video information downloaded by thecapture device 40 to match it with the MOS-ID in the metadata received from theserver 24. In this way, thestorage mechanism 42 can know what to do with such metadata. More importantly, theserver 42 can update the status information associated with audio and/or video information (e.g., news stories) created via the system 10 based on the updating of the metadata created by theprocessor 12 with edited metadata from thecapture device 40. - Many benefits result from using the metadata created and/or edited entered by one or both of the
reporter 32 andvideographer 34 to update the original metadata stored in theserver 24. For example, an editor using a news edit block, comprised of editing software running on a processor with an associated terminal, can edit audio and/or video information downloaded to thestorage mechanism 42 using the metadata information to determine not only the identity of the reporter and videographer assigned to the story, but also the status of the story, including, but not limited to the estimated time of arrival (ETA) for material not yet downloaded. - Extraction and use of the metadata can occur elsewhere in production process. For example, consider the circumstance when a
reporter 32 andvideographer 34 have the task of interviewing a certain individual, say John Doe. During the interview process, thecameraman 32 and/orvideographer 34 can check John Doe's name against the list of names in the assignment information comprising part of the metadata transmitted to thefield device 30. If the name matches, then thecameraman 32 or videographer can add the name to one or more of the video frames captured by thecapture device 40. The non-linear editor (NLE) program running on thenews edit block 48 or elsewhere in the system 10 ofFIG. 1 can enter the name into a character generator (CG) template with one more commands, thus avoiding spelling or transcription errors. This will also reduce the likelihood of misidentification of an interview subject. - The
server 42 serves as the source of metadata for thenews edit block 48. Additionally, theserver 42 as serves as the metadata source for a simple data base (SDB), not shown that stories a play list of stories edited via thenews edit block 48. Additionally theserver 42 can also provide metadata to a news room computer system, an example of which is the Electronic News Room Product System (ENPS) available from Associated Press. Thenews edit block 48 not only receives the both audio and/or video information and associated metadata from theserver 42 but also receives metadata associated with stories, slugs running order computer graphics production techniques and the like, as indicated by the legend inblock 51. Such metadata can originate from one of several sources (not shown) such as an news room computer system (not shown) and or an integrated television production system (IPS) (not shown). The new edit terminal can provide edited news story clips (e.g., edited audio and/or video information), to a playback unit 50. In addition, thenew edit terminal 48 can supply edited news clips (e.g., edited audio and/or video information) and accompanying metadata to the IPS. The edited news clips (and accompany metadata) provided by theserver 42 to the Integrated production System can include information, including metadata, from a character generator (not shown) as indicated by the legend inblock 53. Thenew edit block 48 can also supply the edited news story clips to an asset management system (not shown) for other distribution. Such an asset management system can comprise the “NewsBrowse” system available from Thomson Grass Valley. - The playback unit 50 not only receives the news clips from the
new block 48 but also receives audio and/or video information from theserver 42 and from the IPS. The information from the IPS can also include metadata representing status information, as indicated by the legend inblock 54. The playback unit 50 will feed edited news clips from thenew edit terminal 48, as well as audio and/or video information from theserver 42 and the IPS, to one or more of thenews edit block 48, the news room computer system, the asset management system (e.g., NewsBrowse), an editor (not shown) and a switcher/production system (PS) 57. The clips provided by the playback system 50 to thenews edit block 48 can include metadata associated with the slugs and their running order as indicated by the legend inblock 56. The clips provided by the playback system 50 to the news room computer system, the editor and/or asset management system (NewsBrowse) can include status metadata, as indicated by the legend in block 58. Metadata typically in the form of status information, as indicated by the legend inblock 59, can accompany the audio and/or video information received by the switcher/production system 57 from thepayback unit 52. Metadata, typically containing formation related to stories, slugs, running order on-air talent graphics special effects, character generator data, production techniques, camera data and the like, as indicated by the legend inblock 60, can accompany assignment information and other data received by the switcher/production system 57 from theserver 24. Metadata, typically in the form of character generator proxy and insertion information, as indicated by the legend inblock 62, can accompany the audio and/or video information, and other metadata, received by the switcher/production system 56 from thenews edit terminal 48. - The switcher/
production system 57 can supply audio and/or video information (and accompanying metadata) to the playback unit 50 and to other system devices (not shown). The audio and/or video information supplied by the switcher/production system 57 to the playback unit 50 can include metadata containing status information; such information associated with released audio and/or video information, as indicated by the legend inblock 64. The audio and/or video information and accompanying metadata supplied from the switcher/production system 57 to the other system devices, can include metadata related to Global Positioning Satellite data and lens information for thecapture device 40 or other such devices (not shown), graphics and character generator information, aspect ratio to the switcher/production system, for example, as indicated by the legend inblock 66. - The foregoing describes a technique capturing and processing metadata during production of an audio visual program.
Claims (14)
1. A method for associating metadata with audio visual information, comprising the steps of:
transmitting first metadata to a field device prior to capture of audio visual information;
transmitting second metadata to a server destined to receive the audio visual information along with edited first metadata from the field device following capture of the audio visual information;
updating the second metadata received at the server in accordance with the first edited metadata.
2. The method according to claim 1 wherein the first and second metadata are the same.
3. The method according to claim 1 further comprises the step of operating the field device to edit the first metadata transmitted to the field device.
4. The method according to claim 1 wherein the first transmitting step comprises the step of communicating the first metadata to the field device over a wireless communications network.
5. The method according to claim 1 wherein the step of updating the second metadata received at the server further comprises the steps of:
assigning an identification to the first metadata transmitted to the field device;
assigning the same identification to the second metadata transmitted to the server; and
matching the edited first metadata from the field device with the second metadata transmitted to the server using the identification.
6. The method according to claim 1 wherein the first metadata includes information related to at least one of assignment data, location data, personnel data, story data, and individuals for interview data.
7. The method according to claim 3 wherein the step of operating the field device to edit the first metadata transmitted thereto comprises the step of modifying the first metadata to add information gathered by a news reporter at a story location.
8. The method according to claim 3 wherein the step of operating the field device to edit the first metadata comprises the step of modifying the first metadata to add information regarding capture of an image.
9. The method according to claim 1 further including the steps of:
incorporating edited first metadata from the field device into one of audio and/or video information from a capture device; and
transmitting the one of the audio and/or video information incorporating the first edited metadata to the server.
10. A method for associating metadata with audio visual information, comprising the steps of:
transmitting metadata to a field device prior to capture of associated with audio visual information; the metadata undergoing editing at the field device by an operator;
transmitting the metadata to a server destined to receive the captured audio visual information along with edited metadata from the field device;
updating the metadata received at the server in accordance with the edited metadata from the field device.
11. The method according to claim 10 wherein the first transmitting step comprises the step of communicating the metadata to the field device over a wireless communications network.
12. The method according to claim 10 wherein the step of updating the metadata received at the server further comprises the steps of:
assigning an identification to the metadata transmitted to the field device;
assigning the same identification to the metadata transmitted to the server; and
matching the edited first metadata from the field device with the second metadata transmitted to the server using the identification.
13. The method according to claim 10 wherein the metadata transmitted to the field device includes information related to at least one of assignment data, location data, personnel data, story data, and individuals for interview data.
14. The method according to claim 1 further including the steps of:
incorporating edited metadata from the field device into one of audio and/or video information from a capture device; and
transmitting the one of the audio and/or video information incorporating the first edited metadata to the server.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/990,804 US20090254562A1 (en) | 2005-09-02 | 2006-07-29 | Automatic Metadata Extraction and Metadata Controlled Production Process |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US71384805P | 2005-09-02 | 2005-09-02 | |
US11/990,804 US20090254562A1 (en) | 2005-09-02 | 2006-07-29 | Automatic Metadata Extraction and Metadata Controlled Production Process |
PCT/US2006/033528 WO2007027605A2 (en) | 2005-09-02 | 2006-08-29 | Automatic metadata extraction and metadata controlled production process |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2006/033528 A-371-Of-International WO2007027605A2 (en) | 2005-09-02 | 2006-08-29 | Automatic metadata extraction and metadata controlled production process |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/982,650 Continuation US9420231B2 (en) | 2005-09-02 | 2010-12-30 | Automatic metadata extraction and metadata controlled production process |
Publications (1)
Publication Number | Publication Date |
---|---|
US20090254562A1 true US20090254562A1 (en) | 2009-10-08 |
Family
ID=37698337
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/990,804 Abandoned US20090254562A1 (en) | 2005-09-02 | 2006-07-29 | Automatic Metadata Extraction and Metadata Controlled Production Process |
US12/982,650 Active US9420231B2 (en) | 2005-09-02 | 2010-12-30 | Automatic metadata extraction and metadata controlled production process |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/982,650 Active US9420231B2 (en) | 2005-09-02 | 2010-12-30 | Automatic metadata extraction and metadata controlled production process |
Country Status (6)
Country | Link |
---|---|
US (2) | US20090254562A1 (en) |
EP (2) | EP3133809A3 (en) |
JP (1) | JP5221351B2 (en) |
CN (1) | CN101253769A (en) |
CA (1) | CA2620154A1 (en) |
WO (1) | WO2007027605A2 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110173196A1 (en) * | 2005-09-02 | 2011-07-14 | Thomson Licensing Inc. | Automatic metadata extraction and metadata controlled production process |
WO2012135804A3 (en) * | 2011-04-01 | 2012-11-29 | Mixaroo, Inc. | System and method for real-time processing, storage, indexing, and delivery of segmented video |
US20170366828A1 (en) * | 2012-04-27 | 2017-12-21 | Comcast Cable Communications, Llc | Processing and delivery of segmented video |
US10372883B2 (en) | 2016-06-24 | 2019-08-06 | Scripps Networks Interactive, Inc. | Satellite and central asset registry systems and methods and rights management systems |
US10452714B2 (en) | 2016-06-24 | 2019-10-22 | Scripps Networks Interactive, Inc. | Central asset registry system and method |
US11159709B2 (en) * | 2016-12-27 | 2021-10-26 | Sony Corporation | Camera, camera processing method, server, server processing method, and information processing apparatus |
US11868445B2 (en) | 2016-06-24 | 2024-01-09 | Discovery Communications, Llc | Systems and methods for federated searches of assets in disparate dam repositories |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4650516B2 (en) * | 2008-04-09 | 2011-03-16 | ソニー株式会社 | Imaging data management method and imaging apparatus |
US9648230B2 (en) | 2014-05-27 | 2017-05-09 | Tribune Broadcasting Company, Llc | Use of wireless connection loss to facilitate identifying and recording video capture location |
US9667824B2 (en) | 2014-05-27 | 2017-05-30 | Tribune Broadcasting Company, Llc | Use of location lulls to facilitate identifying and recording video capture location |
US10805577B2 (en) * | 2016-10-25 | 2020-10-13 | Owl Cameras, Inc. | Video-based data collection, image capture and analysis configuration |
CN114630051B (en) * | 2022-03-28 | 2023-06-09 | 阿里巴巴(中国)有限公司 | Video processing method and system |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6018738A (en) * | 1998-01-22 | 2000-01-25 | Microsft Corporation | Methods and apparatus for matching entities and for predicting an attribute of an entity based on an attribute frequency value |
US20020088000A1 (en) * | 2001-01-03 | 2002-07-04 | Morris Robert Paul | Controlled access to image metadata |
US20020087601A1 (en) * | 2000-12-29 | 2002-07-04 | Anderson Eric C. | Client-server system for merging of metadata with images |
US20030154178A1 (en) * | 2002-02-08 | 2003-08-14 | Eastman Kodak Company | Method of automatically updating non-image data associated with a digital image |
US20030177130A1 (en) * | 2002-03-12 | 2003-09-18 | International Business Machines Corporation | Method, system, program, and data structures for maintaining metadata in a storage system |
US6629104B1 (en) * | 2000-11-22 | 2003-09-30 | Eastman Kodak Company | Method for adding personalized metadata to a collection of digital images |
US20030221164A1 (en) * | 2002-03-25 | 2003-11-27 | Williams Michael John | System |
US20040003415A1 (en) * | 2002-06-28 | 2004-01-01 | Koninklijke Philips Electronics N.V. | Removable memory information management |
US20040039755A1 (en) * | 2002-06-05 | 2004-02-26 | Matthew Kunze | Metadata relationships |
US20040126038A1 (en) * | 2002-12-31 | 2004-07-01 | France Telecom Research And Development Llc | Method and system for automated annotation and retrieval of remote digital content |
US20040128697A1 (en) * | 1999-11-15 | 2004-07-01 | Wood Lisa T. | Media acquisition, processing and distribution system for the internet |
US20040174434A1 (en) * | 2002-12-18 | 2004-09-09 | Walker Jay S. | Systems and methods for suggesting meta-information to a camera user |
US20040199387A1 (en) * | 2000-07-31 | 2004-10-07 | Wang Avery Li-Chun | Method and system for purchasing pre-recorded music |
US20040208061A1 (en) * | 2003-04-16 | 2004-10-21 | Kabushiki Kaisha Toshiba, Tokyo, Japan | Non-volatile semiconductor memory device and electric device with the same |
US20050018057A1 (en) * | 2003-07-25 | 2005-01-27 | Bronstein Kenneth H. | Image capture device loaded with image metadata |
US20050044112A1 (en) * | 2003-08-19 | 2005-02-24 | Canon Kabushiki Kaisha | Metadata processing method, metadata storing method, metadata adding apparatus, control program and recording medium, and contents displaying apparatus and contents imaging apparatus |
US6877134B1 (en) * | 1997-08-14 | 2005-04-05 | Virage, Inc. | Integrated data and real-time metadata capture system and method |
US20050134707A1 (en) * | 2003-12-18 | 2005-06-23 | Eastman Kodak Company | Image metadata attachment |
US20050203927A1 (en) * | 2000-07-24 | 2005-09-15 | Vivcom, Inc. | Fast metadata generation and delivery |
US20050210501A1 (en) * | 2004-03-19 | 2005-09-22 | Microsoft Corporation | Method and apparatus for handling metadata |
US20070067447A1 (en) * | 2001-06-28 | 2007-03-22 | Microsoft Corporation | Transportable identifier and system and method to facilitate access to broadcast data |
Family Cites Families (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6833865B1 (en) * | 1998-09-01 | 2004-12-21 | Virage, Inc. | Embedded metadata engines in digital capture devices |
US20030001880A1 (en) * | 2001-04-18 | 2003-01-02 | Parkervision, Inc. | Method, system, and computer program product for producing and distributing enhanced media |
US6608930B1 (en) * | 1999-08-09 | 2003-08-19 | Koninklijke Philips Electronics N.V. | Method and system for analyzing video content using detected text in video frames |
US6707950B1 (en) | 1999-06-22 | 2004-03-16 | Eastman Kodak Company | Method for modification of non-image data in an image processing chain |
GB2359918A (en) | 2000-03-01 | 2001-09-05 | Sony Uk Ltd | Audio and/or video generation apparatus having a metadata generator |
JP4792686B2 (en) * | 2000-02-07 | 2011-10-12 | ソニー株式会社 | Image processing apparatus, image processing method, and recording medium |
JP4438177B2 (en) * | 2000-04-10 | 2010-03-24 | ソニー株式会社 | Data transmission / reception system |
WO2001078385A1 (en) * | 2000-04-10 | 2001-10-18 | Sony Corporation | Asset management system and asset management method |
US6760042B2 (en) * | 2000-09-15 | 2004-07-06 | International Business Machines Corporation | System and method of processing MPEG streams for storyboard and rights metadata insertion |
US20020083124A1 (en) | 2000-10-04 | 2002-06-27 | Knox Christopher R. | Systems and methods for supporting the delivery of streamed content |
US6873344B2 (en) * | 2001-02-22 | 2005-03-29 | Sony Corporation | Media production system using flowgraph representation of operations |
US20020170068A1 (en) * | 2001-03-19 | 2002-11-14 | Rafey Richter A. | Virtual and condensed television programs |
JP2002318807A (en) * | 2001-04-19 | 2002-10-31 | Matsushita Electric Ind Co Ltd | Meta-data generation device and meta-data generation method |
US6868383B1 (en) * | 2001-07-12 | 2005-03-15 | At&T Corp. | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US6947947B2 (en) * | 2001-08-17 | 2005-09-20 | Universal Business Matrix Llc | Method for adding metadata to data |
US7950033B2 (en) * | 2001-10-10 | 2011-05-24 | Opentv, Inc. | Utilization of relational metadata in a television system |
GB2387086A (en) * | 2002-03-25 | 2003-10-01 | Sony Uk Ltd | System |
GB2387087A (en) | 2002-03-25 | 2003-10-01 | Sony Uk Ltd | System |
US20030187820A1 (en) * | 2002-03-29 | 2003-10-02 | Michael Kohut | Media management system and process |
US7073193B2 (en) * | 2002-04-16 | 2006-07-04 | Microsoft Corporation | Media content descriptions |
US7680393B2 (en) * | 2002-11-13 | 2010-03-16 | Sony Corporation | Content editing assistance system, video processing apparatus, playback apparatus, editing apparatus, computer program, and content processing method |
JP4404247B2 (en) * | 2002-11-22 | 2010-01-27 | ソニー株式会社 | Video program production support system, processing apparatus, video program processing method and program |
JP3989365B2 (en) | 2002-12-02 | 2007-10-10 | トチセン化成工業株式会社 | Inner bag for containers. |
KR100511785B1 (en) * | 2002-12-20 | 2005-08-31 | 한국전자통신연구원 | A System and A Method for Authoring Multimedia Content Description Metadata |
JP2004295568A (en) * | 2003-03-27 | 2004-10-21 | Sony Corp | Information processor, information processing method, and computer program |
JP4332364B2 (en) * | 2003-04-04 | 2009-09-16 | ソニー株式会社 | Video recording system and video recording method |
KR20050006565A (en) | 2003-07-09 | 2005-01-17 | 주식회사 픽스트리 | System And Method For Managing And Editing Multimedia Data |
JP2005039354A (en) | 2003-07-16 | 2005-02-10 | Matsushita Electric Ind Co Ltd | Metadata input method and editing system |
US7885391B2 (en) * | 2003-10-30 | 2011-02-08 | Hewlett-Packard Development Company, L.P. | System and method for call center dialog management |
US7685134B2 (en) | 2003-12-31 | 2010-03-23 | Nokia Corporation | Media file sharing, correlation of metadata related to shared media files and assembling shared media file collections |
WO2005069172A1 (en) * | 2004-01-14 | 2005-07-28 | Mitsubishi Denki Kabushiki Kaisha | Summarizing reproduction device and summarizing reproduction method |
JP4304108B2 (en) * | 2004-03-31 | 2009-07-29 | 株式会社東芝 | METADATA DISTRIBUTION DEVICE, VIDEO REPRODUCTION DEVICE, AND VIDEO REPRODUCTION SYSTEM |
JP4491620B2 (en) | 2004-04-22 | 2010-06-30 | 学校法人東京電機大学 | Wave extinguishing method, water quality purification method, and shallow water area environment creation support method using the same |
US8230467B2 (en) * | 2004-04-29 | 2012-07-24 | Harris Corporation | Media asset management system for managing video segments from an aerial sensor platform and associated method |
US7617109B2 (en) * | 2004-07-01 | 2009-11-10 | Dolby Laboratories Licensing Corporation | Method for correcting metadata affecting the playback loudness and dynamic range of audio information |
JP4852967B2 (en) * | 2005-06-03 | 2012-01-11 | ソニー株式会社 | Content management system, management server, management information processing device, and computer program |
EP3133809A3 (en) * | 2005-09-02 | 2017-03-01 | GVBB Holdings S.A.R.L | Automatic metadata extraction and metadata controlled production process |
US7893999B2 (en) * | 2006-05-22 | 2011-02-22 | Broadcom Corporation | Simultaneous video and sub-frame metadata capture system |
-
2006
- 2006-07-29 EP EP16187907.7A patent/EP3133809A3/en not_active Withdrawn
- 2006-07-29 US US11/990,804 patent/US20090254562A1/en not_active Abandoned
- 2006-08-29 CA CA002620154A patent/CA2620154A1/en not_active Abandoned
- 2006-08-29 CN CNA2006800321612A patent/CN101253769A/en active Pending
- 2006-08-29 JP JP2008529165A patent/JP5221351B2/en not_active Expired - Fee Related
- 2006-08-29 WO PCT/US2006/033528 patent/WO2007027605A2/en active Application Filing
- 2006-08-29 EP EP06802472A patent/EP1932344A2/en not_active Ceased
-
2010
- 2010-12-30 US US12/982,650 patent/US9420231B2/en active Active
Patent Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6877134B1 (en) * | 1997-08-14 | 2005-04-05 | Virage, Inc. | Integrated data and real-time metadata capture system and method |
US6018738A (en) * | 1998-01-22 | 2000-01-25 | Microsft Corporation | Methods and apparatus for matching entities and for predicting an attribute of an entity based on an attribute frequency value |
US20040128697A1 (en) * | 1999-11-15 | 2004-07-01 | Wood Lisa T. | Media acquisition, processing and distribution system for the internet |
US20050203927A1 (en) * | 2000-07-24 | 2005-09-15 | Vivcom, Inc. | Fast metadata generation and delivery |
US20040199387A1 (en) * | 2000-07-31 | 2004-10-07 | Wang Avery Li-Chun | Method and system for purchasing pre-recorded music |
US6629104B1 (en) * | 2000-11-22 | 2003-09-30 | Eastman Kodak Company | Method for adding personalized metadata to a collection of digital images |
US20020087601A1 (en) * | 2000-12-29 | 2002-07-04 | Anderson Eric C. | Client-server system for merging of metadata with images |
US20020088000A1 (en) * | 2001-01-03 | 2002-07-04 | Morris Robert Paul | Controlled access to image metadata |
US20070067447A1 (en) * | 2001-06-28 | 2007-03-22 | Microsoft Corporation | Transportable identifier and system and method to facilitate access to broadcast data |
US20030154178A1 (en) * | 2002-02-08 | 2003-08-14 | Eastman Kodak Company | Method of automatically updating non-image data associated with a digital image |
US20030177130A1 (en) * | 2002-03-12 | 2003-09-18 | International Business Machines Corporation | Method, system, program, and data structures for maintaining metadata in a storage system |
US20030221164A1 (en) * | 2002-03-25 | 2003-11-27 | Williams Michael John | System |
US20040039755A1 (en) * | 2002-06-05 | 2004-02-26 | Matthew Kunze | Metadata relationships |
US20040003415A1 (en) * | 2002-06-28 | 2004-01-01 | Koninklijke Philips Electronics N.V. | Removable memory information management |
US20040174434A1 (en) * | 2002-12-18 | 2004-09-09 | Walker Jay S. | Systems and methods for suggesting meta-information to a camera user |
US20040126038A1 (en) * | 2002-12-31 | 2004-07-01 | France Telecom Research And Development Llc | Method and system for automated annotation and retrieval of remote digital content |
US6847555B2 (en) * | 2003-04-16 | 2005-01-25 | Kabushiki Kaisha Toshiba | Non-volatile semiconductor memory device reading and writing multi-value data from and into pair-cells |
US20040208061A1 (en) * | 2003-04-16 | 2004-10-21 | Kabushiki Kaisha Toshiba, Tokyo, Japan | Non-volatile semiconductor memory device and electric device with the same |
US20050018057A1 (en) * | 2003-07-25 | 2005-01-27 | Bronstein Kenneth H. | Image capture device loaded with image metadata |
US20050044112A1 (en) * | 2003-08-19 | 2005-02-24 | Canon Kabushiki Kaisha | Metadata processing method, metadata storing method, metadata adding apparatus, control program and recording medium, and contents displaying apparatus and contents imaging apparatus |
US20050134707A1 (en) * | 2003-12-18 | 2005-06-23 | Eastman Kodak Company | Image metadata attachment |
US20050210501A1 (en) * | 2004-03-19 | 2005-09-22 | Microsoft Corporation | Method and apparatus for handling metadata |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110173196A1 (en) * | 2005-09-02 | 2011-07-14 | Thomson Licensing Inc. | Automatic metadata extraction and metadata controlled production process |
US9420231B2 (en) * | 2005-09-02 | 2016-08-16 | Gvbb Holdings S.A.R.L. | Automatic metadata extraction and metadata controlled production process |
WO2012135804A3 (en) * | 2011-04-01 | 2012-11-29 | Mixaroo, Inc. | System and method for real-time processing, storage, indexing, and delivery of segmented video |
US8769576B2 (en) | 2011-04-01 | 2014-07-01 | Mixaroo, Inc. | System and method for real-time processing, storage, indexing, and delivery of segmented video |
US9100679B2 (en) | 2011-04-01 | 2015-08-04 | Mixaroo, Inc. | System and method for real-time processing, storage, indexing, and delivery of segmented video |
US20170366828A1 (en) * | 2012-04-27 | 2017-12-21 | Comcast Cable Communications, Llc | Processing and delivery of segmented video |
US10372883B2 (en) | 2016-06-24 | 2019-08-06 | Scripps Networks Interactive, Inc. | Satellite and central asset registry systems and methods and rights management systems |
US10452714B2 (en) | 2016-06-24 | 2019-10-22 | Scripps Networks Interactive, Inc. | Central asset registry system and method |
US10769248B2 (en) | 2016-06-24 | 2020-09-08 | Discovery, Inc. | Satellite and central asset registry systems and methods and rights management systems |
US11868445B2 (en) | 2016-06-24 | 2024-01-09 | Discovery Communications, Llc | Systems and methods for federated searches of assets in disparate dam repositories |
US11159709B2 (en) * | 2016-12-27 | 2021-10-26 | Sony Corporation | Camera, camera processing method, server, server processing method, and information processing apparatus |
Also Published As
Publication number | Publication date |
---|---|
EP1932344A2 (en) | 2008-06-18 |
WO2007027605A2 (en) | 2007-03-08 |
CA2620154A1 (en) | 2007-03-08 |
JP5221351B2 (en) | 2013-06-26 |
WO2007027605A3 (en) | 2007-05-03 |
EP3133809A3 (en) | 2017-03-01 |
US20110173196A1 (en) | 2011-07-14 |
JP2009507421A (en) | 2009-02-19 |
WO2007027605A8 (en) | 2008-03-27 |
CN101253769A (en) | 2008-08-27 |
US9420231B2 (en) | 2016-08-16 |
EP3133809A2 (en) | 2017-02-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9420231B2 (en) | Automatic metadata extraction and metadata controlled production process | |
US10592075B1 (en) | System and method for media content collaboration throughout a media production process | |
US7835920B2 (en) | Director interface for production automation control | |
US7895248B2 (en) | Information processing apparatus, information processing method, recording medium and program | |
US20020054244A1 (en) | Method, system and computer program product for full news integration and automation in a real time video production environment | |
JP4580925B2 (en) | Imaging apparatus and imaging system | |
US7836127B2 (en) | Dynamically triggering notifications to human participants in an integrated content production process | |
AU2009236622B2 (en) | Method and apparatus for associating metadata with content for live production | |
JP2003529990A (en) | Method of identifying and processing audio and / or video material | |
US8156532B2 (en) | Video program creation system, table providing device, terminal device, terminal processing method, program, recording medium | |
US20140074938A1 (en) | Providing excess resources as a service | |
WO2002010963A2 (en) | Ethod for networking data and content management | |
JP2007158874A (en) | Recording apparatus and recording method of broadcasting material | |
JP2007194696A (en) | Content management system | |
US20070005386A1 (en) | Content production maintenance tool for human and non-human activity tracking | |
Trepess et al. | User-Centric Development of New Metadata Tools to Enhance File-Based Newsgathering Workflows | |
Spasić et al. | Framework for Software-Intensive Ingest System: One Behavioural Description | |
JP2001298682A (en) | Production system and its method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GVBB HOLDINGS S.A.R.L., LUXEMBOURG Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:THOMSON LICENSING;REEL/FRAME:026028/0071 Effective date: 20101231 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |