US20090307227A1 - Methods and apparatuses for combining and distributing user enhanced video/audio content - Google Patents

Methods and apparatuses for combining and distributing user enhanced video/audio content Download PDF

Info

Publication number
US20090307227A1
US20090307227A1 US12/239,596 US23959608A US2009307227A1 US 20090307227 A1 US20090307227 A1 US 20090307227A1 US 23959608 A US23959608 A US 23959608A US 2009307227 A1 US2009307227 A1 US 2009307227A1
Authority
US
United States
Prior art keywords
content
user
recited
input information
tagged
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/239,596
Inventor
Kyle Prestenback
Evan Tahler
Brian Kwan
David Jessen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Disney Enterprises Inc
Original Assignee
Disney Enterprises Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Disney Enterprises Inc filed Critical Disney Enterprises Inc
Priority to US12/239,596 priority Critical patent/US20090307227A1/en
Assigned to DISNEY ENTERPRISES, INC. reassignment DISNEY ENTERPRISES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JESSEN, DAVID, KWAN, BRIAN, PRESTENBACK, KYLE, TAHLER, EVAN
Priority to PCT/US2009/003129 priority patent/WO2009148513A1/en
Publication of US20090307227A1 publication Critical patent/US20090307227A1/en
Priority to US16/450,032 priority patent/US11582507B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/27Server based end-user applications
    • H04N21/274Storing end-user multimedia data in response to end-user request, e.g. network recorder
    • H04N21/2743Video hosting of uploaded data from client
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • H04N21/43072Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/458Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47202End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/173Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
    • H04N7/17309Transmission or handling of upstream communications
    • H04N7/17318Direct or substantially direct transmission and handling of requests

Definitions

  • the subject matter disclosed herein relates to video and/or audio data content processing, and more particularly to methods and apparatuses for use in combining and distributing user enhanced video/audio content.
  • Standard definition televisions are being replaced with higher definition video monitor devices and high fidelity audio monitor devices in an attempt to provide an improved entertainment experience.
  • standard definition video tape players and/or DVD player devices are being replaced with higher definition optical disc player devices, such as, for example, Blu-Ray disc player devices.
  • content establishing devices such as web cameras, video cameras, still image cameras, sound recording devices, etc., are also being replaced or upgraded to provide higher definition content.
  • Computing and communication systems and networks also continue to advance in sophistication and popularity.
  • the Internet and related computing and communication infrastructure continues to improve and expand in content, access, speed, variety, etc.
  • Electronic entertainment systems may be further enhanced by interfacing with such networked resources.
  • certain set-top boxes may be adapted to access networked resources and provide information and/or content received there from for use in an electronic entertainment system.
  • Some set-top boxes may also be adapted to allow for information and/or content to be provided from one or more devices in an electronic entertainment system to one or more networked resources.
  • the sharing and distribution of content, and in particular video/audio content continues to increase as more and more users upload and download such content to and from various network resource devices.
  • FIG. 1 is a block diagram illustrating an exemplary implementation of a computing environment that includes various networked computing devices and an electronic entertainment system adapted for use in combining and distributing user enhanced video/audio content.
  • FIG. 2 is a block diagram illustrating certain features of an exemplary system that may be implemented to combine and distribute user enhanced video/audio content.
  • FIG. 3 is a block diagram illustrating certain features of an implementation of an optical disc and an optical disc player device as may be implemented for use in combining and distributing user enhanced video/audio content.
  • FIG. 4 is a flow diagram illustrating an exemplary implementation of a method that may be adapted for use in combining and distributing user enhanced video/audio content.
  • FIG. 5 is a time line diagram illustrating some exemplary representative combined user enhanced video/audio content.
  • FIG. 6 is a flow diagram illustrating an exemplary implementation of a method that may be adapted for use in combining and distributing user enhanced video/audio content.
  • Such exemplary methods and apparatuses may, for example, be adapted to provide or otherwise support a community of users having the ability to insert video messages and overlays to enhance or otherwise augment the electronic entertainment experience. Such ability may be used in both an interpersonal messaging sense, and a larger group publishing sense.
  • the methods and apparatuses provided herein may be adapted to provide users a user-friendly space in which they may create their own video tracks (user content) to augment the presentation of a packaged movie, or other like reference content. These video tracks can be used to create combined content for both friends (selected users) and/or the community at large. Such combined content may, in certain implementations, take the form of “movie messages” that may be used for user-made commentary, in-movie messaging, relational notes, fan art, fan videos, fan-enhancement, and other like uses.
  • the methods and apparatuses herein may be adapted to provide or otherwise support the creation of videos that may be displayed as a composite on top of an existing packaged movie (e.g., Picture in Picture, etc.), as a simultaneous video feed (e.g., presented on the same screen as the movie or a separate screen/device), and/or as video that will be inserted into the movie's playback (e.g. an interrupt, interlude, etc.).
  • an existing packaged movie e.g., Picture in Picture, etc.
  • a simultaneous video feed e.g., presented on the same screen as the movie or a separate screen/device
  • video that will be inserted into the movie's playback e.g. an interrupt, interlude, etc.
  • the methods and apparatuses herein may be adapted to provide or otherwise support an integrated system that includes video creation, video storage, video transmission, and/or video playback processes.
  • the methods and apparatuses herein may be adapted to interface with on an existing fundamental architecture for a social networking web service that has the ability to be connected to a website, set-top media players (e.g., Blu-ray player devices, etc.), and/or the like.
  • Such architecture may include, for example, certain features such as inter-network messaging, buddy lists, etc.
  • An exemplary video for use in such an architecture may include a number of attributes, such as, tags or a collection of access-tags.
  • tags may, for example, be managed by the social networking service provider, and include a way to note the video's owner and a list of those intended to be allowed to view the video. This may be useful when considering both privacy concerns and decency concerns.
  • a video's creator may be noted, so that that user may edit and remove the video, including the ability to manage the rest of the video's access-tags.
  • the rest of these tags may be associated with a collection of users who may view the video.
  • a user may select an individual user or groups of users, or select all of their buddies. In the case of ‘all buddies’, the social networking service may be adapted to actively update such a list to be current with the user's buddy list.
  • a user may also opt to submit a video for public consumption.
  • a user When a user uploads a video, they may be given the option to send a message to one or more other users to inform them of this video's existence. In some implementations, other users may browse videos to which they may have access.
  • a user-created video may be access-tagged and uploaded to an existing social-networking web-service.
  • Such user content may, for example, be captured live (e.g., via a webcam), and/or uploaded from a previously captured and/or edited source by the user.
  • a user may have the ability provide user input (sender input) to select certain video presentation options. For example, a user may choose a location on the screen for the tagged user content as a pop-up, weather or not the tagged user content interrupts the movie (reference content), or appear along side of, or on top of it, etc.
  • a user may be given the option to ‘frame’ their video using the images.
  • a networked laptop may be used in conjunction with a connected Blu-ray player.
  • a user may synchronize both devices, and record himself making a commentary track while a movie plays.
  • a website may be provided or created that gives the user's computer control of the Blu-ray player's playback commands, and will note the player's timecode, and automatically link created video to the appropriate spot in the movie.
  • a user may seek to upload a pre-existing video while still connected to a Blu-ray player.
  • the user may scrub through the movie on their player, and select the video's start point.
  • the user may then see a live preview of their video (e.g., streaming from the social network service) on top of their player's reference content.
  • a use may be presented with a computer but no local Blu-ray player. This may, for example, become common in a Kiosk type of display.
  • a user may be presented with a webcam and the ability to record commentary on the fly.
  • the system may be adapted to provide the user with pre-selected popular scenes or other like portions of reference content to choose from, and over/along which to record their user content.
  • a reference content owner/manager may choose which movies to provide.
  • a user may be allowed by a reference content owner/manager to access movies per some account registration agreement or the like.
  • such a system may provide users with sample clips or portions from the movie that will playback as they record their message, and/or still images or frames (e.g., thumbnail images, etc.).
  • a reference content owner/manager and/or web service may include a varied sample of a movie to provide context for features such as this one.
  • a user may preview of user content as overlay or the like (e.g., perhaps presented in a lower resolution format) as generated by the web service and sent to the computer.
  • a preview of the combined content may be presented in a version as a flash application or the like.
  • a web application or the like may be adapted to manage the storage of copies of uploaded tagged user content.
  • the videos that users create may have to be encoded into a few different formats.
  • a version for website playback may, for example, be created using flash, etc.
  • Versions for download to Blu-ray players may, for example, be created using an MPEG-II and/or other like standard, both for download and progressive download for streaming.
  • a receiver user may initiate playback of combined content.
  • a user may choose to view the video within the context of the entire movie, or to view only the portion or portions of the movie that are relevant to the new video created.
  • FIG. 1 illustrates an exemplary implementation of a computing environment 100 that may include various networked computing devices and an electronic entertainment system, and which may be adapted for use in combining and distributing user enhanced video/audio content.
  • environment 100 may include an optical disc 102 adapted for use in an electronic entertainment system 104 having an optical disc player device 106 coupled to a video monitor device 108 and audio monitor device 110 .
  • optical disc 102 may include content information 112 and instructional information 114 .
  • Electronic entertainment system 104 may, for example, be operatively coupled to communicate with at least one computing device, such as, network resource device 120 through a network 116 .
  • At least a portion of instructional information 114 may be implemented to adapt optical disc player device 106 to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein.
  • optical disc 102 and/or other user content may be accessed by a sender user computing device 122 .
  • Computing device 122 may, for example, include any computer or other like appliance that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein.
  • computing device 122 may be adapted to access user content and provide tagged user content to network resource device 120 .
  • computing device 122 may include a personal computer, desk top computer, a lap top computer, a camera, a personal digital assistant, a smart phone, a set top box, etc.
  • computing device 122 may be connected to electronic entertainment system 104 , and/or network 116 .
  • System 100 may also include one or more receiver user computing devices 124 .
  • Computing device 124 may, for example, include any computer or other like appliance that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein.
  • computing device 124 may access combined content from network resource device 120 through network 116 .
  • computing device 122 may include a personal computer, desk top computer, a lap top computer, a camera, a personal digital assistant, a smart phone, an electronic entertainment system, a set top box, etc.
  • Network resource device 120 may include one or more computing devices that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein.
  • network resource device 120 may receive tagged user content from one or more other computing devices or the like and provide combined content that may be established based, at least in part, on the tagged user content.
  • the combined content may, for example, be provided (e.g., downloaded, streamed, etc.) to one or more computing devices or the like through network 116 .
  • FIG. 2 is a block diagram illustrating certain features of an exemplary system 200 that may be implemented to combine and distribute user enhanced video/audio content, for example, as in FIG. 1 .
  • System 200 may include a content establishing device 202 .
  • Content establishing device 202 may be adapted to establish user content 204 .
  • User content 204 may, for example, include one or more video and/or audio tracks.
  • Content establishing device 202 may include one or more devices for capturing, recording, and/or otherwise generating video and/or audio tracks.
  • content establishing device 202 may include a camera, a microphone, a computing device, and/or the like.
  • the resulting user content 204 may, for example, be stored on a computer readable medium.
  • User content 204 may be accessed by a user content tagging device 206 .
  • User content tagging device 206 may, for example, include one or more computing devices, an optical disc reading device, and/or the like.
  • User content tagging device 206 may also be adapted to access at least a portion of reference content 112 .
  • reference content 112 may be accessed from an optical disc 102 , memory, and/or other like computer readable media.
  • Reference content 112 may include, for example, data associated with at least portion of one or more video and/or audio tracks.
  • User content tagging device 206 may be adapted to establish tagged user content 208 and/or sender input information 210 .
  • user content tagging device 206 may be adapted to establish tagged user content 208 based, at least in part, on at least a portion of user content 204 and/or at least a portion of reference content 112 and/or other information associated with reference content 112 .
  • User content tagging device 206 may be adapted to establish sender input information 210 based, at least in part, on user input.
  • Tagged user content 208 and/or sender input information 210 may be provided to a content combining device 212 .
  • Content combining device 212 may, for example, include one or more computing devices and/or the like.
  • content combining device 212 may include network resource device 120 .
  • Content combining device 212 may be adapted to access at least a portion of reference content 112 (e.g., another copy thereof).
  • Content combining device 212 may be adapted to establish at least one version of combined content 214 (e.g., user enhanced) based, at least in part, on tagged user content 208 and at least a portion of reference content 112 .
  • content combining device 212 may be adapted to establish such combined content 214 based, at least in part, on tagged user content 208 , at least a portion of reference content 112 , and at least a portion of sender input information 210 .
  • Content combining device 212 may be adapted to provide a combined content availability notice 216 to one or more computing devices.
  • content combining device 212 may be adapted to provide such an availability notice 216 to one or more computing devices based, at least in part, on sender input information 210 and/or tagged user content 208 .
  • user content tagging device 206 may be adapted to provide such an availability notice 216 to one or more computing devices, e.g., as illustrated in FIG. 2 by the dashed line.
  • such an availability notice 216 may be provided to a combined content accessing device 218 , which may include one or more computing devices and/or the like.
  • combined content accessing device 218 may include receiver user computing device 124 .
  • Combined content accessing device 218 may be adapted to access combined content 214 and/or otherwise provide such to a combined content player device 220 . While illustrated in the example as being separate devices, in certain implementations, combined content accessing device 218 and combined content player device 220 may be implemented using one or more devices.
  • Combined content player device 220 may, for example include one or more computing devices, entertainment systems, smart phones, MP3 Players, etc., that may be adapted to replay or otherwise reproduce at least a portion of at least one of an audio and/or video track of combined content 214 .
  • user content tagging device 206 may be associated with a first user (e.g., a sender user) and combined content accessing device may be associated with a second user (e.g., a receiver user), and content combining device 212 may be associated with a service provider, web entity, and/or the like.
  • system 200 and the associated parties e.g., sender user, receiver user(s), and service provider
  • the users may open accounts and specify other users as “buddies” with whom, through the service provider may share combined content 214 .
  • FIG. 3 is a block diagram illustrating certain example features that may be implemented in an exemplary optical disc 102 and/or an exemplary optical disc player device 106 , and/or implemented to adapt in some manner a function of a computing device such as, e.g., computing devices 122 , 124 , and/or network resource device 120 .
  • a computing device such as, e.g., computing devices 122 , 124 , and/or network resource device 120 .
  • Optical disc player device 106 may, for example, include at least one processing unit 302 , some form of a memory 304 , at least one user interface 306 , an optical disc reading mechanism 330 , and a network interface 326 .
  • at least one bus 350 or the like may be adapted to operatively couple two or more of processing unit 302 , memory 304 , user interface 306 , optical disc reading mechanism 330 , and/or network interface 326 .
  • optical disc 102 may be operatively coupled to (and accessed through) optical disc reading mechanism 330 .
  • Optical disc 102 may have stored thereon content information 112 , including video data 332 and audio data 334 .
  • Optical disc 102 may, for example, have stored thereon instructional information 114 including computer implementable instructions 348 .
  • Memory 304 may, for example, include primary memory 304 A.
  • memory 304 may also include secondary memory 304 B (e.g., a hard disk drive, solid state drive, or other like non-volatile storage devices and/or media).
  • secondary memory 304 B e.g., a hard disk drive, solid state drive, or other like non-volatile storage devices and/or media.
  • primary memory 304 A may include tagged user content 208 and sender input 210 which may be at least a partially associated with user input that may be received through user interface 306 .
  • user interface 306 may include and/or otherwise be operatively coupled to a user input device 308 (e.g., a keypad, a touch screen, a joystick, a mouse, remote control, a camera, a microphone, etc.).
  • user interface 306 may include a wireless communication interface adapted to receive wireless signals (not shown) from a remotely controlled user input device 308 .
  • wireless signals may, for example, include infrared, radio frequency, or other like electromagnetic signals.
  • optical disc reading mechanism 330 may, for example, include a receptacle adapted for receiving optical disc 102 , an optical disc rotation mechanism adapted for rotating optical disc 102 , a optical reading mechanism (e.g., a laser read head, etc.) adapted for reading information stored on optical disc 102 in the form of data, and supporting control circuitry adapted for controlling the operation of optical disc reading mechanism 330 and providing the information read from optical disc 102 to at least bus 350 .
  • optical disc reading mechanism 330 may, for example, include a receptacle adapted for receiving optical disc 102 , an optical disc rotation mechanism adapted for rotating optical disc 102 , a optical reading mechanism (e.g., a laser read head, etc.) adapted for reading information stored on optical disc 102 in the form of data, and supporting control circuitry adapted for controlling the operation of optical disc reading mechanism 330 and providing the information read from optical disc 102 to at least bus 350 .
  • optical disc reading mechanism 330 may be adapted to read one or more of a compact disc (CD), a digital versatile disc (DVD), a Blu-Ray disc, and/or the like. In certain exemplary implementations, optical disc reading mechanism 330 may be adapted to hold and/or otherwise manipulate multiple optical discs. In certain exemplary implementations, optical disc 102 may be adapted as a read only disc, a read/writable disc, a read/rewritable disc, etc. Thus, in certain implementations, optical disc reading mechanism 330 may be adapted to only read from optical disc 102 , while in other implementations optical disc reading mechanism 330 may be adapted to read and write to optical disc 102 . Such mechanisms and techniques are well known.
  • Processing unit 302 may, for example, be implemented using hardware, firmware, software, and/or any combination thereof. Processing unit 302 may, for example, be implemented using digital and/or analog circuitry. Processing unit 302 may, for example, be implemented using a central processing unit, a microprocessor, a microcontroller, an application specific integrated circuit (ASIC), a graphical processing unit (GPU), and/or other like circuits.
  • ASIC application specific integrated circuit
  • GPU graphical processing unit
  • Processing unit 302 may include any form of circuitry that may be adapted based, at least in part, on at least a portion of instructional information 114 . Processing unit 302 may, for example, be adapted to perform and/or to otherwise support at least a portion one or more of exemplary methods 400 ( FIG. 4 ) and/or 600 ( FIG. 6 ).
  • Network interface 326 may, for example, include any circuitry adapted to provide for communication with at least network 116 .
  • Network interface 326 may, for example, be adapted to provide such communication over one or more wired and/or wireless communication links.
  • Network 116 may, for example, include one or more communication resources adapted to provide communication between network interface 326 and at least one computing device, such as, e.g., network resource device 120 .
  • network 116 may include a wired and/or wireless local area network (LAN and/or WLAN), a wired and/or wireless telephone network, the Internet, and/or the like.
  • network interface 226 may include a wired and/or wireless modem, an Ethernet interface, a telephone interface, an antenna, a transmitter, a receiver, a transceiver, and/or the like.
  • FIG. 4 is a flow diagram illustrating an exemplary method 400 that may, for example, be adapted for use in combining and distributing user enhanced video/audio content.
  • Method 400 may be implemented to provide for the services presented in the various examples herein.
  • a web client in a computing device or the like associated with a sender user may be tied to an account associated with a service provider or the like.
  • a player may be loaded with reference content.
  • a Blu-ray player may be activated and synchronized with the web client.
  • timecode or other like temporal indicator may be noted and user content recorded or otherwise generated.
  • a Blu-ray player it may be determined if a Blu-ray player is available. If there is no Blu-ray player, then at block 412 the user may select reference content. At block 414 , the user may select from presented content portions, such as, e.g., clips, frames, thumbnails, etc. If at block 410 it is determined that there is a Blu-ray player, then at block 416 a player may be loaded with reference content (e.g., an optical disc). At block 418 , at least one content timecode or the like may be selected by user input.
  • reference content e.g., an optical disc
  • a web client may not be tied to a user account.
  • a new user account may be created.
  • reference content may be selected.
  • the user may select from presented content portions, such as, e.g., clips, frames, thumbnails, etc.
  • the user may provide input to customize a presentation following, either block 408 , 414 , 418 , or 426 .
  • user content may be tagged and provided along with sender input.
  • combined content may be generated.
  • one or more receiver users may be notified that the combined content is available.
  • the combined content may be accessed by one or more receiver users, e.g., through download and/or streaming.
  • at block 440 at least a portion of the accessed combined content may be played or otherwise displayed.
  • FIG. 5 is a time line diagram illustrating some exemplary representative combined user enhanced video/audio content.
  • user content 204 is illustrated as a track, which may be tagged and/or otherwise adapted in some manner to produce track 502 .
  • Track 503 may have duration of d B , for example.
  • Reference content 112 is illustrated as a track that starts at t A0 and ends at t A2 .
  • a user selected timecode or the like is illustrated by the dashed line at t A1 .
  • An exemplary combined content 212 A is illustrated as a track that starts at t A0 and ends at t A2 , but which also includes user content as combined in some manner at t A1 and ending at t A1 +d B .
  • the combined content is the same length as the reference content.
  • the combined content may have a shorter or longer length then the reference content.
  • combined content 212 B is illustrated as a track that starts at t A1 and ends at about t A1 +d B , making it about the same length as the tagged user content, and/or user content.
  • combined content 212 C is illustrated as a track that starts at t D0 and ends at t D1 , making it the longer than the tagged user content, and/or user content, but shorter than the reference content.
  • FIG. 6 is a flow diagram illustrating an exemplary implementation of a method 600 that may be adapted for use in combining and distributing user enhanced video/audio content in one or more networked devices.
  • user content may be accessed.
  • reference content may be accessed.
  • at least a portion of the user content may be associated with at least a portion of the reference content to produce tagged user content.
  • combined content may be established based, at least in part, on at least a portion of the reference content and the tagged user content.
  • at least a portion of the combined content may be accessed.
  • at least a portion of the combined content may be played.

Abstract

Methods and apparatuses are provided, which may be implemented to combine and distribute user enhanced video and/or audio content.

Description

    RELATED APPLICATIONS
  • This patent application claims benefit of and priority to U.S. Provisional Patent Application 61/059,510, filed Jun. 6, 2008, and titled “Methods And Apparatuses For Combining And Distributing User Enhanced Video/Audio Content”, and which is incorporated in its entirety by reference herein.
  • BACKGROUND
  • 1. Field
  • The subject matter disclosed herein relates to video and/or audio data content processing, and more particularly to methods and apparatuses for use in combining and distributing user enhanced video/audio content.
  • 2. Information
  • Electronic entertainment systems continue to advance in sophistication and popularity. Standard definition televisions are being replaced with higher definition video monitor devices and high fidelity audio monitor devices in an attempt to provide an improved entertainment experience. Likewise, standard definition video tape players and/or DVD player devices are being replaced with higher definition optical disc player devices, such as, for example, Blu-Ray disc player devices. Furthermore, content establishing devices, such as web cameras, video cameras, still image cameras, sound recording devices, etc., are also being replaced or upgraded to provide higher definition content.
  • Computing and communication systems and networks also continue to advance in sophistication and popularity. The Internet and related computing and communication infrastructure, for example, continues to improve and expand in content, access, speed, variety, etc. Electronic entertainment systems may be further enhanced by interfacing with such networked resources. By way of example, certain set-top boxes may be adapted to access networked resources and provide information and/or content received there from for use in an electronic entertainment system. Some set-top boxes may also be adapted to allow for information and/or content to be provided from one or more devices in an electronic entertainment system to one or more networked resources.
  • The sharing and distribution of content, and in particular video/audio content continues to increase as more and more users upload and download such content to and from various network resource devices.
  • BRIEF DESCRIPTION OF DRAWINGS
  • Non-limiting and non-exhaustive aspects are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various figures unless otherwise specified.
  • FIG. 1 is a block diagram illustrating an exemplary implementation of a computing environment that includes various networked computing devices and an electronic entertainment system adapted for use in combining and distributing user enhanced video/audio content.
  • FIG. 2 is a block diagram illustrating certain features of an exemplary system that may be implemented to combine and distribute user enhanced video/audio content.
  • FIG. 3 is a block diagram illustrating certain features of an implementation of an optical disc and an optical disc player device as may be implemented for use in combining and distributing user enhanced video/audio content.
  • FIG. 4 is a flow diagram illustrating an exemplary implementation of a method that may be adapted for use in combining and distributing user enhanced video/audio content.
  • FIG. 5 is a time line diagram illustrating some exemplary representative combined user enhanced video/audio content.
  • FIG. 6 is a flow diagram illustrating an exemplary implementation of a method that may be adapted for use in combining and distributing user enhanced video/audio content.
  • DETAILED DESCRIPTION
  • In accordance with certain aspects, some exemplary methods and apparatuses are presented below that may be adapted for use in an electronic environment that may support community-based creation, distribution, and playback of combined content.
  • Such exemplary methods and apparatuses may, for example, be adapted to provide or otherwise support a community of users having the ability to insert video messages and overlays to enhance or otherwise augment the electronic entertainment experience. Such ability may be used in both an interpersonal messaging sense, and a larger group publishing sense.
  • In the following sections, certain implementations and aspects are presented which may be adapted to provide for an integrated system that includes the video creation, the video storage, the video transmission, and the video playback processes. While the term “video” is used in these sections, it should be understood that the term may include video and/or audio.
  • In accordance with one aspect, the methods and apparatuses provided herein may be adapted to provide users a user-friendly space in which they may create their own video tracks (user content) to augment the presentation of a packaged movie, or other like reference content. These video tracks can be used to create combined content for both friends (selected users) and/or the community at large. Such combined content may, in certain implementations, take the form of “movie messages” that may be used for user-made commentary, in-movie messaging, relational notes, fan art, fan videos, fan-enhancement, and other like uses.
  • In certain example implementations, the methods and apparatuses herein may be adapted to provide or otherwise support the creation of videos that may be displayed as a composite on top of an existing packaged movie (e.g., Picture in Picture, etc.), as a simultaneous video feed (e.g., presented on the same screen as the movie or a separate screen/device), and/or as video that will be inserted into the movie's playback (e.g. an interrupt, interlude, etc.).
  • In certain example implementations, the methods and apparatuses herein may be adapted to provide or otherwise support an integrated system that includes video creation, video storage, video transmission, and/or video playback processes.
  • In accordance with certain aspects, the methods and apparatuses herein may be adapted to interface with on an existing fundamental architecture for a social networking web service that has the ability to be connected to a website, set-top media players (e.g., Blu-ray player devices, etc.), and/or the like. Such architecture may include, for example, certain features such as inter-network messaging, buddy lists, etc.
  • An exemplary video for use in such an architecture may include a number of attributes, such as, tags or a collection of access-tags. Such tags may, for example, be managed by the social networking service provider, and include a way to note the video's owner and a list of those intended to be allowed to view the video. This may be useful when considering both privacy concerns and decency concerns. A video's creator may be noted, so that that user may edit and remove the video, including the ability to manage the rest of the video's access-tags. The rest of these tags may be associated with a collection of users who may view the video. A user may select an individual user or groups of users, or select all of their buddies. In the case of ‘all buddies’, the social networking service may be adapted to actively update such a list to be current with the user's buddy list. In certain implementations, a user may also opt to submit a video for public consumption.
  • When a user uploads a video, they may be given the option to send a message to one or more other users to inform them of this video's existence. In some implementations, other users may browse videos to which they may have access.
  • A user-created video may be access-tagged and uploaded to an existing social-networking web-service. Such user content may, for example, be captured live (e.g., via a webcam), and/or uploaded from a previously captured and/or edited source by the user. In certain example implementations, a user may have the ability provide user input (sender input) to select certain video presentation options. For example, a user may choose a location on the screen for the tagged user content as a pop-up, weather or not the tagged user content interrupts the movie (reference content), or appear along side of, or on top of it, etc. In other example implementations, a user may be given the option to ‘frame’ their video using the images. These and/or other like options may, for example, be presented to the user in a number of ways, and on many devices (computer, player, etc), and make use of the web service framework to further manage the same data.
  • In one example, a networked laptop may be used in conjunction with a connected Blu-ray player. Here, for example, a user may synchronize both devices, and record himself making a commentary track while a movie plays. A website may be provided or created that gives the user's computer control of the Blu-ray player's playback commands, and will note the player's timecode, and automatically link created video to the appropriate spot in the movie.
  • In another example, a user may seek to upload a pre-existing video while still connected to a Blu-ray player. In this situation, once the video is uploaded, the user may scrub through the movie on their player, and select the video's start point. The user may then see a live preview of their video (e.g., streaming from the social network service) on top of their player's reference content.
  • In yet another example, a use may be presented with a computer but no local Blu-ray player. This may, for example, become common in a Kiosk type of display. A user may be presented with a webcam and the ability to record commentary on the fly. The system may be adapted to provide the user with pre-selected popular scenes or other like portions of reference content to choose from, and over/along which to record their user content. In the case of a kiosk or other like device, a reference content owner/manager may choose which movies to provide. In the case of a user's computer being employed, a user may be allowed by a reference content owner/manager to access movies per some account registration agreement or the like.
  • In certain example implementations, such a system may provide users with sample clips or portions from the movie that will playback as they record their message, and/or still images or frames (e.g., thumbnail images, etc.). In certain example implementations, a reference content owner/manager and/or web service may include a varied sample of a movie to provide context for features such as this one. Thus, a user may preview of user content as overlay or the like (e.g., perhaps presented in a lower resolution format) as generated by the web service and sent to the computer. For example, such a preview of the combined content may be presented in a version as a flash application or the like.
  • In accordance with certain exemplary aspects, a web application or the like may be adapted to manage the storage of copies of uploaded tagged user content. The videos that users create may have to be encoded into a few different formats. A version for website playback may, for example, be created using flash, etc. Versions for download to Blu-ray players may, for example, be created using an MPEG-II and/or other like standard, both for download and progressive download for streaming.
  • There may be varying methods with which a receiver user may initiate playback of combined content. By way of example, if a user has been sent a message notifying them of a new video creation by a sender user, they may choose to view the video within the context of the entire movie, or to view only the portion or portions of the movie that are relevant to the new video created.
  • With these and other exemplary aspects in mind, attention is drawn to FIG. 1, which illustrates an exemplary implementation of a computing environment 100 that may include various networked computing devices and an electronic entertainment system, and which may be adapted for use in combining and distributing user enhanced video/audio content.
  • As shown, environment 100 may include an optical disc 102 adapted for use in an electronic entertainment system 104 having an optical disc player device 106 coupled to a video monitor device 108 and audio monitor device 110. In this example, optical disc 102 may include content information 112 and instructional information 114. Electronic entertainment system 104 may, for example, be operatively coupled to communicate with at least one computing device, such as, network resource device 120 through a network 116.
  • At least a portion of instructional information 114 may be implemented to adapt optical disc player device 106 to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein.
  • In certain implementations, optical disc 102 and/or other user content may be accessed by a sender user computing device 122. Computing device 122 may, for example, include any computer or other like appliance that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein. For example, computing device 122 may be adapted to access user content and provide tagged user content to network resource device 120. By way of example but not limitation, computing device 122 may include a personal computer, desk top computer, a lap top computer, a camera, a personal digital assistant, a smart phone, a set top box, etc. As illustrated, computing device 122 may be connected to electronic entertainment system 104, and/or network 116.
  • System 100 may also include one or more receiver user computing devices 124. Computing device 124 may, for example, include any computer or other like appliance that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein. For example, computing device 124 may access combined content from network resource device 120 through network 116. By way of example but not limitation, computing device 122 may include a personal computer, desk top computer, a lap top computer, a camera, a personal digital assistant, a smart phone, an electronic entertainment system, a set top box, etc.
  • Network resource device 120 may include one or more computing devices that may be adapted to support at least a portion of the processes for combining and distributing user enhanced video/audio content as provided herein. For example, network resource device 120 may receive tagged user content from one or more other computing devices or the like and provide combined content that may be established based, at least in part, on the tagged user content. The combined content may, for example, be provided (e.g., downloaded, streamed, etc.) to one or more computing devices or the like through network 116.
  • Attention is drawn next to FIG. 2, which is a block diagram illustrating certain features of an exemplary system 200 that may be implemented to combine and distribute user enhanced video/audio content, for example, as in FIG. 1.
  • System 200 may include a content establishing device 202. Content establishing device 202 may be adapted to establish user content 204. User content 204 may, for example, include one or more video and/or audio tracks. Content establishing device 202 may include one or more devices for capturing, recording, and/or otherwise generating video and/or audio tracks. By way of example but not limitation, content establishing device 202 may include a camera, a microphone, a computing device, and/or the like. The resulting user content 204 may, for example, be stored on a computer readable medium.
  • User content 204 may be accessed by a user content tagging device 206. User content tagging device 206 may, for example, include one or more computing devices, an optical disc reading device, and/or the like. User content tagging device 206 may also be adapted to access at least a portion of reference content 112. For example, in certain implementations, reference content 112 may be accessed from an optical disc 102, memory, and/or other like computer readable media. Reference content 112 may include, for example, data associated with at least portion of one or more video and/or audio tracks.
  • User content tagging device 206 may be adapted to establish tagged user content 208 and/or sender input information 210. For example, user content tagging device 206 may be adapted to establish tagged user content 208 based, at least in part, on at least a portion of user content 204 and/or at least a portion of reference content 112 and/or other information associated with reference content 112. User content tagging device 206 may be adapted to establish sender input information 210 based, at least in part, on user input.
  • Tagged user content 208 and/or sender input information 210 may be provided to a content combining device 212. Content combining device 212 may, for example, include one or more computing devices and/or the like. By way of example but not limitation, content combining device 212 may include network resource device 120. Content combining device 212 may be adapted to access at least a portion of reference content 112 (e.g., another copy thereof). Content combining device 212 may be adapted to establish at least one version of combined content 214 (e.g., user enhanced) based, at least in part, on tagged user content 208 and at least a portion of reference content 112. In certain implementations, content combining device 212 may be adapted to establish such combined content 214 based, at least in part, on tagged user content 208, at least a portion of reference content 112, and at least a portion of sender input information 210.
  • Content combining device 212 may be adapted to provide a combined content availability notice 216 to one or more computing devices. For example, content combining device 212 may be adapted to provide such an availability notice 216 to one or more computing devices based, at least in part, on sender input information 210 and/or tagged user content 208. In certain implementations, user content tagging device 206 may be adapted to provide such an availability notice 216 to one or more computing devices, e.g., as illustrated in FIG. 2 by the dashed line.
  • As in this illustrated example, such an availability notice 216 may be provided to a combined content accessing device 218, which may include one or more computing devices and/or the like. For example, combined content accessing device 218 may include receiver user computing device 124. Combined content accessing device 218 may be adapted to access combined content 214 and/or otherwise provide such to a combined content player device 220. While illustrated in the example as being separate devices, in certain implementations, combined content accessing device 218 and combined content player device 220 may be implemented using one or more devices.
  • Combined content player device 220 may, for example include one or more computing devices, entertainment systems, smart phones, MP3 Players, etc., that may be adapted to replay or otherwise reproduce at least a portion of at least one of an audio and/or video track of combined content 214.
  • In the example system 200 above, user content tagging device 206 may be associated with a first user (e.g., a sender user) and combined content accessing device may be associated with a second user (e.g., a receiver user), and content combining device 212 may be associated with a service provider, web entity, and/or the like. In certain implementations, system 200 and the associated parties (e.g., sender user, receiver user(s), and service provider) may be part of a web-based or other like affiliated community. For example, the users may open accounts and specify other users as “buddies” with whom, through the service provider may share combined content 214.
  • Reference is made next to FIG. 3, which is a block diagram illustrating certain example features that may be implemented in an exemplary optical disc 102 and/or an exemplary optical disc player device 106, and/or implemented to adapt in some manner a function of a computing device such as, e.g., computing devices 122, 124, and/or network resource device 120.
  • Optical disc player device 106 may, for example, include at least one processing unit 302, some form of a memory 304, at least one user interface 306, an optical disc reading mechanism 330, and a network interface 326. As illustrated in FIG. 3, at least one bus 350 or the like may be adapted to operatively couple two or more of processing unit 302, memory 304, user interface 306, optical disc reading mechanism 330, and/or network interface 326.
  • As illustrated, optical disc 102 may be operatively coupled to (and accessed through) optical disc reading mechanism 330. Optical disc 102 may have stored thereon content information 112, including video data 332 and audio data 334. Optical disc 102 may, for example, have stored thereon instructional information 114 including computer implementable instructions 348.
  • Memory 304 may, for example, include primary memory 304A. In certain example implementations memory 304 may also include secondary memory 304B (e.g., a hard disk drive, solid state drive, or other like non-volatile storage devices and/or media).
  • As shown in FIG. 3, primary memory 304A may include tagged user content 208 and sender input 210 which may be at least a partially associated with user input that may be received through user interface 306. By way of example but not limitation, user interface 306 may include and/or otherwise be operatively coupled to a user input device 308 (e.g., a keypad, a touch screen, a joystick, a mouse, remote control, a camera, a microphone, etc.). In certain exemplary implementations, user interface 306 may include a wireless communication interface adapted to receive wireless signals (not shown) from a remotely controlled user input device 308. Such wireless signals may, for example, include infrared, radio frequency, or other like electromagnetic signals.
  • While not shown in FIG. 3, optical disc reading mechanism 330 may, for example, include a receptacle adapted for receiving optical disc 102, an optical disc rotation mechanism adapted for rotating optical disc 102, a optical reading mechanism (e.g., a laser read head, etc.) adapted for reading information stored on optical disc 102 in the form of data, and supporting control circuitry adapted for controlling the operation of optical disc reading mechanism 330 and providing the information read from optical disc 102 to at least bus 350.
  • In certain exemplary implementations, optical disc reading mechanism 330 may be adapted to read one or more of a compact disc (CD), a digital versatile disc (DVD), a Blu-Ray disc, and/or the like. In certain exemplary implementations, optical disc reading mechanism 330 may be adapted to hold and/or otherwise manipulate multiple optical discs. In certain exemplary implementations, optical disc 102 may be adapted as a read only disc, a read/writable disc, a read/rewritable disc, etc. Thus, in certain implementations, optical disc reading mechanism 330 may be adapted to only read from optical disc 102, while in other implementations optical disc reading mechanism 330 may be adapted to read and write to optical disc 102. Such mechanisms and techniques are well known.
  • Processing unit 302 may, for example, be implemented using hardware, firmware, software, and/or any combination thereof. Processing unit 302 may, for example, be implemented using digital and/or analog circuitry. Processing unit 302 may, for example, be implemented using a central processing unit, a microprocessor, a microcontroller, an application specific integrated circuit (ASIC), a graphical processing unit (GPU), and/or other like circuits.
  • Processing unit 302 may include any form of circuitry that may be adapted based, at least in part, on at least a portion of instructional information 114. Processing unit 302 may, for example, be adapted to perform and/or to otherwise support at least a portion one or more of exemplary methods 400 (FIG. 4) and/or 600 (FIG. 6).
  • Network interface 326 may, for example, include any circuitry adapted to provide for communication with at least network 116. Network interface 326 may, for example, be adapted to provide such communication over one or more wired and/or wireless communication links. Network 116 may, for example, include one or more communication resources adapted to provide communication between network interface 326 and at least one computing device, such as, e.g., network resource device 120.
  • By way of example but not limitation, in certain implementations, network 116 may include a wired and/or wireless local area network (LAN and/or WLAN), a wired and/or wireless telephone network, the Internet, and/or the like. By way of example but not limitation, in certain implementations, network interface 226 may include a wired and/or wireless modem, an Ethernet interface, a telephone interface, an antenna, a transmitter, a receiver, a transceiver, and/or the like.
  • Reference is now made to FIG. 4, which is a flow diagram illustrating an exemplary method 400 that may, for example, be adapted for use in combining and distributing user enhanced video/audio content. Method 400 may be implemented to provide for the services presented in the various examples herein.
  • At block 402, a web client in a computing device or the like associated with a sender user may be tied to an account associated with a service provider or the like. At block 404, a player may be loaded with reference content. At block 406, a Blu-ray player may be activated and synchronized with the web client. At block 408, timecode or other like temporal indicator may be noted and user content recorded or otherwise generated.
  • At block 410, it may be determined if a Blu-ray player is available. If there is no Blu-ray player, then at block 412 the user may select reference content. At block 414, the user may select from presented content portions, such as, e.g., clips, frames, thumbnails, etc. If at block 410 it is determined that there is a Blu-ray player, then at block 416 a player may be loaded with reference content (e.g., an optical disc). At block 418, at least one content timecode or the like may be selected by user input.
  • At block 420, a web client may not be tied to a user account. Hence, at block 422 a new user account may be created. At block 424, reference content may be selected. At block 426, the user may select from presented content portions, such as, e.g., clips, frames, thumbnails, etc.
  • At block 430 the user may provide input to customize a presentation following, either block 408, 414, 418, or 426. At block 432, user content may be tagged and provided along with sender input. At block 434, combined content may be generated.
  • At block 436, one or more receiver users may be notified that the combined content is available. At block 438, the combined content may be accessed by one or more receiver users, e.g., through download and/or streaming. At block 440, at least a portion of the accessed combined content may be played or otherwise displayed.
  • Reference is now made to FIG. 5, which is a time line diagram illustrating some exemplary representative combined user enhanced video/audio content. Here, for example, user content 204 is illustrated as a track, which may be tagged and/or otherwise adapted in some manner to produce track 502. Track 503 may have duration of dB, for example. Reference content 112 is illustrated as a track that starts at tA0 and ends at tA2. Here, a user selected timecode or the like is illustrated by the dashed line at tA1.
  • An exemplary combined content 212A is illustrated as a track that starts at tA0 and ends at tA2, but which also includes user content as combined in some manner at tA1 and ending at tA1+dB. Hence, in this example, the combined content is the same length as the reference content.
  • In other implementations, the combined content may have a shorter or longer length then the reference content. For example, combined content 212B is illustrated as a track that starts at tA1 and ends at about tA1+dB, making it about the same length as the tagged user content, and/or user content. In anther example, combined content 212C is illustrated as a track that starts at tD0 and ends at tD1, making it the longer than the tagged user content, and/or user content, but shorter than the reference content.
  • FIG. 6 is a flow diagram illustrating an exemplary implementation of a method 600 that may be adapted for use in combining and distributing user enhanced video/audio content in one or more networked devices. At block 602, user content may be accessed. At block 604, reference content may be accessed. At block 606, at least a portion of the user content may be associated with at least a portion of the reference content to produce tagged user content. At block 608, combined content may be established based, at least in part, on at least a portion of the reference content and the tagged user content. At block 610, at least a portion of the combined content may be accessed. At block 612, at least a portion of the combined content may be played.
  • While certain exemplary techniques have been described and shown herein using various systems and methods, it should be understood by those skilled in the art that various other modifications may be made, and equivalents may be substituted, without departing from claimed subject matter. Additionally, many modifications may be made to adapt a particular situation to the teachings of claimed subject matter without departing from the central concept described herein. Therefore, it is intended that claimed subject matter not be limited to the particular examples disclosed, but that such claimed subject matter may also include all implementations falling within the scope of the appended claims, and equivalents thereof.

Claims (54)

1. A method comprising:
accessing user content, said user content comprising at least one of a user video track and/or a user audio track;
accessing reference content, said reference content comprising at least one of a reference video track and/or a reference audio track;
establishing tagged user content by temporally associating at least a portion of said user content with at least a portion of said reference content, said tagged user content further comprising sender input information;
providing said tagged user content to at least one network resource device; and
with said at least one network resource device:
establishing combined content by combining at least a portion of said tagged user content with at least a portion of said reference content, and
providing access to said combined content based, at least on part, and at least a portion of said sender input information.
2. The method as recited in claim 1, wherein accessing said user content comprises using at least one content establishing device to establish said user content.
3. The method as recited in claim 1, wherein accessing said reference content comprises accessing said reference content from an optical disc.
4. The method as recited in claim 3, wherein said optical disc comprises a Blu-ray disc.
5. The method as recited in claim 1, wherein establishing said tagged user content comprises accessing at least one temporal identifier associated with said reference content
6. The method as recited in claim 5, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
7. The method as recited in claim 5, wherein accessing said at least one temporal identifier associated with said reference content comprises establishing said at least one temporal identifier based, at least in part, on user input.
8. The method as recited in claim 1, wherein sender input information is identifiably associated with at least a first user.
9. The method as recited in claim 1, wherein sender input information comprises at least one presentation parameter.
10. The method as recited in claim 1, wherein establishing said combined content comprises at least one of:
combining at least a portion of said user video track with at least a portion of said reference video track, and/or
combining at least a portion of said user audio track with at least a portion of said reference audio track.
11. The method as recited in claim 1, wherein providing access to said combined content comprises sending at least one notification to a computing device associated with at least one other user, said notification identifying that said combined content is available for access.
12. The method as recited in claim 11, wherein providing access to said combined content comprises identifying said at least one other user based, at least in part, on said sender input information.
13. A system comprising:
a network
a first device connected to said network and adapted to:
access user content and reference content, said user content comprising at least one of a user video track and/or a user audio track, and said reference content comprising at least one of a reference video track and/or a reference audio track,
establish tagged user content by temporally associating at least a portion of said user content with at least a portion of said reference content, said tagged user content further comprising sender input information, and
provide said tagged user content to said network; and
at least one network resource device connected to said network and adapted to:
establish combined content by combining at least a portion of said tagged user content with at least a portion of said reference content, and
provide access to said combined content based, at least on part, and at least a portion of said sender input information.
14. The system as recited in claim 13, further comprising:
at least one content establishing device adapted to establish said user content.
15. The system as recited in claim 13, further comprising:
an optical disc reading mechanism connected to said first device and adapted to accept an optical disc comprising said reference content.
16. The system as recited in claim 15, wherein said optical disc comprises a Blu-ray disc.
17. The system as recited in claim 13, wherein said tagged user content is based, at least in part, on and/or comprises at least one temporal identifier associated with said reference content
18. The system as recited in claim 17, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
19. The system as recited in claim 17, wherein said at least one temporal identifier is established based, at least in part, on user input.
20. The system as recited in claim 13, wherein sender input information is identifiably associated with at least a first user.
21. The system as recited in claim 13, wherein sender input information comprises at least one presentation parameter.
22. The system as recited in claim 13, wherein said at least one network resource device is adapted to combine at least a portion of said user video track with at least a portion of said reference video track, and/or at least a portion of said user audio track with at least a portion of said reference audio track.
23. The system as recited in claim 13, further comprising:
at least one computing device connected to said network and adapted to access said combined content, and wherein said at least one network resource device is adapted to send at least one notification to said at least one computing device, said at least one computing device being associated with at least one other user, said notification identifying that said combined content is available for access.
24. The system as recited in claim 23, wherein said at least one network resource device is adapted to identify said at least one other user based, at least in part, on said sender input information.
25. An apparatus for use in a networked computing environment, the apparatus comprising:
means for accessing tagged user content, said tagged user content being established by temporally associating at least a portion of user content with at least a portion of reference content, said user content comprising at least one of a user video track and/or a user audio track, and said reference content comprising at least one of a reference video track and/or a reference audio track, and wherein said tagged user content further comprises sender input information identifiably associated with at least a first user, and
means for establishing combined content by combining at least a portion of said tagged user content with at least a portion of said reference content;
means for notifying at least one computing device associated with at least one other user, that said combined content is available for access;
means for providing access to said combined content based, at least on part, and at least a portion of said sender input information.
26. The apparatus as recited in claim 25, wherein said tagged user content is based, at least in part, on and/or comprises at least one temporal identifier associated with said reference content
27. The apparatus as recited in claim 26, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
28. The apparatus as recited in claim 25, wherein sender input information comprises at least one presentation parameter.
29. The apparatus as recited in claim 25, wherein said means for establishing said combined content is adapted to combine at least a portion of said user video track with at least a portion of said reference video track, and/or at least a portion of said user audio track with at least a portion of said reference audio track.
30. The apparatus as recited in claim 25, wherein said means for notifying said at least one computing device is adapted to identify said at least one other user based, at least in part, on said sender input information.
31. An apparatus comprising:
means for accessing user content and reference content, said user content comprising at least one of a user video track and/or a user audio track, and said reference content comprising at least one of a reference video track and/or a reference audio track;
means for establishing tagged user content by temporally associating at least a portion of said user content with at least a portion of said reference content, said tagged user content further comprising sender input information; and
means for providing said tagged user content to a network.
32. The apparatus as recited in claim 31, wherein said means for accessing user content and reference content comprises means for reading an optical disc comprising said reference content.
33. The apparatus as recited in claim 32, wherein said optical disc comprises a Blu-ray disc.
34. The apparatus as recited in claim 31, wherein said tagged user content is based, at least in part, on and/or comprises at least one temporal identifier associated with said reference content
35. The apparatus as recited in claim 34, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
36. The apparatus as recited in claim 34, wherein said at least one temporal identifier is established based, at least in part, on user input.
37. The apparatus as recited in claim 31, wherein sender input information is identifiably associated with at least a first user.
38. The apparatus as recited in claim 31, wherein sender input information comprises at least one presentation parameter.
39. The apparatus as recited in claim 31, wherein said sender input information is based, at least in part, on and/or comprises at least one identifier associated with at least one other user.
40. A computer readable medium comprising having stored thereon computer implementable instructions that if implemented using at least one processing unit within a device, adapt the device to:
access tagged user content, said tagged user content being established by temporally associating at least a portion of user content with at least a portion of reference content, said user content comprising at least one of a user video track and/or a user audio track, and said reference content comprising at least one of a reference video track and/or a reference audio track, and wherein said tagged user content further comprises sender input information identifiably associated with at least a first user, and
establish combined content by combining at least a portion of said tagged user content with at least a portion of said reference content;
notify at least one computing device associated with at least one other user, that said combined % content is available for access;
provide access to said combined content based, at least on part, and at least a portion of said sender input information.
41. The computer readable medium as recited in claim 40, wherein said tagged user content is based, at least in part, on and/or comprises at least one temporal identifier associated with said reference content
42. The computer readable medium as recited in claim 41, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
43. The computer readable medium as recited in claim 41, wherein sender input information comprises at least one presentation parameter.
44. The computer readable medium as recited in claim 40, further comprising computer implementable instructions that if implemented using the at least one processing unit within the device, adapt the device to combine at least a portion of said user video track with at least a portion of said reference video track, and/or at least a portion of said user audio track with at least a portion of said reference audio track.
45. The computer readable medium as recited in claim 40, further comprising computer implementable instructions that if implemented using the at least one processing unit within the device, adapt the device to identify said at least one other user based, at least in part, on said sender input information.
46. A computer readable medium comprising having stored thereon computer implementable instructions that if implemented using at least one processing unit within a device, adapt the device to:
access user content and reference content, said user content comprising at least one of a user video track and/or a user audio track, and said reference content comprising at least one of a reference video track and/or a reference audio track;
establish tagged user content by temporally associating at least a portion of said user content with at least a portion of said reference content, said tagged user content further comprising sender input information; and
provide said tagged user content to a network.
47. The computer readable medium as recited in claim 46, further comprising computer implementable instructions that if implemented using the at least one processing unit within the device, adapt the device to access an optical disc comprising said reference content.
48. The computer readable medium as recited in claim 47, wherein said optical disc comprises a Blu-ray disc.
49. The computer readable medium as recited in claim 46, wherein said tagged user content is based, at least in part, on and/or comprises at least one temporal identifier associated with said reference content
50. The computer readable medium as recited in claim 49, wherein said at least one temporal identifier comprises at least one of a timecode value, a chapter identifier, and/or a sampled frame.
51. The computer readable medium as recited in claim 49, wherein said at least one temporal identifier is established based, at least in part, on user input.
52. The computer readable medium as recited in claim 46, wherein sender input information is identifiably associated with at least a first user.
53. The computer readable medium as recited in claim 46, wherein sender input information comprises at least one presentation parameter.
54. The computer readable medium as recited in claim 46, wherein said sender input information is based, at least in part, on and/or comprises at least one identifier associated with at least one other user.
US12/239,596 2008-06-06 2008-09-26 Methods and apparatuses for combining and distributing user enhanced video/audio content Abandoned US20090307227A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US12/239,596 US20090307227A1 (en) 2008-06-06 2008-09-26 Methods and apparatuses for combining and distributing user enhanced video/audio content
PCT/US2009/003129 WO2009148513A1 (en) 2008-06-06 2009-05-19 Methods and apparatuses for combining and distributing user enhanced video/audio content
US16/450,032 US11582507B2 (en) 2008-06-06 2019-06-24 Methods and apparatuses for combining and distributing user enhanced video/audio content

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US5951008P 2008-06-06 2008-06-06
US12/239,596 US20090307227A1 (en) 2008-06-06 2008-09-26 Methods and apparatuses for combining and distributing user enhanced video/audio content

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/450,032 Continuation US11582507B2 (en) 2008-06-06 2019-06-24 Methods and apparatuses for combining and distributing user enhanced video/audio content

Publications (1)

Publication Number Publication Date
US20090307227A1 true US20090307227A1 (en) 2009-12-10

Family

ID=40845789

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/239,596 Abandoned US20090307227A1 (en) 2008-06-06 2008-09-26 Methods and apparatuses for combining and distributing user enhanced video/audio content
US16/450,032 Active 2031-03-18 US11582507B2 (en) 2008-06-06 2019-06-24 Methods and apparatuses for combining and distributing user enhanced video/audio content

Family Applications After (1)

Application Number Title Priority Date Filing Date
US16/450,032 Active 2031-03-18 US11582507B2 (en) 2008-06-06 2019-06-24 Methods and apparatuses for combining and distributing user enhanced video/audio content

Country Status (2)

Country Link
US (2) US20090307227A1 (en)
WO (1) WO2009148513A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110047248A1 (en) * 2009-08-21 2011-02-24 Samsung Electronics Co., Ltd. Shared data transmitting method, server, and system
WO2012094432A1 (en) * 2011-01-04 2012-07-12 Related Content Databases, Inc. System and method for interfacing content playback devices with network sites to supplement content playback
US20130086143A1 (en) * 2011-09-30 2013-04-04 Google Inc. Music On Video Site
US8838680B1 (en) 2011-02-08 2014-09-16 Google Inc. Buffer objects for web-based configurable pipeline media processing
US9172740B1 (en) 2013-01-15 2015-10-27 Google Inc. Adjustable buffer remote access
US9225979B1 (en) 2013-01-30 2015-12-29 Google Inc. Remote access encoding
US9311692B1 (en) 2013-01-25 2016-04-12 Google Inc. Scalable buffer remote access
US20160241627A1 (en) * 2002-01-29 2016-08-18 FiveOpenBooks, LLC Method and System for Delivering Media Data
US9762967B2 (en) 2011-06-14 2017-09-12 Comcast Cable Communications, Llc System and method for presenting content with time based metadata
US10645460B2 (en) * 2016-12-30 2020-05-05 Facebook, Inc. Real-time script for live broadcast

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6011548A (en) * 1998-09-04 2000-01-04 Cyberstar, L.P. System for integrating satellite boardband data distributed over a cable TV network with legacy corporate local area networks
US6061717A (en) * 1993-03-19 2000-05-09 Ncr Corporation Remote collaboration system with annotation and viewer capabilities
US20030172381A1 (en) * 2002-01-25 2003-09-11 Koninklijke Philips Electronics N.V. Digital television system having personalized addressable content
US20040064835A1 (en) * 2002-09-26 2004-04-01 International Business Machines Corporation System and method for content based on-demand video media overlay
US20040117819A1 (en) * 2002-12-03 2004-06-17 Ming-He Yu Apparatus for producing TV advertising contents and inserting interstitial advertisements on TV programs
US20040133647A1 (en) * 1998-12-23 2004-07-08 Canon Kabushiki Kaisha Method and system for conveying video messages
US7069310B1 (en) * 2000-11-10 2006-06-27 Trio Systems, Llc System and method for creating and posting media lists for purposes of subsequent playback
US20060174311A1 (en) * 2004-11-23 2006-08-03 Palo Alto Research Center Incorporated Method, apparatus, and program products for socially synchronizing an experiential data stream
US20060265657A1 (en) * 2005-05-23 2006-11-23 Gilley Thomas S Distributed scalable media environment
US20070162927A1 (en) * 2004-07-23 2007-07-12 Arun Ramaswamy Methods and apparatus for monitoring the insertion of local media content into a program stream
US20070265855A1 (en) * 2006-05-09 2007-11-15 Nokia Corporation mCARD USED FOR SHARING MEDIA-RELATED INFORMATION
US7312803B2 (en) * 2004-06-01 2007-12-25 X20 Media Inc. Method for producing graphics for overlay on a video source
US20070300260A1 (en) * 2006-06-22 2007-12-27 Nokia Corporation Method, system, device and computer program product for generating and distributing media diary podcasts
US20080016114A1 (en) * 2006-07-14 2008-01-17 Gerald Thomas Beauregard Creating a new music video by intercutting user-supplied visual data with a pre-existing music video
US20080060084A1 (en) * 2006-06-30 2008-03-06 Library Video Company Distribution and management of content using playlists
US20090157502A1 (en) * 2007-12-17 2009-06-18 Microsoft Corporation Ad-bearing bookmarks for electronic media documents
US7813228B2 (en) * 2004-08-31 2010-10-12 Koninklijke Philips Electronics N.V. Accessing pre-recorded content on a record carrier
US9407940B1 (en) * 2008-03-20 2016-08-02 Sprint Communications Company L.P. User-targeted ad insertion in streaming media

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8819719B1 (en) * 2006-12-06 2014-08-26 Google Inc. Real-time video commenting
US20080282191A1 (en) * 2007-05-07 2008-11-13 Bellsouth Intellectual Property Corporation Methods, systems, and computer program products for providing centralized management and delivery services for media content
US8316302B2 (en) * 2007-05-11 2012-11-20 General Instrument Corporation Method and apparatus for annotating video content with metadata generated using speech recognition technology
US7953255B2 (en) * 2008-05-01 2011-05-31 At&T Intellectual Property I, L.P. Avatars in social interactive television

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6061717A (en) * 1993-03-19 2000-05-09 Ncr Corporation Remote collaboration system with annotation and viewer capabilities
US6011548A (en) * 1998-09-04 2000-01-04 Cyberstar, L.P. System for integrating satellite boardband data distributed over a cable TV network with legacy corporate local area networks
US20040133647A1 (en) * 1998-12-23 2004-07-08 Canon Kabushiki Kaisha Method and system for conveying video messages
US7069310B1 (en) * 2000-11-10 2006-06-27 Trio Systems, Llc System and method for creating and posting media lists for purposes of subsequent playback
US20030172381A1 (en) * 2002-01-25 2003-09-11 Koninklijke Philips Electronics N.V. Digital television system having personalized addressable content
US20040064835A1 (en) * 2002-09-26 2004-04-01 International Business Machines Corporation System and method for content based on-demand video media overlay
US20040117819A1 (en) * 2002-12-03 2004-06-17 Ming-He Yu Apparatus for producing TV advertising contents and inserting interstitial advertisements on TV programs
US7312803B2 (en) * 2004-06-01 2007-12-25 X20 Media Inc. Method for producing graphics for overlay on a video source
US20070162927A1 (en) * 2004-07-23 2007-07-12 Arun Ramaswamy Methods and apparatus for monitoring the insertion of local media content into a program stream
US7813228B2 (en) * 2004-08-31 2010-10-12 Koninklijke Philips Electronics N.V. Accessing pre-recorded content on a record carrier
US20060174311A1 (en) * 2004-11-23 2006-08-03 Palo Alto Research Center Incorporated Method, apparatus, and program products for socially synchronizing an experiential data stream
US20060265657A1 (en) * 2005-05-23 2006-11-23 Gilley Thomas S Distributed scalable media environment
US20070265855A1 (en) * 2006-05-09 2007-11-15 Nokia Corporation mCARD USED FOR SHARING MEDIA-RELATED INFORMATION
US20070300260A1 (en) * 2006-06-22 2007-12-27 Nokia Corporation Method, system, device and computer program product for generating and distributing media diary podcasts
US20080060084A1 (en) * 2006-06-30 2008-03-06 Library Video Company Distribution and management of content using playlists
US20080016114A1 (en) * 2006-07-14 2008-01-17 Gerald Thomas Beauregard Creating a new music video by intercutting user-supplied visual data with a pre-existing music video
US20090157502A1 (en) * 2007-12-17 2009-06-18 Microsoft Corporation Ad-bearing bookmarks for electronic media documents
US9407940B1 (en) * 2008-03-20 2016-08-02 Sprint Communications Company L.P. User-targeted ad insertion in streaming media

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160241627A1 (en) * 2002-01-29 2016-08-18 FiveOpenBooks, LLC Method and System for Delivering Media Data
US10277656B2 (en) * 2002-01-29 2019-04-30 FiveOpenBooks, LLC Method and system for delivering media data
US10193972B2 (en) * 2009-08-21 2019-01-29 Samsung Electronics Co., Ltd Shared data transmitting method, server, and system
US9686354B2 (en) * 2009-08-21 2017-06-20 Samsung Electronics Co., Ltd Shared data transmitting method, server, and system
US20110047248A1 (en) * 2009-08-21 2011-02-24 Samsung Electronics Co., Ltd. Shared data transmitting method, server, and system
US20170272517A1 (en) * 2009-08-21 2017-09-21 Samsung Electronics Co. , Ltd. Shared data transmitting method, server, and system
WO2012094432A1 (en) * 2011-01-04 2012-07-12 Related Content Databases, Inc. System and method for interfacing content playback devices with network sites to supplement content playback
US8838680B1 (en) 2011-02-08 2014-09-16 Google Inc. Buffer objects for web-based configurable pipeline media processing
US9762967B2 (en) 2011-06-14 2017-09-12 Comcast Cable Communications, Llc System and method for presenting content with time based metadata
US20170339462A1 (en) 2011-06-14 2017-11-23 Comcast Cable Communications, Llc System And Method For Presenting Content With Time Based Metadata
US10306324B2 (en) 2011-06-14 2019-05-28 Comcast Cable Communication, Llc System and method for presenting content with time based metadata
USRE48546E1 (en) 2011-06-14 2021-05-04 Comcast Cable Communications, Llc System and method for presenting content with time based metadata
US20130086143A1 (en) * 2011-09-30 2013-04-04 Google Inc. Music On Video Site
US9172740B1 (en) 2013-01-15 2015-10-27 Google Inc. Adjustable buffer remote access
US9311692B1 (en) 2013-01-25 2016-04-12 Google Inc. Scalable buffer remote access
US9225979B1 (en) 2013-01-30 2015-12-29 Google Inc. Remote access encoding
US10645460B2 (en) * 2016-12-30 2020-05-05 Facebook, Inc. Real-time script for live broadcast

Also Published As

Publication number Publication date
US20190313148A1 (en) 2019-10-10
US11582507B2 (en) 2023-02-14
WO2009148513A1 (en) 2009-12-10

Similar Documents

Publication Publication Date Title
US11582507B2 (en) Methods and apparatuses for combining and distributing user enhanced video/audio content
US11081143B2 (en) Providing enhanced content
US10504558B2 (en) Method, system and computer program product for distributed video editing
US7739584B2 (en) Electronic messaging synchronized to media presentation
EP2569750B1 (en) Editable bookmarks shared via a social network
BR112012005499B1 (en) METHOD FOR SHARING DIGITAL MEDIA CONTENT
CN110740386B (en) Live broadcast switching method and device and storage medium
JP2006054898A (en) Multimedia content display system equipped with schedule function, and its content reproducing method
JP5043711B2 (en) Video evaluation apparatus and method
TWI423041B (en) Av playing method capable of improving multimedia interactive mechanism and related apparatus
Miller Sams teach yourself YouTube in 10 Minutes
US8200717B2 (en) Revision of multimedia content
JP2006279843A (en) Content distribution system and content reproduction device
JP5746865B2 (en) Rich media providing system and control method thereof
JP5263399B2 (en) Content upload system, content upload method, content transmission / reception device
JP5178947B2 (en) Server, data distribution system, and information distribution method
JP2013021430A (en) Thumbnail image provision apparatus, method and system
JP2010226523A (en) Content server device, content transmission method, and content transmission program
JP2008305495A (en) Contents transmitting device
JP4876939B2 (en) Information processing apparatus, information processing method, and program
JP6256105B2 (en) Information processing apparatus, playlist generation method and program in information processing apparatus
Shaw Introduction to digital media and Windows Media 9 Series
JP4826677B2 (en) Recording medium and reproducing apparatus
JP2009130499A (en) Contents reproduction device, contents processing system, and contents processing method
JP2010073100A (en) Distribution server, content distribution method, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: DISNEY ENTERPRISES, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PRESTENBACK, KYLE;TAHLER, EVAN;KWAN, BRIAN;AND OTHERS;REEL/FRAME:021595/0719;SIGNING DATES FROM 20080923 TO 20080924

STCV Information on status: appeal procedure

Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS

STCV Information on status: appeal procedure

Free format text: BOARD OF APPEALS DECISION RENDERED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION