US20140215334A1 - Systems and methods for multi-context media control and playback - Google Patents

Systems and methods for multi-context media control and playback Download PDF

Info

Publication number
US20140215334A1
US20140215334A1 US14/231,203 US201414231203A US2014215334A1 US 20140215334 A1 US20140215334 A1 US 20140215334A1 US 201414231203 A US201414231203 A US 201414231203A US 2014215334 A1 US2014215334 A1 US 2014215334A1
Authority
US
United States
Prior art keywords
media content
electronic device
implementations
media
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/231,203
Inventor
Sten Garmark
Quenton Cook
Gustav Soderstrom
Ivo Silva
Michelle Kadir
Peter Stromberg
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Spotify AB
Original Assignee
Spotify AB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Spotify AB filed Critical Spotify AB
Priority to US14/231,203 priority Critical patent/US20140215334A1/en
Publication of US20140215334A1 publication Critical patent/US20140215334A1/en
Assigned to MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT reassignment MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT SUPPLEMENTAL PATENT SECURITY AGREEMENT Assignors: SPOTIFY AB
Assigned to SPOTIFY AB reassignment SPOTIFY AB RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: MORGAN STANLEY SENIOR FUNDING, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/231Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
    • H04N21/23109Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion by placing content in organized collections, e.g. EPG data repository
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42224Touch pad or touch panel provided on the remote control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4396Processing of audio elementary streams by muting the audio signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/482End-user interface for program selection
    • H04N21/4825End-user interface for program selection using a list of items to be played back in a given order, e.g. playlists
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6125Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • H04N21/8113Monomedia components thereof involving special audio data, e.g. different tracks for different languages comprising music, e.g. song in MP3 format
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8549Creating video summaries, e.g. movie trailer

Definitions

  • the disclosed implementations relate generally to controlling and playing back media content.
  • Modern users have a variety of options for devices and ways to consume media content. For example, multiple different types of media, including text, still images, video, audio, and other digital media, can be consumed on personal computers, laptops, mp3 players, mobile phones, tablet computers, televisions, stereos, and many other electronic devices. These electronic devices typically have the ability to store digital media files in memory that is integrated into the devices. But acquiring and storing all the media content to which a user desires access can be prohibitively difficult and expensive.
  • Streaming services allow users to have access to a much larger collection of media than would otherwise be possible.
  • Streaming services store media content on servers remote from the electronic devices, and then send the media content to the electronic devices when it is requested. The media content is then cached on the electronic device for presentation.
  • the content may be temporarily buffered/stored before presentation, but typically the data is deleted from memory after presentation, allowing an almost unlimited amount of media to be streamed to a device without running out of storage space on the electronic device.
  • Some streaming services also allow streamed content to be stored on the local device.
  • media content can be browsed, previewed, and selected for presentation.
  • previews of media content may be available so that the user can listen to a portion of a song or watch a portion of a video before choosing to purchase it or otherwise select it for presentation.
  • previewing media content may be inconvenient when media content is already being presented, such as when a user is browsing an online music store while listening to other music. Accordingly, it would be advantageous to provide systems and methods that allow a user to preview media content in a convenient and streamlined manner.
  • a method of previewing media content is disclosed.
  • the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the method includes detecting a first input.
  • the first input is a contact with a touch-sensitive surface.
  • the first input is a mouse click event.
  • first media content is added to a playlist.
  • the change to the first input is a liftoff from a touch-sensitive surface.
  • the change to the first input is a mouse click-release event.
  • the change to the first input is an addition of another input.
  • the first media content is caused to be presented.
  • the first media content is presented at the same device that received the input (e.g., a computer, television, or mobile phone).
  • the first media content is presented at a different device (e.g., a home stereo system or wireless speaker system).
  • a method of previewing media content is disclosed.
  • the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the method includes, during presentation of a first song, detecting a first input.
  • the first input is a touch input, mouse input, or other appropriate input.
  • the audibility of the first song is reduced.
  • the audibility is reduced by lowering the volume of, muting, or pausing the first song.
  • a second song is presented. After a first time period, presentation of the second song is ceased.
  • the first time period corresponds to the duration of the first input.
  • the first input can correspond to a touch- or click-and-hold input
  • the second song is presented as long as the input is maintained.
  • the audibility of the first song is then increased.
  • the first song is presented at the same or substantially the same audibility as it was prior to presenting the second song.
  • a method of previewing media content is disclosed.
  • the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the method includes, during presentation of currently presented media content, displaying an affordance to a user, where selection of the affordance will cause secondary media content to be presented.
  • the affordance may be a button, icon, or text that, when selected, causes secondary media content to be previewed or otherwise presented.
  • at least a portion of secondary media content is obtained.
  • the portion of the secondary media content is obtained at the same device on which the affordance is displayed. In some implementations, it is obtained at a different device.
  • a first input corresponding to a selection of the affordance is detected.
  • the audibility of the currently presented media content is reduced, and the secondary media content is presented.
  • the presentation of the secondary media content is ceased, and the audibility of the currently presented media content is increased.
  • a method of previewing media content is disclosed.
  • the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the method includes detecting the presence of a cursor in a first area of a display.
  • the first area of the display corresponds to an icon, button, graphic, text, or other type of control, and can be displayed in a media content description area, including, for example, album art, movie stills, content metadata, and the like.
  • Media content is presented in response to detecting the presence of the cursor within the first area.
  • the removal of the cursor from a second area of the display region is detected, where the second area is larger than and contiguous with the first area.
  • the second area corresponds to the media content description area. In some implementations, it is the same size as the media content description area.
  • the media content is presented in response to detecting the removal of the cursor from the second area.
  • a method of controlling media presentation is disclosed.
  • the method is performed at a first electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors.
  • the first electronic device receives a media control command for a second electronic device.
  • the server sends a server media control request to a server system, and sends a local media control request to a second electronic device within a local network to which both the first electronic device and the second electronic device are connected.
  • Local networks may be Local Area Networks (LANs), Personal Area Networks (PANs), ad-hoc computer networks, peer-to-peer computer networks, and the like, and may be wired or wireless.
  • Local networks may use one or more communication standards, protocols, or technologies, such as Internet Protocol (e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.), BLUETOOTH, Wired Ethernet (e.g., IEEE 802.3), Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • Internet Protocol e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.
  • BLUETOOTH Wired Ethernet
  • Wired Ethernet e.g., IEEE 802.3
  • Wi-Fi Wireless Fidelity
  • IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • the server media control request is sent to the server system over the Internet.
  • the first electronic device provides a single user interface that allows a user both to select media content for presentation by the first electronic device and to generate media control requests configured to cause the media content to be presented by the second electronic device.
  • Media content is selected from the group consisting of: an audio track, a video, an image, an audio playlist, and a video playlist.
  • the first electronic device is selected from the group consisting of: a computer, a mobile phone, a remote control device, and a portable media player.
  • the second electronic device is selected from the group consisting of: a television, a home media receiver/player, a computer, a home audio/visual system, and a speaker.
  • the server and the local media control requests are both configured to cause a single media control operation to be implemented at the second electronic device.
  • the media control operation is selected from the group consisting of: play, pause, skip, fast-forward, rewind, adjust an audio volume, change an order of items in a playlist, add items to a playlist, remove items from a playlist, adjust audio equalizer settings, and set a user setting.
  • another method of controlling media presentation is disclosed.
  • the method is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors.
  • the second electronic device receives, from a server system, a server media control request corresponding to a media control operation to be implemented at the second electronic device, wherein the server media control request was sent from the server in response to the server receiving the server media control request from a first electronic device.
  • the second electronic device receives, from the first electronic device, a local media control request sent within a local network to which both the first electronic device and the second electronic device are connected, wherein the local media control request corresponds to the media control operation.
  • the server media control request is received prior to the local media control request.
  • the second electronic device performs the media control operation.
  • the second electronic devices determines whether the server media control request and the local media control request correspond to the same media control operation. If the server media control request and the local media control request correspond to the same media control operation, the server ignores the local media control request.
  • the media control operation is a request to begin presenting media content at the second electronic device, and the media content is being presented by the first electronic device.
  • the second electronic device Prior to receiving the server media control request or the local media control request, the second electronic device buffers/caches an upcoming portion of the media content being presented by the first electronic device.
  • the second electronic device receives one of the server media control request or the local media control request.
  • the second electronic device initiates presentation of the buffered portion of the media content.
  • another method of controlling media presentation is disclosed.
  • the method is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors.
  • the second electronic device receives, from a first electronic device, a local media control request corresponding to a media control operation, wherein the local media control request is sent within a local network to which both the first electronic device and the second electronic device are connected.
  • the second electronic device receives a media stream from a server system, wherein the media stream was sent from the server in response to the server receiving a server media control request from the first electronic device, and wherein the server media control request corresponds to the media control operation.
  • the media stream is received prior to receiving the local media control request.
  • the second electronic devices ignores the local media control request.
  • another method of controlling media presentation is disclosed.
  • the method is performed at a server system having one or more processors and memory storing one or more programs for execution by the one or more processors.
  • the server system receives, from a first electronic device, a server media control request, wherein the first electronic device also sends a local media control request to the second electronic device within a local network to which both the first electronic device and the second electronic device are connected, and wherein the server media control request and the local media control request are both configured to initiate a same media control operation by the second electronic device.
  • the server system sends at least one of the server media control request or a media stream corresponding to the server media control request to the second electronic device.
  • FIG. 1 is a block diagram illustrating a server client environment in accordance with some implementations.
  • FIG. 2 is a block diagram illustrating a client-server environment in accordance with some implementations.
  • FIG. 3 is a block diagram illustrating a server system in accordance with some implementations.
  • FIG. 4 is a block diagram illustrating an electronic device in accordance with some implementations.
  • FIG. 5 is a flow diagram illustrating a method of controlling media presentation, performed by a first electronic device, in accordance with some implementations.
  • FIGS. 6A-6B are flow diagrams illustrating a method of controlling media presentation, performed by a second electronic device, in accordance with some implementations.
  • FIG. 7 is a flow diagram illustrating another method of controlling media presentation, performed by a second electronic device, in accordance with some implementations.
  • FIG. 8 is a flow diagram illustrating a method of controlling media presentation, performed by a server, in accordance with some implementations.
  • FIGS. 9-12 are flow diagrams illustrating methods of previewing media content, in accordance with some implementations.
  • FIG. 13 is a diagram illustrating an exemplary electronic device and user interface, according to some implementations.
  • FIG. 14 is a diagram illustrating another exemplary user interface, according to some implementations.
  • the disclosed systems and methods enable a user with a first electronic device to use the first electronic device to playback media content, and also to use the first electronic device to cause media content to be presented by a second electronic device.
  • a user may use a mobile phone to listen to music.
  • the user may also use the same mobile phone as a remote control to control the playback (and/or other functionalities) of a second device, such as a laptop computer, a desktop computer, a home media receiver/player, or a home audio/visual system.
  • the first device which is separate and distinct from the second device, can both present media and control other media devices.
  • the first electronic device sends media control requests to the second electronic device.
  • Media control requests are communications configured to cause some media control operation, such as initiating or controlling media content playback (e.g., play, stop, pause, change volume, skip, etc.).
  • media content playback e.g., play, stop, pause, change volume, skip, etc.
  • homes, businesses, and other locations provide local communications networks, such as a wireless or wired Ethernet network, a BLUETOOTH network, or the like, to which multiple devices are connected.
  • a person's mobile phone, home media system, computer, television, etc. may all be connected to the same local communications network, allowing these devices to communicate with each other and with other devices outside the local communications network (e.g., on the Internet) to send and receive the media control requests.
  • the first device sends media control requests to the second device through multiple channels, such as through a central server as well as via a local network. This may increase the speed and reliability of the media control requests, thus creating a more responsive and robust remote-control system for the user. Additionally, sending the media control requests via the different paths (e.g., through a local network as well as through the wider Internet) ensures that the remote control functionality operates properly under many different network connectivity situations.
  • media control requests are still sent from the handheld device, through the Internet (e.g., through a cellular data network), to the other media playing device.
  • the handheld device does not have access to the Internet, but the user's local wireless network is configured to allow direct communication between devices, the media control requests are still sent to the other media playing device despite the lack of Internet access.
  • media control requests are sent from a first electronic device to a server system located remotely from the first electronic device. The server system then causes a second electronic device to present media in accordance with the request from the first electronic device.
  • the first and second electronic devices are located on the same local network.
  • the local network is a Local Area Network (LAN).
  • a LAN is a type of local network that creates IP address space behind a traffic routing device (e.g., a wired or wireless router) and limits incoming connections based on the specific requirements for the Local Area Network.
  • LANs facilitate connections between devices on the LAN without using a public domain name system.
  • devices on a LAN communicate using TCP/IP protocols.
  • Local Area Networks are confined to a single building or a group of buildings in relatively small geographical proximity.
  • the local network is a Personal Area Network (PAN).
  • PAN Personal Area Network
  • a PAN is a network of devices that spans a relatively small geographical distance, such as a few inches or yards.
  • a PAN is established using a peer-to-peer and/or an ad-hoc network.
  • devices in a peer-to-peer and/or ad-hoc network communicate using BLUETOOTH communications technology, wireless universal serial bus, infrared communication protocols, etc.
  • Local networks may be wired or wireless.
  • the server system is located remotely from both the first and the second electronic device.
  • the server system is not located on the same local network (e.g., it is not in the same LAN, PAN, peer-to-peer network, ad-hoc network, etc.) as either the first electronic device or the second electronic device.
  • the first and second electronic devices can communicate over a local network, as described above, while, in contrast, communication between the server system and the first and second electronic device is over a network such as the Internet.
  • GUI graphical user interface
  • a media control request is configured to change the active presentation device for media content that is currently being presented. For example, a user who is listening to a music playlist on a mobile phone can request to change the active listening device to a home stereo system.
  • the request to change media presentation devices is initiated by a user (e.g., via a user interface on the mobile phone).
  • the request is initiated automatically and without user intervention by automatically detecting a predetermined condition, such as location, date, time of day, and/or day of week.
  • the user's home stereo system may begin playing back the audio instead of (or in addition to) the mobile phone once it is determined that she has arrived at her home (e.g., using a positioning system (e.g., a GPS, cell-tower triangulation, etc.), proximity sensors, inertial monitoring devices, etc.).
  • a positioning system e.g., a GPS, cell-tower triangulation, etc.
  • the switching occurs in response to the server system receiving notification from the first electronic device that a specific predetermined condition has been met.
  • the first electronic device may monitor its geographic location, and when it determines that it has moved into (or out of) a predefined area, the first electronic device will send a notification to the server system that the predetermined condition has been met.
  • the condition is met when the server detects a change in the state of the first electronic device, such as when the first electronic device establishes communication with a particular wireless network, or when the first electronic device pairs with or detects the second electronic device.
  • the server system ceases transmitting the media content stream to the first electronic device and begins transmitting the media content stream to a second electronic device. In this way, the user is able to establish conditions for automatically changing the device to which media content is being streamed or stream between devices.
  • Automatic switching may also be conditioned on additional factors, such as whether a device on which presentation is to be initiated is already presenting other media content, a time of day, a time of week, etc. For example, continuing the previous example, if the home stereo system is already presenting media content when the user returns home, the home stereo system will not switch to the playlist that was being presented on the user's mobile phone. This, in this case, the home stereo is available to be automatically activated only when it is idle and/or not presenting other media content.
  • the active presentation device is configured to automatically change only during predetermined time or date ranges. For example, in some implementations, a home stereo system will automatically begin presenting media content only during daytime hours (e.g., between 10:00 AM and 8:00 PM), and/or during weekends. Such rules may
  • different users have different priority levels for automatic device switching. This can help avoid frequent and annoying changes in media presentation due to different users returning home, entering a particular area, etc.
  • the user's priority level is logged or stored. When another user attempts to initiate presentation of different media content at that device, the other user may do so only if his or her priority level is above the first user.
  • priority levels are only used to determine whether automatic device switching may occur.
  • media presentation devices can present media content regardless of the origin of the content.
  • content in the audio playlist in the preceding example may be stored on the mobile phone, and streamed to the home stereo system from the mobile phone (e.g., via a local connection, such as a wireless LAN, peer-to-peer connection, a public network such as the Internet, etc.).
  • a local connection such as a wireless LAN, peer-to-peer connection, a public network such as the Internet, etc.
  • the source can simply switch the destination of the audio stream from the mobile phone to the home stereo system.
  • media content is delivered to the active presentation device regardless of the origin of the content.
  • a song in an active playlist is stored on the user's mobile phone (but presented by the user's home stereo system)
  • the song will be streamed or otherwise provided by the mobile phone
  • a song is stored on a user's network accessible storage (NAS) system
  • the song will be streamed or otherwise provided by the NAS
  • a song in the playlist is stored in a remote server, that song will be streamed or otherwise provided by the remote server.
  • Songs that are stored in different locations can be included in the same playlist, and can be streamed to whatever device is playing the playlist.
  • the context of the content stream when the active presentation device is changed, the context of the content stream, originally available on the first presentation device, becomes available on the second presentation device.
  • the context of a content stream includes various information about the content stream, including the position within the currently playing media content, the previously presented media content (e.g., a play history), the position within previously presented media content (if unfinished), the media content to be played next, metadata of the current or upcoming media content (e.g., artist, album, track), etc.
  • the media content once playback is initiated at the second electronic device, will retain the position in the current media content item and the position of the active playlist. In this way the media stream being presented on the first device is seamlessly transferred to a new device while maintaining the same state as on the first device.
  • users may easily switch between devices without fear of losing their place in a song, a movie, a playlist, etc.
  • one or more other devices when media content is being presented at one device, buffer/cache a portion of that media content, for example, in case the user should choose to present the media content with one of the other devices instead of or in addition to the current device.
  • the other devices are able to begin presenting the media content quickly, as the media presentation can begin at the other device(s) as soon as they receive a request to do so, rather than waiting for communications to be established between a media content source and the presentation device, for initial buffering of streaming content prior to presentation, and the like.
  • a user can select media content for preview. For example, a user can select, at one device, a song, playlist, or album to preview on the same device or on a different device.
  • any media content that is being presented when the preview is requested is muted, dimmed, or paused (or otherwise made less audible/visible) while the preview is presented. Accordingly, the user can preview the media content without the inconvenience of having to manually restart the earlier content playback and/or without losing their place in the earlier content playback.
  • Content for a preview can be pre-fetched by downloading and caching/buffering at least a portion of the content in response to a preview control (e.g., a button) being presented to a user, such as when a user views media content in a media browser on an electronic device.
  • a preview control e.g., a button
  • various user interfaces and techniques can be used to initiate a media content preview and/or to add media to an existing playlist. For example, touch-and-hold inputs on a touch screen interface can be used to initiate a preview, and touch-and-release inputs can be used to add content to a playlist.
  • the client-server environment 100 includes one or more client environments ( 108 - 1 . . . 108 - n ) and a server system 120 that are connected through a network 110 .
  • the client environment 108 - 1 includes one or more electronic devices (e.g., first electronic device 102 - 1 and second electronic device 106 - 1 ).
  • the server system 120 is associated with a media content provider to which users (and their electronic devices) may have accounts that enable the user to access media content from the server system 120 .
  • the network 110 includes any of a variety of networks, including wide area networks (WAN), wireless networks, wired networks, the Internet, or a combination of such networks.
  • WAN wide area networks
  • the client environment 108 - 1 includes a first electronic device 102 - 1 .
  • the first electronic device 102 - 1 is one of the group of: a personal computer, a mobile electronic device, a laptop, a tablet computer, a mobile phone, a digital media player, or any other electronic device able to present media content.
  • the client environment 108 - 1 also includes a second electronic device 106 - 1 .
  • the second electronic device 106 - 1 is one of the group of: a computer, a home audio/visual system, a home media receiver/player, or any other electronic device able to present media content.
  • both the first electronic device 102 - 1 and the second electronic device 106 - 1 are associated with a common user account (or associated user accounts) provided by a content provider with which the server system 120 is associated.
  • the server system 120 is operated and/or provided by a subscription-based media streaming service to which a user may have an account, and the first and second electronic devices 102 - 1 , 106 - 1 are each associated with account credentials that enable them to communicate with and receive content from the server system 120 .
  • both the first electronic device 102 - 1 and the second electronic device 106 - 1 are on the same local network.
  • the local network is a Local Area Network.
  • the server system 120 is not located on the same Local Area Network as either the first electronic device 102 - 1 or the second electronic device 106 - 1 .
  • the traffic routing device uses network address translation to connect devices within the LAN to devices outside the LAN.
  • Network address translation is the process of readdressing packets by modifying the IP address contained in each packet so that the packets reach their intended destination device.
  • Network address translation allows a plurality of electronic devices on the Local Area Network to connect to the Internet through a single IP address (e.g., the IP address of the traffic routing device).
  • the traffic routing device i.e. a router
  • a Local Area Network has 9 devices with local IP addresses 192.168.0.1-192.168.0.9 that all connect to a router.
  • the router connects to the Internet and has an IP address of 12.162.29.2.
  • the router translates the source address for all communications sent from any of the 9 devices and intended for destinations in the Internet to be 12.162.29.2 (the router's IP address).
  • the router collects all packets incoming from the Internet, determines the intended recipient device based upon the contents of each packet, and translates the destination IP address to the address of the correct device on the Local Area Network. So when the intended device is the device with Local Area Network IP address 196.168.0.2, the router would change the destination address to this address.
  • Local Area Networks also commonly use firewalls to limit incoming connections. In this way, computer devices outside of the Local Area Network are generally not able to communicate directly with the devices on a Local Area Network. Indeed, in some Local Area Networks the devices in the network are not contactable even by other devices in the Local Area Network.
  • both the first electronic device 102 - 1 and the second electronic device 106 - 1 are on the same Personal Area Network.
  • the Personal Area Network uses BLUETOOTH communication technology.
  • the server system 120 is not located on the same Personal Area Network as either the first electronic device 102 - 1 or the second electronic device 106 - 1 .
  • the first electronic device 102 - 1 includes a media content presentation and control application 104 (hereinafter “media application”).
  • the media application 104 is able to control the presentation of media by the first electronic device 102 - 1 .
  • the media application 104 enables a user to navigate media content items, select media content items for playback on the first electronic device 102 - 1 , create and edit playlists, etc.
  • media content is stored by the first electronic device 102 - 1 itself.
  • the media content is stored by a server system 120 , which may be located remotely from the first electronic device 102 - 1 . The media content is then streamed from the server system 120 to the first electronic device 102 - 1 over the network 110 .
  • the data streamed from the server system 120 is temporarily stored/cached by the first electronic device 102 - 1 in the media content buffer 105 in the memory of the first electronic device 102 - 1 .
  • media content stored in the media content buffer 105 is removed after the media content is presented by the first electronic device 102 - 1 , allowing new media content data to be stored in the buffer 110 .
  • at least some of the media content stored in the media content buffer 105 is retained for a predetermined amount of time after the content is presented by the first electronic device 102 - 1 and/or until other predetermined conditions are satisfied.
  • the media application 104 is also able to control media content presentation by the second electronic device 106 - 1 , which is distinct from the first electronic device 102 - 1 .
  • the user is able to use the media application 104 to cause the electronic device 102 - 1 to act both as a media presentation device as well as a remote control for other media presentation devices. This allows a user to control media presentation on multiple electronic devices from within a single application 104 , and/or using a single user interface.
  • a media control request (e.g., server media control request 112 , FIG. 2 ) to the server system 120 .
  • the server system 120 receives the media control request over the network 110 .
  • the user may press a button on a touchscreen of the first electronic device 102 - 1 in order to send the media control request to the server system 120 and/or directly to the second electronic device 106 - 1 .
  • a media control request is, for example, a request to begin presentation of media content by the second electronic device 106 - 1 .
  • media control requests also include requests and/or signals to control other aspects of the media that is being presented on the second electronic device 106 - 1 , including but not limited to commands to pause, skip, fast-forward, rewind, adjust volume, change the order of items in a playlist, add or remove items from a playlist, adjust audio equalizer settings, change or set user settings or preferences, provide information about the currently presented content, and the like.
  • the client-server environment 100 also includes a server system 120 .
  • the server system 120 includes a media streaming module 122 , a media content database 124 , and a context database 126 .
  • the media content database 124 stores media content that can be presented by an electronic device.
  • the media content database 124 stores audio (e.g., music, audiobooks, etc.), video (e.g., movies, television shows, etc.), images, or other content that can be streamed to other electronic devices.
  • the media content database includes data stored in different formats and file types to allow a variety of different devices and/or applications to receive streamed content.
  • the data is stored in a single file format and is converted/transcribed to the appropriate data type before or as it is streamed to a remote device.
  • the server system 120 includes a media streaming module 122 .
  • the media streaming module 122 receives media control requests from electronic devices and streams media content in response.
  • the media streaming module 122 receives media control requests from a first electronic device 102 - 1 and forwards the request to a second electronic device 106 - 1 , which then makes the final request to the server system 120 for the media content.
  • a user sends a media control request to the server using a mobile phone (a first electronic device) requesting that media be presented by a home stereo system (a second electronic device).
  • the server system 120 then sends the requested media (and/or the media control request) to the home stereo system. This and other techniques are discussed in greater detail below with respect to FIGS. 5-8 .
  • the received media control request includes information identifying the electronic device to which the server system 120 should forward the media control request.
  • a user may have multiple electronic devices that can present media from the server system 120 , such as a mobile phone, a computer system, a television, a home stereo, etc.
  • the identifying information is a unique or semi-unique device identifier, such as an IP address, a Media Access Control address (MAC address), a user-specified device name, an International Mobile Equipment Identity number (IMEI number), or the like. Accordingly, the media control request will identify that a request is intended for the home stereo, for example, so that the server system 120 can send the requested media and/or the media control request to the home stereo.
  • MAC address Media Access Control address
  • IMEI number International Mobile Equipment Identity number
  • the server system 120 includes a context database 126 .
  • the context database 126 stores data associated with the presentation of media content by an electronic device.
  • the context database 126 includes, among other things, the current position in a media content stream that is being actively presented by an electronic device, a playlist associated with the media content stream, previously played content, skipped pieces of media content, and previously indicated user preferences.
  • the context database may include information that a content stream to an electronic device currently is presenting a song, at 1 minute and 23 seconds into the song, as well as all the songs played in the last hour and the next 20 songs in the playlist.
  • the server system 120 transmits the context associated with a media content stream to the device that is presenting the content stream so that one or more items of context information can be used by the device, such as for display to the user. In some implementations, when the device to which the media content is being streamed changes, the server system 120 transmits the context associated with the active media content to the newly active device.
  • FIG. 2 is a block diagram illustrating a client-server environment 100 in accordance with some implementations.
  • the client-server environment 100 includes a local network 202 that encompasses at least a first electronic device 102 - 1 and a second electronic device 106 - 1 .
  • Client-server environment 100 also includes a server system 120 that is located remotely from and is not integrated into the local network.
  • the server system 120 stores media content and streams the stored media content to remote electronic devices upon request.
  • the local network 202 is a Local Area Network
  • the server system 120 is not within the IP address space of the Local Area Network.
  • the server system 120 communicates with either the first or the second electronic devices ( 102 - 1 , 106 - 1 ) via the Internet.
  • the media content streams 114 - 1 , 114 - 2 , and the server media control requests 112 may be transmitted via the Internet.
  • a user of the first electronic device 102 - 1 uses an application running on the first electronic device 102 - 1 to send a server media control request 112 to the server system 120 .
  • the server media control request 112 includes information identifying a second electronic device to which the control request is addressed, as well as a payload that indicates what media control operation is to be performed by the second electronic device.
  • media control requests include requests to control aspects of the media that is being presented on the second electronic device 106 - 1 , including but not limited to commands to initiate media presentation, cease media presentation, pause, skip, fast-forward, rewind, adjust volume, change the order of items in a playlist, add or remove items from a playlist, adjust audio equalizer settings, change or set user settings or preferences, provide information about the currently presented content (e.g., metadata), and the like.
  • the server system 120 in response to receiving the server media control request 112 , forwards the server media control request 112 to the second electronic device 106 - 1 .
  • the server media control request 112 causes the second electronic device to initiate the media control operation indicated by the request. In some cases, this will include communicating with the server system 120 to perform any of the operations described above that may be requested by a media control request. In some cases, the second electronic device 106 - 1 will not need to establish any subsequent communication in response to receiving the server media control request 112 , such as when the control request is for changing the volume, muting the speakers, changing equalizer settings, etc., as these operations can often be performed by the second electronic device 106 - 1 alone.
  • the server system 120 in response to receiving the server media control request 112 , the server system 120 will perform an operation requested by the server media control request 112 without forwarding the request to the second electronic device 106 - 1 .
  • the server system 120 may simply begin sending the media content to the second electronic device 106 - 1 (e.g., via the media content stream 114 ) without also forwarding the server media control request 112 to the second electronic device 106 - 1 .
  • the server system 120 executes some types of media control operations upon receiving the request from the first electronic device 102 - 1 (and without forwarding the request to the second electronic device 106 - 1 ), the server system 120 still forwards some types of media control requests to the second electronic device 106 - 1 .
  • a server media control request 112 that corresponds to a request to increase or decrease speaker volume, or any other request that can only be executed by the second electronic device 106 - 1 , are forwarded to the second electronic device 106 - 1 , even though requests to initiate presentation of media content may not be forwarded.
  • the first electronic device 102 - 1 also sends a local media control request 116 directly to the second electronic device 106 - 1 at substantially the same time that it sends the server media control request 112 to the server system 120 .
  • the local media control request 116 is configured to initiate the same media control operation as the server media control request 112 .
  • the local media control request 116 is sent via the local network 202 (e.g., a Local Area Network, a Personal Area Network, a peer-to-peer connection, etc.).
  • the local network 202 is a Local Area Network that establishes IP address space behind a traffic routing device
  • the local media control request 116 is sent to the second electronic device 106 - 1 without sending the local media control request 116 to devices that are not within the IP address space of the Local Area Network, and/or without sending the local media control request 116 through a public domain name system.
  • Sending a local media control request 116 to the second electronic device 106 - 1 over the Local Area Network allows for two separate, redundant pathways for delivering the media control request to the second electronic device. This can allow the second device to react more quickly and more reliably to the media control request of the first electronic device 102 - 1 . For example, if either connection path between the first and second electronic devices is not available, is improperly configured, or the like, the other connection path can ensure that remote-control functionality is still available.
  • this allows a service provider (e.g., a provider of software and/or devices that are configured to present media content) to provide a robust remote control solution that reduces the necessary technical knowledge of the end user, and is more likely to work in many different communication networks and environments (e.g., where a user's internal router is not configured for internal LAN communications, where an active Internet connection does not exist, etc.).
  • a service provider e.g., a provider of software and/or devices that are configured to present media content
  • the server media control request 112 and the local media control request 116 are identical. For example, they may contain the same message, such as “Device ⁇ UserX/Home Stereo ⁇ ; Operation ⁇ Play: UserX/Party Mix ⁇ .” Upon receiving this message, each device may respond by performing a predefined set of operations. For example, in some implementations, the server system 120 begins streaming media content associated with User X's “Party Mix” playlist to User X's Home Stereo system. Similarly, the second electronic device 106 - 1 (here, the home stereo system), upon receiving the same message in the local media control request 116 , sends a request to the server system 120 to initiate playback of the same media content.
  • some implementations provide systems and methods for handling duplicate or conflicting control requests, so that media control operations are not repeated, and to ensure that new and/or different media control requests are not ignored. Some examples of this are described below with reference to FIGS. 7-8 .
  • the server system 120 performs the media control request without forwarding it to the second electronic device 106 - 1 .
  • executing the requested operation includes streaming media content associated with User X's “Party Mix” playlist to User X's Home Stereo system.
  • the server system 120 forwards the message to the second electronic device 106 - 1 .
  • the forwarded message may be the identical message, or it may be modified or translated before it is forwarded to the second electronic device 106 - 1 . In implementations where the message is forwarded, it may be forwarded instead of or in addition to the server system 120 executing the requested operation.
  • the second electronic device 106 - 1 may then send a request to the server system 120 to initiate playback of the requested content.
  • the server system 120 may be configured to initiate a media control request (e.g., streaming media to the second electronic device 106 - 1 ) only if the request originates from the device at which the media is to be presented, and not if it originates from a different device. For example, in contrast to the example above where the server system 120 begins streaming media to the second electronic device 106 - 1 in response to a request from the first electronic device 102 - 1 , the server system 120 instead would only begin streaming media to the second electronic device 106 - 1 if the request came directly from the second electronic device itself.
  • a media control request e.g., streaming media to the second electronic device 106 - 1
  • the server 120 may be configured to forward all media control requests to the second electronic device 106 - 1 prior to executing them, as described above, so that the second electronic device 106 - 1 can determine whether, when, and how to execute the request.
  • the second electronic device 106 - 1 (and/or the server system 120 ) determines the authenticity, trustworthiness, and/or origin of a media control request before it executes the request or initiates the operation.
  • the server media control request 112 and the local media control request 116 are not identical (e.g., the message is in a different form, or contains more, less, or different information), but are both configured to initiate the same media control operation by the second electronic device.
  • the server media control request 112 includes a command to begin delivering content to the second electronic device 106 - 1
  • the local media control request 116 includes a command to cause the second electronic device 106 - 1 to issue a request to the server system 120 .
  • the request from the second electronic device 106 - 1 to the server system 120 in response to receiving one of the server media control request 112 and the local media control request 116 may be described as a third media control request (not shown).
  • the form and content of the third media control request may be the same as or different from the server media control request 112 and the local media control request 116 .
  • the third media control request will be configured to cause the same media control operation as those requests. Accordingly, though these requests are configured to cause different processes to be performed by different devices (e.g., causing the server to begin streaming media to the second electronic device, or cause the second electronic device to request that media from the server), they ultimately cause the same ultimate media control operation to be initiated (e.g., presentation of the media by the second electronic device).
  • a media control request is configured to switch the active media presentation device from one device (e.g., a user's mobile phone) to another (e.g., a user's home stereo system). In some implementations, this includes changing the device to which the server system 120 is streaming media content. For example, a user who is listening to a music playlist on a mobile phone can request that the music be played through a home stereo system instead of (or in addition to) the mobile phone.
  • the request to change media presentation devices is initiated by a user (e.g., via a user interface on the mobile phone), and in some implementations the request is initiated automatically (e.g., based on the location of the mobile phone, the proximity of the mobile phone to the home stereo system, the time of day, the day of the week, and/or user identities).
  • FIG. 2 also illustrates how the destination of a media content stream can be changed from one device to another in a client-server environment.
  • the first electronic device 102 - 1 is receiving a media content stream 114 - 1 from the server system 120 .
  • the first electronic device 102 - 1 may be a mobile phone receiving streaming music from the server system 120 .
  • a device within the client-server environment 100 requests that the media also, or instead, be presented by the second electronic device 106 - 1 . This request may come from the first and/or the second electronic device (or any other device within the environment), and may be initiated by a user, or it may be initiated automatically.
  • second electronic devices 106 - n buffer/cache a portion of the same media stream that is being presented by the first electronic device 102 - 1 , so that when the second electronic device 106 - n is selected to begin presentation of the media content, the second electronic device 106 - n begins presenting the media content seamlessly and without interruption. Buffering and/or caching of media content is discussed in greater detail below with reference to FIGS. 6A-6B .
  • the context of the media content stream 114 - 1 is available to the second electronic device 106 - 1 once it begins receiving the media content stream 114 - 2 .
  • information such as the current location within a playlist, recent play history, recent changes to a playlist, etc., are all available to the second electronic device 106 - 1 , and may be viewed, used, modified, or otherwise operated on by a user through a user interface associated with the second electronic device 106 - 1 .
  • the context information is stored in the context database 126 of the server system 120 .
  • the second electronic device 106 - 1 buffers/caches a portion of the content that is being streamed to the first electronic device 102 - 1 , in case the user should choose to begin media presentation by the second device instead of or in addition to the first device.
  • the second electronic device 106 - 1 caches/buffers a predetermined portion of the media content that is being streamed to the first electronic device 102 - 1 , such as the next 10 seconds, 20 seconds, 30 seconds, 1 minute, etc.
  • the second electronic device 106 - 1 receives information about the current media playback location from the first electronic device 102 - 1 and/or the server system 120 , such that when the second electronic device 106 - 1 receives a command to begin presenting the media content, it begins presentation at substantially the same position as the first electronic device 102 - 1 .
  • the server system 120 , the first electronic device 102 - 1 , and the second electronic device 106 - 1 are configured to present media content regardless of the particular device on which the media content is stored.
  • media content can be stored at the server system 120 (e.g., in the media content database 124 ), and streamed to the first and/or the second electronic devices.
  • media content may also be stored at the first and/or the second electronic devices (or other storage devices accessible to the first and/or second electronic devices, such as a NAS). This media content may be made accessible for presentation by any capable device associated with a user, even though the media content may not be available from the server system 120 .
  • a playlist on a first electronic device may include media content that is stored on that device's own memory, as well as media content that must be streamed from a different source, such as a remote server (e.g., server system 120 ) or storage device within the local network 202 (e.g., a NAS).
  • a remote server e.g., server system 120
  • storage device within the local network 202 (e.g., a NAS).
  • the second electronic device may not have the same media content stored in its memory as the first electronic device.
  • media content that is stored on one device in the local network 202 e.g., on first or second electronic devices 102 - n , 106 - n and/or on any device within a client environment 108 - n , such as a NAS
  • music stored locally on the first electronic device is streamed to the second electronic device via local communication paths (e.g., within a Local Area Network) when that music is not available from a remote source associated with a media content provider (e.g., the server system 120 ).
  • music stored on a first electronic device, but not available from the media content provider is streamed to the server associated with the media content provider, and then streamed from the server associated with the media content provider to the second electronic device.
  • media that is stored on one device may be accessible to another device seamlessly and without having to download, transfer, or otherwise duplicate the media content on multiple devices. Also, this process may be seamless to the user, such that the same controls and inputs are used both for causing a second device to stream media from a remote source, and for causing locally stored media to be streamed to the second device via local communication paths.
  • a user can listen to a portion of a media content item before deciding whether to purchase it or otherwise select it for presentation in its entirety. For example, users may wish to browse music choices in an electronic marketplace, and hear a snippet of a song that they come across in order to determine if they want to add it to a playlist or select it for immediate playback.
  • Previewing media content can be a nuisance, though. If a user is already listening to a song, for example, and requests a preview of another song, playback of the first song may simply be terminated. Once the preview is over, then, the user may have to take additional steps to resume playback of the first song.
  • the user may have to navigate to a different area of a user interface, locate the song that was previously playing, and either restart the song from the beginning, or use trial-and-error to locate the point in the song where it was stopped. Accordingly, implementations discussed herein provide convenient previewing techniques to address these and/or other issues.
  • a preview may be presented, such as podcasts, audiobooks, radio (e.g., broadcast radio, Internet radio, etc.), video (e.g., movies, music videos, television programming, short-form video, etc.), and the like. It will be understood to one of ordinary skill in the art that other types of media content can be substituted in various aspects of the implementations described below. For example, a user may preview a video during playback of a song, or may preview a broadcast radio or television stream during playback of a locally stored podcast.
  • radio e.g., broadcast radio, Internet radio, etc.
  • video e.g., movies, music videos, television programming, short-form video, etc.
  • a user may preview a video during playback of a song, or may preview a broadcast radio or television stream during playback of a locally stored podcast.
  • a media content preview interrupts currently played media content and does not restart it when the preview finishes.
  • a method is provided that allows a user to preview media content even when the user is consuming other media content, and restarts the other media content once the preview is complete.
  • a user may be listening to a song through a computer while also browsing other music on the computer, such as in an online music store or a media player.
  • the user finds a song, artist, or album that he would like to preview, he can select the content for preview using a preview button or another input technique.
  • the song that is currently playing can be paused, muted, or otherwise made less audible, while the content to be previewed is presented to the user.
  • a preview of media content should not interrupt presentation of currently playing media content by pausing, muting, or lowering the volume of the currently playing content. For example, if a user is presenting music via a home stereo system, he or she may want to preview other music to add to the current playlist without pausing or muting the music being played from the home stereo system. This situation may occur, for example, when the user is entertaining guests and wants to add music to a playlist without interrupting the current music.
  • the preview may be presented to the user via a different device than the active presentation device, or via a subset of a group of active presentation devices (e.g., one speaker of a group speakers), or using a different audio output of the active presentation device.
  • a subset of a group of active presentation devices e.g., one speaker of a group speakers
  • a different audio output of the active presentation device e.g., a playlist
  • media content e.g., a playlist
  • a media preview may be presented to a user at a computer or mobile phone.
  • a preview can be presented through another output means (e.g., a headphone jack or a built-in speaker).
  • a user manually selects a particular device for previewing media content and a particular device for presenting media content (e.g., regular playback).
  • these and/or other features are automatically enabled when a user selects a “DJ” mode on a media content browsing device.
  • the “DJ” mode causes media content from a playlist to be routed to one preselected device (or device output), and media content for presentation to be routed to a different preselected device (or device output).
  • a user can select the affordance by clicking on the affordance with a cursor (e.g., via a mouse or touchpad interface), by touching on the area of a touch screen associated with the displayed affordance, hovering a cursor over the affordance, or the like.
  • a cursor e.g., via a mouse or touchpad interface
  • Different inputs or interactions with a single affordance can cause different actions.
  • a preview of media content associated with that affordance is presented. If the user selects the affordance with a different type of input or interaction, however, other actions can be taken, such as adding the media content associated with the affordance to a current playlist, or terminating any currently played media content and presenting the selected media content instead.
  • an input representing a user selection of media content is detected (e.g., a touch contact or mouse click-down on a description of some media content), and different actions are taken depending on whether a change to the input is detected before or after expiration of a first time period. For example, if the change to the input is detected before expiration of the time period, the media content may be added to a playlist (e.g., as a next song, at the end of the playlist, or replacing a currently played song). If, on the other hand, no change to the input is detected before the expiration of the time period, the media content may be previewed. As described in several examples below, the input and the change to the input may be many different inputs and/or interactions.
  • the input may be a contact with a touch-sensitive surface, and the change to the input may be a liftoff from the touch-sensitive surface.
  • the input may be a contact with a first area of a touch-sensitive surface, and the change to the first input may be a contact with a second area of the touch-sensitive surface.
  • the contact with the first area may correspond to a user placing one finger on the touch-sensitive surface
  • the contact with the second area may correspond to the user placing a second finger at a different point on the touch sensitive surface.
  • an input can correspond to a “hover” input, where a cursor (e.g., controlled by a mouse, trackpad, etc.) is placed within a certain predefined area on a display.
  • the predefined area may correspond to an icon, a graphic, text, or the like.
  • a change to this input includes, for example, a mouse click or removal of the cursor from the predefined area.
  • Other inputs and touch gestures may also act as the first input and/or the change to the first input, such as tap, pinch-together, pinch-apart, touch-and-hold, touch-and-release, click-and-hold, click-and-release, click-release, double-click, double-tap, sliding gestures, multi-touch gestures, drag-and-drop, slide-and-release, right mouse click, selection of menu items (e.g., stop, pause, play now, play next, add to queue, increase/decrease volume, etc.), and the like.
  • menu items e.g., stop, pause, play now, play next, add to queue, increase/decrease volume, etc.
  • a “touch-and-hold” or a “click-and-hold” input (also referred to as a maintained input) initiates a preview of the selected media content, while a “touch-and-release” or “click-and-release” input (also referred to as a momentary input) adds the selected media content to a current playlist.
  • a “touch-and-hold” or a “click-and-hold” input (also referred to as a maintained input) initiates a preview of the selected media content
  • a “touch-and-release” or “click-and-release” input also referred to as a momentary input
  • an initial contact on a touch screen corresponds to the input
  • a liftoff from the touch screen (or a mouse click-release) corresponds to the change to the input.
  • the selected media content is previewed (e.g., played back) for as long as the input is maintained. For example, in a touch-screen environment, after it is determined that a contact with the touch screen corresponds to a maintained input (i.e., no liftoff has been detected for a certain time period), the media content is previewed until the contact is removed (i.e., when the liftoff is finally detected). In some implementations, the selected media content is previewed for a shorter or longer time than the duration of the maintained input. For example, in some implementations, the maintained input begins a preview clip of a predetermined length (e.g., 5, 10, 15, 20 seconds, etc.).
  • a predetermined length e.g., 5, 10, 15, 20 seconds, etc.
  • a timer is started when a contact or click-down input is detected. If a liftoff or click-release event is detected before the timer expires, the input corresponds to a momentary input. If a liftoff or click-release event is not detected before the time expires, the input corresponds to a maintained input. In this case, the expiration of the timer (possibly in conjunction with additional signals or logic) initiates the media content preview. In some implementations, a liftoff or click-release event that occurs after the time expires causes the preview to end.
  • a change to the input such as moving a cursor with a mouse, or making a “swipe” gesture on a touch screen—changes a playback attribute of the preview (or causes other appropriate actions).
  • the change to the input may include various interactions and gestures, such as pinch-together, pinch-apart, touch-and-hold, touch-and-release, click-and-hold, click-and-release, click-release, double-click, sliding gestures, multi-touch gestures, etc.
  • the change to the input causes the preview to skip forward or backward within a song, a playlist (e.g., including an album, a user generated playlist, a series of podcasts, etc), or any other media content or collection of media content items.
  • a playlist e.g., including an album, a user generated playlist, a series of podcasts, etc
  • the change to the input includes a directional component, such as a “swipe” gesture on a touch screen, and actions can be taken based on the particular direction of the input. For example, in some implementations, a user first initiates a preview by touching and holding an area of the touch screen with a finger. The user can then slide his finger forward to skip to a next track in the playlist, or slide his finger backwards to skip to a previous track in the playlist. In some implementations, the actions taken (e.g., skipping tracks, fast-forwarding within a track) are triggered if the input has a sufficient directional component in a certain direction.
  • Various techniques can be used to determine the directional components of a swipe input.
  • a swipe input path with a length, direction, and/or angle (e.g., as measured from the point of initial contact) that falls within a particular range may trigger the action.
  • a swipe input in a right-hand direction (e.g., substantially horizontal and from left to right) causes a first action
  • a swipe input in a left-hand direction (e.g., substantially horizontal and from right to left) causes a second action.
  • a swipe input in a down direction e.g., substantially vertical and from top to bottom
  • a swipe input in an up direction (e.g., substantially vertical and from bottom to top) causes a second action.
  • the first and second actions are a skip forward action and a skip back action, respectively.
  • a first contact corresponding to maintained input may cause a preview to begin, and a second contact may cause the preview to skip forward to a next or previous track, or fast-forward or rewind the preview.
  • the second contact may be a maintained input or a momentary input.
  • a second contact corresponding to a maintained input will cause the preview to fast-forward for as long as the maintained input is held.
  • Other directional inputs or gestures that may be used in the above implementations include, but are not limited to, diagonal swipes, loops, curves, multi-touch inputs, tap sequences, and other complex touch paths.
  • FIG. 13 illustrates an exemplary device 1300 showing a part of a user interface, in accordance with some implementations.
  • the device 1300 includes a display area 1302 .
  • the device is a mobile phone, mobile audio player, or the like.
  • a media preview area 1304 is displayed on the display 1302 , and can include any appropriate information.
  • the media preview area 1304 includes an artist name, album title, album art, and a preview button 1306 .
  • the preview button 1306 is configured, in some implementations, so that a tap input (e.g., a momentary input) causes the album and/or individual songs on the album to be presented by adding it to a playlist.
  • a tap input e.g., a momentary input
  • the preview button 1306 is pressed and held (e.g., a maintained input)
  • one or more tracks from the album are previewed as long as the input is maintained.
  • media content can be previewed in response to a cursor being “hovered” over a particular area of a display screen.
  • a media browsing environment presented in a web page or media player application may include multiple graphical items (also referred to as “tiles”) corresponding to media content, such as albums, playlists, songs, audiobooks, movies, videos, and the like.
  • the tiles include depictions of album art, media content metadata (e.g., album title, track listings, authors, actor names, artist names, artist photos, etc.), and the like.
  • the tile may also include a “preview” area that, when selected, causes a preview of the media content to be presented.
  • the preview area corresponds to an icon, graphic, text, picture, or other discernible area within or associated with the tile.
  • a preview of the media content associated with the tile is initiated.
  • the preview is initiated once the user has hovered over the area for a length of time, such as 0.5 seconds, 1 second, 2 seconds, or any other appropriate duration.
  • the preview continues until the user removes the cursor from the area (e.g., until the cursor is no longer within the area defined by the icon).
  • the second area corresponds to a tile, as described above.
  • the user can remove the cursor from the preview area (e.g., a preview icon) and the preview will continue until the cursor is removed from the larger tile area.
  • the second area e.g., the tile
  • the first area e.g., the preview icon.
  • a preview icon may be displayed at least partially within a tile, graphically connected to the tile (e.g., having at least a single common border), or fully within (i.e., completely surrounded by) a tile.
  • additional control elements are displayed to the user.
  • the playback controls may include controls (e.g., icons, graphics, etc.) to skip forward, skip backward, fast forward, rewind, pause, stop, add media content to a playlist, jump to specific track (e.g., track numbers), play now, play next, etc.
  • the controls may be selected in any appropriate way, such as by hovering over the control (e.g., moving the cursor from the preview icon to the control icon), or clicking on the control (e.g., left- or right-clicking).
  • the controls allow the user to skip between the individual content items during the preview.
  • the controls display numbers associated with individual tracks of an album, and selecting a number (e.g., by hovering over the number or clicking on the number) initiates a preview of that particular album track.
  • FIG. 14 illustrates an exemplary user interface 1400 , in accordance with some implementations.
  • the user interface 1400 is divided into three sections with dotted lines. The segments may be understood as depicting a portion of the user interface at different times during a user interaction.
  • the user interface 1400 includes tiles, such as tile 1402 in the left-most segment of the user interface 1400 .
  • the tiles include information about media content, including artist name, album title, and album art.
  • the tiles include a preview icon, such as preview icon 1404 in the left-most segment of the user interface 1400 .
  • preview icons may define or be associated with areas that, when a cursor is detected therein, cause a preview of the media content associated with the tile to begin.
  • the middle segment of the user interface 1400 illustrates a tile 1411 with a preview icon 1412 before any preview associated with this tile has been activated.
  • the outline of the preview icon 1412 is shown in dashed lines to illustrate the area of the display that is “active” such that if the cursor 1408 is detected within that area, a preview associated with that tile will be presented.
  • the dashed line border of the preview icon 1412 is provided here for illustrative purposes; borders or other graphical features or elements of a preview icon may or may not change during an interaction with a user.
  • the border of a preview icon and/or a media information tile flashes, moves, is made bold, is animated, changes color, changes brightness, or otherwise changes to grab the user's attention and, in some implementations, indicate that an interaction with the border will cause the device to take some action.
  • the right-most segment of the user interface 1400 illustrates a tile 1414 after a preview has been initiated in response to detecting the cursor 1408 within the area of the preview icon.
  • the outline of the tile 1414 is shown in dashed lines to illustrate the area of the display that is “active” such that if the cursor 1408 is detected outside that area, the preview that was initiated by detecting the cursor 1408 within the preview icon will cease.
  • the right-most segment includes media control icons 1410 that were displayed in response to detecting the cursor 1408 within the preview icon.
  • the media control icons 1410 correspond to track numbers of the album associated with the tile 1414 . Because the “active” region has been expanded to the border of the tile 1414 , the preview will continue even if the user moves the cursor 1408 around within the area of the tile 1414 . Thus, the user may move the cursor 1408 to the media control icons to select different tracks to preview, e.g., by clicking on or hovering over them.
  • Other types of controls may also be provided instead of or in addition to the track numbers shown in FIG. 14 .
  • the media control icons may include fast forward, rewind, skip, play, pause, add to playlist, play next, and the like.
  • previews of media content are initiated by a user selection of a control that is displayed to the user.
  • a control such as a button, icon, graphic, text, etc.
  • a web page e.g., a web page associated with a social networking service, such as FACEBOOK
  • controls may be displayed on any type of device, such as a computer (e.g., laptop, desktop, tablet), mobile phone, etc.
  • media content for a preview is pre-fetched when a control associated with a media content preview is displayed to a user. Accordingly, at least a portion of each media content that could be previewed at a given time—as determined by those controls that are displayed to the user and hence are selectable by the user—is downloaded and buffered/cached so that it can be presented to the user without undue delays if selected.
  • pre-fetching a portion of the album can include pre-fetching a part of each track or a subset of tracks of the media content.
  • pre-fetched media content associated with that control can be deleted or marked for deletion.
  • the various inputs, outputs, and actions described may be performed by one device or combinations of devices.
  • a user may browse media content and select media content for preview on a first electronic device 102 - 1 (e.g., a laptop computer), and use a second electronic device 106 - 1 (e.g., a home stereo system) to present the preview or other media content.
  • a user may browse and select media content on one device and also preview the media content on that same device (e.g., a first electronic device 102 - 1 ).
  • the media content can be downloaded and/or cached to the device that will present the media content (e.g., the second electronic device 106 - 1 ), even if the control that triggers the pre-fetching is displayed at a different device (e.g., the first electronic device 102 - 1 ).
  • the browsing device and the presentation device are in communication with each other so that the browsing device can cause the presentation device to pre-fetch media content associated with controls that are displayed on the browsing device at that time.
  • the mobile phone can send a request to the home audio system to download a portion of “American Pie.”
  • the mobile phone sends a request to a remote server (discussed below), and the server causes the home audio system to download the portion of the song.
  • the browsing device and the playback device are the same device.
  • the device detects that a control for initiating a preview is displayed, it can request (or receive without issuing a request) a portion of the media content to be presented if the user selects the control.
  • a remote server (e.g., server system 120 , FIG. 1 ) can determine when a control for initiating a preview of media content is displayed to a user, and send or otherwise enable access to the media content in response.
  • the remote server itself causes the control to be displayed to the user (e.g., by serving a web page or content for a web page to the user).
  • the server can also download a portion of the media content to a device in case the user should initiate presentation of the media content.
  • the device to which the portion of the media content is downloaded depends on several factors. For example, it may be downloaded to a device that is currently presenting other media content to the user. Or, if no device is currently presenting other media content to the user, it may be downloaded to the browsing device. Or it may be downloaded to a device that the user most frequently or most recently used to present media content.
  • the preview control itself is configured to initiate the pre-fetching when it is displayed.
  • a “preview” button in a web page may be configured to cause a download request (e.g., an HTTP request) to be sent to the remote server when the button is displayed in a web browser (or other application).
  • the download request includes a user or account identifier that associates the download request with a particular user or a particular set of devices.
  • the download request is sent to the same server that delivers media content to a user's presentation device (or a server that is associated with the service that delivers media content).
  • the server system (or, more generally, a content provider) can deliver the portion of the media content to the same device that that is currently presenting media content. Then, if the user selects the control to initiate presentation of the new media content (e.g., to preview it), the presentation device can simply begin presenting the pre-fetched portion of the content.
  • media content is provided by a remote server (e.g., server system 120 ).
  • the remote server may be associated with a content provider that provides access to media content such as music, movies, podcasts, audiobooks, etc.
  • the remote server communicates with browsing devices (e.g., mobile phones, computers, or any device on which a user may browse media content and select media content for presentation and/or preview) and with presentation devices (e.g., televisions, home audio systems, computers, mobile phones, or any device that can present media content).
  • browsing devices e.g., mobile phones, computers, or any device on which a user may browse media content and select media content for presentation and/or preview
  • presentation devices e.g., televisions, home audio systems, computers, mobile phones, or any device that can present media content.
  • Browsing devices, presentation devices, and remote servers communicate by any appropriate communication standards, protocols, or technologies, such as Internet Protocol (e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.), BLUETOOTH, Wired Ethernet (e.g., IEEE 802.3), Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • Internet Protocol e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.
  • BLUETOOTH e.g., Wired Ethernet (e.g., IEEE 802.3), Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n), or any other suitable communication protocol, including
  • a preview control can be displayed to a user when a currently presented song includes samples of other songs. When selected, the preview control initiates a preview of the song (or songs) that are sampled in the currently presented song.
  • the preview control may be presented to the user on any appropriate device, such as a first electronic device 102 - 1 , or a second electronic device 106 - 1 .
  • a user may be presented with a preview control (e.g., a button or icon, as described above).
  • a preview control e.g., a button or icon, as described above.
  • the original Led Zeppelin version of the song “Kashmir” is then presented to the user.
  • the various techniques described above for selecting a media content preview, seamlessly integrating the preview into currently presented media, and pre-fetching media content are used in this technique as well.
  • the songs that are sampled in a currently played song are identified using a service or a lookup table that identifies the songs that are sampled in other songs.
  • a server system associated with a media content provider e.g., server system 120
  • the lookup table may be stored and/or maintained by the content provider, or by a third party.
  • FIG. 3 is a block diagram illustrating a server system 120 , in accordance with some implementations.
  • the server system 120 typically includes one or more processing units (CPUs) 302 , one or more network interfaces 304 , memory 306 , and one or more communication buses 308 for interconnecting these components.
  • CPUs processing units
  • network interfaces 304 one or more network interfaces 304
  • memory 306 one or more communication buses 308 for interconnecting these components.
  • Memory 306 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM, or other random access solid state memory devices; and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Memory 306 may optionally include one or more storage devices remotely located from the CPU(s) 302 . Memory 306 , or alternately the non-volatile memory device(s) within memory 306 , includes a non-transitory computer readable storage medium. In some implementations, memory 306 or the computer readable storage medium of memory 306 stores the following programs, modules and data structures, or a subset thereof:
  • FIG. 4 is a block diagram illustrating a client system 101 , in accordance with some implementations.
  • the client system 101 represents first electronic devices 102 - n and/or second electronic devices 106 - n .
  • the client system 101 typically includes one or more processing units (CPUs) 402 , one or more network interfaces 410 , memory 412 , and one or more communication buses 414 for interconnecting these components.
  • the client system 101 includes a user interface 404 .
  • the user interface 404 includes user interface elements that enable output 406 to be presented to a user, including via speakers or a visual display.
  • the user interface 404 includes user interface components that facilitate user input 408 such as a keyboard, a mouse, a voice-command input unit, a touch sensitive display, or other input buttons 408 .
  • the client system 101 is a wireless device, such as a mobile phone.
  • some client systems 101 use a microphone and voice recognition to supplement or replace the keyboard.
  • Memory 412 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM, or other random access solid state memory devices; and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Memory 412 may optionally include one or more storage devices remotely located from the CPU(s) 402 . Memory 412 , or alternately the non-volatile memory device(s) within memory 412 , includes a non-transitory computer readable storage medium. In some implementations, memory 412 or the computer readable storage medium of memory 412 stores the following programs, modules and data structures, or a subset thereof:
  • FIGS. 5-8 are flow diagrams illustrating a process of controlling media presentation at an electronic device, in accordance with some implementations. Each of the operations shown in FIGS. 5-8 may correspond to instructions stored in a computer memory or computer readable storage medium. Optional operations are indicated by dashed lines (e.g., boxes with dashed-line borders).
  • the first electronic device corresponds to the first electronic device 102 - 1 , discussed above (also described with respect to FIG. 4 ).
  • a media control command for a second electronic device is received ( 502 ).
  • the media control command corresponds to a user input.
  • the user input is received via user interface 404 of the first electronic device 102 - 1 .
  • the user may select a song for playback by pressing a representation of the song (e.g., an icon, title, etc.) on a touchscreen of a mobile phone.
  • the first electronic device is selected from the group consisting of: a computer, a mobile phone, a remote control device, a portable media player, and the like.
  • the second electronic device is selected from the group consisting of: a television, a home media receiver/player, a computer, a home audio/visual system, a speaker, and the like.
  • any of the identified devices, and other devices not explicitly listed may be a first electronic device or a second electronic device.
  • a server media control request is sent to a server system ( 504 ).
  • the server system corresponds to the server system 120 , discussed above.
  • the server media control request is sent to the server system over the Internet ( 505 ).
  • the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected ( 506 ).
  • a local media control request is sent to a second electronic device within a local network to which both the first electronic device and the second electronic device are connected ( 508 ).
  • the local network is a Local Area Network, as discussed above.
  • the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork ( 510 ).
  • the local media control request is sent over any communication type that allows communication between the two electronic devices without having to leave the local network.
  • the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • the server and the local media control requests are both configured to cause a single media control operation to be implemented at the second electronic device. For example, if a user inputs a media control command into a mobile phone requesting that the song “Jessie's Girl” be presented by a second electronic device, such as a home media system, the mobile phone will send two media control requests that are each configured to cause the home media system to begin playback of “Jessie's Girl.” Sending both media control requests increases the speed with which the second electronic device responds to the request and the reliability of the request being received.
  • the second electronic device has already buffered some of the requested media content and can therefore begin presenting that content without needing to wait for the arrival of a media content stream from the server system, as discussed below.
  • the local and server media control requests need not contain identical data, but they are configured to cause the same media control operation to occur at the second electronic device.
  • the media control operation is selected from the group consisting of: play, pause, skip, fast-forward, rewind, adjust an audio volume, change an order of items in a playlist, add items to a playlist, remove items from a playlist, adjust audio equalizer settings, set a user setting, and the like.
  • the local and server media control requests include information specifying a particular electronic device from a group of available electronic devices to be controlled as the second electronic device.
  • the first electronic device provides a single user interface that allows a user both to select media content for presentation by the first electronic device and to generate media control requests configured to cause the media content to be presented by the second electronic device ( 512 ).
  • a user can control playback of media content at the device that is presenting the user interface (e.g., a mobile phone that is displaying a media player application) and also control playback of media content at a different device (e.g., a home media system or remote speaker) using the same application and/or user interface.
  • the user need not switch between applications or user interface environments in order to control the various devices, including the device on which the user interface is presented.
  • the media content is selected from the group consisting of, but not limited to: an audio track, a video, an image, an audio playlist, and a video playlist.
  • the video is presented on a display associated with the second electronic device ( 514 ).
  • the audio track is presented on a speaker associated with the second electronic device ( 516 ).
  • FIGS. 6A-6B are flow diagrams illustrating a method 600 of controlling media presentation at an electronic device, in accordance with some implementations.
  • the method 600 is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors (e.g., second electronic device 106 - n , FIG. 1 ).
  • the second electronic device may be a home media system, television, computer, or the like.
  • the method 600 corresponds to a method implemented in conjunction with and/or in response to the method 500 .
  • the method 600 corresponds to an implementation where media control requests that are sent to the server system from the first electronic device (e.g., the server media control request 112 , FIG. 2 ) are forwarded to the second electronic device without the server system attempting to perform the media control operation.
  • the server system acts as a relay or forwarding service for a media control request from the first electronic device. This may be used in implementations where a media control request is only performed if the request originates from the device that is to perform the action (or that is to be impacted by the action). For example, a device acting as a remote control cannot alone cause the server to begin streaming media to a second electronic device, because the control request did not originate from the second electronic device.
  • the second electronic device must itself make the request to the server. Accordingly, the server forwards the media control request to the second electronic device, and the second electronic device can then perform an operation to satisfy the request (e.g., requesting media content from the server, changing a media presentation setting, etc.).
  • the server forwards the media control request to the second electronic device, and the second electronic device can then perform an operation to satisfy the request (e.g., requesting media content from the server, changing a media presentation setting, etc.).
  • a server media control request corresponding to a media control operation to be implemented at the second electronic device is received from a server system, wherein the server media control request was sent from the server in response to the server receiving the server media control request from a first electronic device ( 602 ).
  • the server media control request is sent from the server system to the second electronic device over the Internet ( 603 ).
  • the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected ( 604 ).
  • a local media control request sent within a local network to which both the first electronic device and the second electronic device are connected is received from the first electronic device, wherein the local media control request corresponds to the media control operation ( 606 ).
  • the local network is a Local Area Network, as discussed above.
  • the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork ( 608 ).
  • the local media control request is sent over any communication type that allows peer-to-peer communication between the two electronic devices without having to leave the local network.
  • the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • the second electronic device may receive the server media control request and the local media control request from different sources and via different communication paths, the requests may not arrive in a consistent or predictable order. That is, in some cases, the local media control request may arrive before the server media control request, while in other cases, the order is reversed. However, a subsequently received media control request may not be a duplicate of the prior request, and so it should not be ignored simply because it was received shortly after another request. Accordingly, in some implementations, the second electronic device will determine whether to perform or ignore the media control requests based on various factors, as described with respect to FIG. 6B .
  • Method 600 continues on FIG. 6B .
  • the media control operation is performed in response to receiving a server media control request ( 610 ).
  • After receiving a local media control request it is determined whether the server media control request and the local media control request correspond to the same media control operation ( 612 ). If the server media control request and the local media control request correspond to the same media control operation, the local media control request is ignored ( 614 ). In some implementations, if the server media control request and the local media control request correspond to different media control operations, the local media control request is performed ( 616 ).
  • steps similar to ( 610 )-( 616 ) are employed to determine whether to ignore or perform the server media control request.
  • the second electronic device is prevented from performing the same media control operation twice in succession, thus preventing possible pauses, jumps, or other interruptions in the media content presentation caused by the processing of both the local and the server media control requests.
  • whether the server media control request and the local media control request correspond to the same media control operation is determined using timestamps.
  • the server media control request includes a first timestamp and the local media control request includes a second timestamp.
  • the first and the second timestamp both correspond to a time at which the first electronic device issued the media control requests.
  • the first timestamp corresponds to a time at which the server system received a server media control request from the first electronic device
  • the second timestamp corresponds to a time at which the first electronic device issued the local media control request.
  • determining that the server media control request and the local media control request correspond to the same media control operation includes comparing the first and the second timestamps. If the timestamps are the same or substantially similar, it is determined that the local media control request and the server media control request correspond to the same media control operation, and the local media control request is ignored. In some implementations, timestamps are substantially similar if they are within 1 second. In some implementations, they are substantially similar if they are within 5 seconds. Other times may be used depending on the needs of any specific implementation. In some implementations, other time ranges are used to determine if timestamps are considered substantially similar. If the timestamps are not the same or substantially similar, it is determined that server media control request and the local media control request do not correspond to the same media control operation, and the local media control request is performed.
  • the second electronic device e.g., the device that is to begin presenting media content
  • the first electronic device e.g., the device that is currently presenting the media content
  • the second electronic device is configured to buffer the upcoming portion of media content that is being presented at the first electronic device when the first electronic device is connected to the same local network (e.g., a LAN, a BLUETOOTH connection, etc.), when the first electronic device is within a predetermined proximity to the second electronic device (e.g., within the same building, room, floor, etc., as determined by a GPS, for example), and the like.
  • one or more of the devices, including the first and second electronic devices and the server system are configured to determine when and whether to buffer/cache media content at the second electronic device. Buffering media content at a second electronic device is discussed further with reference to steps ( 617 )-( 622 ).
  • the media control operation corresponding to one or more media control requests is a request to begin presenting media content at the second electronic device while the media content is being presented by the first electronic device ( 617 ).
  • an upcoming portion of the media content being presented by the first electronic device is buffered at the second electronic device ( 618 ).
  • the second electronic device may continuously receive and store a portion of the media content that is being presented at the first electronic device.
  • the second electronic device receives the media content for buffering from the server system.
  • the media content may be sent from the server system via the Internet.
  • the second electronic device receives the media content for buffering from the first electronic device.
  • the media content may be sent from the first electronic device via the local network.
  • the buffered media content has a lower bit-rate than media content that is being presented and/or streamed for current presentation, thus helping to reduce bandwidth and network usage for the buffering technique.
  • the media content is audio
  • it is presented at a bit-rate of either about 160 kbps or 320 kbps, and it is buffered at about 96 kbps.
  • Other bit-rates may be used for the presentation bit-rate and the buffering bit-rate, depending on factors such as available network speeds and capabilities, distances between devices, device capabilities, user preferences, and the like.
  • one of the server media control request or the local media control request is received at the second electronic device ( 620 ), and, in response, presentation of the buffered portion of the media content is initiated at the second electronic device ( 622 ).
  • the presentation can begin without the additional time necessary to initialize a media content stream or otherwise download or receive the media content in a way suitable for continuous presentation.
  • the second electronic device buffers about 5 seconds of the media content that is being presented by the first electronic device. In some implementations, it buffers about 10 seconds of the media content. In some implementations, it buffers more or less time, depending on the needs of any specific implementation.
  • the media content continues to be presented by the first electronic device in substantial synchronization with the presentation by the second electronic device. In some implementations, the media content ceases to be presented by the first electronic device when presentation begins at the second electronic device (although it may still be cached/buffered at the first electronic device, in some implementations).
  • FIG. 7 is a flow diagram illustrating a method 700 of controlling media presentation at an electronic device in accordance with some implementations.
  • the method 700 is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors.
  • the second electronic device may be a home media system, television, computer, or the like.
  • the method 700 corresponds to a method implemented in conjunction with and/or in response to the method 500 .
  • the method 700 corresponds to an implementation where media control requests that are sent to the server system (e.g., server system 120 , FIG. 2 ) from the first electronic device (e.g., the server media control request 112 , FIG.
  • the server system e.g., server system 120 , FIG. 2
  • the first electronic device e.g., the server media control request 112 , FIG.
  • local media control requests (e.g., the local media control request 116 , FIG. 2 ) are also received by the second electronic device after having been sent from the first electronic device, and are processed by the second electronic device when received, if appropriate.
  • a local media control request corresponding to a media control operation is received from a first electronic device, wherein the local media control request is sent within a local network to which both the first electronic device and the second electronic device are connected ( 702 ).
  • the local network is a Local Area Network, as discussed above.
  • the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork ( 703 ).
  • the local media control request is sent using any communication type that allows communication between the two electronic devices without having to leave the local network.
  • the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • a media stream is received from a server system, wherein the media stream was sent from the server in response to the server receiving a server media control request from the first electronic device, and wherein the server media control request corresponds to the media control operation ( 704 ).
  • the first electronic device requests a media control operation (e.g., initiated by a user or automatically)
  • the second electronic device receives a local media control request, as well as a media stream that corresponds to that same media control request.
  • the media stream is sent from the server system to the second electronic device over the Internet ( 705 ).
  • the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected ( 706 ).
  • the second electronic device Because the second electronic device receives both a media stream and a local media control request that may correspond to the same media control operation, the second electronic device should determine whether to perform a subsequently received local media control request. Accordingly, in some implementations, when the media stream is received prior to receiving the local media control request, the second electronic device determines whether to perform or ignore the local media control request. In some implementations, in response to a determination that the local media control request corresponds to a request to initiate presentation of the media stream, the local media control request is ignored ( 708 ). In some implementations, in response to a determination that the local media control request does not correspond to the request to initiate presentation of the media stream, the local media control request is performed ( 710 ). Thus, the second electronic device is prevented from performing the same media control operation twice in succession, thus preventing possible pauses, jumps, or other interruptions in the media content presentation.
  • whether the local media control request corresponds to a request to initiate presentation of the media stream is determined using timestamps.
  • the media stream includes a first timestamp
  • the local media control request includes a second timestamp.
  • the first and the second timestamp both correspond to a time at which the first electronic device issued the media control requests.
  • the first timestamp corresponds to a time at which the server system received a media control request from the first electronic device
  • the second timestamp corresponds to a time at which the first electronic device issued the local media control request. Accordingly, determining that the local media control request corresponds to a request to initiate presentation of the media stream includes comparing the first and the second timestamps.
  • timestamps are the same or substantially similar, it is determined that the local media control request corresponds to the media stream, and the local media control request is ignored. In some implementations, timestamps are substantially similar if they are within 1 second. In some implementations, they are substantially similar if they are within 5 seconds. In some implementations, other time ranges are used to determine if timestamps are considered substantially similar. If the timestamps are not the same or substantially similar, it is determined that the local media control request corresponds to the media stream, and the local media control request is performed.
  • FIG. 8 is a flow diagram illustrating a method 800 of controlling media presentation, in accordance with some implementations.
  • the method 800 is performed at a server system having one or more processors and memory storing one or more programs for execution by the one or more processors (e.g., server system 120 , FIGS. 1-3 ).
  • a server media control request is received from a first electronic device ( 802 ).
  • the first electronic device also sends a local media control request to the second electronic device within a local network to which both the first electronic device and the second electronic device are connected.
  • the server media control request and the local media control request are both configured to initiate a same media control operation by the second electronic device.
  • the server media control request is sent to the server system over the Internet ( 803 ).
  • the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected ( 804 ).
  • the local network is a Local Area Network.
  • the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork ( 805 ).
  • At least one of the server media control request or a media stream corresponding to the server media control request is sent to the second electronic device ( 806 ).
  • the server forwards all the server media control requests to the second electronic device.
  • the server performs the media control operation requested by the server media control request. For example, if the server media control request requests playback of media content at the second electronic device (e.g., a home media system), the server streams the media content to the second electronic device.
  • the server performs some media control requests, but forwards others to the second electronic device.
  • requests to control playback of media content are performed by the server, while requests that relate specifically to the second electronic device (e.g., change the volume, change an input source, etc.) are forwarded to the second electronic device.
  • requests to control playback of media content including commands such as play, pause, skip, repeat, etc.
  • requests that relate specifically to the second electronic device e.g., change the volume, change an input source, etc.
  • FIGS. 9-12 are flow diagrams illustrating a method for previewing media content, in accordance with some implementations.
  • Each of the operations shown in FIGS. 9-12 may correspond to instructions stored in a computer memory or computer readable storage medium. Optional operations are indicated by dashed lines (e.g., boxes with dashed-line borders).
  • each of the operations shown in FIGS. 9-12 may be executed by any appropriate device or combination of devices, including first electronic devices 102 - n , second electronic devices 106 - n , and/or a server system 120 .
  • the method 900 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the electronic device corresponds to the first electronic device 102 - 1 , discussed above.
  • a first input is detected ( 902 ).
  • the first input is a touch input including a contact with a touch-sensitive surface.
  • the first input is a mouse click-down event.
  • first media content is added to a playlist ( 904 ).
  • the first input is detected ( 902 ) during presentation of second media content, and the playlist includes the second media content.
  • the change to the first input causes the first media content to be added to a playlist that is currently being presented.
  • adding the first media content to the playlist comprises adding the first media content to an end of the playlist.
  • adding the first media content to the playlist comprises adding the first media content as next media content in the playlist.
  • adding the first media content to the playlist comprises replacing existing content in the playlist with the first media content.
  • the change to the first input is a liftoff from a touch-sensitive surface, such as a touchscreen.
  • the change to the first input is an additional contact with the touch-sensitive surface, wherein the contact and the additional contact correspond to distinct contact areas of the touch-sensitive surface.
  • the change to the first input is a mouse click-release event.
  • the change to the first input is a gesture, swipe, cursor movement, or multi-touch input.
  • the first time period is 0.5 seconds. In some implementations, the first time period is 1 second. In some implementations, the first time period is 2 seconds.
  • the first media content is caused to be presented to the user ( 906 ).
  • causing the first media content to be presented to the user includes presenting the media content on the same device that received the first input. In some implementations, it includes sending a request to another device to initiate presentation of the media content.
  • causing the first media content to be presented includes reducing a volume of the second media content while the first media content is presented ( 908 ). In some implementations, reducing the volume includes muting the second media content. In some implementations, reducing the volume includes pausing the second media content. In some implementations, causing the first media content to be presented also includes ceasing presentation of the first media content in response to detecting a change to the first input after the first time period.
  • the first media content comprises a plurality of media content items.
  • a directional component of the first input can be used to navigate among individual media content items of the first media content.
  • the first media content is a playlist, album, listing of podcasts, or the like.
  • the method further includes detecting a direction of the first input ( 910 ).
  • a next media content item of the plurality of media content items is caused to be presented in response to detecting that that the direction is a first direction ( 912 ).
  • a previous media content item of the plurality of media content items is caused to be presented in response to determining that the direction is a second direction ( 914 ).
  • the first direction corresponds to an input movement having a predetermined directional component in a right-hand direction.
  • the second direction corresponds to an input movement having a predetermined directional component in a left-hand direction.
  • the first direction corresponds to an input movement having a predetermined directional component in a down direction.
  • the second direction corresponds to an input movement having a predetermined directional component in an up direction.
  • the method 1000 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the electronic device corresponds to the first electronic device 102 - 1 , discussed above.
  • a first input is detected during presentation of a first song ( 1002 ). Various inputs that may be detected are described above.
  • the audibility of the first song is reduced ( 1004 ).
  • reducing the audibility of the first song includes reducing the volume of the first song; in some embodiments it includes muting the first song; in some implementations, it includes pausing the first song.
  • reducing the audibility of the first song includes sending a command to a remote device, where the command is configured to cause that device to reduce the audibility of the first song.
  • a second song is caused to be presented ( 1006 ).
  • causing the second song to be presented includes beginning playback of the second song at the device.
  • causing the second song to be presented includes sending a request to a remote device to initiate playback of the second song.
  • the method also includes, after a first time period, causing the presentation of the second song to be ceased ( 1008 ), and causing the audibility of the first song to be increased ( 1010 ).
  • the first time period corresponds to the duration of a maintained input.
  • the end of the first time period corresponds to an expiration of a timer.
  • the end of the first time period corresponds to an end of a preview portion of the second song.
  • the end of the first time period corresponds to detecting a change to the first input, such as a liftoff from a touch-sensitive surface or a mouse click-release event.
  • the method 1100 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the electronic device corresponds to the first electronic device 102 - 1 , discussed above.
  • an affordance is displayed to a user, wherein selection of the affordance will cause secondary media content to be presented ( 1102 ).
  • the affordance corresponds to an icon (e.g., a thumbnail image of an album cover or movie poster), graphic, or text.
  • at least a portion of secondary media content is obtained ( 1104 ).
  • the portion of the secondary media content is received by the same device that displayed the affordance. In some implementations, it is received by a different device, such as a second electronic device 106 - n . In some implementations, it is sent by a server system remote from the device that displayed the affordance, such as server system 120 .
  • the secondary media content includes multiple media content items, and obtaining at least a portion of the secondary media content includes obtaining at least a portion of each media content item of the secondary media content. For example, if the secondary content corresponds to an album, at least a part of each track is obtained for caching/buffering by a presentation device. In some implementations, obtaining at least a portion of the secondary media content includes obtaining an initial media content item and the next media content item in the playlist. In some implementations, obtaining at least a portion of the secondary media content includes obtaining an initial media content item and the next two media content items in the playlist.
  • a first input corresponding to a selection of the affordance is detected ( 1106 ).
  • the first input corresponds to any of those inputs discussed above (e.g., mouse clicks or hovers, touch events and gestures, etc.).
  • the audibility of the currently presented media content is reduced ( 1108 ). In some implementations, this includes muting, lowering the volume or brightness of, and/or pausing the currently presented media content.
  • the secondary media content is presented ( 1110 ). After a first time period, the presentation of the secondary media content is ceased ( 1112 ), and the audibility of the currently presented media content is increased ( 1114 ). In some implementations, the first time period corresponds to the duration of a maintained input.
  • the end of the first time period corresponds to an expiration of a timer. In some implementations, the end of the first time period corresponds to an end of a preview portion of the second song. In some implementations, the end of the first time period corresponds to detecting a change to the first input, such as a liftoff from a touch-sensitive surface or a mouse click-release event.
  • the method 1200 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors.
  • the electronic device corresponds to the first electronic device 102 - 1 , discussed above.
  • the presence of a cursor is detected in a first area of a display ( 1202 ).
  • the first area of the display corresponds to an icon (e.g., a thumbnail image of an album cover or movie poster), graphic, or text.
  • the first area is displayed partially or entirely within, adjacent to, and/or contiguous with a second area of the display region.
  • the second area of the display region corresponds to a media content description area (e.g., a tile, as discussed above), including, for example, album art, movie stills, content metadata, artist information, artist photographs, and the like.
  • Media content is presented in response to detecting the presence of the cursor within the first area ( 1204 ).
  • the media content corresponds to a preview of the media content identified and/or described in the media content description area.
  • the media content is presented after the cursor has been detected in the first area for a time period, such as 0.5 seconds, 1 second, 2 seconds, or any other appropriate time period.
  • the removal of the cursor from a second area of the display region is detected ( 1206 ), where the second area is larger than and contiguous with the first area.
  • the second area shares at least one border with the first area.
  • the second area completely encompasses the first area on all sides.
  • Presentation of the media content is ceased in response to detecting the removal of the cursor from the second area ( 1208 ).
  • the media content is presented—in response to detecting the cursor in the first area—it will continue to be presented until the cursor is removed from the larger second area.
  • one or more additional icons are displayed within the second area of the display ( 1210 ).
  • the one or more additional icons include media playback controls.
  • the media playback controls are selected from the group consisting of: play, pause, stop, skip forward, skip backward, add to playlist (e.g., add as next, add to end, replace playlist), and audio track indices.
  • the methods illustrated in FIGS. 5-12 may be governed by instructions that are stored in a computer readable storage medium and that are executed by at least one processor of at least one server. Each of the operations shown in FIGS. 5-12 may correspond to instructions stored in a non-transitory computer memory or computer readable storage medium.
  • the non-transitory computer readable storage medium includes a magnetic or optical disk storage device, solid state storage devices, such as Flash memory, or other non-volatile memory device or devices.
  • the computer readable instructions stored on the non-transitory computer readable storage medium may be in source code, assembly language code, object code, or other instruction format that is interpreted and/or executable by one or more processors.
  • first first
  • second second
  • first contact first contact
  • first contact second contact
  • first contact second contact
  • the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting,” that a stated condition precedent is true, depending on the context.
  • the phrase “if it is determined (that a stated condition precedent is true)” or “if (a stated condition precedent is true)” or “when (a stated condition precedent is true)” may be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.

Abstract

An electronic device is disclosed. The electronic device includes a display, one or more processors, memory, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors. The one or more programs include instructions for detecting a first input; and determining whether a change to the first input is detected within a first time period after detecting the first input. In accordance with a determination that a change to the first input is detected within the first time period after detecting the first input, first media content is added to a playlist. In accordance with a determination that a change to the first input is not detected within the first time period after detecting the first input, the first media content is caused to be presented.

Description

    RELATED APPLICATION
  • This application is a continuation of U.S. patent application Ser. No. 14/053,443, filed Oct. 14, 2013, which claims priority and benefit to U.S. Provisional Application No. 61/713,444, filed Oct. 12, 2012 entitled “SYSTEMS AND METHODS FOR MULTI-CONTEXT MEDIA CONTROL AND PLAYBACK.” All above-identified patent applications are hereby incorporated by reference in their entireties.
  • TECHNICAL FIELD
  • The disclosed implementations relate generally to controlling and playing back media content.
  • BACKGROUND
  • Modern users have a variety of options for devices and ways to consume media content. For example, multiple different types of media, including text, still images, video, audio, and other digital media, can be consumed on personal computers, laptops, mp3 players, mobile phones, tablet computers, televisions, stereos, and many other electronic devices. These electronic devices typically have the ability to store digital media files in memory that is integrated into the devices. But acquiring and storing all the media content to which a user desires access can be prohibitively difficult and expensive.
  • Current technology also allows users to stream digital media content to their electronic devices over local networks or the Internet, for example, including over wireless connections. Streaming services allow users to have access to a much larger collection of media than would otherwise be possible. Streaming services store media content on servers remote from the electronic devices, and then send the media content to the electronic devices when it is requested. The media content is then cached on the electronic device for presentation. In some cases, the content may be temporarily buffered/stored before presentation, but typically the data is deleted from memory after presentation, allowing an almost unlimited amount of media to be streamed to a device without running out of storage space on the electronic device. Some streaming services also allow streamed content to be stored on the local device.
  • As more and more devices are capable of streaming media content from remote servers, as well as storing media content locally, it is often cumbersome to manage all of the different sources of media to which a user has access. Moreover, even where a user has multiple devices that can access the same content (for example, a handheld device and a laptop computer that can both access a media streaming service), each device must typically be controlled with its own dedicated interface. Accordingly, it would be advantageous to provide systems and methods that allow a user to remotely control various electronic devices in a simple and intuitive manner.
  • Furthermore, as users are increasingly using computers and other electronic devices to browse and consume media content, it is beneficial to increase the ease and convenience with which media content can be browsed, previewed, and selected for presentation. For example, when browsing in a media store, previews of media content may be available so that the user can listen to a portion of a song or watch a portion of a video before choosing to purchase it or otherwise select it for presentation. However, previewing media content may be inconvenient when media content is already being presented, such as when a user is browsing an online music store while listening to other music. Accordingly, it would be advantageous to provide systems and methods that allow a user to preview media content in a convenient and streamlined manner.
  • SUMMARY
  • In accordance with some implementations, a method of previewing media content is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors. The method includes detecting a first input. In some implementations, the first input is a contact with a touch-sensitive surface. In some implementations, the first input is a mouse click event. In response to determining that a change to the first input has been detected within a first time period after detecting the first input, first media content is added to a playlist. In some implementations, the change to the first input is a liftoff from a touch-sensitive surface. In some implementations, the change to the first input is a mouse click-release event. In some implementations, the change to the first input is an addition of another input. In response to determining that a change to the first input has not been detected within the first time period after detecting the first input, the first media content is caused to be presented. In some implementations, the first media content is presented at the same device that received the input (e.g., a computer, television, or mobile phone). In some implementations, the first media content is presented at a different device (e.g., a home stereo system or wireless speaker system).
  • In accordance with some implementations, a method of previewing media content is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors. The method includes, during presentation of a first song, detecting a first input. In some implementations, the first input is a touch input, mouse input, or other appropriate input. The audibility of the first song is reduced. In some implementations, the audibility is reduced by lowering the volume of, muting, or pausing the first song. A second song is presented. After a first time period, presentation of the second song is ceased. In some implementations, the first time period corresponds to the duration of the first input. For example, the first input can correspond to a touch- or click-and-hold input, and the second song is presented as long as the input is maintained. The audibility of the first song is then increased. In some implementations, the first song is presented at the same or substantially the same audibility as it was prior to presenting the second song.
  • In accordance with some implementations, a method of previewing media content is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors. The method includes, during presentation of currently presented media content, displaying an affordance to a user, where selection of the affordance will cause secondary media content to be presented. For example, the affordance may be a button, icon, or text that, when selected, causes secondary media content to be previewed or otherwise presented. In response to displaying the affordance, at least a portion of secondary media content is obtained. In some implementations, the portion of the secondary media content is obtained at the same device on which the affordance is displayed. In some implementations, it is obtained at a different device. A first input corresponding to a selection of the affordance is detected. The audibility of the currently presented media content is reduced, and the secondary media content is presented. After a first time period, the presentation of the secondary media content is ceased, and the audibility of the currently presented media content is increased.
  • In accordance with some implementations, a method of previewing media content is disclosed. In some implementations, the method is performed at an electronic device having one or more processors and memory storing instructions for execution by the one or more processors. The method includes detecting the presence of a cursor in a first area of a display. In some implementations, the first area of the display corresponds to an icon, button, graphic, text, or other type of control, and can be displayed in a media content description area, including, for example, album art, movie stills, content metadata, and the like. Media content is presented in response to detecting the presence of the cursor within the first area. The removal of the cursor from a second area of the display region is detected, where the second area is larger than and contiguous with the first area. In some implementations, the second area corresponds to the media content description area. In some implementations, it is the same size as the media content description area. The media content is presented in response to detecting the removal of the cursor from the second area.
  • In accordance with some implementations, a method of controlling media presentation is disclosed. In some implementations, the method is performed at a first electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The first electronic device receives a media control command for a second electronic device. In response to receiving the media control command, the server sends a server media control request to a server system, and sends a local media control request to a second electronic device within a local network to which both the first electronic device and the second electronic device are connected. Local networks may be Local Area Networks (LANs), Personal Area Networks (PANs), ad-hoc computer networks, peer-to-peer computer networks, and the like, and may be wired or wireless. Local networks may use one or more communication standards, protocols, or technologies, such as Internet Protocol (e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.), BLUETOOTH, Wired Ethernet (e.g., IEEE 802.3), Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document. For example, in some implementations, two nearby devices in communication with one another via BLUETOOTH communication technologies constitutes a local network.
  • In some implementations, the server media control request is sent to the server system over the Internet. In some implementations, the first electronic device provides a single user interface that allows a user both to select media content for presentation by the first electronic device and to generate media control requests configured to cause the media content to be presented by the second electronic device.
  • Media content is selected from the group consisting of: an audio track, a video, an image, an audio playlist, and a video playlist. In some implementations, the first electronic device is selected from the group consisting of: a computer, a mobile phone, a remote control device, and a portable media player. In some implementations, the second electronic device is selected from the group consisting of: a television, a home media receiver/player, a computer, a home audio/visual system, and a speaker. In some implementations, the server and the local media control requests are both configured to cause a single media control operation to be implemented at the second electronic device. In some implementations, the media control operation is selected from the group consisting of: play, pause, skip, fast-forward, rewind, adjust an audio volume, change an order of items in a playlist, add items to a playlist, remove items from a playlist, adjust audio equalizer settings, and set a user setting.
  • In accordance with some implementations, another method of controlling media presentation is disclosed. In some implementations, the method is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The second electronic device receives, from a server system, a server media control request corresponding to a media control operation to be implemented at the second electronic device, wherein the server media control request was sent from the server in response to the server receiving the server media control request from a first electronic device. The second electronic device receives, from the first electronic device, a local media control request sent within a local network to which both the first electronic device and the second electronic device are connected, wherein the local media control request corresponds to the media control operation.
  • In some implementations, the server media control request is received prior to the local media control request. In response to receiving the server media control request, the second electronic device performs the media control operation. After receiving the local media control request, the second electronic devices determines whether the server media control request and the local media control request correspond to the same media control operation. If the server media control request and the local media control request correspond to the same media control operation, the server ignores the local media control request.
  • In some implementations, the media control operation is a request to begin presenting media content at the second electronic device, and the media content is being presented by the first electronic device. Prior to receiving the server media control request or the local media control request, the second electronic device buffers/caches an upcoming portion of the media content being presented by the first electronic device. The second electronic device receives one of the server media control request or the local media control request. The second electronic device initiates presentation of the buffered portion of the media content.
  • In accordance with some implementations, another method of controlling media presentation is disclosed. In some implementations, the method is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. The second electronic device receives, from a first electronic device, a local media control request corresponding to a media control operation, wherein the local media control request is sent within a local network to which both the first electronic device and the second electronic device are connected. The second electronic device receives a media stream from a server system, wherein the media stream was sent from the server in response to the server receiving a server media control request from the first electronic device, and wherein the server media control request corresponds to the media control operation.
  • In some implementations, the media stream is received prior to receiving the local media control request. In response to a determination that the local media control request corresponds to a request to initiate presentation of the media stream, the second electronic devices ignores the local media control request.
  • In accordance with some implementations, another method of controlling media presentation is disclosed. In some implementations, the method is performed at a server system having one or more processors and memory storing one or more programs for execution by the one or more processors. The server system receives, from a first electronic device, a server media control request, wherein the first electronic device also sends a local media control request to the second electronic device within a local network to which both the first electronic device and the second electronic device are connected, and wherein the server media control request and the local media control request are both configured to initiate a same media control operation by the second electronic device. The server system sends at least one of the server media control request or a media stream corresponding to the server media control request to the second electronic device.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The implementations disclosed herein are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings. Like reference numerals refer to corresponding parts throughout the drawings.
  • FIG. 1 is a block diagram illustrating a server client environment in accordance with some implementations.
  • FIG. 2 is a block diagram illustrating a client-server environment in accordance with some implementations.
  • FIG. 3 is a block diagram illustrating a server system in accordance with some implementations.
  • FIG. 4 is a block diagram illustrating an electronic device in accordance with some implementations.
  • FIG. 5 is a flow diagram illustrating a method of controlling media presentation, performed by a first electronic device, in accordance with some implementations.
  • FIGS. 6A-6B are flow diagrams illustrating a method of controlling media presentation, performed by a second electronic device, in accordance with some implementations.
  • FIG. 7 is a flow diagram illustrating another method of controlling media presentation, performed by a second electronic device, in accordance with some implementations.
  • FIG. 8 is a flow diagram illustrating a method of controlling media presentation, performed by a server, in accordance with some implementations.
  • FIGS. 9-12 are flow diagrams illustrating methods of previewing media content, in accordance with some implementations.
  • FIG. 13 is a diagram illustrating an exemplary electronic device and user interface, according to some implementations.
  • FIG. 14 is a diagram illustrating another exemplary user interface, according to some implementations.
  • DETAILED DESCRIPTION
  • The various implementations described herein enable a first electronic device to remotely control media presentation at a second electronic device. In some implementations, the disclosed systems and methods enable a user with a first electronic device to use the first electronic device to playback media content, and also to use the first electronic device to cause media content to be presented by a second electronic device. For example, a user may use a mobile phone to listen to music. The user may also use the same mobile phone as a remote control to control the playback (and/or other functionalities) of a second device, such as a laptop computer, a desktop computer, a home media receiver/player, or a home audio/visual system. Thus, the first device, which is separate and distinct from the second device, can both present media and control other media devices.
  • In order to cause media control operations to be performed at a second electronic device, the first electronic device sends media control requests to the second electronic device. Media control requests are communications configured to cause some media control operation, such as initiating or controlling media content playback (e.g., play, stop, pause, change volume, skip, etc.). Increasingly, homes, businesses, and other locations provide local communications networks, such as a wireless or wired Ethernet network, a BLUETOOTH network, or the like, to which multiple devices are connected. Accordingly, a person's mobile phone, home media system, computer, television, etc., may all be connected to the same local communications network, allowing these devices to communicate with each other and with other devices outside the local communications network (e.g., on the Internet) to send and receive the media control requests.
  • In some implementations, the first device sends media control requests to the second device through multiple channels, such as through a central server as well as via a local network. This may increase the speed and reliability of the media control requests, thus creating a more responsive and robust remote-control system for the user. Additionally, sending the media control requests via the different paths (e.g., through a local network as well as through the wider Internet) ensures that the remote control functionality operates properly under many different network connectivity situations. For example, if a user's local wireless network (e.g., a wireless LAN, PAN, etc.) is not configured to allow a handheld device to communicate directly to another media playing device, media control requests are still sent from the handheld device, through the Internet (e.g., through a cellular data network), to the other media playing device. If, on the other hand, the handheld device does not have access to the Internet, but the user's local wireless network is configured to allow direct communication between devices, the media control requests are still sent to the other media playing device despite the lack of Internet access. Accordingly, in some implementations described herein, media control requests are sent from a first electronic device to a server system located remotely from the first electronic device. The server system then causes a second electronic device to present media in accordance with the request from the first electronic device.
  • As noted above, in some implementations, the first and second electronic devices are located on the same local network. In some implementations, the local network is a Local Area Network (LAN). In some implementations, a LAN is a type of local network that creates IP address space behind a traffic routing device (e.g., a wired or wireless router) and limits incoming connections based on the specific requirements for the Local Area Network. In some implementations, LANs facilitate connections between devices on the LAN without using a public domain name system. In some implementations, devices on a LAN communicate using TCP/IP protocols. In some cases, Local Area Networks are confined to a single building or a group of buildings in relatively small geographical proximity. In some implementations, the local network is a Personal Area Network (PAN). In some implementations, a PAN is a network of devices that spans a relatively small geographical distance, such as a few inches or yards. In some implementations, a PAN is established using a peer-to-peer and/or an ad-hoc network. In some implementations, devices in a peer-to-peer and/or ad-hoc network communicate using BLUETOOTH communications technology, wireless universal serial bus, infrared communication protocols, etc. Local networks may be wired or wireless.
  • In some implementations, the server system is located remotely from both the first and the second electronic device. For example, in some implementations, the server system is not located on the same local network (e.g., it is not in the same LAN, PAN, peer-to-peer network, ad-hoc network, etc.) as either the first electronic device or the second electronic device. Thus, the first and second electronic devices can communicate over a local network, as described above, while, in contrast, communication between the server system and the first and second electronic device is over a network such as the Internet.
  • Another feature of the disclosed ideas is that the same application, and even the same graphical user interface (GUI), can be used for both presenting media content by the first electronic device and for controlling the presentation of media by the second electronic device. This provides a seamless experience for the user, as the user does not need to use different control interfaces (e.g., buttons, touchscreens, remote controls, etc.) for each device, thus obviating the need for the user to learn and use different control interfaces, keyboard layouts, media control functionalities, and the like, for different devices.
  • Furthermore, in some implementations, a media control request is configured to change the active presentation device for media content that is currently being presented. For example, a user who is listening to a music playlist on a mobile phone can request to change the active listening device to a home stereo system. In some implementations, the request to change media presentation devices is initiated by a user (e.g., via a user interface on the mobile phone). In some implementations, the request is initiated automatically and without user intervention by automatically detecting a predetermined condition, such as location, date, time of day, and/or day of week. For example, if a user is listening to an audio playlist on a mobile phone while walking home, the user's home stereo system may begin playing back the audio instead of (or in addition to) the mobile phone once it is determined that she has arrived at her home (e.g., using a positioning system (e.g., a GPS, cell-tower triangulation, etc.), proximity sensors, inertial monitoring devices, etc.).
  • In some implementations, the switching occurs in response to the server system receiving notification from the first electronic device that a specific predetermined condition has been met. For example, the first electronic device may monitor its geographic location, and when it determines that it has moved into (or out of) a predefined area, the first electronic device will send a notification to the server system that the predetermined condition has been met. In another example, the condition is met when the server detects a change in the state of the first electronic device, such as when the first electronic device establishes communication with a particular wireless network, or when the first electronic device pairs with or detects the second electronic device. In response, the server system ceases transmitting the media content stream to the first electronic device and begins transmitting the media content stream to a second electronic device. In this way, the user is able to establish conditions for automatically changing the device to which media content is being streamed or stream between devices.
  • Automatic switching may also be conditioned on additional factors, such as whether a device on which presentation is to be initiated is already presenting other media content, a time of day, a time of week, etc. For example, continuing the previous example, if the home stereo system is already presenting media content when the user returns home, the home stereo system will not switch to the playlist that was being presented on the user's mobile phone. This, in this case, the home stereo is available to be automatically activated only when it is idle and/or not presenting other media content.
  • In some implementations, the active presentation device is configured to automatically change only during predetermined time or date ranges. For example, in some implementations, a home stereo system will automatically begin presenting media content only during daytime hours (e.g., between 10:00 AM and 8:00 PM), and/or during weekends. Such rules may
  • In some implementations, different users have different priority levels for automatic device switching. This can help avoid frequent and annoying changes in media presentation due to different users returning home, entering a particular area, etc. In some implementations, when a user initiates presentation of media content at a device, the user's priority level is logged or stored. When another user attempts to initiate presentation of different media content at that device, the other user may do so only if his or her priority level is above the first user. In some implementations, priority levels are only used to determine whether automatic device switching may occur. Thus, if a lower-priority user returns home while listening to a certain playlist, but a higher-priority user is already listening to music (or presenting other media content) via the home stereo, the lower-priority user's playlist will not be automatically presented via the home stereo system. On the other hand, manual requests to change the media content being presented by the home stereo system may be implemented regardless of the priority level of the requestor.
  • In some implementations, media presentation devices can present media content regardless of the origin of the content. For example, content in the audio playlist in the preceding example may be stored on the mobile phone, and streamed to the home stereo system from the mobile phone (e.g., via a local connection, such as a wireless LAN, peer-to-peer connection, a public network such as the Internet, etc.). If the content in the audio playlist is being streamed from a separate source (e.g., a remote server associated with a media content provider, a local media server, etc.), then the source can simply switch the destination of the audio stream from the mobile phone to the home stereo system.
  • In some implementations, when the active presentation device is changed, media content is delivered to the active presentation device regardless of the origin of the content. For example, when a song in an active playlist is stored on the user's mobile phone (but presented by the user's home stereo system), the song will be streamed or otherwise provided by the mobile phone; when a song is stored on a user's network accessible storage (NAS) system, the song will be streamed or otherwise provided by the NAS; when a song in the playlist is stored in a remote server, that song will be streamed or otherwise provided by the remote server. Songs that are stored in different locations (e.g., on the mobile phone and the remote server) can be included in the same playlist, and can be streamed to whatever device is playing the playlist.
  • In some implementations, when the active presentation device is changed, the context of the content stream, originally available on the first presentation device, becomes available on the second presentation device. The context of a content stream includes various information about the content stream, including the position within the currently playing media content, the previously presented media content (e.g., a play history), the position within previously presented media content (if unfinished), the media content to be played next, metadata of the current or upcoming media content (e.g., artist, album, track), etc. For example, the media content, once playback is initiated at the second electronic device, will retain the position in the current media content item and the position of the active playlist. In this way the media stream being presented on the first device is seamlessly transferred to a new device while maintaining the same state as on the first device. Thus, users may easily switch between devices without fear of losing their place in a song, a movie, a playlist, etc.
  • In some implementations, when media content is being presented at one device, one or more other devices buffer/cache a portion of that media content, for example, in case the user should choose to present the media content with one of the other devices instead of or in addition to the current device. In this way, the other devices are able to begin presenting the media content quickly, as the media presentation can begin at the other device(s) as soon as they receive a request to do so, rather than waiting for communications to be established between a media content source and the presentation device, for initial buffering of streaming content prior to presentation, and the like.
  • In some implementations, a user can select media content for preview. For example, a user can select, at one device, a song, playlist, or album to preview on the same device or on a different device. In some implementations, any media content that is being presented when the preview is requested is muted, dimmed, or paused (or otherwise made less audible/visible) while the preview is presented. Accordingly, the user can preview the media content without the inconvenience of having to manually restart the earlier content playback and/or without losing their place in the earlier content playback. Content for a preview can be pre-fetched by downloading and caching/buffering at least a portion of the content in response to a preview control (e.g., a button) being presented to a user, such as when a user views media content in a media browser on an electronic device. Moreover, various user interfaces and techniques can be used to initiate a media content preview and/or to add media to an existing playlist. For example, touch-and-hold inputs on a touch screen interface can be used to initiate a preview, and touch-and-release inputs can be used to add content to a playlist. Various implementations of these and other ideas are discussed below.
  • Attention is now directed to the figures, and in particular to FIG. 1, which is a block diagram of a client-server environment, according to some implementations. The client-server environment 100 includes one or more client environments (108-1 . . . 108-n) and a server system 120 that are connected through a network 110. In some implementations, the client environment 108-1 includes one or more electronic devices (e.g., first electronic device 102-1 and second electronic device 106-1). In some implementations, the server system 120 is associated with a media content provider to which users (and their electronic devices) may have accounts that enable the user to access media content from the server system 120. The network 110 includes any of a variety of networks, including wide area networks (WAN), wireless networks, wired networks, the Internet, or a combination of such networks.
  • In accordance with some implementations, the client environment 108-1 includes a first electronic device 102-1. In some implementations, the first electronic device 102-1 is one of the group of: a personal computer, a mobile electronic device, a laptop, a tablet computer, a mobile phone, a digital media player, or any other electronic device able to present media content.
  • In accordance with some implementations, the client environment 108-1 also includes a second electronic device 106-1. In some implementations, the second electronic device 106-1 is one of the group of: a computer, a home audio/visual system, a home media receiver/player, or any other electronic device able to present media content. In some implementations, both the first electronic device 102-1 and the second electronic device 106-1 are associated with a common user account (or associated user accounts) provided by a content provider with which the server system 120 is associated. For example, in some implementations, the server system 120 is operated and/or provided by a subscription-based media streaming service to which a user may have an account, and the first and second electronic devices 102-1, 106-1 are each associated with account credentials that enable them to communicate with and receive content from the server system 120.
  • In accordance with some implementations, both the first electronic device 102-1 and the second electronic device 106-1 are on the same local network. In some implementations, the local network is a Local Area Network. In some implementations, the server system 120 is not located on the same Local Area Network as either the first electronic device 102-1 or the second electronic device 106-1.
  • As noted above, Local Area Networks are often associated with a relatively small geographic area (e.g., one house or building) and create IP address space behind a traffic routing device. In some implementations, the traffic routing device uses network address translation to connect devices within the LAN to devices outside the LAN. Network address translation is the process of readdressing packets by modifying the IP address contained in each packet so that the packets reach their intended destination device. Network address translation allows a plurality of electronic devices on the Local Area Network to connect to the Internet through a single IP address (e.g., the IP address of the traffic routing device). The traffic routing device (i.e. a router) receives incoming packets, determines which device on the Local Area Network is the intended recipient, and modifies the IP address to correctly identify that device. For example, a Local Area Network has 9 devices with local IP addresses 192.168.0.1-192.168.0.9 that all connect to a router. The router connects to the Internet and has an IP address of 12.162.29.2. Using network address translation, the router translates the source address for all communications sent from any of the 9 devices and intended for destinations in the Internet to be 12.162.29.2 (the router's IP address). On the other hand, the router collects all packets incoming from the Internet, determines the intended recipient device based upon the contents of each packet, and translates the destination IP address to the address of the correct device on the Local Area Network. So when the intended device is the device with Local Area Network IP address 196.168.0.2, the router would change the destination address to this address. Local Area Networks also commonly use firewalls to limit incoming connections. In this way, computer devices outside of the Local Area Network are generally not able to communicate directly with the devices on a Local Area Network. Indeed, in some Local Area Networks the devices in the network are not contactable even by other devices in the Local Area Network.
  • In some implementations, both the first electronic device 102-1 and the second electronic device 106-1 are on the same Personal Area Network. In some implementations, the Personal Area Network uses BLUETOOTH communication technology. In some implementations, the server system 120 is not located on the same Personal Area Network as either the first electronic device 102-1 or the second electronic device 106-1.
  • In some implementations, the first electronic device 102-1 includes a media content presentation and control application 104 (hereinafter “media application”). The media application 104 is able to control the presentation of media by the first electronic device 102-1. For example, the media application 104 enables a user to navigate media content items, select media content items for playback on the first electronic device 102-1, create and edit playlists, etc. In some implementations, media content is stored by the first electronic device 102-1 itself. In other implementations, the media content is stored by a server system 120, which may be located remotely from the first electronic device 102-1. The media content is then streamed from the server system 120 to the first electronic device 102-1 over the network 110.
  • In some implementations, the data streamed from the server system 120 is temporarily stored/cached by the first electronic device 102-1 in the media content buffer 105 in the memory of the first electronic device 102-1. In some implementations, media content stored in the media content buffer 105 is removed after the media content is presented by the first electronic device 102-1, allowing new media content data to be stored in the buffer 110. In other implementations, at least some of the media content stored in the media content buffer 105 is retained for a predetermined amount of time after the content is presented by the first electronic device 102-1 and/or until other predetermined conditions are satisfied.
  • In some implementations, the media application 104 is also able to control media content presentation by the second electronic device 106-1, which is distinct from the first electronic device 102-1. Thus, the user is able to use the media application 104 to cause the electronic device 102-1 to act both as a media presentation device as well as a remote control for other media presentation devices. This allows a user to control media presentation on multiple electronic devices from within a single application 104, and/or using a single user interface.
  • In some implementations, when a user wants to use the first electronic device 102-1 to control media presentation by the second electronic device 106-1, the user interacts with the media application 104 to send a media control request (e.g., server media control request 112, FIG. 2) to the server system 120. The server system 120 receives the media control request over the network 110. For example, the user may press a button on a touchscreen of the first electronic device 102-1 in order to send the media control request to the server system 120 and/or directly to the second electronic device 106-1. As described below, a media control request is, for example, a request to begin presentation of media content by the second electronic device 106-1. Though often used herein to describe requests to initiate or begin presentation of media by the second electronic device 106-1, in some implementations, media control requests also include requests and/or signals to control other aspects of the media that is being presented on the second electronic device 106-1, including but not limited to commands to pause, skip, fast-forward, rewind, adjust volume, change the order of items in a playlist, add or remove items from a playlist, adjust audio equalizer settings, change or set user settings or preferences, provide information about the currently presented content, and the like.
  • The client-server environment 100 also includes a server system 120. In some implementations, the server system 120 includes a media streaming module 122, a media content database 124, and a context database 126. The media content database 124 stores media content that can be presented by an electronic device. For example, in some implementations, the media content database 124 stores audio (e.g., music, audiobooks, etc.), video (e.g., movies, television shows, etc.), images, or other content that can be streamed to other electronic devices. In some implementations, the media content database includes data stored in different formats and file types to allow a variety of different devices and/or applications to receive streamed content. In some implementations, the data is stored in a single file format and is converted/transcribed to the appropriate data type before or as it is streamed to a remote device.
  • In some implementations, the server system 120 includes a media streaming module 122. In some implementations, the media streaming module 122 receives media control requests from electronic devices and streams media content in response. In some implementations, the media streaming module 122 receives media control requests from a first electronic device 102-1 and forwards the request to a second electronic device 106-1, which then makes the final request to the server system 120 for the media content. For example, a user sends a media control request to the server using a mobile phone (a first electronic device) requesting that media be presented by a home stereo system (a second electronic device). The server system 120 then sends the requested media (and/or the media control request) to the home stereo system. This and other techniques are discussed in greater detail below with respect to FIGS. 5-8.
  • In some implementations, the received media control request includes information identifying the electronic device to which the server system 120 should forward the media control request. For example, a user may have multiple electronic devices that can present media from the server system 120, such as a mobile phone, a computer system, a television, a home stereo, etc. In some implementations, the identifying information is a unique or semi-unique device identifier, such as an IP address, a Media Access Control address (MAC address), a user-specified device name, an International Mobile Equipment Identity number (IMEI number), or the like. Accordingly, the media control request will identify that a request is intended for the home stereo, for example, so that the server system 120 can send the requested media and/or the media control request to the home stereo.
  • In some implementations, the server system 120 includes a context database 126. The context database 126 stores data associated with the presentation of media content by an electronic device. In some implementations, the context database 126 includes, among other things, the current position in a media content stream that is being actively presented by an electronic device, a playlist associated with the media content stream, previously played content, skipped pieces of media content, and previously indicated user preferences. For example, the context database may include information that a content stream to an electronic device currently is presenting a song, at 1 minute and 23 seconds into the song, as well as all the songs played in the last hour and the next 20 songs in the playlist. In some implementations, the server system 120 transmits the context associated with a media content stream to the device that is presenting the content stream so that one or more items of context information can be used by the device, such as for display to the user. In some implementations, when the device to which the media content is being streamed changes, the server system 120 transmits the context associated with the active media content to the newly active device.
  • FIG. 2 is a block diagram illustrating a client-server environment 100 in accordance with some implementations. The client-server environment 100 includes a local network 202 that encompasses at least a first electronic device 102-1 and a second electronic device 106-1. Client-server environment 100 also includes a server system 120 that is located remotely from and is not integrated into the local network. In some implementations, the server system 120 stores media content and streams the stored media content to remote electronic devices upon request. In some implementations, the local network 202 is a Local Area Network, and the server system 120 is not within the IP address space of the Local Area Network. Thus, the server system 120 communicates with either the first or the second electronic devices (102-1, 106-1) via the Internet. For example, the media content streams 114-1, 114-2, and the server media control requests 112, discussed below, may be transmitted via the Internet.
  • In some implementations, a user of the first electronic device 102-1 uses an application running on the first electronic device 102-1 to send a server media control request 112 to the server system 120. In some implementations, the server media control request 112 includes information identifying a second electronic device to which the control request is addressed, as well as a payload that indicates what media control operation is to be performed by the second electronic device. As noted above, in some implementations, media control requests include requests to control aspects of the media that is being presented on the second electronic device 106-1, including but not limited to commands to initiate media presentation, cease media presentation, pause, skip, fast-forward, rewind, adjust volume, change the order of items in a playlist, add or remove items from a playlist, adjust audio equalizer settings, change or set user settings or preferences, provide information about the currently presented content (e.g., metadata), and the like.
  • In some implementations, in response to receiving the server media control request 112, the server system 120 forwards the server media control request 112 to the second electronic device 106-1. When received by the second electronic device 106-1, the server media control request 112 causes the second electronic device to initiate the media control operation indicated by the request. In some cases, this will include communicating with the server system 120 to perform any of the operations described above that may be requested by a media control request. In some cases, the second electronic device 106-1 will not need to establish any subsequent communication in response to receiving the server media control request 112, such as when the control request is for changing the volume, muting the speakers, changing equalizer settings, etc., as these operations can often be performed by the second electronic device 106-1 alone.
  • In some implementations, in response to receiving the server media control request 112, the server system 120 will perform an operation requested by the server media control request 112 without forwarding the request to the second electronic device 106-1. For example, if the server media control request 112 contains a request to initiate presentation of media content at the second electronic device 106-1, the server system 120 may simply begin sending the media content to the second electronic device 106-1 (e.g., via the media content stream 114) without also forwarding the server media control request 112 to the second electronic device 106-1. However, even where the server system 120 executes some types of media control operations upon receiving the request from the first electronic device 102-1 (and without forwarding the request to the second electronic device 106-1), the server system 120 still forwards some types of media control requests to the second electronic device 106-1. For example, a server media control request 112 that corresponds to a request to increase or decrease speaker volume, or any other request that can only be executed by the second electronic device 106-1, are forwarded to the second electronic device 106-1, even though requests to initiate presentation of media content may not be forwarded.
  • In some implementations, the first electronic device 102-1 also sends a local media control request 116 directly to the second electronic device 106-1 at substantially the same time that it sends the server media control request 112 to the server system 120. In some implementations, the local media control request 116 is configured to initiate the same media control operation as the server media control request 112.
  • In some implementations, the local media control request 116 is sent via the local network 202 (e.g., a Local Area Network, a Personal Area Network, a peer-to-peer connection, etc.). In implementations where the local network 202 is a Local Area Network that establishes IP address space behind a traffic routing device, the local media control request 116 is sent to the second electronic device 106-1 without sending the local media control request 116 to devices that are not within the IP address space of the Local Area Network, and/or without sending the local media control request 116 through a public domain name system. Sending a local media control request 116 to the second electronic device 106-1 over the Local Area Network, in addition to sending the server media control request 112 over the Internet to the server system 120, allows for two separate, redundant pathways for delivering the media control request to the second electronic device. This can allow the second device to react more quickly and more reliably to the media control request of the first electronic device 102-1. For example, if either connection path between the first and second electronic devices is not available, is improperly configured, or the like, the other connection path can ensure that remote-control functionality is still available. Moreover, this allows a service provider (e.g., a provider of software and/or devices that are configured to present media content) to provide a robust remote control solution that reduces the necessary technical knowledge of the end user, and is more likely to work in many different communication networks and environments (e.g., where a user's internal router is not configured for internal LAN communications, where an active Internet connection does not exist, etc.).
  • In some implementations, the server media control request 112 and the local media control request 116 are identical. For example, they may contain the same message, such as “Device{UserX/Home Stereo}; Operation{Play: UserX/Party Mix}.” Upon receiving this message, each device may respond by performing a predefined set of operations. For example, in some implementations, the server system 120 begins streaming media content associated with User X's “Party Mix” playlist to User X's Home Stereo system. Similarly, the second electronic device 106-1 (here, the home stereo system), upon receiving the same message in the local media control request 116, sends a request to the server system 120 to initiate playback of the same media content. Because multiple control requests that are configured to initiate the same media control operation are sent from the first electronic device, some implementations provide systems and methods for handling duplicate or conflicting control requests, so that media control operations are not repeated, and to ensure that new and/or different media control requests are not ignored. Some examples of this are described below with reference to FIGS. 7-8.
  • In some implementations, the server system 120 performs the media control request without forwarding it to the second electronic device 106-1. In this case, executing the requested operation includes streaming media content associated with User X's “Party Mix” playlist to User X's Home Stereo system. In some implementations, the server system 120 forwards the message to the second electronic device 106-1. The forwarded message may be the identical message, or it may be modified or translated before it is forwarded to the second electronic device 106-1. In implementations where the message is forwarded, it may be forwarded instead of or in addition to the server system 120 executing the requested operation. Upon receipt of the forwarded message, the second electronic device 106-1 may then send a request to the server system 120 to initiate playback of the requested content.
  • In some implementations, the server system 120 may be configured to initiate a media control request (e.g., streaming media to the second electronic device 106-1) only if the request originates from the device at which the media is to be presented, and not if it originates from a different device. For example, in contrast to the example above where the server system 120 begins streaming media to the second electronic device 106-1 in response to a request from the first electronic device 102-1, the server system 120 instead would only begin streaming media to the second electronic device 106-1 if the request came directly from the second electronic device itself. In such implementations, the server 120 may be configured to forward all media control requests to the second electronic device 106-1 prior to executing them, as described above, so that the second electronic device 106-1 can determine whether, when, and how to execute the request. In some implementations, the second electronic device 106-1 (and/or the server system 120) determines the authenticity, trustworthiness, and/or origin of a media control request before it executes the request or initiates the operation.
  • In some implementations, the server media control request 112 and the local media control request 116 are not identical (e.g., the message is in a different form, or contains more, less, or different information), but are both configured to initiate the same media control operation by the second electronic device. For example, in some implementations, the server media control request 112 includes a command to begin delivering content to the second electronic device 106-1, while the local media control request 116 includes a command to cause the second electronic device 106-1 to issue a request to the server system 120. The request from the second electronic device 106-1 to the server system 120 in response to receiving one of the server media control request 112 and the local media control request 116 may be described as a third media control request (not shown). The form and content of the third media control request may be the same as or different from the server media control request 112 and the local media control request 116. In either case, the third media control request will be configured to cause the same media control operation as those requests. Accordingly, though these requests are configured to cause different processes to be performed by different devices (e.g., causing the server to begin streaming media to the second electronic device, or cause the second electronic device to request that media from the server), they ultimately cause the same ultimate media control operation to be initiated (e.g., presentation of the media by the second electronic device).
  • As noted above, in some implementations, a media control request is configured to switch the active media presentation device from one device (e.g., a user's mobile phone) to another (e.g., a user's home stereo system). In some implementations, this includes changing the device to which the server system 120 is streaming media content. For example, a user who is listening to a music playlist on a mobile phone can request that the music be played through a home stereo system instead of (or in addition to) the mobile phone. In some implementations, the request to change media presentation devices is initiated by a user (e.g., via a user interface on the mobile phone), and in some implementations the request is initiated automatically (e.g., based on the location of the mobile phone, the proximity of the mobile phone to the home stereo system, the time of day, the day of the week, and/or user identities).
  • FIG. 2 also illustrates how the destination of a media content stream can be changed from one device to another in a client-server environment. In some implementations, the first electronic device 102-1 is receiving a media content stream 114-1 from the server system 120. For example, the first electronic device 102-1 may be a mobile phone receiving streaming music from the server system 120. A device within the client-server environment 100 then requests that the media also, or instead, be presented by the second electronic device 106-1. This request may come from the first and/or the second electronic device (or any other device within the environment), and may be initiated by a user, or it may be initiated automatically. For example, the user may, upon returning home, request that the music that she is listening to on the mobile phone instead by presented by the second electronic device 106-1. In some implementations, the request to change the active presentation device is sent via the local and server media control requests 116, 112. Upon receiving and/or processing a request, the server system 120 begins sending the media control stream 114-2 to the second electronic device 106-1. The media content stream 114-1 that was being sent to the first electronic device 102-1 may continue or may be terminated, depending on factors such as an explicit user request, historical usage patterns, predefined preferences, etc. In some implementations, second electronic devices 106-n buffer/cache a portion of the same media stream that is being presented by the first electronic device 102-1, so that when the second electronic device 106-n is selected to begin presentation of the media content, the second electronic device 106-n begins presenting the media content seamlessly and without interruption. Buffering and/or caching of media content is discussed in greater detail below with reference to FIGS. 6A-6B.
  • In some implementations, as noted above, the context of the media content stream 114-1 is available to the second electronic device 106-1 once it begins receiving the media content stream 114-2. For example, information such as the current location within a playlist, recent play history, recent changes to a playlist, etc., are all available to the second electronic device 106-1, and may be viewed, used, modified, or otherwise operated on by a user through a user interface associated with the second electronic device 106-1. In some implementations, the context information is stored in the context database 126 of the server system 120.
  • Also, as noted above, in some implementations, the second electronic device 106-1 buffers/caches a portion of the content that is being streamed to the first electronic device 102-1, in case the user should choose to begin media presentation by the second device instead of or in addition to the first device. In some implementations, the second electronic device 106-1 caches/buffers a predetermined portion of the media content that is being streamed to the first electronic device 102-1, such as the next 10 seconds, 20 seconds, 30 seconds, 1 minute, etc. In some embodiments, the second electronic device 106-1 receives information about the current media playback location from the first electronic device 102-1 and/or the server system 120, such that when the second electronic device 106-1 receives a command to begin presenting the media content, it begins presentation at substantially the same position as the first electronic device 102-1.
  • In some implementations, the server system 120, the first electronic device 102-1, and the second electronic device 106-1 are configured to present media content regardless of the particular device on which the media content is stored. For example, as described above, media content can be stored at the server system 120 (e.g., in the media content database 124), and streamed to the first and/or the second electronic devices. However, media content may also be stored at the first and/or the second electronic devices (or other storage devices accessible to the first and/or second electronic devices, such as a NAS). This media content may be made accessible for presentation by any capable device associated with a user, even though the media content may not be available from the server system 120. For example, in some implementations, a playlist on a first electronic device (e.g., a user's mobile phone) may include media content that is stored on that device's own memory, as well as media content that must be streamed from a different source, such as a remote server (e.g., server system 120) or storage device within the local network 202 (e.g., a NAS). When the user plays back the playlist at the first electronic device, that device can switch between retrieving media content from its own memory and the different source, depending on where the media content item is stored. However, if the user wished to instead playback the playlist from a second electronic device (e.g., the user's home audio system), the second electronic device may not have the same media content stored in its memory as the first electronic device. Thus, in accordance with some implementations, media content that is stored on one device in the local network 202 (e.g., on first or second electronic devices 102-n, 106-n and/or on any device within a client environment 108-n, such as a NAS) may be streamed to another device within the local network 202 when its turn in the playlist arrives. Accordingly, the user need not worry about where any given media content item is stored, because the media may be streamed from either remote or local sources, or both.
  • In some implementations, music stored locally on the first electronic device is streamed to the second electronic device via local communication paths (e.g., within a Local Area Network) when that music is not available from a remote source associated with a media content provider (e.g., the server system 120). In some implementations, music stored on a first electronic device, but not available from the media content provider, is streamed to the server associated with the media content provider, and then streamed from the server associated with the media content provider to the second electronic device. Accordingly, media that is stored on one device may be accessible to another device seamlessly and without having to download, transfer, or otherwise duplicate the media content on multiple devices. Also, this process may be seamless to the user, such that the same controls and inputs are used both for causing a second device to stream media from a remote source, and for causing locally stored media to be streamed to the second device via local communication paths.
  • As described above, it would be beneficial to allow users to preview media content on an electronic device. This way, a user can listen to a portion of a media content item before deciding whether to purchase it or otherwise select it for presentation in its entirety. For example, users may wish to browse music choices in an electronic marketplace, and hear a snippet of a song that they come across in order to determine if they want to add it to a playlist or select it for immediate playback. Previewing media content can be a nuisance, though. If a user is already listening to a song, for example, and requests a preview of another song, playback of the first song may simply be terminated. Once the preview is over, then, the user may have to take additional steps to resume playback of the first song. For example, the user may have to navigate to a different area of a user interface, locate the song that was previously playing, and either restart the song from the beginning, or use trial-and-error to locate the point in the song where it was stopped. Accordingly, implementations discussed herein provide convenient previewing techniques to address these and/or other issues.
  • Moreover, modern media consumers access media on many different devices and use many different types of interfaces. For example, users browse and consume media on personal computers, laptops, mp3 players, portable media players, mobile phones, tablet computers, televisions, stereos, etc. These devices use various input devices and techniques, such as touch screens, pointing devices (e.g., computer mice or trackballs), touch-sensitive track-pads, keyboards, remote controls, and the like. Because of the differences between these various types of devices and user interfaces, various user interface techniques are also discussed below that provide convenient ways of initiating media content previews and other media controls.
  • While the following discussion frequently refers to previewing and/or presenting music, this is merely exemplary of any media content for which a preview may be presented, such as podcasts, audiobooks, radio (e.g., broadcast radio, Internet radio, etc.), video (e.g., movies, music videos, television programming, short-form video, etc.), and the like. It will be understood to one of ordinary skill in the art that other types of media content can be substituted in various aspects of the implementations described below. For example, a user may preview a video during playback of a song, or may preview a broadcast radio or television stream during playback of a locally stored podcast.
  • As noted above, it is inconvenient if a media content preview interrupts currently played media content and does not restart it when the preview finishes. Thus, a method is provided that allows a user to preview media content even when the user is consuming other media content, and restarts the other media content once the preview is complete. For example, a user may be listening to a song through a computer while also browsing other music on the computer, such as in an online music store or a media player. When the user finds a song, artist, or album that he would like to preview, he can select the content for preview using a preview button or another input technique. In response to the selection, the song that is currently playing can be paused, muted, or otherwise made less audible, while the content to be previewed is presented to the user. If the user was watching a video, the video can be dimmed, muted, lowered in volume, made translucent, paused, reduced in size, split-screened with the preview, or the like. Then, when the preview is over (e.g., because the user terminated the preview, or because a preview time expired), the original song or video is resumed. For example, if an originally played song was muted, or if the volume was otherwise lowered, the volume is returned to its previous level. If the original song was paused, it is un-paused. By resuming playback of the original song, the previewing experience is conveniently integrated into the overall listening experience, and the user is not burdened with additional navigation and selection tasks in order to return to his or her earlier content playback.
  • Furthermore, throughout the preview, any position and/or playback information of the originally playing media content can be maintained, such as the position in a song, the position in a playlist, a play history associated with the current playback session, etc. For example, if a user is listening to a playlist or album on a “shuffle” setting, the play history is maintained so that, when the original playback is resumed, songs are not replayed or re-ordered.
  • In some cases, a preview of media content should not interrupt presentation of currently playing media content by pausing, muting, or lowering the volume of the currently playing content. For example, if a user is presenting music via a home stereo system, he or she may want to preview other music to add to the current playlist without pausing or muting the music being played from the home stereo system. This situation may occur, for example, when the user is entertaining guests and wants to add music to a playlist without interrupting the current music. Thus, instead of muting, pausing, or otherwise interrupting the music, the preview may be presented to the user via a different device than the active presentation device, or via a subset of a group of active presentation devices (e.g., one speaker of a group speakers), or using a different audio output of the active presentation device. For example, if media content (e.g., a playlist) is being presented via a home audio system, a media preview may be presented to a user at a computer or mobile phone. In another example, if media content is being presented through one output means of a device (e.g., a charging port, data port, BLUETOOTH link, etc.), a preview can be presented through another output means (e.g., a headphone jack or a built-in speaker). In some implementations, a user manually selects a particular device for previewing media content and a particular device for presenting media content (e.g., regular playback). In some implementations, these and/or other features are automatically enabled when a user selects a “DJ” mode on a media content browsing device. In some implementations, the “DJ” mode causes media content from a playlist to be routed to one preselected device (or device output), and media content for presentation to be routed to a different preselected device (or device output).
  • Various ways of initiating a media content preview may also be provided. In some implementations, a user may select an affordance displayed on an electronic device in order to begin the preview. The affordance may be an icon, graphic, image, area, button (e.g., hardware button or touch-screen button), text, or any other control that can be selected by a user. For example, in some implementations, the affordance is an icon of a “play” button (or any other appropriate image, such as a thumbnail image of an album cover or movie poster), or the text “preview now” (or any other appropriate text). A user can select the affordance by clicking on the affordance with a cursor (e.g., via a mouse or touchpad interface), by touching on the area of a touch screen associated with the displayed affordance, hovering a cursor over the affordance, or the like.
  • Different inputs or interactions with a single affordance can cause different actions. In some implementations, when a user selects an affordance using one type of input or interaction, a preview of media content associated with that affordance is presented. If the user selects the affordance with a different type of input or interaction, however, other actions can be taken, such as adding the media content associated with the affordance to a current playlist, or terminating any currently played media content and presenting the selected media content instead.
  • In some implementations, an input representing a user selection of media content is detected (e.g., a touch contact or mouse click-down on a description of some media content), and different actions are taken depending on whether a change to the input is detected before or after expiration of a first time period. For example, if the change to the input is detected before expiration of the time period, the media content may be added to a playlist (e.g., as a next song, at the end of the playlist, or replacing a currently played song). If, on the other hand, no change to the input is detected before the expiration of the time period, the media content may be previewed. As described in several examples below, the input and the change to the input may be many different inputs and/or interactions. For example, the input may be a contact with a touch-sensitive surface, and the change to the input may be a liftoff from the touch-sensitive surface. In another example, the input may be a contact with a first area of a touch-sensitive surface, and the change to the first input may be a contact with a second area of the touch-sensitive surface. In this case, the contact with the first area may correspond to a user placing one finger on the touch-sensitive surface, and the contact with the second area may correspond to the user placing a second finger at a different point on the touch sensitive surface. In yet another example, an input can correspond to a “hover” input, where a cursor (e.g., controlled by a mouse, trackpad, etc.) is placed within a certain predefined area on a display. The predefined area may correspond to an icon, a graphic, text, or the like. A change to this input includes, for example, a mouse click or removal of the cursor from the predefined area. Other inputs and touch gestures may also act as the first input and/or the change to the first input, such as tap, pinch-together, pinch-apart, touch-and-hold, touch-and-release, click-and-hold, click-and-release, click-release, double-click, double-tap, sliding gestures, multi-touch gestures, drag-and-drop, slide-and-release, right mouse click, selection of menu items (e.g., stop, pause, play now, play next, add to queue, increase/decrease volume, etc.), and the like.
  • In some implementations, a “touch-and-hold” or a “click-and-hold” input (also referred to as a maintained input) initiates a preview of the selected media content, while a “touch-and-release” or “click-and-release” input (also referred to as a momentary input) adds the selected media content to a current playlist. In this case, an initial contact on a touch screen (or mouse click-down) corresponds to the input, and a liftoff from the touch screen (or a mouse click-release) corresponds to the change to the input.
  • In some implementations, where a maintained input is used to initiate a media content preview, the selected media content is previewed (e.g., played back) for as long as the input is maintained. For example, in a touch-screen environment, after it is determined that a contact with the touch screen corresponds to a maintained input (i.e., no liftoff has been detected for a certain time period), the media content is previewed until the contact is removed (i.e., when the liftoff is finally detected). In some implementations, the selected media content is previewed for a shorter or longer time than the duration of the maintained input. For example, in some implementations, the maintained input begins a preview clip of a predetermined length (e.g., 5, 10, 15, 20 seconds, etc.). In some implementations, the preview will continue beyond the point when the user ends the maintained input, such as to provide an audio/video fade-out or other transition, or until the full preview clip is over. In some implementations, the preview will end at the earlier of (i) the expiration of a time period (e.g., the preview clip length) or (ii) the end of the maintained input. Thus, once a user provides a maintained input, the preview will be presented until the user releases the maintained input or until the preview clip is over. In some implementations, a preview is only limited by the length of the actual media content, so it is possible to preview an entire song, album, playlist, or other media content.
  • In some implementations, in order to determine whether an input corresponds to a maintained input or a momentary input, a timer is started when a contact or click-down input is detected. If a liftoff or click-release event is detected before the timer expires, the input corresponds to a momentary input. If a liftoff or click-release event is not detected before the time expires, the input corresponds to a maintained input. In this case, the expiration of the timer (possibly in conjunction with additional signals or logic) initiates the media content preview. In some implementations, a liftoff or click-release event that occurs after the time expires causes the preview to end.
  • In some implementations, other actions can be taken in response to detecting changes to the input. For example, in some implementations, once an input initiates a preview, a change to the input—such as moving a cursor with a mouse, or making a “swipe” gesture on a touch screen—changes a playback attribute of the preview (or causes other appropriate actions). Here too, the change to the input may include various interactions and gestures, such as pinch-together, pinch-apart, touch-and-hold, touch-and-release, click-and-hold, click-and-release, click-release, double-click, sliding gestures, multi-touch gestures, etc. In some implementations, the change to the input causes the preview to skip forward or backward within a song, a playlist (e.g., including an album, a user generated playlist, a series of podcasts, etc), or any other media content or collection of media content items.
  • In some implementations, the change to the input includes a directional component, such as a “swipe” gesture on a touch screen, and actions can be taken based on the particular direction of the input. For example, in some implementations, a user first initiates a preview by touching and holding an area of the touch screen with a finger. The user can then slide his finger forward to skip to a next track in the playlist, or slide his finger backwards to skip to a previous track in the playlist. In some implementations, the actions taken (e.g., skipping tracks, fast-forwarding within a track) are triggered if the input has a sufficient directional component in a certain direction. Various techniques can be used to determine the directional components of a swipe input. In some implementations, a swipe input path with a length, direction, and/or angle (e.g., as measured from the point of initial contact) that falls within a particular range may trigger the action. In some implementations, a swipe input in a right-hand direction (e.g., substantially horizontal and from left to right) causes a first action, and a swipe input in a left-hand direction (e.g., substantially horizontal and from right to left) causes a second action. In some implementations, a swipe input in a down direction (e.g., substantially vertical and from top to bottom) causes a first action, and a swipe input in an up direction (e.g., substantially vertical and from bottom to top) causes a second action. In some implementations, the first and second actions are a skip forward action and a skip back action, respectively.
  • In some implementations, other directional inputs or gestures cause these or other actions. For example, a first contact corresponding to maintained input (i.e., a “touch-and-hold” input on a touch screen) may cause a preview to begin, and a second contact may cause the preview to skip forward to a next or previous track, or fast-forward or rewind the preview. The second contact may be a maintained input or a momentary input. In some implementations, a second contact corresponding to a maintained input will cause the preview to fast-forward for as long as the maintained input is held. Other directional inputs or gestures that may be used in the above implementations include, but are not limited to, diagonal swipes, loops, curves, multi-touch inputs, tap sequences, and other complex touch paths.
  • FIG. 13 illustrates an exemplary device 1300 showing a part of a user interface, in accordance with some implementations. The device 1300 includes a display area 1302. In some implementations, the device is a mobile phone, mobile audio player, or the like. A media preview area 1304 is displayed on the display 1302, and can include any appropriate information. As shown, the media preview area 1304 includes an artist name, album title, album art, and a preview button 1306. As described above, the preview button 1306 is configured, in some implementations, so that a tap input (e.g., a momentary input) causes the album and/or individual songs on the album to be presented by adding it to a playlist. Alternatively, if the preview button 1306 is pressed and held (e.g., a maintained input), one or more tracks from the album are previewed as long as the input is maintained.
  • As noted above, in some implementations, media content can be previewed in response to a cursor being “hovered” over a particular area of a display screen. For example, a media browsing environment presented in a web page or media player application may include multiple graphical items (also referred to as “tiles”) corresponding to media content, such as albums, playlists, songs, audiobooks, movies, videos, and the like. In some implementations, the tiles include depictions of album art, media content metadata (e.g., album title, track listings, authors, actor names, artist names, artist photos, etc.), and the like. The tile may also include a “preview” area that, when selected, causes a preview of the media content to be presented. In some implementations, the preview area corresponds to an icon, graphic, text, picture, or other discernible area within or associated with the tile. In some implementations, when the user hovers over this area with a cursor, a preview of the media content associated with the tile is initiated. In some implementations, the preview is initiated once the user has hovered over the area for a length of time, such as 0.5 seconds, 1 second, 2 seconds, or any other appropriate duration. In some implementations, the preview continues until the user removes the cursor from the area (e.g., until the cursor is no longer within the area defined by the icon).
  • In some implementations, once the preview has been initiated, it continues until the user removes the cursor from a second, larger area. In some implementations, the second area corresponds to a tile, as described above. Once the preview is initiated, then, the user can remove the cursor from the preview area (e.g., a preview icon) and the preview will continue until the cursor is removed from the larger tile area. In some implementations, the second area (e.g., the tile) is contiguous with the first area (e.g., the preview icon). For example, a preview icon may be displayed at least partially within a tile, graphically connected to the tile (e.g., having at least a single common border), or fully within (i.e., completely surrounded by) a tile.
  • In some implementations, when the preview is initiated, additional control elements are displayed to the user. For example, in some implementations, once a user hovers over a preview icon for a sufficient time, the preview starts and one or more playback controls appear with which the user can control the preview. The playback controls may include controls (e.g., icons, graphics, etc.) to skip forward, skip backward, fast forward, rewind, pause, stop, add media content to a playlist, jump to specific track (e.g., track numbers), play now, play next, etc. The controls may be selected in any appropriate way, such as by hovering over the control (e.g., moving the cursor from the preview icon to the control icon), or clicking on the control (e.g., left- or right-clicking). In some implementations, where the media content to be previewed includes multiple media content items, such as where the media content is an album or a playlist, the controls allow the user to skip between the individual content items during the preview. In some implementations, the controls display numbers associated with individual tracks of an album, and selecting a number (e.g., by hovering over the number or clicking on the number) initiates a preview of that particular album track.
  • FIG. 14 illustrates an exemplary user interface 1400, in accordance with some implementations. The user interface 1400 is divided into three sections with dotted lines. The segments may be understood as depicting a portion of the user interface at different times during a user interaction. The user interface 1400 includes tiles, such as tile 1402 in the left-most segment of the user interface 1400. The tiles include information about media content, including artist name, album title, and album art. The tiles include a preview icon, such as preview icon 1404 in the left-most segment of the user interface 1400. As described above, preview icons may define or be associated with areas that, when a cursor is detected therein, cause a preview of the media content associated with the tile to begin.
  • The middle segment of the user interface 1400 illustrates a tile 1411 with a preview icon 1412 before any preview associated with this tile has been activated. The outline of the preview icon 1412 is shown in dashed lines to illustrate the area of the display that is “active” such that if the cursor 1408 is detected within that area, a preview associated with that tile will be presented. The dashed line border of the preview icon 1412 is provided here for illustrative purposes; borders or other graphical features or elements of a preview icon may or may not change during an interaction with a user. (In some implementations, the border of a preview icon and/or a media information tile flashes, moves, is made bold, is animated, changes color, changes brightness, or otherwise changes to grab the user's attention and, in some implementations, indicate that an interaction with the border will cause the device to take some action.) The right-most segment of the user interface 1400 illustrates a tile 1414 after a preview has been initiated in response to detecting the cursor 1408 within the area of the preview icon. The outline of the tile 1414 is shown in dashed lines to illustrate the area of the display that is “active” such that if the cursor 1408 is detected outside that area, the preview that was initiated by detecting the cursor 1408 within the preview icon will cease. Also, the right-most segment includes media control icons 1410 that were displayed in response to detecting the cursor 1408 within the preview icon. In this example, the media control icons 1410 correspond to track numbers of the album associated with the tile 1414. Because the “active” region has been expanded to the border of the tile 1414, the preview will continue even if the user moves the cursor 1408 around within the area of the tile 1414. Thus, the user may move the cursor 1408 to the media control icons to select different tracks to preview, e.g., by clicking on or hovering over them. Other types of controls may also be provided instead of or in addition to the track numbers shown in FIG. 14. For example, as described above, the media control icons may include fast forward, rewind, skip, play, pause, add to playlist, play next, and the like.
  • When a user is browsing or is otherwise presented with media content that can be previewed, it is also beneficial to ensure that the media content can be presented to the user quickly and without delays. Thus, portions of media content that may be previewed by a user can be pre-fetched (e.g., downloaded and buffered/cached) so as to reduce or minimize the delay between the selection of the preview and the presentation of the media content. In some implementations, previews of media content are initiated by a user selection of a control that is displayed to the user. A control, such as a button, icon, graphic, text, etc., can be displayed, for example, in a web page (e.g., a web page associated with a social networking service, such as FACEBOOK), or in a media player/browser application. Moreover, controls may be displayed on any type of device, such as a computer (e.g., laptop, desktop, tablet), mobile phone, etc.
  • In some implementations, media content for a preview is pre-fetched when a control associated with a media content preview is displayed to a user. Accordingly, at least a portion of each media content that could be previewed at a given time—as determined by those controls that are displayed to the user and hence are selectable by the user—is downloaded and buffered/cached so that it can be presented to the user without undue delays if selected. (Where previewable media content includes several individual media content items, such as when the media content is an album or a playlist, pre-fetching a portion of the album can include pre-fetching a part of each track or a subset of tracks of the media content.) In some implementations, when a control ceases to be displayed to the user, then, the pre-fetched media content associated with that control can be deleted or marked for deletion.
  • In the various implementations described above, the various inputs, outputs, and actions described may be performed by one device or combinations of devices. For example, a user may browse media content and select media content for preview on a first electronic device 102-1 (e.g., a laptop computer), and use a second electronic device 106-1 (e.g., a home stereo system) to present the preview or other media content. Or, a user may browse and select media content on one device and also preview the media content on that same device (e.g., a first electronic device 102-1).
  • In implementations where media content is pre-fetched, the media content can be downloaded and/or cached to the device that will present the media content (e.g., the second electronic device 106-1), even if the control that triggers the pre-fetching is displayed at a different device (e.g., the first electronic device 102-1). For example, if a user is browsing media content on a mobile phone while listening to other media content via a home audio system, media content can be downloaded by the home audio system in response to a control being displayed on the mobile phone. In some implementations, the browsing device and the presentation device are in communication with each other so that the browsing device can cause the presentation device to pre-fetch media content associated with controls that are displayed on the browsing device at that time. For example, if a user is presented with a control in a web page or an application on a mobile phone that would allow him to preview the song “American Pie” by Don McLean, the mobile phone can send a request to the home audio system to download a portion of “American Pie.” In some implementations, the mobile phone sends a request to a remote server (discussed below), and the server causes the home audio system to download the portion of the song.
  • In some implementations, the browsing device and the playback device are the same device. When the device detects that a control for initiating a preview is displayed, it can request (or receive without issuing a request) a portion of the media content to be presented if the user selects the control.
  • A remote server (e.g., server system 120, FIG. 1) can determine when a control for initiating a preview of media content is displayed to a user, and send or otherwise enable access to the media content in response. For example, in some implementations, the remote server itself causes the control to be displayed to the user (e.g., by serving a web page or content for a web page to the user). When the remote server detects that the control is sent to the user's device or otherwise displayed to the user, the server can also download a portion of the media content to a device in case the user should initiate presentation of the media content. The device to which the portion of the media content is downloaded depends on several factors. For example, it may be downloaded to a device that is currently presenting other media content to the user. Or, if no device is currently presenting other media content to the user, it may be downloaded to the browsing device. Or it may be downloaded to a device that the user most frequently or most recently used to present media content.
  • In some implementations, the preview control itself is configured to initiate the pre-fetching when it is displayed. For example, a “preview” button in a web page may be configured to cause a download request (e.g., an HTTP request) to be sent to the remote server when the button is displayed in a web browser (or other application). In some implementations, the download request includes a user or account identifier that associates the download request with a particular user or a particular set of devices. In some implementations, the download request is sent to the same server that delivers media content to a user's presentation device (or a server that is associated with the service that delivers media content). Thus, when a download request is received from a user who is already receiving media from the content provider, the server system (or, more generally, a content provider) can deliver the portion of the media content to the same device that that is currently presenting media content. Then, if the user selects the control to initiate presentation of the new media content (e.g., to preview it), the presentation device can simply begin presenting the pre-fetched portion of the content.
  • Various techniques for providing media controls to a user are discussed in commonly owned U.S. application Ser. No. 13/443,781, entitled “Systems and Methods for Controlling a Local Application Through a Web Page,” which is hereby incorporated by reference in its entirety.
  • In some implementations, media content is provided by a remote server (e.g., server system 120). The remote server may be associated with a content provider that provides access to media content such as music, movies, podcasts, audiobooks, etc. In some implementations, the remote server communicates with browsing devices (e.g., mobile phones, computers, or any device on which a user may browse media content and select media content for presentation and/or preview) and with presentation devices (e.g., televisions, home audio systems, computers, mobile phones, or any device that can present media content). Browsing devices, presentation devices, and remote servers, communicate by any appropriate communication standards, protocols, or technologies, such as Internet Protocol (e.g., including Transmission Control Protocol, User Datagram Protocol, Internet Control Message Protocol, Hypertext Transfer Protocol, etc.), BLUETOOTH, Wired Ethernet (e.g., IEEE 802.3), Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.11a, IEEE 802.11b, IEEE 802.11g, and/or IEEE 802.11n), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • Various implementations described above may also be advantageously used to facilitate previewing of individual songs or tracks that are sampled in a currently played song. For example, some songs include small portions of other songs mixed in to create a composite work. This and/or similar techniques are known as “sampling.” In some implementations, a preview control can be displayed to a user when a currently presented song includes samples of other songs. When selected, the preview control initiates a preview of the song (or songs) that are sampled in the currently presented song. The preview control may be presented to the user on any appropriate device, such as a first electronic device 102-1, or a second electronic device 106-1.
  • For example, if the user is listening to the song “Come with Me” by Sean Combs (which includes a sample of the Led Zeppelin song “Kashmir”), a user may be presented with a preview control (e.g., a button or icon, as described above). When selected, the original Led Zeppelin version of the song “Kashmir” is then presented to the user. In some implementations, the various techniques described above for selecting a media content preview, seamlessly integrating the preview into currently presented media, and pre-fetching media content are used in this technique as well.
  • In some implementations, the songs that are sampled in a currently played song are identified using a service or a lookup table that identifies the songs that are sampled in other songs. For example, a server system associated with a media content provider (e.g., server system 120) may access a lookup table to determine if a song that is being presented to a user contains any samples of other songs. If so, original versions of the sampled songs may be made available for preview by the user. The lookup table may be stored and/or maintained by the content provider, or by a third party.
  • FIG. 3 is a block diagram illustrating a server system 120, in accordance with some implementations. The server system 120 typically includes one or more processing units (CPUs) 302, one or more network interfaces 304, memory 306, and one or more communication buses 308 for interconnecting these components.
  • Memory 306 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM, or other random access solid state memory devices; and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Memory 306 may optionally include one or more storage devices remotely located from the CPU(s) 302. Memory 306, or alternately the non-volatile memory device(s) within memory 306, includes a non-transitory computer readable storage medium. In some implementations, memory 306 or the computer readable storage medium of memory 306 stores the following programs, modules and data structures, or a subset thereof:
      • an operating system 310 that includes procedures for handling various basic system services and for performing hardware dependent tasks;
      • a network communication module 312 that is used for connecting the server system 120 to other computers via the one or more communication network interfaces 304 (wired or wireless) and one or more communication networks, such as the Internet, other Wide Area Networks, Local Area Networks, Personal Area Networks, metropolitan area networks, VPNs, and so on;
      • one or more server application module(s) 314 for enabling the server system 120 to perform the functions offered by the server system 120, including but not limited to:
        • a streaming module 122 for streaming media content to an electronic device (e.g., first and second electronic devices 102-n, 106-n, FIG. 1) remote from the server system 120;
        • a request processing module 316 for receiving requests from electronic devices (e.g., first and/or second electronic devices 102-n, 106-n, FIG. 1), wherein the requests include requests to stream specific media content to the electronic devices and/or requests to change the destination of the media content stream (e.g., media content stream 114, FIG. 2) from a first electronic device to a second electronic device;
        • a context tracking module 318 for tracking and storing the context of a media content stream (e.g., media content streams 114-1, 114-2, FIG. 2), including storing, among other data, the current playback position in a media content stream that is currently being presented by an electronic device (e.g., first and/or second electronic devices 102-n, 106-n, FIG. 1), the position in a current playlist, the play history of a user, the preferences of a user, previously skipped media content, whether media content items were “liked” or “disliked” (e.g., via “starred,” “thumbs-up,” and/or “thumbs-down” indications), and the like;
        • a context access module 320 for allowing electronic devices (e.g., first and/or second electronic devices 102-n, 106-n, FIG. 1) associated with a given user account to access the current context for media content streams associated with the given user account; and
      • one or more server data module(s) 330 for storing data related to the data server system 120, including but not limited to:
        • media content database 124 including a library of media content;
        • a context database 126 including information associated with one or more media content streams (e.g., media content streams 114-1, 114-2, FIG. 2), wherein context information includes the current playback position in a media content stream, metadata relating to the media, a position in a playlist, play history of a user, user preferences, skipped media, and user settings; and
        • a user profile database 332 including account information for a plurality of users, each account including user media histories, user preferences, and determined user interests.
  • FIG. 4 is a block diagram illustrating a client system 101, in accordance with some implementations. In some implementations, the client system 101 represents first electronic devices 102-n and/or second electronic devices 106-n. The client system 101 typically includes one or more processing units (CPUs) 402, one or more network interfaces 410, memory 412, and one or more communication buses 414 for interconnecting these components. The client system 101 includes a user interface 404. The user interface 404 includes user interface elements that enable output 406 to be presented to a user, including via speakers or a visual display. The user interface 404 includes user interface components that facilitate user input 408 such as a keyboard, a mouse, a voice-command input unit, a touch sensitive display, or other input buttons 408. In some implementations, the client system 101 is a wireless device, such as a mobile phone. Furthermore, some client systems 101 use a microphone and voice recognition to supplement or replace the keyboard.
  • Memory 412 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM, or other random access solid state memory devices; and may include non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Memory 412 may optionally include one or more storage devices remotely located from the CPU(s) 402. Memory 412, or alternately the non-volatile memory device(s) within memory 412, includes a non-transitory computer readable storage medium. In some implementations, memory 412 or the computer readable storage medium of memory 412 stores the following programs, modules and data structures, or a subset thereof:
      • an operating system 416 that includes procedures for handling various basic system services and for performing hardware dependent tasks;
      • a network communication module 418 that is used for connecting the client system 101 to other computers via the one or more communication network interfaces 410 (wired or wireless) and one or more communication networks, such as the Internet, other Wide Area Networks, Local Area Networks, Personal Area Networks, metropolitan area networks, VPNs, local peer-to-peer and/or ad-hoc connections, and so on;
      • a presentation module 420 for enabling presentation of media content at the client system 101 through the output mechanisms 406 associated with the user interface 404 (e.g., screens, speakers, headphones, etc.);
      • one or more client system applications module(s) 422 for enabling the client system 101 to perform the functions offered by the client system 101, the application module(s) 422 including but not limited to:
        • a media stream reception module 424 for receiving (and, in some implementations, transcoding) a media content stream (e.g., media content stream 114-n, FIG. 2) from a server system (e.g., server system 120, FIG. 2), the media content stream including data needed to present media content at the client system 101 and context data for the media content stream, wherein the media content stream can include a variety of media types including, but not limited to, audio media such as music, radio, audio books, and podcasts, video media such as internet videos, television programming, movies, digital photographs, and any other type of digitally transmissible media (e.g., games, images, advertisements, etc.);
        • a media presentation and control application 104 for presenting media content by the client system 101, for controlling media presentation by other electronic devices, and for interacting with the server system 120;
        • a control request generation module 428 for, in response to user input, generating a server media control request for transmission to the server system (e.g., server media control request 112, FIG. 2) and a local media control request for transmission to the second electronic device (e.g., local media control request 116, FIG. 2) via a Local Area Network, VPN, BLUETOOTH, peer-to-peer, etc.;
        • a webpage display module 430 (including, e.g., a web browser) for displaying a webpage published by a third party;
        • a condition evaluation module 432 for determining whether a predetermined condition is met; in some implementations, the condition evaluation module 432 periodically determines whether any predetermined condition has been met; and
        • an automatic request generation module 434 for, in response to a determination by the condition evaluation module 432 that a predetermined condition is met, generating a request for transmission to the server system without any direct user input;
      • an electronic device data module 440 for storing data, including but not limited to:
        • media content buffer(s) 105 including media content data received from a stream from the server system (e.g., server system 120, FIG. 2) and stored in the memory of the client system 101 until it is presented by the client system 101;
        • predetermined condition database 442 including, but not limited to, one or more predetermined conditions, established by the user of the client system 101, that determine whether a media control request to change the destination of the media content stream (e.g., media content stream 114-1, FIG. 2) from the server system should be automatically generated;
        • media content database 444 for storing, on the local device, media content that is part of the user's personal library of media content; and
        • a user profile database 446 including account information for a specific user of the client system 101 including user media history, user preferences, determined user interests, and account credentials.
  • FIGS. 5-8 are flow diagrams illustrating a process of controlling media presentation at an electronic device, in accordance with some implementations. Each of the operations shown in FIGS. 5-8 may correspond to instructions stored in a computer memory or computer readable storage medium. Optional operations are indicated by dashed lines (e.g., boxes with dashed-line borders).
  • In some implementations, the first electronic device corresponds to the first electronic device 102-1, discussed above (also described with respect to FIG. 4). A media control command for a second electronic device is received (502). In some implementations, the media control command corresponds to a user input. In some implementations, the user input is received via user interface 404 of the first electronic device 102-1. For example, the user may select a song for playback by pressing a representation of the song (e.g., an icon, title, etc.) on a touchscreen of a mobile phone. In some implementations, the first electronic device is selected from the group consisting of: a computer, a mobile phone, a remote control device, a portable media player, and the like. In some implementations, the second electronic device is selected from the group consisting of: a television, a home media receiver/player, a computer, a home audio/visual system, a speaker, and the like. In some implementations, any of the identified devices, and other devices not explicitly listed, may be a first electronic device or a second electronic device.
  • In response to receiving the media control command, a server media control request is sent to a server system (504). In some implementations, the server system corresponds to the server system 120, discussed above. In some implementations, the server media control request is sent to the server system over the Internet (505). In some implementations, the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected (506).
  • Also in response to receiving the media control command, a local media control request is sent to a second electronic device within a local network to which both the first electronic device and the second electronic device are connected (508). In some implementations, the local network is a Local Area Network, as discussed above. In some implementations, the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork (510). The local media control request is sent over any communication type that allows communication between the two electronic devices without having to leave the local network. In some implementations, the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • In some implementations, the server and the local media control requests are both configured to cause a single media control operation to be implemented at the second electronic device. For example, if a user inputs a media control command into a mobile phone requesting that the song “Jessie's Girl” be presented by a second electronic device, such as a home media system, the mobile phone will send two media control requests that are each configured to cause the home media system to begin playback of “Jessie's Girl.” Sending both media control requests increases the speed with which the second electronic device responds to the request and the reliability of the request being received. In some implementations, the second electronic device has already buffered some of the requested media content and can therefore begin presenting that content without needing to wait for the arrival of a media content stream from the server system, as discussed below. Also, the local and server media control requests need not contain identical data, but they are configured to cause the same media control operation to occur at the second electronic device. In some implementations, the media control operation is selected from the group consisting of: play, pause, skip, fast-forward, rewind, adjust an audio volume, change an order of items in a playlist, add items to a playlist, remove items from a playlist, adjust audio equalizer settings, set a user setting, and the like. In some implementations, the local and server media control requests include information specifying a particular electronic device from a group of available electronic devices to be controlled as the second electronic device.
  • In some implementations, the first electronic device provides a single user interface that allows a user both to select media content for presentation by the first electronic device and to generate media control requests configured to cause the media content to be presented by the second electronic device (512). For example, as described above, a user can control playback of media content at the device that is presenting the user interface (e.g., a mobile phone that is displaying a media player application) and also control playback of media content at a different device (e.g., a home media system or remote speaker) using the same application and/or user interface. Thus, the user need not switch between applications or user interface environments in order to control the various devices, including the device on which the user interface is presented.
  • In some implementations, the media content is selected from the group consisting of, but not limited to: an audio track, a video, an image, an audio playlist, and a video playlist. In some implementations, when the media content is a video, the video is presented on a display associated with the second electronic device (514). In some implementations, when the media content is an audio track, the audio track is presented on a speaker associated with the second electronic device (516).
  • FIGS. 6A-6B are flow diagrams illustrating a method 600 of controlling media presentation at an electronic device, in accordance with some implementations. In some implementations, the method 600 is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors (e.g., second electronic device 106-n, FIG. 1). For example, the second electronic device may be a home media system, television, computer, or the like. In some implementations, the method 600 corresponds to a method implemented in conjunction with and/or in response to the method 500.
  • In some implementations, the method 600 corresponds to an implementation where media control requests that are sent to the server system from the first electronic device (e.g., the server media control request 112, FIG. 2) are forwarded to the second electronic device without the server system attempting to perform the media control operation. Thus, the server system acts as a relay or forwarding service for a media control request from the first electronic device. This may be used in implementations where a media control request is only performed if the request originates from the device that is to perform the action (or that is to be impacted by the action). For example, a device acting as a remote control cannot alone cause the server to begin streaming media to a second electronic device, because the control request did not originate from the second electronic device. Rather, the second electronic device must itself make the request to the server. Accordingly, the server forwards the media control request to the second electronic device, and the second electronic device can then perform an operation to satisfy the request (e.g., requesting media content from the server, changing a media presentation setting, etc.).
  • With reference to FIG. 6A, a server media control request corresponding to a media control operation to be implemented at the second electronic device is received from a server system, wherein the server media control request was sent from the server in response to the server receiving the server media control request from a first electronic device (602). In some implementations, the server media control request is sent from the server system to the second electronic device over the Internet (603). In some implementations, the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected (604).
  • A local media control request sent within a local network to which both the first electronic device and the second electronic device are connected is received from the first electronic device, wherein the local media control request corresponds to the media control operation (606). In some implementations, the local network is a Local Area Network, as discussed above. In some implementations, the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork (608). The local media control request is sent over any communication type that allows peer-to-peer communication between the two electronic devices without having to leave the local network. In some implementations, the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • Because the second electronic device may receive the server media control request and the local media control request from different sources and via different communication paths, the requests may not arrive in a consistent or predictable order. That is, in some cases, the local media control request may arrive before the server media control request, while in other cases, the order is reversed. However, a subsequently received media control request may not be a duplicate of the prior request, and so it should not be ignored simply because it was received shortly after another request. Accordingly, in some implementations, the second electronic device will determine whether to perform or ignore the media control requests based on various factors, as described with respect to FIG. 6B.
  • Method 600 continues on FIG. 6B. In some implementations, when the server media control request is received prior to the local media control request, the following steps are performed. The media control operation is performed in response to receiving a server media control request (610). After receiving a local media control request, it is determined whether the server media control request and the local media control request correspond to the same media control operation (612). If the server media control request and the local media control request correspond to the same media control operation, the local media control request is ignored (614). In some implementations, if the server media control request and the local media control request correspond to different media control operations, the local media control request is performed (616). In cases where the local media control request is received prior to the server media control request, steps similar to (610)-(616) are employed to determine whether to ignore or perform the server media control request. Thus, the second electronic device is prevented from performing the same media control operation twice in succession, thus preventing possible pauses, jumps, or other interruptions in the media content presentation caused by the processing of both the local and the server media control requests.
  • In some implementations, whether the server media control request and the local media control request correspond to the same media control operation is determined using timestamps. For example, in some implementations, the server media control request includes a first timestamp and the local media control request includes a second timestamp. In some implementations, the first and the second timestamp both correspond to a time at which the first electronic device issued the media control requests. In some implementations, the first timestamp corresponds to a time at which the server system received a server media control request from the first electronic device, and the second timestamp corresponds to a time at which the first electronic device issued the local media control request. Accordingly, determining that the server media control request and the local media control request correspond to the same media control operation includes comparing the first and the second timestamps. If the timestamps are the same or substantially similar, it is determined that the local media control request and the server media control request correspond to the same media control operation, and the local media control request is ignored. In some implementations, timestamps are substantially similar if they are within 1 second. In some implementations, they are substantially similar if they are within 5 seconds. Other times may be used depending on the needs of any specific implementation. In some implementations, other time ranges are used to determine if timestamps are considered substantially similar. If the timestamps are not the same or substantially similar, it is determined that server media control request and the local media control request do not correspond to the same media control operation, and the local media control request is performed.
  • As discussed above, it may be desirable for users to cause media content that is being presented at one device to be transferred to another device for presentation at that device. For example, a user listening to music on a mobile phone may decide to switch the playback device from the phone to a home stereo system. In the process of terminating presentation at one device and initiating presentation at another, however, processing and communication delays can cause pauses or gaps in presentation that diminish the overall user experience. Accordingly, in some implementations, the second electronic device (e.g., the device that is to begin presenting media content) buffers an upcoming portion of media content that is being presented at the first electronic device (e.g., the device that is currently presenting the media content). In some implementations, the second electronic device is configured to buffer the upcoming portion of media content that is being presented at the first electronic device when the first electronic device is connected to the same local network (e.g., a LAN, a BLUETOOTH connection, etc.), when the first electronic device is within a predetermined proximity to the second electronic device (e.g., within the same building, room, floor, etc., as determined by a GPS, for example), and the like. In some implementations, one or more of the devices, including the first and second electronic devices and the server system, are configured to determine when and whether to buffer/cache media content at the second electronic device. Buffering media content at a second electronic device is discussed further with reference to steps (617)-(622).
  • In some implementations, the media control operation corresponding to one or more media control requests is a request to begin presenting media content at the second electronic device while the media content is being presented by the first electronic device (617). Prior to receiving the server media control request or the local media control request, an upcoming portion of the media content being presented by the first electronic device is buffered at the second electronic device (618). For example, the second electronic device may continuously receive and store a portion of the media content that is being presented at the first electronic device. In some implementations, the second electronic device receives the media content for buffering from the server system. The media content may be sent from the server system via the Internet. In some implementations, the second electronic device receives the media content for buffering from the first electronic device. In these implementations, the media content may be sent from the first electronic device via the local network. In some implementations, the buffered media content has a lower bit-rate than media content that is being presented and/or streamed for current presentation, thus helping to reduce bandwidth and network usage for the buffering technique. In some implementations, where the media content is audio, it is presented at a bit-rate of either about 160 kbps or 320 kbps, and it is buffered at about 96 kbps. Other bit-rates may be used for the presentation bit-rate and the buffering bit-rate, depending on factors such as available network speeds and capabilities, distances between devices, device capabilities, user preferences, and the like.
  • In some implementations, one of the server media control request or the local media control request is received at the second electronic device (620), and, in response, presentation of the buffered portion of the media content is initiated at the second electronic device (622).
  • Accordingly, when the second electronic device receives a command to begin presenting the media content, the presentation can begin without the additional time necessary to initialize a media content stream or otherwise download or receive the media content in a way suitable for continuous presentation. In some implementations, the second electronic device buffers about 5 seconds of the media content that is being presented by the first electronic device. In some implementations, it buffers about 10 seconds of the media content. In some implementations, it buffers more or less time, depending on the needs of any specific implementation.
  • In some implementations, the media content continues to be presented by the first electronic device in substantial synchronization with the presentation by the second electronic device. In some implementations, the media content ceases to be presented by the first electronic device when presentation begins at the second electronic device (although it may still be cached/buffered at the first electronic device, in some implementations).
  • FIG. 7 is a flow diagram illustrating a method 700 of controlling media presentation at an electronic device in accordance with some implementations. In some implementations, the method 700 is performed at a second electronic device having one or more processors and memory storing one or more programs for execution by the one or more processors. For example, the second electronic device may be a home media system, television, computer, or the like. In some implementations, the method 700 corresponds to a method implemented in conjunction with and/or in response to the method 500. In some implementations, the method 700 corresponds to an implementation where media control requests that are sent to the server system (e.g., server system 120, FIG. 2) from the first electronic device (e.g., the server media control request 112, FIG. 2) are performed without first being forwarded to the second electronic device. Thus, the server system will take action when it receives a control request, even if the request originated from a different device than that which is to be acted upon. In some implementations, local media control requests (e.g., the local media control request 116, FIG. 2) are also received by the second electronic device after having been sent from the first electronic device, and are processed by the second electronic device when received, if appropriate.
  • With reference to FIG. 7, a local media control request corresponding to a media control operation is received from a first electronic device, wherein the local media control request is sent within a local network to which both the first electronic device and the second electronic device are connected (702). In some implementations, the local network is a Local Area Network, as discussed above. In some implementations, the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork (703). The local media control request is sent using any communication type that allows communication between the two electronic devices without having to leave the local network. In some implementations, the local media control request is sent via Wi-Fi, BLUETOOTH, or the like.
  • A media stream is received from a server system, wherein the media stream was sent from the server in response to the server receiving a server media control request from the first electronic device, and wherein the server media control request corresponds to the media control operation (704). Accordingly, in this case, when the first electronic device requests a media control operation (e.g., initiated by a user or automatically), the second electronic device receives a local media control request, as well as a media stream that corresponds to that same media control request.
  • In some implementations, the media stream is sent from the server system to the second electronic device over the Internet (705). In some implementations, the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected (706).
  • Because the second electronic device receives both a media stream and a local media control request that may correspond to the same media control operation, the second electronic device should determine whether to perform a subsequently received local media control request. Accordingly, in some implementations, when the media stream is received prior to receiving the local media control request, the second electronic device determines whether to perform or ignore the local media control request. In some implementations, in response to a determination that the local media control request corresponds to a request to initiate presentation of the media stream, the local media control request is ignored (708). In some implementations, in response to a determination that the local media control request does not correspond to the request to initiate presentation of the media stream, the local media control request is performed (710). Thus, the second electronic device is prevented from performing the same media control operation twice in succession, thus preventing possible pauses, jumps, or other interruptions in the media content presentation.
  • In some implementations, whether the local media control request corresponds to a request to initiate presentation of the media stream is determined using timestamps. For example, in some implementations, the media stream includes a first timestamp, and the local media control request includes a second timestamp. In some implementations, the first and the second timestamp both correspond to a time at which the first electronic device issued the media control requests. In some implementations, the first timestamp corresponds to a time at which the server system received a media control request from the first electronic device, and the second timestamp corresponds to a time at which the first electronic device issued the local media control request. Accordingly, determining that the local media control request corresponds to a request to initiate presentation of the media stream includes comparing the first and the second timestamps. If the timestamps are the same or substantially similar, it is determined that the local media control request corresponds to the media stream, and the local media control request is ignored. In some implementations, timestamps are substantially similar if they are within 1 second. In some implementations, they are substantially similar if they are within 5 seconds. In some implementations, other time ranges are used to determine if timestamps are considered substantially similar. If the timestamps are not the same or substantially similar, it is determined that the local media control request corresponds to the media stream, and the local media control request is performed.
  • FIG. 8 is a flow diagram illustrating a method 800 of controlling media presentation, in accordance with some implementations. In some implementations, the method 800 is performed at a server system having one or more processors and memory storing one or more programs for execution by the one or more processors (e.g., server system 120, FIGS. 1-3).
  • A server media control request is received from a first electronic device (802). The first electronic device also sends a local media control request to the second electronic device within a local network to which both the first electronic device and the second electronic device are connected. The server media control request and the local media control request are both configured to initiate a same media control operation by the second electronic device. As described above, in some implementations, the server media control request is sent to the server system over the Internet (803). In some implementations, the server system is associated with an IP address outside an IP address space of a local network to which both the first electronic device and the second electronic device are connected (804). In some implementations, the local network is a Local Area Network. In some implementations, the first electronic device and the second electronic device are associated with IP addresses within the same subnetwork (805).
  • At least one of the server media control request or a media stream corresponding to the server media control request is sent to the second electronic device (806). In some cases, the server forwards all the server media control requests to the second electronic device. In some cases, the server performs the media control operation requested by the server media control request. For example, if the server media control request requests playback of media content at the second electronic device (e.g., a home media system), the server streams the media content to the second electronic device. In some implementations, the server performs some media control requests, but forwards others to the second electronic device. For example, requests to control playback of media content (including commands such as play, pause, skip, repeat, etc.) are performed by the server, while requests that relate specifically to the second electronic device (e.g., change the volume, change an input source, etc.) are forwarded to the second electronic device.
  • FIGS. 9-12 are flow diagrams illustrating a method for previewing media content, in accordance with some implementations. Each of the operations shown in FIGS. 9-12 may correspond to instructions stored in a computer memory or computer readable storage medium. Optional operations are indicated by dashed lines (e.g., boxes with dashed-line borders). Moreover, each of the operations shown in FIGS. 9-12 may be executed by any appropriate device or combination of devices, including first electronic devices 102-n, second electronic devices 106-n, and/or a server system 120.
  • With reference to FIG. 9, in some implementations, the method 900 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors. In some implementations, the electronic device corresponds to the first electronic device 102-1, discussed above. A first input is detected (902). In some implementations, the first input is a touch input including a contact with a touch-sensitive surface. In some implementations, the first input is a mouse click-down event.
  • In response to determining that a change to the first input has been detected within a first time period after detecting the first input, first media content is added to a playlist (904). In some implementations, the first input is detected (902) during presentation of second media content, and the playlist includes the second media content. Thus, in these implementations, the change to the first input causes the first media content to be added to a playlist that is currently being presented. In some implementations, adding the first media content to the playlist comprises adding the first media content to an end of the playlist. In some implementations, adding the first media content to the playlist comprises adding the first media content as next media content in the playlist. In some implementations, adding the first media content to the playlist comprises replacing existing content in the playlist with the first media content.
  • In some implementations, the change to the first input is a liftoff from a touch-sensitive surface, such as a touchscreen. In some implementations, the change to the first input is an additional contact with the touch-sensitive surface, wherein the contact and the additional contact correspond to distinct contact areas of the touch-sensitive surface. In some implementations, the change to the first input is a mouse click-release event. In some implementations, the change to the first input is a gesture, swipe, cursor movement, or multi-touch input.
  • In some implementations, the first time period is 0.5 seconds. In some implementations, the first time period is 1 second. In some implementations, the first time period is 2 seconds.
  • In response to determining that a change to the first input has not been detected within the first time period after detecting the first input, the first media content is caused to be presented to the user (906). In some implementations, causing the first media content to be presented to the user includes presenting the media content on the same device that received the first input. In some implementations, it includes sending a request to another device to initiate presentation of the media content.
  • In some implementations, causing the first media content to be presented includes reducing a volume of the second media content while the first media content is presented (908). In some implementations, reducing the volume includes muting the second media content. In some implementations, reducing the volume includes pausing the second media content. In some implementations, causing the first media content to be presented also includes ceasing presentation of the first media content in response to detecting a change to the first input after the first time period.
  • In some implementations, the first media content comprises a plurality of media content items. In such instances, a directional component of the first input can be used to navigate among individual media content items of the first media content. For example, in some implementations, the first media content is a playlist, album, listing of podcasts, or the like. In some implementations, the method further includes detecting a direction of the first input (910). A next media content item of the plurality of media content items is caused to be presented in response to detecting that that the direction is a first direction (912). In some implementations, a previous media content item of the plurality of media content items is caused to be presented in response to determining that the direction is a second direction (914). In some implementations, the first direction corresponds to an input movement having a predetermined directional component in a right-hand direction. In some implementations, the second direction corresponds to an input movement having a predetermined directional component in a left-hand direction. In some implementations, the first direction corresponds to an input movement having a predetermined directional component in a down direction. In some implementations, the second direction corresponds to an input movement having a predetermined directional component in an up direction.
  • With reference to FIG. 10, in some implementations, the method 1000 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors. In some implementations, the electronic device corresponds to the first electronic device 102-1, discussed above. A first input is detected during presentation of a first song (1002). Various inputs that may be detected are described above. The audibility of the first song is reduced (1004). In some implementations, reducing the audibility of the first song includes reducing the volume of the first song; in some embodiments it includes muting the first song; in some implementations, it includes pausing the first song. In some implementations, reducing the audibility of the first song includes sending a command to a remote device, where the command is configured to cause that device to reduce the audibility of the first song.
  • A second song is caused to be presented (1006). In some implementations, causing the second song to be presented includes beginning playback of the second song at the device. In some implementations, causing the second song to be presented includes sending a request to a remote device to initiate playback of the second song.
  • The method also includes, after a first time period, causing the presentation of the second song to be ceased (1008), and causing the audibility of the first song to be increased (1010). In some implementations, the first time period corresponds to the duration of a maintained input. In some implementations, the end of the first time period corresponds to an expiration of a timer. In some implementations, the end of the first time period corresponds to an end of a preview portion of the second song. In some implementations, the end of the first time period corresponds to detecting a change to the first input, such as a liftoff from a touch-sensitive surface or a mouse click-release event.
  • With reference to FIG. 11, in some implementations, the method 1100 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors. In some implementations, the electronic device corresponds to the first electronic device 102-1, discussed above. During presentation of currently presented media content, an affordance is displayed to a user, wherein selection of the affordance will cause secondary media content to be presented (1102). In some implementations, the affordance corresponds to an icon (e.g., a thumbnail image of an album cover or movie poster), graphic, or text. In response to displaying the affordance, at least a portion of secondary media content is obtained (1104). In some implementations, the portion of the secondary media content is received by the same device that displayed the affordance. In some implementations, it is received by a different device, such as a second electronic device 106-n. In some implementations, it is sent by a server system remote from the device that displayed the affordance, such as server system 120.
  • In some implementations, the secondary media content includes multiple media content items, and obtaining at least a portion of the secondary media content includes obtaining at least a portion of each media content item of the secondary media content. For example, if the secondary content corresponds to an album, at least a part of each track is obtained for caching/buffering by a presentation device. In some implementations, obtaining at least a portion of the secondary media content includes obtaining an initial media content item and the next media content item in the playlist. In some implementations, obtaining at least a portion of the secondary media content includes obtaining an initial media content item and the next two media content items in the playlist.
  • A first input corresponding to a selection of the affordance is detected (1106). In some implementations, the first input corresponds to any of those inputs discussed above (e.g., mouse clicks or hovers, touch events and gestures, etc.). The audibility of the currently presented media content is reduced (1108). In some implementations, this includes muting, lowering the volume or brightness of, and/or pausing the currently presented media content. The secondary media content is presented (1110). After a first time period, the presentation of the secondary media content is ceased (1112), and the audibility of the currently presented media content is increased (1114). In some implementations, the first time period corresponds to the duration of a maintained input. In some implementations, the end of the first time period corresponds to an expiration of a timer. In some implementations, the end of the first time period corresponds to an end of a preview portion of the second song. In some implementations, the end of the first time period corresponds to detecting a change to the first input, such as a liftoff from a touch-sensitive surface or a mouse click-release event.
  • With reference to FIG. 12, in some implementations, the method 1200 is performed at a first electronic device having one or more processors and memory storing instructions for execution by the one or more processors. In some implementations, the electronic device corresponds to the first electronic device 102-1, discussed above. The presence of a cursor is detected in a first area of a display (1202). In some implementations, the first area of the display corresponds to an icon (e.g., a thumbnail image of an album cover or movie poster), graphic, or text. In some implementations, the first area is displayed partially or entirely within, adjacent to, and/or contiguous with a second area of the display region. In some implementations, the second area of the display region corresponds to a media content description area (e.g., a tile, as discussed above), including, for example, album art, movie stills, content metadata, artist information, artist photographs, and the like.
  • Media content is presented in response to detecting the presence of the cursor within the first area (1204). In some implementations, the media content corresponds to a preview of the media content identified and/or described in the media content description area. In some implementations, the media content is presented after the cursor has been detected in the first area for a time period, such as 0.5 seconds, 1 second, 2 seconds, or any other appropriate time period.
  • The removal of the cursor from a second area of the display region is detected (1206), where the second area is larger than and contiguous with the first area. In some implementations, the second area shares at least one border with the first area.
  • In some implementations, the second area completely encompasses the first area on all sides.
  • Presentation of the media content is ceased in response to detecting the removal of the cursor from the second area (1208). Thus, in some implementations, after the media content is presented—in response to detecting the cursor in the first area—it will continue to be presented until the cursor is removed from the larger second area.
  • In some implementations, in response to detecting the presence of the cursor in the first area of the display, one or more additional icons are displayed within the second area of the display (1210). In some implementations the one or more additional icons include media playback controls. In some implementations, the media playback controls are selected from the group consisting of: play, pause, stop, skip forward, skip backward, add to playlist (e.g., add as next, add to end, replace playlist), and audio track indices.
  • The methods illustrated in FIGS. 5-12 may be governed by instructions that are stored in a computer readable storage medium and that are executed by at least one processor of at least one server. Each of the operations shown in FIGS. 5-12 may correspond to instructions stored in a non-transitory computer memory or computer readable storage medium. In various implementations, the non-transitory computer readable storage medium includes a magnetic or optical disk storage device, solid state storage devices, such as Flash memory, or other non-volatile memory device or devices. The computer readable instructions stored on the non-transitory computer readable storage medium may be in source code, assembly language code, object code, or other instruction format that is interpreted and/or executable by one or more processors.
  • Plural instances may be provided for components, operations, or structures described herein as a single instance. Finally, boundaries between various components, operations, and data stores are somewhat arbitrary, and particular operations are illustrated in the context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within the scope of the implementation(s). In general, structures and functionality presented as separate components in the example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the implementation(s).
  • It will also be understood that, although the terms “first,” “second,” etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first contact could be termed a second contact, and, similarly, a second contact could be termed a first contact, which changing the meaning of the description, so long as all occurrences of the “first contact” are renamed consistently and all occurrences of the second contact are renamed consistently. The first contact and the second contact are both contacts, but they are not the same contact.
  • The terminology used herein is for the purpose of describing particular implementations only and is not intended to be limiting of the claims. As used in the description of the implementations and the appended claims, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
  • As used herein, the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting,” that a stated condition precedent is true, depending on the context. Similarly, the phrase “if it is determined (that a stated condition precedent is true)” or “if (a stated condition precedent is true)” or “when (a stated condition precedent is true)” may be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.
  • The foregoing description included example systems, methods, techniques, instruction sequences, and computing machine program products that embody illustrative implementations. For purposes of explanation, numerous specific details were set forth in order to provide an understanding of various implementations of the inventive subject matter. It will be evident, however, to those skilled in the art that implementations of the inventive subject matter may be practiced without these specific details. In general, well-known instruction instances, protocols, structures and techniques have not been shown in detail.
  • The foregoing description, for purpose of explanation, has been described with reference to specific implementations. However, the illustrative discussions above are not intended to be exhaustive or to limit the implementations to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The implementations were chosen and described in order to best explain the principles and their practical applications, to thereby enable others skilled in the art to best utilize the implementations and various implementations with various modifications as are suited to the particular use contemplated.

Claims (28)

What is claimed is:
1. A server computer, comprising:
one or more processors;
memory; and
one or more programs stored in the memory, the one or more programs including instructions that, when transmitted to a client computer and executed by the client computer, cause the client computer to:
during presentation of first media content, display an affordance to a user, wherein the affordance is configured such that selection of the affordance will cause second media content to be presented;
during presentation of the first media content, detect a first input corresponding to a selection of the affordance; and
in response to detecting the first input:
cause the audibility of the first media content to be reduced;
cause the second media content to be presented; and
at an end of a first time period:
cause the presentation of the second media content to be ceased; and
cause the audibility of the first media content to be increased.
2. The server computer of claim 1, the one or more programs further including instructions that, when transmitted to a client computer and executed by the client computer, cause the client computer to, in response to displaying the affordance, obtain at least a portion of the second media content.
3. The server computer of claim 1, wherein the end of the first time period corresponds to the earlier of:
a determination that a cursor is detected outside of a display region associated with the second media content;
a determination that a change to the first input is detected after a certain time period has passed since detecting the first input; and
a determination that a preview duration is expired.
4. The server computer of claim 3, wherein the first input corresponds to a click-down event.
5. The server computer of claim 4, wherein the change to the first input corresponds to a click-release event.
6. The server computer of claim 1, wherein the instructions to cause the audibility of the first media content to be reduced include instructions for pausing the first media content.
7. The server computer of claim 6, wherein the instructions to cause the audibility of the first media content to be increased include instructions for unpausing the first media content.
8. A non-transitory computer readable storage medium storing one or more programs for execution by an electronic device having at least one processor and memory, the one or more programs including instructions for:
during presentation of first media content, displaying an affordance to a user, wherein the affordance is configured such that selection of the affordance will cause second media content to be presented;
during presentation of the first media content, detecting a first input corresponding to a selection of the affordance;
in response to detecting the first input:
causing the audibility of the first media content to be reduced;
causing the second media content to be presented; and
at an end of a first time period:
causing the presentation of the second media content to be ceased; and
causing the audibility of the first media content to be increased.
9. The non-transitory computer readable storage medium of claim 8, the one or more programs further including instructions for, in response to displaying the affordance, obtaining at least a portion of the second media content.
10. The non-transitory computer readable storage medium of claim 8, wherein the end of the first time period corresponds to the earlier of:
a determination that a cursor is detected outside of a display region associated with the second media content;
a determination that a change to the first input is detected after a certain time period has passed since detecting the first input; and
a determination that a preview duration is expired.
11. The non-transitory computer readable storage medium of claim 10, wherein the first input corresponds to a click-down event.
12. The non-transitory computer readable storage medium of claim 11, wherein the change to the first input corresponds to a click-release event.
13. The non-transitory computer readable storage medium of claim 8, wherein the instructions for causing the audibility of the first media content to be reduced include instructions for pausing the first media content.
14. The non-transitory computer readable storage medium of claim 13, wherein the instructions for causing the audibility of the first media content to be increased include instructions for unpausing the first media content.
15. A method for previewing media content, comprising:
at an electronic device with one or more processors and memory storing one or more programs for execution by the one or more processors:
during presentation of first media content, displaying an affordance to a user, wherein the affordance is configured such that selection of the affordance will cause second media content to be presented;
during presentation of the first media content, detecting a first input corresponding to a selection of the affordance;
in response to detecting the first input:
causing the audibility of the first media content to be reduced;
causing the second media content to be presented; and
at an end of a first time period:
causing the presentation of the second media content to be ceased; and
causing the audibility of the first media content to be increased.
16. The method of claim 15, further comprising, in response to displaying the affordance, obtaining at least a portion of the second media content.
17. The method of claim 15, wherein the end of the first time period corresponds to the earlier of:
a determination that a cursor is detected outside of a display region associated with the second media content;
a determination that a change to the first input is detected after a certain time period has passed since detecting the first input; and
a determination that a preview duration is expired.
18. The method of claim 17, wherein the first input corresponds to a click-down event.
19. The method of claim 18, wherein the change to the first input corresponds to a click-release event.
20. The method of claim 15, wherein causing the audibility of the first media content to be reduced includes pausing the first media content.
21. The method of claim 20, wherein causing the audibility of the first media content to be increased including unpausing the first media content.
22. An electronic device, comprising:
a display;
one or more processors;
memory; and
one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for:
during presentation of first media content, displaying an affordance to a user, wherein the affordance is configured such that selection of the affordance will cause second media content to be presented;
during presentation of the first media content, detecting a first input corresponding to a selection of the affordance;
in response to detecting the first input:
causing the audibility of the first media content to be reduced;
causing the second media content to be presented; and
at an end of a first time period:
causing the presentation of the second media content to be ceased; and
causing the audibility of the first media content to be increased.
23. The electronic device of claim 22, the one or more programs further including instructions for, in response to displaying the affordance, obtaining at least a portion of the second media content.
24. The electronic device of claim 22, wherein the end of the first time period corresponds to the earlier of:
a determination that a cursor is detected outside of a display region associated with the second media content;
a determination that a change to the first input is detected after a certain time period has passed since detecting the first input; and
a determination that a preview duration is expired.
25. The electronic device of claim 24, wherein the first input corresponds to a click-down event.
26. The electronic device of claim 25, wherein the change to the first input corresponds to a click-release event.
27. The electronic device of claim 22, the instructions for causing the audibility of the first media content to be reduced including instructions for pausing the first media content.
28. The electronic device of claim 27, the instructions for causing the audibility of the first media content to be increased including instructions for unpausing the first media content.
US14/231,203 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback Abandoned US20140215334A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/231,203 US20140215334A1 (en) 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201261713444P 2012-10-12 2012-10-12
US14/053,443 US9246967B2 (en) 2012-10-12 2013-10-14 Systems, methods, and user interfaces for previewing media content
US14/231,203 US20140215334A1 (en) 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/053,443 Continuation US9246967B2 (en) 2012-10-12 2013-10-14 Systems, methods, and user interfaces for previewing media content

Publications (1)

Publication Number Publication Date
US20140215334A1 true US20140215334A1 (en) 2014-07-31

Family

ID=50476605

Family Applications (4)

Application Number Title Priority Date Filing Date
US14/053,443 Active 2033-12-26 US9246967B2 (en) 2012-10-12 2013-10-14 Systems, methods, and user interfaces for previewing media content
US14/231,170 Abandoned US20140214927A1 (en) 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback
US14/231,203 Abandoned US20140215334A1 (en) 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback
US15/006,816 Abandoned US20160313872A1 (en) 2012-10-12 2016-01-26 Systems, methods, and user interfaces for previewing media content

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US14/053,443 Active 2033-12-26 US9246967B2 (en) 2012-10-12 2013-10-14 Systems, methods, and user interfaces for previewing media content
US14/231,170 Abandoned US20140214927A1 (en) 2012-10-12 2014-03-31 Systems and methods for multi-context media control and playback

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/006,816 Abandoned US20160313872A1 (en) 2012-10-12 2016-01-26 Systems, methods, and user interfaces for previewing media content

Country Status (3)

Country Link
US (4) US9246967B2 (en)
EP (2) EP3151576A1 (en)
WO (1) WO2014057356A2 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140006483A1 (en) * 2012-06-29 2014-01-02 Spotify Ab Systems and methods for multi-context media control and playback
US20140122589A1 (en) * 2012-11-01 2014-05-01 Research In Motion Limited System and method of transferring control of media playback between electronic devices
US9798514B2 (en) * 2016-03-09 2017-10-24 Spotify Ab System and method for color beat display in a media content environment
US10248190B2 (en) 2015-05-19 2019-04-02 Spotify Ab Multi-track playback of media content during repetitive motion activities
US10467998B2 (en) 2015-09-29 2019-11-05 Amper Music, Inc. Automated music composition and generation system for spotting digital media objects and event markers using emotion-type, style-type, timing-type and accent-type musical experience descriptors that characterize the digital music to be automatically composed and generated by the system
US10620797B2 (en) 2012-06-29 2020-04-14 Spotify Ab Systems and methods for multi-context media control and playback
US10854180B2 (en) 2015-09-29 2020-12-01 Amper Music, Inc. Method of and system for controlling the qualities of musical energy embodied in and expressed by digital music to be automatically composed and generated by an automated music composition and generation engine
US10964299B1 (en) 2019-10-15 2021-03-30 Shutterstock, Inc. Method of and system for automatically generating digital performances of music compositions using notes selected from virtual musical instruments based on the music-theoretic states of the music compositions
US11024275B2 (en) 2019-10-15 2021-06-01 Shutterstock, Inc. Method of digitally performing a music composition using virtual musical instruments having performance logic executing within a virtual musical instrument (VMI) library management system
US11037538B2 (en) 2019-10-15 2021-06-15 Shutterstock, Inc. Method of and system for automated musical arrangement and musical instrument performance style transformation supported within an automated music performance system
US11190561B2 (en) * 2018-01-26 2021-11-30 Citrix Systems, Inc. Virtual computing system providing local screen sharing from hosted collaboration applications and related methods
US11481744B2 (en) * 2013-09-30 2022-10-25 Sonos, Inc. Discovery and media control at a point-of-sale display

Families Citing this family (101)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11294618B2 (en) 2003-07-28 2022-04-05 Sonos, Inc. Media player system
US11106424B2 (en) 2003-07-28 2021-08-31 Sonos, Inc. Synchronizing operations among a plurality of independently clocked digital data processing devices
US8086752B2 (en) 2006-11-22 2011-12-27 Sonos, Inc. Systems and methods for synchronizing operations among a plurality of independently clocked digital data processing devices that independently source digital data
US11106425B2 (en) 2003-07-28 2021-08-31 Sonos, Inc. Synchronizing operations among a plurality of independently clocked digital data processing devices
US8234395B2 (en) 2003-07-28 2012-07-31 Sonos, Inc. System and method for synchronizing operations among a plurality of independently clocked digital data processing devices
US8290603B1 (en) 2004-06-05 2012-10-16 Sonos, Inc. User interfaces for controlling and manipulating groupings in a multi-zone media system
US11650784B2 (en) 2003-07-28 2023-05-16 Sonos, Inc. Adjusting volume levels
US9207905B2 (en) 2003-07-28 2015-12-08 Sonos, Inc. Method and apparatus for providing synchrony group status information
US9374607B2 (en) 2012-06-26 2016-06-21 Sonos, Inc. Media playback system with guest access
US9977561B2 (en) 2004-04-01 2018-05-22 Sonos, Inc. Systems, methods, apparatus, and articles of manufacture to provide guest access
US8868698B2 (en) 2004-06-05 2014-10-21 Sonos, Inc. Establishing a secure wireless network with minimum human intervention
US8326951B1 (en) 2004-06-05 2012-12-04 Sonos, Inc. Establishing a secure wireless network with minimum human intervention
US8483853B1 (en) 2006-09-12 2013-07-09 Sonos, Inc. Controlling and manipulating groupings in a multi-zone media system
US9202509B2 (en) 2006-09-12 2015-12-01 Sonos, Inc. Controlling and grouping in a multi-zone media system
US8788080B1 (en) 2006-09-12 2014-07-22 Sonos, Inc. Multi-channel pairing in a media system
US11429343B2 (en) 2011-01-25 2022-08-30 Sonos, Inc. Stereo playback configuration and control
US11265652B2 (en) 2011-01-25 2022-03-01 Sonos, Inc. Playback device pairing
US9654821B2 (en) 2011-12-30 2017-05-16 Sonos, Inc. Systems and methods for networked music playback
US11544750B1 (en) 2012-01-17 2023-01-03 Google Llc Overlaying content items with third-party reviews
US9729115B2 (en) 2012-04-27 2017-08-08 Sonos, Inc. Intelligently increasing the sound level of player
US8908879B2 (en) 2012-05-23 2014-12-09 Sonos, Inc. Audio content auditioning
US9674587B2 (en) 2012-06-26 2017-06-06 Sonos, Inc. Systems and methods for networked music playback including remote add to queue
US9225307B2 (en) 2012-06-28 2015-12-29 Sonos, Inc. Modification of audio responsive to proximity detection
US9008330B2 (en) 2012-09-28 2015-04-14 Sonos, Inc. Crossover frequency adjustments for audio speakers
US9510055B2 (en) 2013-01-23 2016-11-29 Sonos, Inc. System and method for a media experience social interface
US9247363B2 (en) 2013-04-16 2016-01-26 Sonos, Inc. Playback queue transfer in a media playback system
US9501533B2 (en) 2013-04-16 2016-11-22 Sonos, Inc. Private queue for a media playback system
US9361371B2 (en) 2013-04-16 2016-06-07 Sonos, Inc. Playlist update in a media playback system
US9633020B2 (en) * 2013-05-27 2017-04-25 Tencent Technology (Shenzhen) Company Limited Method and device for playing media synchronously
US9735978B2 (en) 2013-05-29 2017-08-15 Sonos, Inc. Playback queue control via a playlist on a mobile device
US9684484B2 (en) 2013-05-29 2017-06-20 Sonos, Inc. Playback zone silent connect
US9953179B2 (en) 2013-05-29 2018-04-24 Sonos, Inc. Private queue indicator
US9703521B2 (en) 2013-05-29 2017-07-11 Sonos, Inc. Moving a playback queue to a new zone
US10715973B2 (en) 2013-05-29 2020-07-14 Sonos, Inc. Playback queue control transition
US9798510B2 (en) 2013-05-29 2017-10-24 Sonos, Inc. Connected state indicator
US9495076B2 (en) 2013-05-29 2016-11-15 Sonos, Inc. Playlist modification
US9066048B2 (en) 2013-06-17 2015-06-23 Spotify Ab System and method for switching between audio content while navigating through video streams
US9298415B2 (en) 2013-07-09 2016-03-29 Sonos, Inc. Systems and methods to provide play/pause content
US9485329B1 (en) * 2013-07-17 2016-11-01 Google Inc. Action-defined conditions for selecting curated content
US9588647B1 (en) 2013-07-17 2017-03-07 Google Inc. Navigation of a list of content sharing platform media items on a client device via gesture controls and contextual synchronization
US9516082B2 (en) 2013-08-01 2016-12-06 Spotify Ab System and method for advancing to a predefined portion of a decompressed media stream
US20150046812A1 (en) * 2013-08-12 2015-02-12 Google Inc. Dynamic resizable media item player
US9906575B2 (en) 2013-08-15 2018-02-27 Comcast Cable Communications, Llc Media fling system
US9654532B2 (en) 2013-09-23 2017-05-16 Spotify Ab System and method for sharing file portions between peers with different capabilities
US9529888B2 (en) 2013-09-23 2016-12-27 Spotify Ab System and method for efficiently providing media and associated metadata
US9355555B2 (en) * 2013-09-27 2016-05-31 Sonos, Inc. System and method for issuing commands in a media playback system
US9063640B2 (en) 2013-10-17 2015-06-23 Spotify Ab System and method for switching between media items in a plurality of sequences of media items
US10873616B1 (en) 2013-12-10 2020-12-22 Google Llc Providing content to co-located devices with enhanced presentation characteristics
US11609689B2 (en) * 2013-12-11 2023-03-21 Given Imaging Ltd. System and method for controlling the display of an image stream
US9826347B2 (en) * 2014-01-09 2017-11-21 Comcast Cable Communications, Llc Methods and systems for content consumption
US20150220498A1 (en) * 2014-02-05 2015-08-06 Sonos, Inc. Remote Creation of a Playback Queue for a Future Event
US9226087B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9226073B2 (en) 2014-02-06 2015-12-29 Sonos, Inc. Audio output balancing during synchronized playback
US9830061B1 (en) * 2014-04-09 2017-11-28 Google Inc. Methods, systems, and media for providing media guidance with contextual controls
US10129599B2 (en) * 2014-04-28 2018-11-13 Sonos, Inc. Media preference database
AU2015280062A1 (en) 2014-06-24 2016-10-13 Apple Inc. Music now playing user interface
US9747011B2 (en) 2014-09-16 2017-08-29 Google Inc. Continuation of playback of media content by different output devices
US10528589B2 (en) 2014-09-26 2020-01-07 Oracle International Corporation Cross visualization interaction between data visualizations
US9971742B2 (en) 2014-09-26 2018-05-15 Oracle International Corporation Semantic distance-based assignment of data elements to visualization edges
US20160196104A1 (en) * 2015-01-07 2016-07-07 Zachary Paul Gordon Programmable Audio Device
US9521496B2 (en) * 2015-02-12 2016-12-13 Harman International Industries, Inc. Media content playback system and method
US9891880B2 (en) 2015-03-31 2018-02-13 Sonos, Inc. Information display regarding playback queue subscriptions
US20160337429A1 (en) * 2015-05-15 2016-11-17 Spotify Ab Method and device for resumed playback of streamed media
US10338794B2 (en) 2015-05-15 2019-07-02 Spotify Ab Playback of media streams in dependence of a time of a day
US10248376B2 (en) 2015-06-11 2019-04-02 Sonos, Inc. Multiple groupings in a playback system
US10671234B2 (en) * 2015-06-24 2020-06-02 Spotify Ab Method and an electronic device for performing playback of streamed media including related media content
US20170060405A1 (en) * 2015-08-28 2017-03-02 Facebook, Inc. Systems and methods for content presentation
US10120531B2 (en) 2015-09-08 2018-11-06 Apple Inc. User interfaces for navigating and playing content
US10897449B2 (en) 2015-10-06 2021-01-19 Twitter, Inc. Social media messaging platform for creating and sharing moments
US10771521B2 (en) 2015-12-22 2020-09-08 Spotify Ab Methods and systems for overlaying and playback of audio data received from distinct sources
US20170214651A1 (en) 2016-01-05 2017-07-27 William McMichael Systems and methods of transmitting and displaying private message data via a text input application
US20170220573A1 (en) 2016-01-05 2017-08-03 William McMichael Systems and methods of performing searches within a text input application
US10348849B2 (en) * 2016-02-22 2019-07-09 At&T Mobility Ii Llc Automatic delivery of media content to a device
US10447748B2 (en) * 2016-05-12 2019-10-15 Apple Inc. Sharing media information between applications on client devices
US10545653B2 (en) * 2016-06-12 2020-01-28 Apple Inc. Device, method, and graphical user interface for media playback
DK179034B1 (en) 2016-06-12 2017-09-04 Apple Inc Devices, methods, and graphical user interfaces for dynamically adjusting presentation of audio outputs
US10712997B2 (en) 2016-10-17 2020-07-14 Sonos, Inc. Room association based on name
USD815127S1 (en) * 2016-10-28 2018-04-10 Spotify Ab Display screen or portion thereof with graphical user interface
USD829742S1 (en) * 2016-10-28 2018-10-02 Spotify Ab Display screen or portion thereof with transitional graphical user interface
USD815128S1 (en) 2016-10-28 2018-04-10 Spotify Ab Display screen or portion thereof with graphical user interface
USD815129S1 (en) * 2016-10-28 2018-04-10 Spotify Ab Display screen or portion thereof with graphical user interface
CN106713316B (en) * 2016-12-22 2020-05-22 西安电子科技大学 Method for acquiring streaming media service transmission resource requirement based on receiving buffer
US10728695B2 (en) 2016-12-30 2020-07-28 Spotify Ab System and method for use of crowdsourced microphone or other information with a digital media content environment
US10489106B2 (en) 2016-12-31 2019-11-26 Spotify Ab Media content playback during travel
US11514098B2 (en) * 2016-12-31 2022-11-29 Spotify Ab Playlist trailers for media content playback during travel
US10747423B2 (en) * 2016-12-31 2020-08-18 Spotify Ab User interface for media content playback
US11474583B2 (en) * 2017-01-09 2022-10-18 Inmusic Brands, Inc. Systems and methods for generating a playback-information display during time compression or expansion of an audio signal
CN107895388B (en) * 2017-11-13 2022-04-12 广州视睿电子科技有限公司 Method and device for filling colors of graph, computer equipment and storage medium
USD876465S1 (en) 2018-02-23 2020-02-25 Samsung Electronics Co., Ltd. Display screen or portion thereof with transitional graphical user interface
US11157130B2 (en) * 2018-02-26 2021-10-26 Adobe Inc. Cursor-based resizing for copied image portions
DK201870353A1 (en) * 2018-05-07 2019-12-04 Apple Inc. User interfaces for recommending and consuming content on an electronic device
CN110456948B (en) * 2018-05-07 2023-04-18 苹果公司 User interface for recommending and consuming content on electronic devices
KR102192934B1 (en) * 2018-08-08 2020-12-18 엔에이치엔 주식회사 Network server and method to provide user terminal with web page
CN110035005B (en) * 2019-02-27 2021-07-09 视联动力信息技术股份有限公司 Data processing method and device
CN110456955B (en) * 2019-08-01 2022-03-29 腾讯科技(深圳)有限公司 Exposed clothing detection method, device, system, equipment and storage medium
CN110691282B (en) * 2019-10-16 2021-11-02 北京字节跳动网络技术有限公司 Video processing method and device, storage medium and electronic equipment
US11656838B2 (en) 2019-11-11 2023-05-23 Apple Inc. User interfaces for time period-based curated playlists
JP7347324B2 (en) * 2020-05-18 2023-09-20 トヨタ自動車株式会社 Agent cooperation device
CN114816603A (en) * 2021-01-27 2022-07-29 北京字跳网络技术有限公司 Work display method and device, electronic equipment and storage medium
US11451870B1 (en) * 2021-08-19 2022-09-20 Rovi Guides, Inc. Methods and systems to dynamically adjust a playlist based on cumulative mood score
CN115426531A (en) * 2022-08-30 2022-12-02 北京字跳网络技术有限公司 Live broadcast room access method, device, equipment and medium

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070016611A1 (en) * 2005-07-13 2007-01-18 Ulead Systems, Inc. Preview method for seeking media content
US20070130159A1 (en) * 2005-12-07 2007-06-07 Ask Jeeves. Inc. Method and system to present video content
US20080107400A1 (en) * 2006-11-06 2008-05-08 Samsung Electronics Co., Ltd. Method and apparatus for reproducing discontinuous av data
US20080271095A1 (en) * 2007-04-24 2008-10-30 Yahoo! Inc. Method and system for previewing media over a network
US20110113331A1 (en) * 2009-11-10 2011-05-12 Tilman Herberger System and method for dynamic visual presentation of digital audio content
US20110234480A1 (en) * 2010-03-23 2011-09-29 Apple Inc. Audio preview of music
US20120078398A1 (en) * 2010-09-28 2012-03-29 Sony Corporation Sound processing device, sound data selecting method and sound data selecting program
US20120173981A1 (en) * 2010-12-02 2012-07-05 Day Alexandrea L Systems, devices and methods for streaming multiple different media content in a digital container
US8321784B1 (en) * 2008-05-30 2012-11-27 Adobe Systems Incorporated Reviewing objects
US20120307150A1 (en) * 2010-02-23 2012-12-06 Echostar Ukraine Llc Electronic device audio volume reduction in response to audio communication request
US20130317635A1 (en) * 2012-05-23 2013-11-28 Sonos, Inc Audio Content Auditioning
US20130328925A1 (en) * 2012-06-12 2013-12-12 Stephen G. Latta Object focus in a mixed reality environment
US20140035831A1 (en) * 2012-07-31 2014-02-06 Apple Inc. Method and System for Scanning Preview of Digital Media

Family Cites Families (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5903727A (en) 1996-06-18 1999-05-11 Sun Microsystems, Inc. Processing HTML to embed sound in a web page
US6363204B1 (en) 1997-09-30 2002-03-26 Compaq Computer Corporation Viewing management for video sources
CA2447787A1 (en) 2001-06-04 2002-12-12 Nct Group, Inc. A system and method for reducing the time to deliver information from a communications network to a user
US20030126605A1 (en) * 2001-12-28 2003-07-03 Betz Steve Craig Method for displaying EPG video-clip previews on demand
US7571014B1 (en) * 2004-04-01 2009-08-04 Sonos, Inc. Method and apparatus for controlling multimedia players in a multi-zone system
US8028323B2 (en) * 2004-05-05 2011-09-27 Dryden Enterprises, Llc Method and system for employing a first device to direct a networked audio device to obtain a media item
US20060069617A1 (en) 2004-09-27 2006-03-30 Scott Milener Method and apparatus for prefetching electronic data for enhanced browsing
US7457915B2 (en) 2005-04-07 2008-11-25 Microsoft Corporation Intelligent media caching based on device state
US7613736B2 (en) * 2005-05-23 2009-11-03 Resonance Media Services, Inc. Sharing music essence in a recommendation system
US20080022229A1 (en) 2005-12-23 2008-01-24 Soujanya Bhumkar Methods and systems for enhancing internet experiences using previews
US9740794B2 (en) 2005-12-23 2017-08-22 Yahoo Holdings, Inc. Methods and systems for enhancing internet experiences
US20070294297A1 (en) * 2006-06-19 2007-12-20 Lawrence Kesteloot Structured playlists and user interface
US20090228919A1 (en) 2007-11-16 2009-09-10 Zott Joseph A Media playlist management and viewing remote control
US8302033B2 (en) * 2007-06-22 2012-10-30 Apple Inc. Touch screen device, method, and graphical user interface for providing maps, directions, and location-based information
KR101398134B1 (en) 2007-10-04 2014-05-20 엘지전자 주식회사 Apparatus and method for playing moving-picture in mobile terminal
JP2009093357A (en) 2007-10-05 2009-04-30 Sony Corp Information processor, content provision server, information processing method and content provision method
US8140974B2 (en) 2007-12-14 2012-03-20 Microsoft Corporation Presenting secondary media objects to a user
US8529264B2 (en) 2008-12-23 2013-09-10 Benjamin J. Hamlin Method facilitating language learning
KR100993064B1 (en) 2009-03-02 2010-11-08 주식회사 팬택 Method for Music Selection Playback in Touch Screen Adopted Music Playback Apparatus
US20100229094A1 (en) * 2009-03-04 2010-09-09 Apple Inc. Audio preview of music
KR101608532B1 (en) * 2009-08-11 2016-04-01 엘지전자 주식회사 Method for displaying data and mobile terminal thereof
EP2323134A1 (en) * 2009-11-06 2011-05-18 Research In Motion Limited Methods and electronic devices for previewing videos
US20110118858A1 (en) 2009-11-13 2011-05-19 Apple Inc. Local storage of a portion of streamed media items
GB201011146D0 (en) 2010-07-02 2010-08-18 Vodafone Ip Licensing Ltd Mobile computing device
US9015639B2 (en) * 2011-02-11 2015-04-21 Linkedin Corporation Methods and systems for navigating a list with gestures
US8467663B2 (en) 2011-02-18 2013-06-18 Apple Inc. Video context popups
US20120311444A1 (en) * 2011-06-05 2012-12-06 Apple Inc. Portable multifunction device, method, and graphical user interface for controlling media playback using gestures
US9843607B2 (en) * 2012-11-01 2017-12-12 Blackberry Limited System and method of transferring control of media playback between electronic devices
US9495076B2 (en) * 2013-05-29 2016-11-15 Sonos, Inc. Playlist modification
US10715973B2 (en) * 2013-05-29 2020-07-14 Sonos, Inc. Playback queue control transition

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070016611A1 (en) * 2005-07-13 2007-01-18 Ulead Systems, Inc. Preview method for seeking media content
US20070130159A1 (en) * 2005-12-07 2007-06-07 Ask Jeeves. Inc. Method and system to present video content
US20080107400A1 (en) * 2006-11-06 2008-05-08 Samsung Electronics Co., Ltd. Method and apparatus for reproducing discontinuous av data
US20080271095A1 (en) * 2007-04-24 2008-10-30 Yahoo! Inc. Method and system for previewing media over a network
US8321784B1 (en) * 2008-05-30 2012-11-27 Adobe Systems Incorporated Reviewing objects
US20110113331A1 (en) * 2009-11-10 2011-05-12 Tilman Herberger System and method for dynamic visual presentation of digital audio content
US20120307150A1 (en) * 2010-02-23 2012-12-06 Echostar Ukraine Llc Electronic device audio volume reduction in response to audio communication request
US20110234480A1 (en) * 2010-03-23 2011-09-29 Apple Inc. Audio preview of music
US20120078398A1 (en) * 2010-09-28 2012-03-29 Sony Corporation Sound processing device, sound data selecting method and sound data selecting program
US20120173981A1 (en) * 2010-12-02 2012-07-05 Day Alexandrea L Systems, devices and methods for streaming multiple different media content in a digital container
US20130317635A1 (en) * 2012-05-23 2013-11-28 Sonos, Inc Audio Content Auditioning
US20130328925A1 (en) * 2012-06-12 2013-12-12 Stephen G. Latta Object focus in a mixed reality environment
US20140035831A1 (en) * 2012-07-31 2014-02-06 Apple Inc. Method and System for Scanning Preview of Digital Media

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9942283B2 (en) 2012-06-29 2018-04-10 Spotify Ab Systems and methods for multi-context media control and playback
US10620797B2 (en) 2012-06-29 2020-04-14 Spotify Ab Systems and methods for multi-context media control and playback
US9195383B2 (en) * 2012-06-29 2015-11-24 Spotify Ab Systems and methods for multi-path control signals for media presentation devices
US9635068B2 (en) 2012-06-29 2017-04-25 Spotify Ab Systems and methods for multi-context media control and playback
US10884588B2 (en) 2012-06-29 2021-01-05 Spotify Ab Systems and methods for multi-context media control and playback
US11294544B2 (en) 2012-06-29 2022-04-05 Spotify Ab Systems and methods for multi-context media control and playback
US20140006483A1 (en) * 2012-06-29 2014-01-02 Spotify Ab Systems and methods for multi-context media control and playback
US10440075B2 (en) 2012-06-29 2019-10-08 Spotify Ab Systems and methods for multi-context media control and playback
US20140122589A1 (en) * 2012-11-01 2014-05-01 Research In Motion Limited System and method of transferring control of media playback between electronic devices
US9843607B2 (en) * 2012-11-01 2017-12-12 Blackberry Limited System and method of transferring control of media playback between electronic devices
US11481744B2 (en) * 2013-09-30 2022-10-25 Sonos, Inc. Discovery and media control at a point-of-sale display
US11818225B1 (en) 2013-09-30 2023-11-14 Sonos, Inc. Automatic discovery and control of a remotely controllable system
US11137826B2 (en) 2015-05-19 2021-10-05 Spotify Ab Multi-track playback of media content during repetitive motion activities
US10248190B2 (en) 2015-05-19 2019-04-02 Spotify Ab Multi-track playback of media content during repetitive motion activities
US10671155B2 (en) 2015-05-19 2020-06-02 Spotify Ab Multi-track playback of media content during repetitive motion activities
US10467998B2 (en) 2015-09-29 2019-11-05 Amper Music, Inc. Automated music composition and generation system for spotting digital media objects and event markers using emotion-type, style-type, timing-type and accent-type musical experience descriptors that characterize the digital music to be automatically composed and generated by the system
US10672371B2 (en) 2015-09-29 2020-06-02 Amper Music, Inc. Method of and system for spotting digital media objects and event markers using musical experience descriptors to characterize digital music to be automatically composed and generated by an automated music composition and generation engine
US11017750B2 (en) 2015-09-29 2021-05-25 Shutterstock, Inc. Method of automatically confirming the uniqueness of digital pieces of music produced by an automated music composition and generation system while satisfying the creative intentions of system users
US11776518B2 (en) 2015-09-29 2023-10-03 Shutterstock, Inc. Automated music composition and generation system employing virtual musical instrument libraries for producing notes contained in the digital pieces of automatically composed music
US11030984B2 (en) 2015-09-29 2021-06-08 Shutterstock, Inc. Method of scoring digital media objects using musical experience descriptors to indicate what, where and when musical events should appear in pieces of digital music automatically composed and generated by an automated music composition and generation system
US11037539B2 (en) 2015-09-29 2021-06-15 Shutterstock, Inc. Autonomous music composition and performance system employing real-time analysis of a musical performance to automatically compose and perform music to accompany the musical performance
US11037541B2 (en) 2015-09-29 2021-06-15 Shutterstock, Inc. Method of composing a piece of digital music using musical experience descriptors to indicate what, when and how musical events should appear in the piece of digital music automatically composed and generated by an automated music composition and generation system
US11657787B2 (en) 2015-09-29 2023-05-23 Shutterstock, Inc. Method of and system for automatically generating music compositions and productions using lyrical input and music experience descriptors
US11037540B2 (en) 2015-09-29 2021-06-15 Shutterstock, Inc. Automated music composition and generation systems, engines and methods employing parameter mapping configurations to enable automated music composition and generation
US11651757B2 (en) 2015-09-29 2023-05-16 Shutterstock, Inc. Automated music composition and generation system driven by lyrical input
US11011144B2 (en) 2015-09-29 2021-05-18 Shutterstock, Inc. Automated music composition and generation system supporting automated generation of musical kernels for use in replicating future music compositions and production environments
US11468871B2 (en) 2015-09-29 2022-10-11 Shutterstock, Inc. Automated music composition and generation system employing an instrument selector for automatically selecting virtual instruments from a library of virtual instruments to perform the notes of the composed piece of digital music
US10854180B2 (en) 2015-09-29 2020-12-01 Amper Music, Inc. Method of and system for controlling the qualities of musical energy embodied in and expressed by digital music to be automatically composed and generated by an automated music composition and generation engine
US11430419B2 (en) 2015-09-29 2022-08-30 Shutterstock, Inc. Automatically managing the musical tastes and preferences of a population of users requesting digital pieces of music automatically composed and generated by an automated music composition and generation system
US11430418B2 (en) 2015-09-29 2022-08-30 Shutterstock, Inc. Automatically managing the musical tastes and preferences of system users based on user feedback and autonomous analysis of music automatically composed and generated by an automated music composition and generation system
US9798514B2 (en) * 2016-03-09 2017-10-24 Spotify Ab System and method for color beat display in a media content environment
US20220086202A1 (en) * 2018-01-26 2022-03-17 Citrix Systems, Inc. Virtual computing system providing local screen sharing from hosted collaboration applications and related methods
US11190561B2 (en) * 2018-01-26 2021-11-30 Citrix Systems, Inc. Virtual computing system providing local screen sharing from hosted collaboration applications and related methods
US11838335B2 (en) * 2018-01-26 2023-12-05 Citrix Systems, Inc. Virtual computing system providing local screen sharing from hosted collaboration applications and related methods
US10964299B1 (en) 2019-10-15 2021-03-30 Shutterstock, Inc. Method of and system for automatically generating digital performances of music compositions using notes selected from virtual musical instruments based on the music-theoretic states of the music compositions
US11037538B2 (en) 2019-10-15 2021-06-15 Shutterstock, Inc. Method of and system for automated musical arrangement and musical instrument performance style transformation supported within an automated music performance system
US11024275B2 (en) 2019-10-15 2021-06-01 Shutterstock, Inc. Method of digitally performing a music composition using virtual musical instruments having performance logic executing within a virtual musical instrument (VMI) library management system

Also Published As

Publication number Publication date
EP3151576A1 (en) 2017-04-05
US20140214927A1 (en) 2014-07-31
US9246967B2 (en) 2016-01-26
US20140108929A1 (en) 2014-04-17
EP2772064A2 (en) 2014-09-03
WO2014057356A2 (en) 2014-04-17
US20160313872A1 (en) 2016-10-27
WO2014057356A3 (en) 2014-07-24

Similar Documents

Publication Publication Date Title
US11343295B2 (en) Systems and methods for providing song samples
US9246967B2 (en) Systems, methods, and user interfaces for previewing media content
US11294544B2 (en) Systems and methods for multi-context media control and playback
US9942283B2 (en) Systems and methods for multi-context media control and playback

Legal Events

Date Code Title Description
AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL

Free format text: SUPPLEMENTAL PATENT SECURITY AGREEMENT;ASSIGNOR:SPOTIFY AB;REEL/FRAME:034709/0364

Effective date: 20141223

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: SPOTIFY AB, SWEDEN

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:038982/0327

Effective date: 20160609