US20120213383A1 - Automatic resource retrieval and use - Google Patents
Automatic resource retrieval and use Download PDFInfo
- Publication number
- US20120213383A1 US20120213383A1 US13/459,497 US201213459497A US2012213383A1 US 20120213383 A1 US20120213383 A1 US 20120213383A1 US 201213459497 A US201213459497 A US 201213459497A US 2012213383 A1 US2012213383 A1 US 2012213383A1
- Authority
- US
- United States
- Prior art keywords
- content
- resource
- commercial location
- cue
- audio signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000005236 sound signal Effects 0.000 claims abstract description 53
- 238000000034 method Methods 0.000 claims description 58
- 238000004891 communication Methods 0.000 claims description 8
- 230000004044 response Effects 0.000 claims description 4
- 230000005540 biological transmission Effects 0.000 description 14
- 230000008569 process Effects 0.000 description 10
- 230000000007 visual effect Effects 0.000 description 9
- 230000011664 signaling Effects 0.000 description 8
- 238000012545 processing Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 5
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000005070 sampling Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 241000282412 Homo Species 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000006855 networking Effects 0.000 description 2
- 230000010363 phase shift Effects 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 238000002592 echocardiography Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 238000013515 script Methods 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/24—Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
- H04N21/2407—Monitoring of transmitted content, e.g. distribution time, number of downloads
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/438—Presentation of query results
- G06F16/4387—Presentation of query results by the use of playlists
- G06F16/4393—Multimedia presentations, e.g. slide shows, multimedia albums
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/432—Query formulation
- G06F16/433—Query formulation using audio data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
Definitions
- This disclosure relates automatic resource retrieval and use. More specifically, this disclosure relates to generating, providing, and processing cues for resource retrieval and use.
- Computers and other computing devices are useful for retrieving and displaying resources such as multimedia content, interactive content, executable programs, and the like. Such resources are often useful in providing information to a user of the computer. Often, that content can augment or enhance other information, performances, or experiences that the user is presented with from a source other than his computer.
- the audio signal has a frequency that renders the signal inaudible to a human; corresponds to the commercial location; and bears information that causes the device to obtain content relevant to the commercial location.
- a content acquisition cue via an audio signal.
- the audio signal has a frequency that renders the signal inaudible to a human and corresponds to the commercial location.
- obtaining content relevant to the commercial location In response to receiving the content acquisition cue, obtaining content relevant to the commercial location.
- Implementations may have one or more of the following features.
- Content is also provided to the device.
- the content includes a coupon for an item sold at the commercial location.
- the content includes an advertisement for an item sold at the commercial location.
- the content acquisition cue is broadcast over a plurality of substantially separate audio channels.
- the content acquisition cue is broadcast over three separate audio channels.
- the commercial location is at an entrance of a store.
- the audio signal further includes a content use cue that causes the content to be used on the device.
- FIG. 1 is schematic depiction of a resource retrieval system.
- FIGS. 2-3 are flowcharts for a resource retrieval method.
- FIG. 4 is a schematic depiction of an encoding process.
- FIG. 5 is a flowchart for encoding a resource identifier
- FIGS. 6A and 6B are schematic depictions of broadcasting an audio signal.
- FIG. 7 is a flowchart for decoding a resource identifier.
- FIG. 8 is a schematic depiction of a venue with several content use sites.
- Computing devices including personal computers, laptops, netbooks, smart phones, special-purpose computing hardware, and other portable or non-portable equipment, collectively referred to in this document as “computers”) are useful for retrieving information.
- computers often can electronically communicate (e.g., via the Internet) with a vast number of information sources, and can therefore retrieve a vast amount of information.
- computers are capable of displaying a vast amount of information in a variety of forms. Even computers with relatively modest capabilities can routinely display a variety of video, audio, document, and other data formats, and can similarly execute a wide variety of computer programs.
- the speaker at a presentation can display a visual aid to the audience. However, this does not allow audience members to individually interact with the visual aid.
- the speaker may disseminate an electronic copy of the visual aid to conference attendants in advance of the presentation. Although this allows audience members to individually interact with the visual aid, it is relatively difficult to focus the audience's collective attention on a desired feature of the visual aid. For example, if the speaker desired audience members to activate a function on a program by clicking on one of many available buttons, the speaker's options would be relatively inefficient. For example, one option is for the speaker to attempt to verbally describe which of the several buttons to click.
- the techniques described in this document allow for the automatic retrieval and/or use of content or other resources. These techniques may advantageously be employed, e.g., by a speaker at a conference to cause the participants' computers to automatically retrieve, load, or use content or other resources. Other exemplary uses are discusses below.
- FIG. 1 is a schematic depiction of a resource retrieval system.
- the resource retrieval system 10 may include one or more client computers 12 and a signaling computer 14 .
- the signaling computer 14 may include a loudspeaker 16
- each of the client computers 12 may include a microphone 20 .
- Each client computer 12 may also be in data communication via a computer network 22 to one or more resource identification servers 24 and resource providers 28 .
- the computer network 22 can include any known networking type or structure, including local area networks, ad-hoc networks, mesh networks, wide-area networks such as the Internet, telephone networks, satellite networks, etc.
- the data communication may be implemented in any known fashion, including hypertext transfer protocol (HTTP), file transfer protocol (FTP), BitTorrent protocol, e-mail, short message service (SMS), or any other standard or proprietary fashion.
- HTTP hypertext transfer protocol
- FTP file transfer protocol
- SMS short message service
- the client computer 12 may more generally include any computing device that includes a microphone, networking communications capability, and processing capability adequate for the signal processing described herein.
- a client computer 12 as described herein may include a laptop computer, a notebook computer, a netbook computer, a cellular phone (including smart phones such as an iPhone or a Blackberry), a personal digital device, a media player (e.g., an iPod), an electronic messaging device, and so forth.
- the presenter may emit suitable audio signals 18 using the loudspeaker 16 , which can subsequently be detected by the microphone 20 of a client computer 12 .
- the audio signal 18 can be decoded to reveal a resource identifier.
- the resource identifier can consist of a relatively small amount of data; e.g., a 30 bit integer.
- the resource identifier can be used to determine a corresponding resource location. For example, this can be accomplished by submitting the resource identifier over the computer network 22 to a resource identification server 24 .
- a server 24 can include a data store 26 that lists a resource location corresponding to each resource identifier. The client computer 12 can then obtain the resource from a resource provider 28 at the specified resource location.
- the word “resource” can refer to any computer-stored information or instructions that can be transferred, described, or implemented over a computer network.
- a “resource” can include audio, video, images, documents, markup language documents, hypertext links, scripts, macros, source code, executable code, or any combinations of the foregoing.
- the words “content” and “resource” are synonymous in this document.
- FIG. 2 is a method of associating an audio signal with a resource.
- the method 30 begins by selecting a resource (step 32 ) with which to associate an audio signal 18 .
- this resource can include a visual aid for use by a speaker during a presentation. Many more examples are given below.
- a resource identifier can be associated with the resource.
- the resource identifier can consist solely of a number; for example, a 30 (or fewer) bit number.
- a resource identifier can also include other information, such as information that describes a context or a geographic location in which the resource will be used. Further considerations pertinent to selecting a resource identifier are discussed in connection with FIGS. 6A and 6B .
- Step 34 can be carried out at any location, on any computer.
- step 34 can be carried out at the signaling computer 14 , a resource identification server 24 , a resource provider 28 , or at another location outside the resource retrieval system 10 .
- the resource is sent to known location at a resource provider 28 .
- this can be done via any electronic communication, such as file transfer protocol, secure copying, via e-mail attachment, etc.
- the performer of the method 30 is also the resource provider 28 , in which case the resource need not be sent to another location.
- the resource identifier and the location of the corresponding resource are stored in a data store 26 within a server 24 .
- This information can be stored in any manner.
- the resource identifier and resource location can be stored in corresponding fields in a record of a database, as corresponding entries in an index or table, as concurrent entries in a list, etc.
- the resource identifier can be encoded as an audio signal 18 . Further details about possible encoding techniques are described below in connection with FIGS. 4 and 5 .
- the encoding can take place at any location on any computer.
- step 38 can be carried out at the signaling computer 14 , a server 24 , a resource provider 28 , or at another location outside the resource retrieval system 10 .
- the audio signal 18 can be broadcasted.
- the audio signal 18 is broadcasted by the signaling computer 14 using the loudspeaker 16 , or other hardware capable of broadcasting an audio signal.
- a computer different from the signaling computer 14 broadcasts the audio signal.
- FIG. 3 is a method for retrieving a resource.
- the method 42 of FIG. 3 can begin with receiving an audio signal (step 44 ). For example, an audio signal 18 that had previously been broadcasted using the method 30 or another method can be received.
- step 46 the audio signal can be decoded. Further details about decoding techniques are described below, in connection with FIG. 7 .
- a resource identifier can be obtained (step 48 ).
- a resource location can be ascertained based on the resource identifier.
- the resource identifier can be submitted to a server 24 , in response to which the server 24 provides the corresponding resource location.
- the resource identifier can be used as a query for a database, as an argument for a lookup routine, etc.
- the resource can be obtained from the resource location.
- the resource can be obtained directly by the performer of the retrieval method 42 , such as a client computer 12 .
- the client computer 12 can initiate a data connection to the resource location and obtain the resource.
- the performer of the retrieval method 42 can indirectly obtain the resource.
- the resource provider can send the resource as an attachment in an e-mail to an e-mail server, from which the performer of the method 42 subsequently obtains the resource.
- the client computer 12 may take a specified action with respect to the resource (e.g., display or execute the resource), or may simply store the resource for later use.
- the resource provider 28 is illustrated separately from the client computer 12 in FIG. 1 , in some implementations the resource provider 28 can be the same as the client computer 12 . In this case, the resource is “local” to the client computer 12 .
- FIGS. 4 and 5 a schematic illustration of encoding a resource identifier 52 is shown ( FIG. 4 ), together with a method 58 of encoding a resource identifier ( FIG. 5 ).
- a resource identifier 52 is illustrated as a string of bits. Although only nine bits are illustrated in FIG. 4 , in principal any number of bits can be used, subject to limitations discussed below.
- the resource identifier 52 can be obtained. For example, this can be the result of step 34 of process 30 ( FIG. 2 ).
- a number of audio channels 54 are identified. Although three audio channels 54 a , 54 b , and 54 c are illustrated in FIG. 4 , in principal any number of audio channels can be used.
- An audio channel is characterized by a central frequency 55 and a channel width (or “band width”) that describes the range of frequencies in the channel.
- the channel width is illustrated in FIG. 4 by the geometric width of the channels 54 a , 54 b , 54 c.
- the audio channels 54 a , 54 b , 54 c do not overlap (i.e., are separate), then there is no interference between the audio signals sent over the audio channels. However, if the audio channels 54 a , 54 b , 54 c do overlap, then the audio signals may, but need not, interfere. In some implementations, the audio channels 54 a , 54 b , and 54 c are separate. In some implementations, one or more pairs of audio channels 54 a , 54 b , and 54 c are only substantially separate, with a small amount of overlap permissible based on an acceptable level of interference-based errors. The rate at which these errors occurs depends on many factors, including the size of the overlap between channels, the type of encoding used, the data being sent during a particular transmission, and other factors.
- the number of audio channels 54 can be chosen by identifying a total frequency range available for transmission, and partitioning the available frequency range into separate (or substantially separate) channels of the minimum-sized bandwidth necessary for the selected encoding technique. This approach maximizes the number of audio channels 54 available for transmission.
- the sampling rate of the microphone 20 can determine the maximum possible frequency it can receive.
- the maximum possible identifiable frequency is equal to half the sampling rate.
- commonly available microphones have a sampling rate of 44.1 kilo-samples per second. Consequently, the maximum available frequency such a microphone can detect is 22.05 kHz.
- the audio signals 18 In principle, there is no lower limit for the total frequency range available for transmission. However, in some implementations, it is desirable for the audio signals 18 to be in inaudible to humans or other animals. This has the advantage that people would not be distracted by audio signals 18 that are broadcast in their vicinity. According to one estimate, the audible range for most humans has an upper limit of 15 kHz. This, in turn, can be used as a lower limit of the total available frequency range.
- the resource identifier 52 can be segmented into as many parts as there are channels (step 64 ). For example, in FIG. 4 , the nine-bit resource identifier 52 is segmented into three parts, one for each of the communication channels 54 a , 54 b , 54 c . If the number of bits in the resource identifier 52 is not evenly divisible by the number of communication channels 54 , then the remainder bits can be associated with the higher-frequency channels. This can be advantageous for reasons of echo prevention, described more fully below (see FIGS. 6A and 6B ).
- each segment can be processed as follows.
- a first segment is identified (step 64 ). That segment is encoded according to a chosen algorithm (step 66 ).
- a chosen algorithm step 66 .
- modulation algorithms There are a number of modulation algorithms that can be employed, including analog modulation, digital modulation, spread spectrum modulation, and others. Each algorithm has potential strengths and weaknesses, depending on the particular constraints that are important in the deployment of a particular resource retrieval system 10 .
- GMSK Gaussian minimum shift keying
- GMSK has been usefully demonstrated to provide adequate data rates and robustness in a variety of physical contexts
- this disclosure is not limited to GMSK coding unless explicitly stated to the contrary.
- other shift-keyed encoding techniques include minimum shift keying, frequency shift keying, phase shift keying, quadrature phase shift keying, and so forth.
- other modulation techniques based upon frequency modulation, amplitude modulation, phase modulation, and various combinations of the foregoing, are generally known in the art and all such techniques that may be usefully adapted to reliably encode data onto a high-frequency audio carrier are intended to fall within the scope of this disclosure.
- FIGS. 6A and 6B concurrently broadcasting several audio signals 18 a , 18 b , 18 c helps reduce the possibility of errors related to echoes as compared to broadcasting a single audio signal 18 .
- a single audio signal 18 is broadcast from a loudspeaker 16 ;
- three audio signals 18 a , 18 b , and 18 c are broadcast from a loudspeaker 16 . If the signal 18 carries the same information content as the signals 18 a , 18 b , and 18 c , then the signal 18 is three times as long as any individual signal 18 a , 18 b , or 18 c.
- An echo-related error may occur when a signal interferes with itself where a microphone 20 is located.
- the signal 18 reflects off a wall 67 , so that both the last portion and the (reflected) initial portion of the signal arrive at the microphone.
- FIG. 6B when the loudspeaker 16 and microphone 20 are in the same relative positions, there are no echo-related errors, because the signal lengths are sufficiently short.
- the resource identifier 52 consists of an integer with a length between twenty four and thirty bits, which are segmented for transmission across three audio channels and encoded using GMSK modulation.
- the resource identifier 52 can also include other data, e.g., data describing a geographic or other context to which the resource relates.
- process 58 determines whether there are any unencoded segments remaining (decision 68 ). If there are any unencoded segments remaining (decision 68 ), then the encoding process 58 returns back to step 64 , where the next unencoded segment is selected. Once all segments have been encoded, process 58 concludes, having produced a number of audio signals that collectively encode the resource identifier.
- FIG. 7 is a flowchart for a decoding process.
- an audio signal can be received (step 72 ).
- the audio signal can be the result of broadcasting step 40 of method 30 .
- the received audio signal can then be decoded (step 74 ).
- the decoding process utilizes the compatible techniques of those used to encode the audio signal. For example, if GMSK techniques were used to encode the audio signal, those techniques are used to decode the signal.
- the step 74 Upon decoding the signal, the step 74 produces one or more resource identifier segments. If more than one segment is produced, those segments are subsequently reassembled (step 76 ), thereby producing a resource identification 52 .
- FIG. 8 is a schematic depiction of a venue with several content display points.
- the venue 80 includes several content display points 82 , each with a corresponding effective display area 84 .
- there may be a signaling computer 14 at each display point 82 and each effective display area 84 corresponds to the range of the signaling computer's loudspeaker 16 .
- Other examples are possible.
- Each content display point 82 provides a content use cue to cause a user's computer to display content relevant to that particular content display point.
- the content use cue can include an audio signal 18 as described above.
- the content use cue can include any transmission that causes the user's computer to display specific content.
- the content use cue can include a visual signal, a conventional data transmission over a network connection, etc.
- one or more content acquisition sites 86 can be established in the venue 80 .
- a content acquisition site can be established near an entrance 88 of the venue 80 .
- a content acquisition site 86 can provide a content acquisition cue to a user.
- the content acquisition cue can include an audio signal 18 as describe above.
- the content acquisition cue can include any transmission that causes the user's computer to acquire pertinent content.
- the content acquisition cue can include a visual signal, a conventional data transmission over a network connection, etc.
- a content acquisition cue can simply include a verbal instruction to load a particular piece of physical media, such as a DVD, that bears the content pertinent to the content display points 82 .
- the user's computer acquires the content pertinent to the various content display sites.
- the user's computer can perform the method 42 of FIG. 3 .
- the user's computer can obtain the content in other ways, such as acquiring the content via a conventional network connection to a local or remote computer, or by loading files from a local storage device.
- Acquiring content prior to use can have advantages. For example, having content locally available when receiving a content use cue tends to allow the user's computer to use the content relatively quickly. Moreover, obtaining all the content for the various content use points 82 in the venue 80 in one network transmission can provide power or bandwidth savings over obtaining content from the different content use points 82 on an as-needed basis.
- the resource retrieval system 10 can be deployed at a retail venue.
- an audio signal source can be deployed to cause computers that detect the audio signal to retrieve content pertinent to one or more products in the vicinity.
- content can include, e.g., advertisement materials, promotions, coupons, product descriptions, product reviews, links to product-related web sites, etc.
- a store-wide audio signal can be broadcast periodically alerting shoppers to promotions or specials occurring in specified departments or locations in the retail venue.
- the resource retrieval system 10 can be deployed in a museum or other venue amenable to self-guided tours. Near points of interest (e.g., exhibits, historical sites, etc.), the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the point of interest. Moreover, different information can be displayed to different users. For example, an “advanced” presentation or a “beginner” presentation can be displayed, depending on the user's preferences.
- the resource retrieval system 10 can be deployed in a sporting event. At various times, the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information pertinent to a particular competitor, a team, a play, etc.
- the resource retrieval system 10 can be deployed in a classroom setting.
- the teacher can broadcast audio signals to cause students' computers to display specific resources pertinent to a lecture.
- the teacher can broadcast audio signals to cause students' computers to begin or end an examination.
- the resource retrieval system 10 can be deployed in any venue.
- the resource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the venue such as a map of the venue or surrounding areas.
- a loudspeaker can be deployed at information kiosks, orientation booths, entrances to the venue, etc.
- the resource retrieval system can be used to cause users' computers to automatically retrieve information about emergency procedures.
- a venue-wide broadcast can be sounded over the venue's alarm system to alert occupants about the emergency, where the venue-wide broadcast includes an audio signal according to the techniques described above.
- Emergency procedures can include evacuation routes, first aid information, safety information, etc.
- a realization of the processes or devices described above may include computer-executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software.
- processing may be distributed across devices such as the various systems described above, or all of the functionality may be integrated into a dedicated, standalone device. All such permutations and combinations are intended to fall within the scope of the present disclosure.
- a computer program product comprising computer-executable code or computer-usable code that, when executing on one or more computing devices (such as the devices/systems described above), performs any and/or all of the steps described above.
- the code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (such as random access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices.
- any of the processes described above may be embodied in any suitable transmission or propagation medium carrying the computer-executable code described above and/or any inputs or outputs from same.
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 12/789,419, filed on May 27, 2010, which claims the benefit of U.S. Provisional Pat. App. 61/181,472 filed May 27, 2009, each of which the entire content is hereby incorporated by reference.
- This disclosure relates automatic resource retrieval and use. More specifically, this disclosure relates to generating, providing, and processing cues for resource retrieval and use.
- Computers and other computing devices are useful for retrieving and displaying resources such as multimedia content, interactive content, executable programs, and the like. Such resources are often useful in providing information to a user of the computer. Often, that content can augment or enhance other information, performances, or experiences that the user is presented with from a source other than his computer.
- In general, in one aspect: At a commercial location, broadcasting a content acquisition cue via an audio signal. The audio signal has a frequency that renders the signal inaudible to a human; corresponds to the commercial location; and bears information that causes the device to obtain content relevant to the commercial location.
- In general, in another aspect, at a commercial location, receiving a content acquisition cue via an audio signal. The audio signal has a frequency that renders the signal inaudible to a human and corresponds to the commercial location. In response to receiving the content acquisition cue, obtaining content relevant to the commercial location.
- Implementations may have one or more of the following features. Content is also provided to the device. The content includes a coupon for an item sold at the commercial location. The content includes an advertisement for an item sold at the commercial location. The content acquisition cue is broadcast over a plurality of substantially separate audio channels. The content acquisition cue is broadcast over three separate audio channels. The commercial location is at an entrance of a store. The audio signal further includes a content use cue that causes the content to be used on the device.
- Other aspects include other combinations of the features recited above and other features, expressed as methods, apparatus, systems, program products, and in other ways. Other features and advantages will be apparent from the description and from the claims.
- Embodiments of the invention described herein may be understood by reference to the following figures, which are provided by way of example and not of limitation:
-
FIG. 1 is schematic depiction of a resource retrieval system. -
FIGS. 2-3 are flowcharts for a resource retrieval method. -
FIG. 4 is a schematic depiction of an encoding process. -
FIG. 5 is a flowchart for encoding a resource identifier -
FIGS. 6A and 6B are schematic depictions of broadcasting an audio signal. -
FIG. 7 is a flowchart for decoding a resource identifier. -
FIG. 8 is a schematic depiction of a venue with several content use sites. - Computing devices (including personal computers, laptops, netbooks, smart phones, special-purpose computing hardware, and other portable or non-portable equipment, collectively referred to in this document as “computers”) are useful for retrieving information. For example, computers often can electronically communicate (e.g., via the Internet) with a vast number of information sources, and can therefore retrieve a vast amount of information.
- Moreover, computers are capable of displaying a vast amount of information in a variety of forms. Even computers with relatively modest capabilities can routinely display a variety of video, audio, document, and other data formats, and can similarly execute a wide variety of computer programs.
- It is often useful or entertaining to employ these capabilities of a computer to enhance events that are occurring in the computer user's environment. For example, if the computer user is attending a presentation on a topic, it may enhance the user's experience of the presentation if the user was simultaneously viewing pertinent content, e.g., on his computer.
- To some degree, this may be accomplished by traditional means. For example, the speaker at a presentation can display a visual aid to the audience. However, this does not allow audience members to individually interact with the visual aid. Alternatively, the speaker may disseminate an electronic copy of the visual aid to conference attendants in advance of the presentation. Although this allows audience members to individually interact with the visual aid, it is relatively difficult to focus the audience's collective attention on a desired feature of the visual aid. For example, if the speaker desired audience members to activate a function on a program by clicking on one of many available buttons, the speaker's options would be relatively inefficient. For example, one option is for the speaker to attempt to verbally describe which of the several buttons to click.
- Among other things, the techniques described in this document allow for the automatic retrieval and/or use of content or other resources. These techniques may advantageously be employed, e.g., by a speaker at a conference to cause the participants' computers to automatically retrieve, load, or use content or other resources. Other exemplary uses are discusses below.
-
FIG. 1 is a schematic depiction of a resource retrieval system. Theresource retrieval system 10 may include one or more client computers 12 and asignaling computer 14. Thesignaling computer 14 may include aloudspeaker 16, and each of the client computers 12 may include amicrophone 20. Each client computer 12 may also be in data communication via acomputer network 22 to one or moreresource identification servers 24 and resource providers 28. Thecomputer network 22 can include any known networking type or structure, including local area networks, ad-hoc networks, mesh networks, wide-area networks such as the Internet, telephone networks, satellite networks, etc. The data communication may be implemented in any known fashion, including hypertext transfer protocol (HTTP), file transfer protocol (FTP), BitTorrent protocol, e-mail, short message service (SMS), or any other standard or proprietary fashion. - The client computer 12 may more generally include any computing device that includes a microphone, networking communications capability, and processing capability adequate for the signal processing described herein. By way of example and not limitation, a client computer 12 as described herein may include a laptop computer, a notebook computer, a netbook computer, a cellular phone (including smart phones such as an iPhone or a Blackberry), a personal digital device, a media player (e.g., an iPod), an electronic messaging device, and so forth.
- As described more thoroughly below, the presenter may emit
suitable audio signals 18 using theloudspeaker 16, which can subsequently be detected by themicrophone 20 of a client computer 12. Theaudio signal 18 can be decoded to reveal a resource identifier. As described more fully below, the resource identifier can consist of a relatively small amount of data; e.g., a 30 bit integer. - The resource identifier can be used to determine a corresponding resource location. For example, this can be accomplished by submitting the resource identifier over the
computer network 22 to aresource identification server 24. Aserver 24 can include adata store 26 that lists a resource location corresponding to each resource identifier. The client computer 12 can then obtain the resource from a resource provider 28 at the specified resource location. - As used in this document, the word “resource” can refer to any computer-stored information or instructions that can be transferred, described, or implemented over a computer network. In particular, a “resource” can include audio, video, images, documents, markup language documents, hypertext links, scripts, macros, source code, executable code, or any combinations of the foregoing. The words “content” and “resource” are synonymous in this document.
-
FIG. 2 is a method of associating an audio signal with a resource. Themethod 30 begins by selecting a resource (step 32) with which to associate anaudio signal 18. For example, this resource can include a visual aid for use by a speaker during a presentation. Many more examples are given below. - In step 34, a resource identifier can be associated with the resource. In some implementations, the resource identifier can consist solely of a number; for example, a 30 (or fewer) bit number. Additionally, a resource identifier can also include other information, such as information that describes a context or a geographic location in which the resource will be used. Further considerations pertinent to selecting a resource identifier are discussed in connection with
FIGS. 6A and 6B . - Step 34 can be carried out at any location, on any computer. For example, step 34 can be carried out at the signaling
computer 14, aresource identification server 24, a resource provider 28, or at another location outside theresource retrieval system 10. - In step 36, the resource is sent to known location at a resource provider 28. In some implementations, this can be done via any electronic communication, such as file transfer protocol, secure copying, via e-mail attachment, etc. In some implementations, the performer of the
method 30 is also the resource provider 28, in which case the resource need not be sent to another location. - In step 37, the resource identifier and the location of the corresponding resource are stored in a
data store 26 within aserver 24. This information can be stored in any manner. For example, the resource identifier and resource location can be stored in corresponding fields in a record of a database, as corresponding entries in an index or table, as concurrent entries in a list, etc. - In step 38, the resource identifier can be encoded as an
audio signal 18. Further details about possible encoding techniques are described below in connection withFIGS. 4 and 5 . The encoding can take place at any location on any computer. For example, step 38 can be carried out at the signalingcomputer 14, aserver 24, a resource provider 28, or at another location outside theresource retrieval system 10. - In step 40, the
audio signal 18 can be broadcasted. In some implementations, theaudio signal 18 is broadcasted by the signalingcomputer 14 using theloudspeaker 16, or other hardware capable of broadcasting an audio signal. In some implementations, a computer different from the signalingcomputer 14 broadcasts the audio signal. -
FIG. 3 is a method for retrieving a resource. Themethod 42 ofFIG. 3 can begin with receiving an audio signal (step 44). For example, anaudio signal 18 that had previously been broadcasted using themethod 30 or another method can be received. - In
step 46, the audio signal can be decoded. Further details about decoding techniques are described below, in connection withFIG. 7 . As a result of the decoding instep 46, a resource identifier can be obtained (step 48). - In
step 49, a resource location can be ascertained based on the resource identifier. For example, the resource identifier can be submitted to aserver 24, in response to which theserver 24 provides the corresponding resource location. In some implementations, the resource identifier can be used as a query for a database, as an argument for a lookup routine, etc. - In
step 50, the resource can be obtained from the resource location. In some implementations, the resource can be obtained directly by the performer of theretrieval method 42, such as a client computer 12. For example, the client computer 12 can initiate a data connection to the resource location and obtain the resource. In some implementations, the performer of theretrieval method 42 can indirectly obtain the resource. For example, the resource provider can send the resource as an attachment in an e-mail to an e-mail server, from which the performer of themethod 42 subsequently obtains the resource. - After obtaining the resource, the client computer 12 may take a specified action with respect to the resource (e.g., display or execute the resource), or may simply store the resource for later use. Although the resource provider 28 is illustrated separately from the client computer 12 in
FIG. 1 , in some implementations the resource provider 28 can be the same as the client computer 12. In this case, the resource is “local” to the client computer 12. - Referring to
FIGS. 4 and 5 , a schematic illustration of encoding aresource identifier 52 is shown (FIG. 4 ), together with amethod 58 of encoding a resource identifier (FIG. 5 ). In the schematic illustration, aresource identifier 52 is illustrated as a string of bits. Although only nine bits are illustrated inFIG. 4 , in principal any number of bits can be used, subject to limitations discussed below. Instep 60, theresource identifier 52 can be obtained. For example, this can be the result of step 34 of process 30 (FIG. 2 ). - In
step 62, a number of audio channels 54 are identified. Although threeaudio channels FIG. 4 , in principal any number of audio channels can be used. An audio channel is characterized by acentral frequency 55 and a channel width (or “band width”) that describes the range of frequencies in the channel. The channel width is illustrated inFIG. 4 by the geometric width of thechannels - If the
audio channels audio channels audio channels audio channels - As discussed below, some encoding techniques require a minimum bandwidth for a communication channel 54. In some implementations, the number of audio channels 54 can be chosen by identifying a total frequency range available for transmission, and partitioning the available frequency range into separate (or substantially separate) channels of the minimum-sized bandwidth necessary for the selected encoding technique. This approach maximizes the number of audio channels 54 available for transmission.
- In principle, there is no upper limit for the total frequency range available for transmission. However, the sampling rate of the
microphone 20 can determine the maximum possible frequency it can receive. In particular, the maximum possible identifiable frequency is equal to half the sampling rate. In some implementations, commonly available microphones have a sampling rate of 44.1 kilo-samples per second. Consequently, the maximum available frequency such a microphone can detect is 22.05 kHz. - In principle, there is no lower limit for the total frequency range available for transmission. However, in some implementations, it is desirable for the audio signals 18 to be in inaudible to humans or other animals. This has the advantage that people would not be distracted by
audio signals 18 that are broadcast in their vicinity. According to one estimate, the audible range for most humans has an upper limit of 15 kHz. This, in turn, can be used as a lower limit of the total available frequency range. - Once the number of channels 54 is identified in
step 62, theresource identifier 52 can be segmented into as many parts as there are channels (step 64). For example, inFIG. 4 , the nine-bit resource identifier 52 is segmented into three parts, one for each of thecommunication channels resource identifier 52 is not evenly divisible by the number of communication channels 54, then the remainder bits can be associated with the higher-frequency channels. This can be advantageous for reasons of echo prevention, described more fully below (seeFIGS. 6A and 6B ). - After segmenting the
resource identifier 52, each segment can be processed as follows. A first segment is identified (step 64). That segment is encoded according to a chosen algorithm (step 66). There are a number of modulation algorithms that can be employed, including analog modulation, digital modulation, spread spectrum modulation, and others. Each algorithm has potential strengths and weaknesses, depending on the particular constraints that are important in the deployment of a particularresource retrieval system 10. - Among the possible encoding techniques is the class of continuous spectrum, frequency shift keying algorithms. One such algorithm, known as Gaussian minimum shift keying (“GMSK”), has the advantageous property of requiring a relatively narrow bandwidth, e.g., approximately 1.5 kHz) for each of three audio channels to transmit about ten bits (or thirty bits in aggregate) within two to five milliseconds (without any cyclic redundancy check). Consequently, using GMSK for encoding accommodates a relatively high number of audio channels 54 that can be used for transmission. It will be understood that other modulation techniques may be suitably adapted to encode data for high-frequency audio transmission. Thus, while GMSK has been usefully demonstrated to provide adequate data rates and robustness in a variety of physical contexts, this disclosure is not limited to GMSK coding unless explicitly stated to the contrary. For example, other shift-keyed encoding techniques include minimum shift keying, frequency shift keying, phase shift keying, quadrature phase shift keying, and so forth. Still more generally, other modulation techniques based upon frequency modulation, amplitude modulation, phase modulation, and various combinations of the foregoing, are generally known in the art and all such techniques that may be usefully adapted to reliably encode data onto a high-frequency audio carrier are intended to fall within the scope of this disclosure.
- Referring to
FIGS. 6A and 6B , concurrently broadcasting severalaudio signals single audio signal 18. InFIG. 6A , asingle audio signal 18 is broadcast from aloudspeaker 16; InFIG. 6B , threeaudio signals loudspeaker 16. If thesignal 18 carries the same information content as thesignals signal 18 is three times as long as anyindividual signal - An echo-related error may occur when a signal interferes with itself where a
microphone 20 is located. InFIG. 6A , for example, thesignal 18 reflects off awall 67, so that both the last portion and the (reflected) initial portion of the signal arrive at the microphone. By contrast, inFIG. 6B when theloudspeaker 16 andmicrophone 20 are in the same relative positions, there are no echo-related errors, because the signal lengths are sufficiently short. - The higher the frequency of an audio signal, the shorter its duration needs to be in order to send a certain fixed amount of data. Thus, allocating any extra bits of the resource identifier to higher-frequency channels helps reduce echo-related errors by keeping signals relatively short.
- Using a larger resource identifier has the advantage of being able to accommodate more distinct resources, but has the disadvantage of requiring longer audio signals 18, thereby increasing the risk of echo-related error, as described above. In some implementations, the
resource identifier 52 consists of an integer with a length between twenty four and thirty bits, which are segmented for transmission across three audio channels and encoded using GMSK modulation. In some implementations, theresource identifier 52 can also include other data, e.g., data describing a geographic or other context to which the resource relates. - Referring back to
FIG. 5 , if there are any unencoded segments remaining (decision 68), then theencoding process 58 returns back to step 64, where the next unencoded segment is selected. Once all segments have been encoded,process 58 concludes, having produced a number of audio signals that collectively encode the resource identifier. - Although the use of several separate (or substantially separate) audio channels to broadcast several audio signals is advantageous for reasons described herein, these techniques can still be successful with a single audio signal broadcast on a single audio channel.
-
FIG. 7 is a flowchart for a decoding process. Initially, an audio signal can be received (step 72). For example, the audio signal can be the result of broadcasting step 40 ofmethod 30. The received audio signal can then be decoded (step 74). The decoding process utilizes the compatible techniques of those used to encode the audio signal. For example, if GMSK techniques were used to encode the audio signal, those techniques are used to decode the signal. - Upon decoding the signal, the
step 74 produces one or more resource identifier segments. If more than one segment is produced, those segments are subsequently reassembled (step 76), thereby producing aresource identification 52. -
FIG. 8 is a schematic depiction of a venue with several content display points. Thevenue 80 includes several content display points 82, each with a correspondingeffective display area 84. For example, there may be a signalingcomputer 14 at eachdisplay point 82, and eacheffective display area 84 corresponds to the range of the signaling computer'sloudspeaker 16. Other examples are possible. - As a person navigates the
venue 80, they encounter the various content display points 82. Eachcontent display point 82 provides a content use cue to cause a user's computer to display content relevant to that particular content display point. For example, the content use cue can include anaudio signal 18 as described above. More generally, the content use cue can include any transmission that causes the user's computer to display specific content. For example, the content use cue can include a visual signal, a conventional data transmission over a network connection, etc. - In the foregoing scenario, it is sometimes convenient for the user to have obtained the pertinent content for the various content display points 82 in advance of encountering those display points. To facilitate this, one or more content acquisition sites 86 can be established in the
venue 80. For example, a content acquisition site can be established near anentrance 88 of thevenue 80. - A content acquisition site 86 can provide a content acquisition cue to a user. For example, the content acquisition cue can include an
audio signal 18 as describe above. More generally, the content acquisition cue can include any transmission that causes the user's computer to acquire pertinent content. For example, the content acquisition cue can include a visual signal, a conventional data transmission over a network connection, etc. A content acquisition cue can simply include a verbal instruction to load a particular piece of physical media, such as a DVD, that bears the content pertinent to the content display points 82. - In response to receiving the content acquisition cue, the user's computer acquires the content pertinent to the various content display sites. For example, the user's computer can perform the
method 42 ofFIG. 3 . The user's computer can obtain the content in other ways, such as acquiring the content via a conventional network connection to a local or remote computer, or by loading files from a local storage device. - Acquiring content prior to use can have advantages. For example, having content locally available when receiving a content use cue tends to allow the user's computer to use the content relatively quickly. Moreover, obtaining all the content for the various content use points 82 in the
venue 80 in one network transmission can provide power or bandwidth savings over obtaining content from the different content use points 82 on an as-needed basis. - In addition to the example discussed above involving a presenter and audience members at a conference utilizing a
resource retrieval system 10 to retrieve visual aides, interactive surveys, and the like, numerous other examples exist. - In one example, the
resource retrieval system 10 can be deployed at a retail venue. At various locations near specified products, an audio signal source can be deployed to cause computers that detect the audio signal to retrieve content pertinent to one or more products in the vicinity. Such content can include, e.g., advertisement materials, promotions, coupons, product descriptions, product reviews, links to product-related web sites, etc. Additionally, a store-wide audio signal can be broadcast periodically alerting shoppers to promotions or specials occurring in specified departments or locations in the retail venue. - In another example, the
resource retrieval system 10 can be deployed in a museum or other venue amenable to self-guided tours. Near points of interest (e.g., exhibits, historical sites, etc.), theresource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the point of interest. Moreover, different information can be displayed to different users. For example, an “advanced” presentation or a “beginner” presentation can be displayed, depending on the user's preferences. - In another example, the
resource retrieval system 10 can be deployed in a sporting event. At various times, theresource retrieval system 10 can be used to cause users' computers to automatically retrieve information pertinent to a particular competitor, a team, a play, etc. - In another example, the
resource retrieval system 10 can be deployed in a classroom setting. The teacher can broadcast audio signals to cause students' computers to display specific resources pertinent to a lecture. Similarly, the teacher can broadcast audio signals to cause students' computers to begin or end an examination. - In another example, the
resource retrieval system 10 can be deployed in any venue. At various locations, theresource retrieval system 10 can be used to cause users' computers to automatically retrieve information about the venue such as a map of the venue or surrounding areas. For example a loudspeaker can be deployed at information kiosks, orientation booths, entrances to the venue, etc. - In another example, at any venue, the resource retrieval system can be used to cause users' computers to automatically retrieve information about emergency procedures. For example, in the event of an emergency, a venue-wide broadcast can be sounded over the venue's alarm system to alert occupants about the emergency, where the venue-wide broadcast includes an audio signal according to the techniques described above. Emergency procedures can include evacuation routes, first aid information, safety information, etc.
- Many of the above systems, devices, methods, processes, and the like may be realized in hardware, software, or any combination of these suitable for the control, data acquisition, and data processing described herein. This includes realization in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices or processing circuitry, along with internal and/or external memory. This may also, or instead, include one or more application specific integrated circuits, programmable gate arrays, programmable array logic components, or any other device or devices that may be configured to process electronic signals. It will further be appreciated that a realization of the processes or devices described above may include computer-executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software. At the same time, processing may be distributed across devices such as the various systems described above, or all of the functionality may be integrated into a dedicated, standalone device. All such permutations and combinations are intended to fall within the scope of the present disclosure.
- In embodiments, disclosed herein are computer program products comprising computer-executable code or computer-usable code that, when executing on one or more computing devices (such as the devices/systems described above), performs any and/or all of the steps described above. The code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (such as random access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices. In another aspect, any of the processes described above may be embodied in any suitable transmission or propagation medium carrying the computer-executable code described above and/or any inputs or outputs from same.
- It will be appreciated that the methods and systems described above are set forth by way of example and not of limitation. Numerous variations, additions, omissions, and other modifications will be apparent to one of ordinary skill in the art. In addition, the order or presentation of method steps in the description and drawings above is not intended to require this order of performing the recited steps unless a particular order is expressly required or otherwise clear from the context.
- While particular embodiments of the present invention have been shown and described, it will be apparent to those skilled in the art that various changes and modifications in form and details may be made therein without departing from the spirit and scope of the invention as defined by the following claims. The claims that follow are intended to include all such variations and modifications that might fall within their scope, and should be interpreted in the broadest sense allowable by law.
Claims (19)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/459,497 US20120213383A1 (en) | 2009-05-27 | 2012-04-30 | Automatic resource retrieval and use |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18147209P | 2009-05-27 | 2009-05-27 | |
US12/789,419 US8805723B2 (en) | 2009-05-27 | 2010-05-27 | Acoustically transmitting a resource identifier in multiple concurrent segments |
US13/459,497 US20120213383A1 (en) | 2009-05-27 | 2012-04-30 | Automatic resource retrieval and use |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,419 Continuation US8805723B2 (en) | 2009-05-27 | 2010-05-27 | Acoustically transmitting a resource identifier in multiple concurrent segments |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120213383A1 true US20120213383A1 (en) | 2012-08-23 |
Family
ID=43221121
Family Applications (8)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,419 Expired - Fee Related US8805723B2 (en) | 2009-05-27 | 2010-05-27 | Acoustically transmitting a resource identifier in multiple concurrent segments |
US12/789,377 Expired - Fee Related US8718805B2 (en) | 2009-05-27 | 2010-05-27 | Audio-based synchronization to media |
US13/096,950 Abandoned US20110208334A1 (en) | 2009-05-27 | 2011-04-28 | Audio-based synchronization server |
US13/096,965 Abandoned US20110202687A1 (en) | 2009-05-27 | 2011-04-28 | Synchronizing audience feedback from live and time-shifted broadcast views |
US13/096,920 Abandoned US20110202156A1 (en) | 2009-05-27 | 2011-04-28 | Device with audio-based media synchronization |
US13/096,943 Abandoned US20110208333A1 (en) | 2009-05-27 | 2011-04-28 | Pre-processing media for audio-based synchronization |
US13/096,977 Expired - Fee Related US8789084B2 (en) | 2009-05-27 | 2011-04-28 | Identifying commercial breaks in broadcast media |
US13/459,497 Abandoned US20120213383A1 (en) | 2009-05-27 | 2012-04-30 | Automatic resource retrieval and use |
Family Applications Before (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/789,419 Expired - Fee Related US8805723B2 (en) | 2009-05-27 | 2010-05-27 | Acoustically transmitting a resource identifier in multiple concurrent segments |
US12/789,377 Expired - Fee Related US8718805B2 (en) | 2009-05-27 | 2010-05-27 | Audio-based synchronization to media |
US13/096,950 Abandoned US20110208334A1 (en) | 2009-05-27 | 2011-04-28 | Audio-based synchronization server |
US13/096,965 Abandoned US20110202687A1 (en) | 2009-05-27 | 2011-04-28 | Synchronizing audience feedback from live and time-shifted broadcast views |
US13/096,920 Abandoned US20110202156A1 (en) | 2009-05-27 | 2011-04-28 | Device with audio-based media synchronization |
US13/096,943 Abandoned US20110208333A1 (en) | 2009-05-27 | 2011-04-28 | Pre-processing media for audio-based synchronization |
US13/096,977 Expired - Fee Related US8789084B2 (en) | 2009-05-27 | 2011-04-28 | Identifying commercial breaks in broadcast media |
Country Status (2)
Country | Link |
---|---|
US (8) | US8805723B2 (en) |
WO (2) | WO2010138776A2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100305730A1 (en) * | 2009-05-27 | 2010-12-02 | Glitsch Hans M | Automatic resource retrieval and use |
Families Citing this family (93)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7644282B2 (en) | 1998-05-28 | 2010-01-05 | Verance Corporation | Pre-processed information embedding system |
US6737957B1 (en) | 2000-02-16 | 2004-05-18 | Verance Corporation | Remote control signaling using audio watermarks |
EP1552454B1 (en) | 2002-10-15 | 2014-07-23 | Verance Corporation | Media monitoring, management and information system |
US20060239501A1 (en) | 2005-04-26 | 2006-10-26 | Verance Corporation | Security enhancements of digital watermarks for multi-media content |
US8020004B2 (en) | 2005-07-01 | 2011-09-13 | Verance Corporation | Forensic marking using a common customization function |
US8781967B2 (en) | 2005-07-07 | 2014-07-15 | Verance Corporation | Watermarking in an encrypted domain |
US8316308B2 (en) | 2007-06-08 | 2012-11-20 | Google Inc. | Adaptive user interface for multi-source systems |
US8321261B2 (en) | 2007-12-14 | 2012-11-27 | John Nicholas and Kristin Gross | Integrated gourmet item data collection, recommender and vending system and method |
US8205148B1 (en) | 2008-01-11 | 2012-06-19 | Bruce Sharpe | Methods and apparatus for temporal alignment of media |
US8259938B2 (en) | 2008-06-24 | 2012-09-04 | Verance Corporation | Efficient and secure forensic marking in compressed |
US20100205628A1 (en) | 2009-02-12 | 2010-08-12 | Davis Bruce L | Media processing methods and arrangements |
US8489774B2 (en) | 2009-05-27 | 2013-07-16 | Spot411 Technologies, Inc. | Synchronized delivery of interactive content |
US20110063503A1 (en) * | 2009-07-06 | 2011-03-17 | Brand Steven M | Synchronizing secondary content to a multimedia presentation |
US8463875B2 (en) * | 2009-08-20 | 2013-06-11 | Google Inc. | Synchronized playback of media players |
US8457118B2 (en) | 2010-05-17 | 2013-06-04 | Google Inc. | Decentralized system and method for voice and video sessions |
US10324605B2 (en) | 2011-02-16 | 2019-06-18 | Apple Inc. | Media-editing application with novel editing tools |
US8965026B2 (en) * | 2011-06-10 | 2015-02-24 | Canopy Co. | Method and apparatus for remote capture of audio in a handheld device |
US9607131B2 (en) | 2010-09-16 | 2017-03-28 | Verance Corporation | Secure and efficient content screening in a networked environment |
EP2625816B1 (en) | 2010-10-08 | 2017-11-29 | Lumi Technologies Limited | Scalable synchronization of events among server and clients with varying lag-times |
US20130331972A1 (en) * | 2010-11-15 | 2013-12-12 | Egonocast | System and method for synchronizing an application with audio content |
US9277292B2 (en) * | 2010-12-09 | 2016-03-01 | Cox Communications, Inc. | System, method and device for providing a mobile application across smartphone platforms to enable consumer connectivity and control of media |
US8910032B2 (en) | 2011-01-28 | 2014-12-09 | Apple Inc. | Media-editing application with automatic background rendering capabilities |
US9251855B2 (en) | 2011-01-28 | 2016-02-02 | Apple Inc. | Efficient media processing |
US8621355B2 (en) | 2011-02-02 | 2013-12-31 | Apple Inc. | Automatic synchronization of media clips |
US11747972B2 (en) | 2011-02-16 | 2023-09-05 | Apple Inc. | Media-editing application with novel editing tools |
US9997196B2 (en) | 2011-02-16 | 2018-06-12 | Apple Inc. | Retiming media presentations |
US8798777B2 (en) * | 2011-03-08 | 2014-08-05 | Packetvideo Corporation | System and method for using a list of audio media to create a list of audiovisual media |
US9160837B2 (en) | 2011-06-29 | 2015-10-13 | Gracenote, Inc. | Interactive streaming content apparatus, systems and methods |
US8923548B2 (en) | 2011-11-03 | 2014-12-30 | Verance Corporation | Extraction of embedded watermarks from a host content using a plurality of tentative watermarks |
US8615104B2 (en) | 2011-11-03 | 2013-12-24 | Verance Corporation | Watermark extraction based on tentative watermarks |
US8682026B2 (en) | 2011-11-03 | 2014-03-25 | Verance Corporation | Efficient extraction of embedded watermarks in the presence of host content distortions |
US8533481B2 (en) | 2011-11-03 | 2013-09-10 | Verance Corporation | Extraction of embedded watermarks from a host content based on extrapolation techniques |
US9111579B2 (en) | 2011-11-14 | 2015-08-18 | Apple Inc. | Media editing with multi-camera media clips |
US8745403B2 (en) | 2011-11-23 | 2014-06-03 | Verance Corporation | Enhanced content management based on watermark extraction records |
US9323902B2 (en) | 2011-12-13 | 2016-04-26 | Verance Corporation | Conditional access using embedded watermarks |
US9547753B2 (en) | 2011-12-13 | 2017-01-17 | Verance Corporation | Coordinated watermarking |
US9418669B2 (en) * | 2012-05-13 | 2016-08-16 | Harry E. Emerson, III | Discovery of music artist and title for syndicated content played by radio stations |
US20130325601A1 (en) * | 2012-06-05 | 2013-12-05 | Yahoo! Inc. | System for providing content |
US8930457B2 (en) * | 2012-06-19 | 2015-01-06 | International Business Machines Corporation | Proximity initiated co-browsing sessions |
US8806518B2 (en) * | 2012-06-27 | 2014-08-12 | Intel Corporation | Performance analysis for combining remote audience responses |
US8843952B2 (en) | 2012-06-28 | 2014-09-23 | Google Inc. | Determining TV program information based on analysis of audio fingerprints |
US9113203B2 (en) | 2012-06-28 | 2015-08-18 | Google Inc. | Generating a sequence of audio fingerprints at a set top box |
CN104584571B (en) * | 2012-06-28 | 2018-01-09 | 谷歌公司 | Audio-frequency fingerprint sequence is produced at set top box |
US9088406B2 (en) * | 2012-07-29 | 2015-07-21 | Qualcomm Incorporated | Frame sync across multiple channels |
US9571606B2 (en) | 2012-08-31 | 2017-02-14 | Verance Corporation | Social media viewing system |
US9226011B2 (en) * | 2012-09-11 | 2015-12-29 | Comcast Cable Communications, Llc | Synchronizing program presentation |
US8869222B2 (en) | 2012-09-13 | 2014-10-21 | Verance Corporation | Second screen content |
US8726304B2 (en) | 2012-09-13 | 2014-05-13 | Verance Corporation | Time varying evaluation of multimedia content |
US9106964B2 (en) | 2012-09-13 | 2015-08-11 | Verance Corporation | Enhanced content distribution using advertisements |
US9661361B2 (en) | 2012-09-19 | 2017-05-23 | Google Inc. | Systems and methods for live media content matching |
US20140122588A1 (en) * | 2012-10-31 | 2014-05-01 | Alain Nimri | Automatic Notification of Audience Boredom during Meetings and Conferences |
US9318116B2 (en) | 2012-12-14 | 2016-04-19 | Disney Enterprises, Inc. | Acoustic data transmission based on groups of audio receivers |
US11375347B2 (en) * | 2013-02-20 | 2022-06-28 | Disney Enterprises, Inc. | System and method for delivering secondary content to movie theater patrons |
US9262793B2 (en) | 2013-03-14 | 2016-02-16 | Verance Corporation | Transactional video marking system |
US9499128B2 (en) | 2013-03-14 | 2016-11-22 | The Crawford Group, Inc. | Mobile device-enhanced user selection of specific rental vehicles for a rental vehicle reservation |
US10567489B2 (en) * | 2013-03-15 | 2020-02-18 | Time Warner Cable Enterprises Llc | System and method for seamless switching between data streams |
US10142700B2 (en) | 2013-03-15 | 2018-11-27 | DISH Technologies L.L.C. | Television service provided social networking service |
US20140336797A1 (en) * | 2013-05-12 | 2014-11-13 | Harry E. Emerson, III | Audio content monitoring and identification of broadcast radio stations |
US20140336799A1 (en) * | 2013-05-13 | 2014-11-13 | Harry E. Emerson, III | Discovery of music artist and title via companionship between a cellular phone and a broadcast radio receiver |
US9930404B2 (en) | 2013-06-17 | 2018-03-27 | Echostar Technologies L.L.C. | Event-based media playback |
US9251549B2 (en) | 2013-07-23 | 2016-02-02 | Verance Corporation | Watermark extractor enhancements based on payload ranking |
US9380383B2 (en) | 2013-09-06 | 2016-06-28 | Gracenote, Inc. | Modifying playback of content using pre-processed profile information |
US9208334B2 (en) | 2013-10-25 | 2015-12-08 | Verance Corporation | Content management using multiple abstraction layers |
EP3117626A4 (en) | 2014-03-13 | 2017-10-25 | Verance Corporation | Interactive content acquisition using embedded codes |
US20150302086A1 (en) | 2014-04-22 | 2015-10-22 | Gracenote, Inc. | Audio identification during performance |
GB2525912B (en) | 2014-05-08 | 2018-02-28 | Mewt Ltd | Synchronisation of audio and video playback |
GB2525914B (en) | 2014-05-08 | 2018-07-18 | Mewt Ltd | Synchronisation of audio and video playback |
GB2541785B (en) | 2014-05-08 | 2018-03-28 | Mewt Ltd | Synchronisation of audio and video playback |
US9786291B2 (en) | 2014-06-18 | 2017-10-10 | Google Technology Holdings LLC | Communicating information between devices using ultra high frequency audio |
US9912973B2 (en) | 2014-08-07 | 2018-03-06 | Echostar Technologies L.L.C. | Systems and methods for facilitating content discovery based on viewer ratings |
US10306319B2 (en) | 2014-09-09 | 2019-05-28 | Google Llc | Collaboration between a broadcaster and an audience for a broadcast |
US9792424B1 (en) | 2014-09-18 | 2017-10-17 | Amazon Technologies, Inc. | Chained authentication using musical transforms |
US9565474B2 (en) | 2014-09-23 | 2017-02-07 | Echostar Technologies L.L.C. | Media content crowdsource |
US10506295B2 (en) | 2014-10-09 | 2019-12-10 | Disney Enterprises, Inc. | Systems and methods for delivering secondary content to viewers |
US10432296B2 (en) | 2014-12-31 | 2019-10-01 | DISH Technologies L.L.C. | Inter-residence computing resource sharing |
US10218684B2 (en) * | 2015-01-30 | 2019-02-26 | Ncr Corporation | Secure data transmission |
KR102051985B1 (en) * | 2015-09-30 | 2019-12-04 | 애플 인크. | Synchronization of Media Rendering in Heterogeneous Networking Environments |
US11494761B2 (en) * | 2015-11-06 | 2022-11-08 | Cable Television Laboratories, Inc. | Systems and methods for digital asset security ecosystems |
US11941588B2 (en) | 2015-11-06 | 2024-03-26 | Cable Television Laboratories, Inc. | Systems and methods for blockchain virtualization and scalability |
US10306332B2 (en) * | 2016-06-14 | 2019-05-28 | DISH Technologies L.L.C. | Use of audio signals to provide interactive content to end users via smart devices |
US10129586B2 (en) * | 2016-12-19 | 2018-11-13 | Google Llc | Detecting and isolating television program content from recordings of television airings |
US10552218B2 (en) * | 2017-01-06 | 2020-02-04 | Microsoft Technology Licensing, Llc | Dynamic context of tasks |
US10397287B2 (en) | 2017-03-01 | 2019-08-27 | Microsoft Technology Licensing, Llc | Audio data transmission using frequency hopping |
US11568038B1 (en) | 2017-09-19 | 2023-01-31 | Amazon Technologies, Inc. | Threshold-based authentication |
US20190208236A1 (en) * | 2018-01-02 | 2019-07-04 | Source Digital, Inc. | Coordinates as ancillary data |
US10582254B2 (en) * | 2018-03-02 | 2020-03-03 | At&T Intellectual Property I, L.P. | Flexible event recording system and method |
CA3093023C (en) | 2018-04-09 | 2023-01-03 | Tunevu Inc. | System and method for management and delivery of secondary syndicated companion conent of discovered primary digital media presentations |
US10771866B2 (en) * | 2018-05-24 | 2020-09-08 | Google Llc | Methods, systems, and media synchronizing audio and video content on multiple media devices |
CN110162350B (en) * | 2019-05-27 | 2022-10-25 | 天津卓漫科技有限公司 | Notification bar information display method, notification bar information display device, server and storage medium |
US11006191B2 (en) * | 2019-08-02 | 2021-05-11 | The Nielsen Company (Us), Llc | Use of watermarking to control abandonment of dynamic content modification |
US11321904B2 (en) | 2019-08-30 | 2022-05-03 | Maxon Computer Gmbh | Methods and systems for context passing between nodes in three-dimensional modeling |
US11714928B2 (en) | 2020-02-27 | 2023-08-01 | Maxon Computer Gmbh | Systems and methods for a self-adjusting node workspace |
US11373369B2 (en) | 2020-09-02 | 2022-06-28 | Maxon Computer Gmbh | Systems and methods for extraction of mesh geometry from straight skeleton for beveled shapes |
Family Cites Families (107)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US734672A (en) * | 1900-10-06 | 1903-07-28 | Jacob C Counts | Grain-cradle. |
USRE37131E1 (en) * | 1991-02-19 | 2001-04-10 | Roy J. Mankovitz | Apparatus and methods for music and lyrics broadcasting |
US7562392B1 (en) * | 1999-05-19 | 2009-07-14 | Digimarc Corporation | Methods of interacting with audio and ambient music |
US6505160B1 (en) * | 1995-07-27 | 2003-01-07 | Digimarc Corporation | Connected audio and other media objects |
US7346472B1 (en) * | 2000-09-07 | 2008-03-18 | Blue Spike, Inc. | Method and device for monitoring and analyzing signals |
US6209028B1 (en) * | 1997-03-21 | 2001-03-27 | Walker Digital, Llc | System and method for supplying supplemental audio information for broadcast television programs |
AU6882998A (en) * | 1997-03-31 | 1998-10-22 | Broadband Associates | Method and system for providing a presentation on a network |
US7490169B1 (en) * | 1997-03-31 | 2009-02-10 | West Corporation | Providing a presentation on a network having a plurality of synchronized media types |
EP2261914A3 (en) * | 1998-02-23 | 2011-03-09 | Kabushiki Kaisha Toshiba | Information storage medium, information playback method and apparatus and information recording method |
US7423983B1 (en) * | 1999-09-20 | 2008-09-09 | Broadcom Corporation | Voice and data exchange over a packet based network |
US20060041639A1 (en) * | 1999-04-21 | 2006-02-23 | Interactual Technologies, Inc. | Platform detection |
US6674993B1 (en) * | 1999-04-30 | 2004-01-06 | Microvision, Inc. | Method and system for identifying data locations associated with real world observations |
US7013301B2 (en) * | 2003-09-23 | 2006-03-14 | Predixis Corporation | Audio fingerprinting system and method |
GB9917985D0 (en) * | 1999-07-30 | 1999-09-29 | Scient Generics Ltd | Acoustic communication system |
US6738803B1 (en) * | 1999-09-03 | 2004-05-18 | Cisco Technology, Inc. | Proxy browser providing voice enabled web application audio control for telephony devices |
US7194752B1 (en) * | 1999-10-19 | 2007-03-20 | Iceberg Industries, Llc | Method and apparatus for automatically recognizing input audio and/or video streams |
US6834308B1 (en) | 2000-02-17 | 2004-12-21 | Audible Magic Corporation | Method and apparatus for identifying media content presented on a media playing device |
US6453252B1 (en) * | 2000-05-15 | 2002-09-17 | Creative Technology Ltd. | Process for identifying audio content |
US7624337B2 (en) | 2000-07-24 | 2009-11-24 | Vmark, Inc. | System and method for indexing, searching, identifying, and editing portions of electronic multimedia files |
US7853664B1 (en) | 2000-07-31 | 2010-12-14 | Landmark Digital Services Llc | Method and system for purchasing pre-recorded music |
US6990453B2 (en) * | 2000-07-31 | 2006-01-24 | Landmark Digital Services Llc | System and methods for recognizing sound and music signals in high noise and distortion |
US7103906B1 (en) | 2000-09-29 | 2006-09-05 | International Business Machines Corporation | User controlled multi-device media-on-demand system |
US6947966B1 (en) * | 2000-10-13 | 2005-09-20 | Road Runner Holdco Llc | System and method for influencing dynamic community shared elements of audio, video, and text programming via a polling system |
US7277766B1 (en) * | 2000-10-24 | 2007-10-02 | Moodlogic, Inc. | Method and system for analyzing digital audio files |
AU2211102A (en) * | 2000-11-30 | 2002-06-11 | Scient Generics Ltd | Acoustic communication system |
US20020072982A1 (en) * | 2000-12-12 | 2002-06-13 | Shazam Entertainment Ltd. | Method and system for interacting with a user in an experiential environment |
DE60228202D1 (en) * | 2001-02-12 | 2008-09-25 | Gracenote Inc | METHOD FOR GENERATING AN IDENTIFICATION HASH FROM THE CONTENTS OF A MULTIMEDIA FILE |
US7359889B2 (en) * | 2001-03-02 | 2008-04-15 | Landmark Digital Services Llc | Method and apparatus for automatically creating database for use in automated media recognition system |
US6907570B2 (en) * | 2001-03-29 | 2005-06-14 | International Business Machines Corporation | Video and multimedia browsing while switching between views |
US6973574B2 (en) * | 2001-04-24 | 2005-12-06 | Microsoft Corp. | Recognizer of audio-content in digital signals |
US7421376B1 (en) | 2001-04-24 | 2008-09-02 | Auditude, Inc. | Comparison of data signals using characteristic electronic thumbprints |
EP1410380B1 (en) * | 2001-07-20 | 2010-04-28 | Gracenote, Inc. | Automatic identification of sound recordings |
US20030036408A1 (en) * | 2001-08-17 | 2003-02-20 | Johansson Lars Olof | High-density radio access system |
US7171367B2 (en) * | 2001-12-05 | 2007-01-30 | Ssi Corporation | Digital audio with parameters for real-time time scaling |
US20030120749A1 (en) * | 2001-12-21 | 2003-06-26 | Kotzin Michael D. | Method and apparatus for obtaining internet content for a wireless device |
US7142565B2 (en) * | 2002-02-21 | 2006-11-28 | Qualcomm, Incorporated | Method and apparatus for radio link control of signaling messages and short messages data services in a communication system |
US20030182663A1 (en) * | 2002-03-25 | 2003-09-25 | Sony Corporation | System and method for sharing user comments on TV screens |
DE60323086D1 (en) | 2002-04-25 | 2008-10-02 | Landmark Digital Services Llc | ROBUST AND INVARIANT AUDIO COMPUTER COMPARISON |
AU2003239385A1 (en) | 2002-05-10 | 2003-11-11 | Richard R. Reisman | Method and apparatus for browsing using multiple coordinated device |
US7231351B1 (en) * | 2002-05-10 | 2007-06-12 | Nexidia, Inc. | Transcript alignment |
US7752648B2 (en) * | 2003-02-11 | 2010-07-06 | Nds Limited | Apparatus and methods for handling interactive applications in broadcast networks |
US7768234B2 (en) * | 2004-02-28 | 2010-08-03 | Janik Craig M | System and method for automatically synchronizing and acquiring content for battery powered devices |
US7359006B1 (en) * | 2003-05-20 | 2008-04-15 | Micronas Usa, Inc. | Audio module supporting audio signature |
US7941554B2 (en) * | 2003-08-01 | 2011-05-10 | Microsoft Corporation | Sparse caching for streaming media |
US20050147256A1 (en) * | 2003-12-30 | 2005-07-07 | Peters Geoffrey W. | Automated presentation of entertainment content in response to received ambient audio |
US8010986B2 (en) * | 2004-02-04 | 2011-08-30 | Ericsson Television, Inc. | Synchronization and automation in an ITV environment |
US7325013B2 (en) * | 2004-04-15 | 2008-01-29 | Id3Man, Inc. | Database with efficient fuzzy matching |
US8688248B2 (en) * | 2004-04-19 | 2014-04-01 | Shazam Investments Limited | Method and system for content sampling and identification |
US20070067297A1 (en) * | 2004-04-30 | 2007-03-22 | Kublickis Peter J | System and methods for a micropayment-enabled marketplace with permission-based, self-service, precision-targeted delivery of advertising, entertainment and informational content and relationship marketing to anonymous internet users |
EP1751672A4 (en) * | 2004-05-03 | 2007-05-09 | Lg Electronics Inc | Method and apparatus for managing bookmark information for content stored in a networked media server |
US20050289531A1 (en) * | 2004-06-08 | 2005-12-29 | Daniel Illowsky | Device interoperability tool set and method for processing interoperability application specifications into interoperable application packages |
US20070276733A1 (en) * | 2004-06-23 | 2007-11-29 | Frank Geshwind | Method and system for music information retrieval |
US8190680B2 (en) | 2004-07-01 | 2012-05-29 | Netgear, Inc. | Method and system for synchronization of digital media playback |
US20090138108A1 (en) * | 2004-07-06 | 2009-05-28 | Kok Keong Teo | Method and System for Identification of Audio Input |
US7725318B2 (en) * | 2004-07-30 | 2010-05-25 | Nice Systems Inc. | System and method for improving the accuracy of audio searching |
US7399006B2 (en) * | 2004-09-22 | 2008-07-15 | Jungers Jon W | Method and apparatus for assembling HVAC ductwork components |
US20060068818A1 (en) | 2004-09-28 | 2006-03-30 | Amir Leitersdorf | Audience participation method and apparatus |
US20060068816A1 (en) * | 2004-09-30 | 2006-03-30 | Lucent Technologies Inc. | Network based contacts with optional DTMF exchange of the same |
US7567899B2 (en) * | 2004-12-30 | 2009-07-28 | All Media Guide, Llc | Methods and apparatus for audio recognition |
WO2006091636A2 (en) * | 2005-02-23 | 2006-08-31 | Digital Intelligence, L.L.C. | Signal decomposition and reconstruction |
US20060200037A1 (en) * | 2005-03-02 | 2006-09-07 | Falasco Marianne R | System and method for selectively collecting exhaled air |
US8356021B2 (en) * | 2005-03-11 | 2013-01-15 | Ross Neil Williams | Method and apparatus for indexing in a reduced-redundancy storage system |
US20070016918A1 (en) * | 2005-05-20 | 2007-01-18 | Alcorn Allan E | Detecting and tracking advertisements |
US8140954B2 (en) * | 2005-06-05 | 2012-03-20 | International Business Machines Corporation | System and method for providing on demand captioning of broadcast programs |
US7599840B2 (en) * | 2005-07-15 | 2009-10-06 | Microsoft Corporation | Selectively using multiple entropy models in adaptive coding and decoding |
US7516074B2 (en) * | 2005-09-01 | 2009-04-07 | Auditude, Inc. | Extraction and matching of characteristic fingerprints from audio signals |
US8290810B2 (en) * | 2005-09-14 | 2012-10-16 | Jumptap, Inc. | Realtime surveying within mobile sponsored content |
US7873259B2 (en) * | 2005-09-15 | 2011-01-18 | International Business Machines Corporation | Synching a recording time of a program to the actual program broadcast time for the program |
CA2628061A1 (en) * | 2005-11-10 | 2007-05-24 | Melodis Corporation | System and method for storing and retrieving non-text-based information |
KR100803206B1 (en) * | 2005-11-11 | 2008-02-14 | 삼성전자주식회사 | Apparatus and method for generating audio fingerprint and searching audio data |
US9319720B2 (en) | 2005-12-13 | 2016-04-19 | Audio Pod Inc. | System and method for rendering digital content using time offsets |
US8285809B2 (en) | 2005-12-13 | 2012-10-09 | Audio Pod Inc. | Segmentation and transmission of audio streams |
US8214516B2 (en) * | 2006-01-06 | 2012-07-03 | Google Inc. | Dynamic media serving infrastructure |
US20100138561A1 (en) * | 2006-01-27 | 2010-06-03 | Michael Wayne Church | Simulcast internet media distribution system and method |
US7627890B2 (en) * | 2006-02-21 | 2009-12-01 | At&T Intellectual Property, I,L.P. | Methods, systems, and computer program products for providing content synchronization or control among one or more devices |
US7735101B2 (en) * | 2006-03-28 | 2010-06-08 | Cisco Technology, Inc. | System allowing users to embed comments at specific points in time into media presentation |
US20070260634A1 (en) * | 2006-05-04 | 2007-11-08 | Nokia Corporation | Apparatus, system, method, and computer program product for synchronizing the presentation of media content |
EP2044772A4 (en) * | 2006-07-07 | 2010-03-31 | Redlasso Corp | Search engine for audio data |
WO2008011422A2 (en) * | 2006-07-21 | 2008-01-24 | Melodis Corporation | System and method for producing scored search results on a database using approximate search queries |
EP2070231B1 (en) * | 2006-10-03 | 2013-07-03 | Shazam Entertainment, Ltd. | Method for high throughput of identification of distributed broadcast content |
EP1933482A1 (en) * | 2006-12-13 | 2008-06-18 | Taylor Nelson Sofres Plc | audience measurement system, fixed and portable monitoring devices |
US9253222B2 (en) * | 2007-02-22 | 2016-02-02 | Match.Com. L.L.C. | Synchronous delivery of media content in a collaborative environment |
US7667126B2 (en) * | 2007-03-12 | 2010-02-23 | The Tc Group A/S | Method of establishing a harmony control signal controlled in real-time by a guitar input signal |
EP2135376A4 (en) * | 2007-03-22 | 2012-12-19 | Nielsen Co Us Llc | Digital rights management and audience measurement systems and methods |
JP4935468B2 (en) * | 2007-04-03 | 2012-05-23 | ソニー株式会社 | Information processing apparatus, information processing method, and computer program |
WO2008129482A2 (en) * | 2007-04-18 | 2008-10-30 | Koninklijke Philips Electronics, N.V. | Apparatus and methods for rendering personal stories to medical patients |
US8054382B2 (en) * | 2007-05-21 | 2011-11-08 | International Business Machines Corporation | Apparatus, method and system for synchronizing a common broadcast signal among multiple television units |
GB0710211D0 (en) | 2007-05-29 | 2007-07-11 | Intrasonics Ltd | AMR Spectrography |
US20080306824A1 (en) * | 2007-06-08 | 2008-12-11 | Parkinson David C | Empty Space Advertising Engine |
US8027365B2 (en) * | 2007-07-27 | 2011-09-27 | Realnetworks, Inc. | System and method for combining media data |
US8314688B2 (en) * | 2007-08-22 | 2012-11-20 | Tagarray, Inc. | Method and apparatus for low power modulation and massive medium access control |
US8311828B2 (en) * | 2007-08-28 | 2012-11-13 | Nexidia Inc. | Keyword spotting using a phoneme-sequence index |
US9662578B2 (en) * | 2007-09-21 | 2017-05-30 | Sony Interactive Entertainment America Llc | Methods and apparatuses for recording and utilizing an audio signature within an application |
WO2009083797A2 (en) * | 2007-10-17 | 2009-07-09 | Marvin Igelman | Synchronized media playback using autonomous clients over standard internet protocols |
US7899807B2 (en) * | 2007-12-20 | 2011-03-01 | Yahoo! Inc. | System and method for crawl ordering by search impact |
US8229160B2 (en) | 2008-01-03 | 2012-07-24 | Apple Inc. | Systems and methods for identifying objects and providing information related to identified objects |
US8161160B2 (en) * | 2008-02-28 | 2012-04-17 | Microsoft Corporation | XML-based web feed for web access of remote resources |
US8019882B2 (en) * | 2008-06-27 | 2011-09-13 | Microsoft Corporation | Content identification for peer-to-peer content retrieval |
KR101008853B1 (en) | 2008-07-28 | 2011-01-19 | 주식회사 삼천리기계 | A Hydraulic Chuck |
US8239359B2 (en) * | 2008-09-23 | 2012-08-07 | Disney Enterprises, Inc. | System and method for visual search in a video media player |
US8635645B2 (en) * | 2008-09-30 | 2014-01-21 | Qualcomm Incorporated | Apparatus and methods of providing and receiving venue level transmissions and services |
US20100088726A1 (en) * | 2008-10-08 | 2010-04-08 | Concert Technology Corporation | Automatic one-click bookmarks and bookmark headings for user-generated videos |
US20100299131A1 (en) | 2009-05-21 | 2010-11-25 | Nexidia Inc. | Transcript alignment |
WO2010138776A2 (en) * | 2009-05-27 | 2010-12-02 | Spot411 Technologies, Inc. | Audio-based synchronization to media |
US20120102233A1 (en) * | 2009-05-27 | 2012-04-26 | Ajay Shah | Synchronized delivery of interactive content using standardized vectors |
US20130097632A1 (en) * | 2009-05-27 | 2013-04-18 | Ajay Shah | Synchronization to broadcast media |
US8489774B2 (en) * | 2009-05-27 | 2013-07-16 | Spot411 Technologies, Inc. | Synchronized delivery of interactive content |
-
2010
- 2010-05-27 WO PCT/US2010/036492 patent/WO2010138776A2/en active Application Filing
- 2010-05-27 US US12/789,419 patent/US8805723B2/en not_active Expired - Fee Related
- 2010-05-27 US US12/789,377 patent/US8718805B2/en not_active Expired - Fee Related
- 2010-05-27 WO PCT/US2010/036493 patent/WO2010138777A1/en active Application Filing
-
2011
- 2011-04-28 US US13/096,950 patent/US20110208334A1/en not_active Abandoned
- 2011-04-28 US US13/096,965 patent/US20110202687A1/en not_active Abandoned
- 2011-04-28 US US13/096,920 patent/US20110202156A1/en not_active Abandoned
- 2011-04-28 US US13/096,943 patent/US20110208333A1/en not_active Abandoned
- 2011-04-28 US US13/096,977 patent/US8789084B2/en not_active Expired - Fee Related
-
2012
- 2012-04-30 US US13/459,497 patent/US20120213383A1/en not_active Abandoned
Non-Patent Citations (1)
Title |
---|
Gerasimov V., and Bender W., "Things that talk: using sound for device-to-device and device-to-human communication," IBM Systems Journal, Vol. 39, Nos. 3 & 4, pp. 530-546, December 2000 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100305730A1 (en) * | 2009-05-27 | 2010-12-02 | Glitsch Hans M | Automatic resource retrieval and use |
US8805723B2 (en) | 2009-05-27 | 2014-08-12 | Iviu Technologies, Llc | Acoustically transmitting a resource identifier in multiple concurrent segments |
Also Published As
Publication number | Publication date |
---|---|
US8789084B2 (en) | 2014-07-22 |
US8718805B2 (en) | 2014-05-06 |
WO2010138776A3 (en) | 2011-02-03 |
US20110208333A1 (en) | 2011-08-25 |
WO2010138777A1 (en) | 2010-12-02 |
US20110208334A1 (en) | 2011-08-25 |
US8805723B2 (en) | 2014-08-12 |
US20100305730A1 (en) | 2010-12-02 |
US20110202949A1 (en) | 2011-08-18 |
US20110202687A1 (en) | 2011-08-18 |
US20100305729A1 (en) | 2010-12-02 |
WO2010138776A2 (en) | 2010-12-02 |
US20110202156A1 (en) | 2011-08-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8805723B2 (en) | Acoustically transmitting a resource identifier in multiple concurrent segments | |
US11562751B2 (en) | Data communication using audio patterns systems and methods | |
US10959040B2 (en) | Venue and event recommendations for a user of a portable media player device | |
US20190173590A1 (en) | Sonic signaling communication for user devices | |
EP2681895B1 (en) | Method and apparatus for grouping client devices based on context similarity | |
US8862156B2 (en) | System effective to demodulate a modulated code and provide content to a user | |
US9065578B2 (en) | Audio watermark detection for delivering contextual content to a user | |
JP6454741B2 (en) | Low power related content providing system, method, and computer-readable recording medium recording program | |
US20140304068A1 (en) | System and method for providing inaudible codes and corresponding information to users via their computing devices | |
US20130078988A1 (en) | System effective to modulate a code and provide content to a user | |
JP2020536423A (en) | High bandwidth sonic tone generation | |
US20160316264A1 (en) | Electronic device, display apparatus, and method of operating the electronic device | |
Nittala et al. | Phoneear: interactions for mobile devices that hear high-frequency sound-encoded data | |
US20160182172A1 (en) | Data communication with acoustic signal communication | |
US10019987B2 (en) | Audible proximity messaging | |
CN104038772B (en) | Generate the method and device of ring signal file | |
US20180018712A1 (en) | Provider identifier announcement insertion | |
Jeon et al. | Developing a location-aware mobile guide system for GLAMs based on TAPIR sound tag: A case study of the Lee Ungno museum | |
CN104038773B (en) | Generate the method and device of ring signal file | |
CN104115475A (en) | Method and apparatus for providing neighborhood-based ringtone advertisements | |
US20160037237A1 (en) | System and method for encoding audio based on psychoacoustics | |
US10592924B1 (en) | Managing third party interactions with venue communications | |
WO2022136933A1 (en) | Method and associated system for information and generic data transmission |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: IVIU TECHNOLOGIES, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHAH, AJAY;SHIOZAKI, RANDOLPH;INTERACTION CAPITAL GROUP, LLC;REEL/FRAME:028310/0018 Effective date: 20100923 Owner name: SHAH, AJAY, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973 Effective date: 20100923 Owner name: INTERACTION CAPITAL GROUP, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973 Effective date: 20100923 Owner name: SHIOZAKI, RANDOLPH, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ARSH TECHNOLOGIES, LLC;REEL/FRAME:028309/0973 Effective date: 20100923 Owner name: ARSH TECHNOLOGIES, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GLITSCH, HANS M.;SHAH, AJAY;REEL/FRAME:028309/0877 Effective date: 20100922 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |