WO2014032709A1 - Audio rendering system - Google Patents
Audio rendering system Download PDFInfo
- Publication number
- WO2014032709A1 WO2014032709A1 PCT/EP2012/066775 EP2012066775W WO2014032709A1 WO 2014032709 A1 WO2014032709 A1 WO 2014032709A1 EP 2012066775 W EP2012066775 W EP 2012066775W WO 2014032709 A1 WO2014032709 A1 WO 2014032709A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- audio rendering
- mobile device
- microphones
- rendering system
- loudspeakers
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/301—Automatic calibration of stereophonic sound system, e.g. with test microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/01—Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
Definitions
- the present invention relates to an audio rendering system, in particular a stereo or multichannel rendering system for rendering room acoustics, a method for calibrating an audio rendering system and a portable electronic device, in particular a Smartphone or a Tablet PC usable for calibration of an audio rendering system.
- the optimization of loudspeaker and/or room rendering is a technology which aims at improving and/or correcting the audio rendering of non-standard loudspeaker layouts.
- Some examples of such optimization can be found in JP2000261900, FR2850183 and EP2378795.
- the methods are based on adaptive modifications, i.e. filtering, gain, delay, equalization, etc. of the audio signal computed and applied for each channel which are used for the optimization of the rendering.
- the adaptation of the rendering is usually based on the measurement of the audio signal which is actually received at the listening position. This optimal position is usually called the "sweet spot". This can be done by a directional microphone system, i.e.
- FIG. 9 illustrates a 5.1 multichannel loudspeaker system 900 according to the recommendation ITU-R BS 775-1.
- Three front loudspeakers (left L, center C, right R) are combined with two rear/side loudspeakers (left surround L s , right surround R s ).
- the left and right frontal loudspeakers are placed at the extremities of an arc subtending 60° at the reference listening point 901.
- Both side/rear loudspeakers L s , Rs should be placed within the sectors from 100° to 120° from the center front reference. Precise location is not necessary.
- Side/rear loudspeakers should be not closer to the listener than the frontal loudspeakers, unless compensating time delay is introduced.
- the frontal loudspeakers should ideally be at a height approximately equal to that of the listener's ears.
- Figure 10 illustrates a non-standardized 5.1 multichannel loudspeaker system 1000 which requires rendering adaptation and correction.
- the three front loudspeakers L, C, R are not placed on a straight line base and the two rear/side loudspeakers L s , Rs are located in different distances from the reference listening point 1001 .
- Both side/rear loudspeakers L s , Rs are not placed within the sectors from 100° to 120° from the center front reference.
- the left and right frontal loudspeakers are not placed at the extremities of the arc subtending 60° at the reference listening point 1001 .
- the invention is based on the finding that by using the available microphones (two or more) of a mobile device, e.g. a mobile device 1 100 depicted in Fig. 1 1 or any other mobile device like a Smartphone or a Tablet PC for example, associated with an adapted user interface allows a flexible and adaptive multi-loudspeaker calibration of the audio rendering system comprising thereof.
- a mobile device e.g. a mobile device 1 100 depicted in Fig. 1 1 or any other mobile device like a Smartphone or a Tablet PC for example
- Such an audio rendering system improves the prior art calibration system by considering the position of the user which holds a mobile device used for the calibration process.
- the audio rendering calibration system based on the mobile device comprises a mobile device with at least two microphones, synchronization means between the rendering system and the mobile device, e.g. WiFi, docking station, etc., rendering means for rendering of test or training signals, analysis means for analysis of the loudspeaker rendering system, e.g. position, frequency response, etc, and/or room characteristics, and compensation means for adaptive compensation of the audio rendering based on the analysis step performed by the analysis means.
- rendering a reproduction technique capable of creating spatial sound fields in an extended area by means of loudspeakers or loudspeaker arrays, sweet
- OS operational system
- App application on a mobile device
- WiFi Wireless Fidelity according to IEEE 802.1 1 standard.
- the invention relates to an audio rendering system for audio rendering room acoustics, comprising: a plurality of loudspeakers; a mobile device comprising at least two microphones, the mobile device being located at a room position; and a control device coupled to the plurality of loudspeakers and coupled to the mobile device, wherein the control device is configured to calibrate the audio rendering system based on the room position of the mobile device and on room positions of the
- the room position of the mobile device corresponds, for example, to a room position of a user using the mobile device.
- the audio rendering system allows a dynamic calibration, in particular a calibration when the user moves or changes his listening position, based on the synchronization between the rendering system and the mobile device, synchronized by WiFi, docking station, etc., for example.
- the audio rendering system provides a control interface to the user and allows the user to control the calibration of the rendering system.
- the user can directly feedback on the performance of the calibration, modify the detected position of the loudspeaker and thus improve the performance of the audio rendering system.
- control device is configured to start calibrating the audio rendering system by sending a training signal through the plurality of loudspeakers.
- the calibration of the audio rendering is started at a predetermined point in time.
- the mobile device is configured to record the training signal and to send the recorded training signal and/or information based thereupon to the control device.
- the control device is able to improve its audio rendering by exploiting data from a position inside the room.
- the room may be a closed room such as a theater, a concert hall or a small office room or it may be an open room such as an arena or a football stadium.
- complexity of the mobile device is low as processing is performed in the control device.
- information based on the recorded training signal is sent to the control device, data being transmitted can be kept low, as pre-processing is performed in the mobile device and only key performance data is sent to the control device.
- the information based on the recorded training signal comprises at least one of the following information: information on positions of the plurality of loudspeakers, information on a room characteristic, information on gains in signal paths between the plurality of loudspeakers and the at least two m rophones, information on delays in signal paths between the plurality of loudspeakers and the at least two microphones and information on transfer function in signal paths between the plurality of loudspeakers and the at least two microphones.
- control device is configured to adapt the audio rendering of the audio rendering system based on the recorded training signal and/or information based thereupon received from the mobile device.
- Audio rendering is not limited to static environments, it performs well in dynamic environments by an adaptive rendering process based on the recorded training signal.
- the control device is located in a docking station of the mobile device.
- the control device does not require a separate unit, it can be integrated in an existing unit such as a docking station. Implementation effort and costs can be kept low when an existing docking station is enhanced by control device functionality.
- each of the at least two microphones is located in the middle of a different edge of the mobile device.
- the invention relates to a portable electronic device, comprising: at least two microphones; and a processor configured to record a training signal received by the at least two microphones and to transmit the recorded training signal and/or information based thereupon via a transmit interface for calibration of an audio rendering system.
- the portable electronic device allows a dynamic calibration of an audio rendering system, in particular a calibration when the user moves or changes his listening position, based on the synchronization between the audio rendering system and the portable electronic device. The user is allowed to control the calibration of the audio rendering system.
- the user can directly feedback on the performance of the calibration, modify the detected position of the loudspeaker and thus improve the performance of the audio rendering system.
- the processor comprises an analysis module configured for analyzing the recorded training signal to provide the information based on the recorded training signal.
- the processor comprises an analysis module for analyzing the recorded training signal
- analyzed information based on the recorded training signal can be sent to the control device, thereby keeping the load of the interface between portable electronic device and the control device low. Only key performance data found by the analysis module is sent to the control device.
- the analysis module is configured to provide as information based on the recorded training signal at least one of the following information: information on a room
- the portable electronic device comprises a synchronization circuit configured to synchronize the recording of the training signal and the transmission of the recorded training signal and/or the information based thereupon with a control device initiating the calibration of the audio rendering system.
- Synchronizing the calibration process improves the accuracy of audio rendering.
- the portable electronic device comprises a graphical user interface configured for allowing a user to control the calibration of the audio rendering system by inputting information on a room characteristic used for adapting the audio rendering system.
- the graphical user interface enables the direct interaction with the user who can indicate if a loudspeaker which is positioned on the front should be actually positioned on the rear. For instance, if the mobile device is equipped with only two microphones, the analysis system can only discriminate the position of the loudspeaker according to one direction, e.g. left/right or front/back depending on the position of the microphones on the mobile device. The final set of loudspeaker positions and necessary rendering adaptation are then determined according to the recording and additional user information. The user, however, cannot provide the sufficient information on potential delay and equalization between channels. Recording is always necessary to achieve the optimal rendering.
- the portable electronic device may comprise an App, i.e., an application tool on a mobile device with a graphic user interface, which can be installed in the mobile OS (operational system) directly.
- the user can directly get the feedback of the rendering system on the display, e.g. information on loudspeaker position, configuration of the room, etc. and control the adaptation of the rendering system without any difficulties.
- the portable electronic device further comprises earphones, wherein the processor is configured to record the training signal received by microphones integrated in the earphones for binaurally capturing the training signal at an ear canal of a user using the portable electronic device or to record the training signal by a combination of the at least two microphones and the microphones integrated in the earphones.
- the invention relates to a method for calibrating an audio rendering system for audio rendering room acoustics, the method comprising: recording a training signal by at least two microphones of a mobile device being located in a room, the training signal being transmitted through a plurality of loudspeakers located in the room; transmitting the recorded training signal and/or information based thereupon to a control device configured for calibrating the audio rendering system.
- the method further comprises: initiating the calibrating by the mobile device by transmitting an initiation signal to the control device.
- the calibration of the audio rendering is acknowledged by the mobile device and thus started at a predetermined point in time.
- DSP Digital Signal Processor
- ASIC application specific integrated circuit
- the invention can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations thereof.
- Fig. 1 shows a schematic diagram of a basic audio rendering system according to an implementation form
- Fig. 2 shows a schematic diagram of an audio rendering system with the user being outside the system according to an implementation form
- Fig. 3 shows a schematic diagram of an audio rendering system with all loudspeakers being positioned in front of the user according to an implementation form
- Fig. 4 shows a schematic diagram of an audio rendering system with calibration based on a docking station according to an implementation form
- Fig. 5 shows a schematic diagram of a microphone arrangement on a Tablet PC according to an implementation form
- Fig. 6 shows a schematic diagram of an audio rendering system with calibration based on a docking station integrating a control device according to an implementation form
- Fig. 7 shows a schematic diagram of an audio rendering system with calibration based on microphones installed in headphones of a listener according to an implementation form
- Fig. 8 shows a schematic diagram of a method for calibrating an audio rendering system according to an implementation form
- Fig. 9 shows a schematic diagram of a 5.1 multichannel loudspeaker system according to the recommendation ITU- BS 775-1 ;
- Fig. 10 shows a schematic diagram of a non-standardized 5.1 multichannel loudspeaker system
- Fig. 1 1 shows a schematic diagram of a microphone configuration of a conventional mobile phone.
- the first one corresponds to the configuration depicted in Fig. 1 1 but the second microphone 1 107 depicted in Fig. 1 1 is missing. It has only one single microphone 1 1 13 which is used for communication and any other mono sound pickup.
- This first configuration is based on omnidirectional microphone and cannot provide any stereo image.
- Fig. 1 1 The other configuration is illustrated in Fig. 1 1 and uses two omnidirectional microphones, the main microphone 1 1 13 and the auxiliary microphone 1 107.
- the main microphone 1 1 13 is used for the sound pick up, i.e. for communication application as well as simple audio/video recording.
- the auxiliary microphone 1 107 is used for noise cancellation and gain control.
- An omnidirectional microphone needs only one hole in the terminal as opposed to directional microphones. Indeed, the omnidirectional microphone offers a uniform directivity pattern in all the direction, equivalent to a sphere, and only one hole is then required for the microphone housing 1 1 15. The size of the hole and the actual microphone housing 1 1 15 will affect the directivity of the complete system, but a single hole is sufficient for perfect sound pickup. Two omnidirectional microphones mounted with a spacing of several centimeters can also be used in order to obtain a stereo recording.
- Fig. 1 shows a schematic diagram of a basic audio rendering system 100 according to an implementation form.
- a user or listener 109 sitting on a sofa holds a mobile device 105, e.g. a Smartphone or a tablet PC including at least two microphones 107.
- the mobile device 105 is connected to a control device 103, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 103 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 100 comprises the following elements:
- a plurality of loudspeakers 101 e.g. a loudspeaker array or a loudspeaker system
- a mobile device 105 e.g. a Smartphone or a tablet PC with multiple microphones
- 107 i.e., at least two omnidirectional or directional microphones
- control device 103 for the adaptive compensation of the audio rendering based on an analysis step of the analysis device.
- the analysis and control devices are implemented in separated devices. In an alternative implementation form, the analysis and control devices are implemented in the same device.
- the adaptation of the multi-loudspeaker audio rendering can be directly calculated in the analysis device if the test signal is known a priori by this device.
- the user 109 of the mobile device 105 performs the following steps which interact with the control device 103:
- mobile device 105 sends indication to control device 103 of starting the procedure through the connection means, e.g. via WiFi, WLAN or docking station;
- control device 103 starts the rendering of training signal through the loudspeaker system, i.e. through the multi-loudspeaker 101 shown in Fig. 1 ;
- mobile device 105 records the training signal with multiple microphones 107; mobile device 105 analyzes the position of loudspeakers 101 and/or room characteristics;
- a set of parameters e.g. delay, gains, filters, etc. are extracted by the mobile device 105 to build a map of the loudspeaker system and to send these parameters to the control device 103.
- at least positions of the loudspeakers 101 are transmitted by applying a specific protocol to exchange this information.
- the gains and delays are also transmitted.
- the mobile device 105 sends the recorded signals to the control device 103 and the analysis is done in the control device 103; and
- control device 103 adapts the audio rendering based on the loudspeaker positions and/or room characteristic analysis.
- the mobile device 105 performs the computation of adaptation parameters such as delay, gains, filters, equalizer, etc. and sends these parameters to the control device 103.
- adaptation parameters such as delay, gains, filters, equalizer, etc.
- the computation of the adaptation parameters are then performed in the rendering system, i.e., in the control device 103 in order to select the most appropriate processing, e.g., post processing of the audio channel signal, or adaptation in the coded domain.
- Fig. 2 shows a schematic diagram of an audio rendering system 200 with the user 209 being outside the system 200 according to an implementation form.
- a user or listener 209 sitting on a sofa outside the audio system holds a mobile device 205, e.g. a Smartphone or a tablet PC including at least two microphones 207.
- the mobile device 205 is connected to a control device 203, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 203 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 200 comprises the following elements: a plurality of loudspeakers 201 , a mobile device 205 with at least two omnidirectional or directional microphones, a synchronization means between the control device 203 and the mobile device 205, an analysis device of the audio rendering system and/or room characteristics and a control device 103 for the adaptive compensation of the audio rendering based on the analysis of the analysis device.
- the calibration is performed analogously to the procedure described with respect to Fig. 1 .
- the audio rendering system 200 is able to calibrate the system if the user 209 is not located at the sweet spot of the audio rendering system even if the user 209 is sitting outside the audio system.
- Fig. 3 shows a schematic diagram of an audio rendering system 300 with all loudspeakers being positioned in front of the user according to an implementation form.
- the user 309 holds a mobile device 305, e.g. a Smartphone or a tablet PC including at least two microphones 307.
- the mobile device 305 is connected to a control device 303, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 303 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 300 comprises the following elements: a plurality of loudspeakers 301 arranged in line in front of the user 309, a mobile device 305 with at least two omnidirectional or directional microphones 307, synchronization means between the control device 303 and the mobile device 305, analysis device of the audio rendering system and/or room characteristics and a control device 303 for the adaptive compensation of the audio rendering based on the analysis of the analysis device.
- the calibration is performed analogously to the procedure described with respect to Fig. 1 .
- the audio rendering system 300 is able to calibrate the system even if the
- loudspeakers 301 are arranged in front of the user 309 and if there are no surround loudspeakers available in the audio system.
- Fig. 4 shows a schematic diagram of an audio rendering system 400 with calibration based on a docking station 401 according to an implementation form.
- a user or listener 409 is sitting on a sofa behind a docking station 401 comprising at least two loudspeakers.
- the user 409 holds a mobile device 405, e.g. a Smartphone or a tablet PC including at least two microphones 407.
- the mobile device 405 is connected to a control device 403, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 403 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 400 comprises the following elements: loudspeakers integrated in a docking station 401 arranged in front of the user 409, a mobile device 405 with at least two omnidirectional or directional microphones 407, synchronization means between the control device 403 and the mobile device 405, analysis device of the audio rendering system and/or room characteristics and a control device 403 for the adaptive compensation of the audio rendering based on the analysis of the analysis device.
- Docking station 401 and control device 403 are separate units.
- Fig. 5 shows a schematic diagram of a microphone arrangement on a Tablet PC 500 according to an implementation form.
- the Tablet PC 500 comprises a number of four microphones 507 arranged in the middle of each edge of the tablet in order to better discriminate or distinguish the directions of the sounds.
- the tablet PC 500 corresponds to the mobile device described above with respect to Figures 1 to 4.
- the table PC 500 is adapted to perform the calibration analogously to the procedure described with respect to Fig. 1.
- Fig. 6 shows a schematic diagram of an audio rendering system 600 with calibration based on a docking station integrating a control device according to an implementation form.
- a user or listener 609 sitting on a sofa holds a mobile device 605, e.g. a Smartphone or a tablet PC including at least two microphones 607.
- the mobile device 605 is connected to a control device 603, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 603 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 600 comprises the following elements:
- mobile device 605 e.g. a Smartphone or a tablet PC with multiple microphones 607, i.e., at least two omnidirectional or directional microphones;
- control device integrated in the docking station 601 for the adaptive compensation of the audio rendering based on an analysis step of the analysis device.
- the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
- mobile device 605 sends indication to the docking station 601 of starting the calibration procedure
- docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
- mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
- the mobile device 605 puts the mobile device 605 on the docking station 601 and the mobile device 605 provides the loudspeaker positions and/or room characteristics to the docking station 605 which adapts the rendering based on the loudspeaker positions and/or room characteristics analysis.
- the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
- mobile device 605 sends indication to the docking station 601 of starting the calibration procedure
- docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
- mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
- the user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 transmits the recorded signal to the docking station 601 and the analysis is done directly in the docking station 601 prior to the adaptation of the rendering based on the loudspeaker positions and/or room characteristics analysis of the control device in the docking station 601 .
- the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
- mobile device 605 sends timing information to the docking station 601 to inform docking station 601 of starting the calibration procedure;
- docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
- mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
- the user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 provides the loudspeaker positions and/or room characteristics to the docking station 605 which adapts the rendering based on the loudspeaker positions and/or room characteristics analysis.
- the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
- mobile device 605 sends timing information to the docking station 601 to inform docking station 601 of starting the calibration procedure;
- docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
- mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
- the mobile device 605 puts the mobile device 605 on the docking station 601 and the mobile device 605 transmits the recorded signal to the docking station 601 and the analysis is done directly in the docking station 601 prior to the adaptation of the rendering based on the loudspeaker positions and/or room characteristics analysis of the control device in the docking station 601.
- Fig. 7 shows a schematic diagram of an audio rendering system 700 with calibration based on microphones installed in headphones of a listener according to an
- a user or listener 709 is sitting on a sofa and holds a mobile device 705, e.g. a
- the Smartphone or a tablet PC including at least two microphones 707.
- the user 709 carries earphones with stereo microphones 713.
- the mobile device 705 is connected to a control device 703, e.g. a set top box or an amplifier through a wireless or a wire line connection.
- the control device 703 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc.
- the audio rendering system 700 comprises the following elements: front loudspeakers 701 and surround loudspeakers 71 1 , a mobile device 705 with at least two omnidirectional or directional microphones 707, synchronization means between the control device 703 and the mobile device 705, analysis device of the audio rendering system and/or room characteristics and a control device 703 for the adaptive
- the audio rendering system 700 is based on the use of stereo microphones 713 installed on the earphones or headphones and which are connected to the mobile device 705.
- this alternative scenario is based on binaural recording made through earphones or headphones with microphones 713 mounted on each side.
- Microphones 713 are directly integrated in the earphones/headphones capturing the audio signal directly at the ear canal.
- the user 709 connects the earphones/headphones to the mobile device 705 and the sound recording is done by this binaural microphone 713.
- This implementation form is advantageously based on recording which is directly representative of human perception of the multi-loudspeaker rendering system.
- the rendering adaptation is directly based on the recording at the user's 709 ears which ensure that the adaptation, if done properly, will perfectly reconstruct the optimal multichannel signal at the ears of the listener 709.
- the analysis device combines the recording by the microphones 713 integrated to the earphones with the recording by the
- FIG. 8 shows a schematic diagram of a method 800 for calibrating an audio rendering system according to an implementation form.
- the method 800 comprises recording 801 a training signal by at least two microphones of a mobile device being located in a room, the training signal being transmitted through a plurality of loudspeakers located in the room.
- the method 800 comprises transmitting 803 the recorded training signal and/or information based thereupon to a control device configured for calibrating the audio rendering system.
- the method 800 comprises initiating the calibrating by the mobile device by transmitting an initiation signal to the control device.
- the present disclosure also supports a system configured to execute the performing and computing steps described herein.
Abstract
The invention relates to an audio rendering system(100) for audio rendering room acoustics, comprising: a plurality of loudspeakers (101); a mobile device (105) comprising at least two microphones (107), the mobile device (105) being located at a room position of a user (109) using the mobile device (105); and a control device (103) coupled to the plurality of loudspeakers (101) and coupled to the mobile device (105), wherein the control device (103) is configured to calibrate the audio rendering system (100) based on the room position of the user (109) and on room positions of the loudspeakers(101) by evaluating information of the at least two microphones (107) of the mobile device (105).
Description
DESCRIPTION
Audio rendering system BACKGROUND OF THE INVENTION
The present invention relates to an audio rendering system, in particular a stereo or multichannel rendering system for rendering room acoustics, a method for calibrating an audio rendering system and a portable electronic device, in particular a Smartphone or a Tablet PC usable for calibration of an audio rendering system.
The optimization of loudspeaker and/or room rendering is a technology which aims at improving and/or correcting the audio rendering of non-standard loudspeaker layouts. Some examples of such optimization can be found in JP2000261900, FR2850183 and EP2378795. The methods are based on adaptive modifications, i.e. filtering, gain, delay, equalization, etc. of the audio signal computed and applied for each channel which are used for the optimization of the rendering. The adaptation of the rendering is usually based on the measurement of the audio signal which is actually received at the listening position. This optimal position is usually called the "sweet spot". This can be done by a directional microphone system, i.e. at least a stereo microphone which aims at detecting the position of the loudspeaker and then provides the necessary information for the adaptation of the rendering. Those methods can be applied to a multichannel audio rendering system, i.e. an audio rendering system having at least two and more channels like a 5.1 , 7.1 , 10.2 or 22.2 system. In prior art, calibration systems are automatic and are not controlled by the user and do not give any freedom to the user for the adaptation. For multichannel audio rendering, the loudspeaker positions are usually standardized in order to have an optimized rendering for this loudspeaker layout. Figure 9 illustrates a 5.1 multichannel loudspeaker system 900 according to the recommendation ITU-R BS 775-1. Three front loudspeakers (left L, center C, right R) are combined with two rear/side loudspeakers (left surround Ls, right surround Rs). The left and right frontal loudspeakers are placed at the extremities of an arc subtending 60° at the reference listening point 901. For reasons of available space, it is preferred to place the frontal loudspeakers L, C, R on a straight line base, then it may be necessary to introduce compensating time delays in the signal feed of the center loudspeaker C. Both side/rear loudspeakers Ls, Rs should be
placed within the sectors from 100° to 120° from the center front reference. Precise location is not necessary. Side/rear loudspeakers should be not closer to the listener than the frontal loudspeakers, unless compensating time delay is introduced. The frontal loudspeakers should ideally be at a height approximately equal to that of the listener's ears.
However, in practical applications, it is often difficult to strictly follow the standardized loudspeaker layout due to non-compatible dimensions of the room or other external constraints. Figure 10 illustrates a non-standardized 5.1 multichannel loudspeaker system 1000 which requires rendering adaptation and correction.
The three front loudspeakers L, C, R are not placed on a straight line base and the two rear/side loudspeakers Ls, Rs are located in different distances from the reference listening point 1001 . Both side/rear loudspeakers Ls, Rs are not placed within the sectors from 100° to 120° from the center front reference. The left and right frontal loudspeakers are not placed at the extremities of the arc subtending 60° at the reference listening point 1001 .
Conventional rendering adaptation systems are based on the configuration depicted in Figure 9, i.e. they can only adapt the rendering to the listening position, i.e. the sweet spot 901 . They cannot be controlled by the user or give any freedom to the user for the adaptation. A calibration process with a dedicated system does not allow an easy and dynamic calibration as it is based on a dedicated microphone and calibration is thus inflexible.
SUMMARY OF THE INVENTION
It is the object of the invention to provide a concept for an audio rendering system providing a flexible and adaptive multi-loudspeaker calibration.
This object is achieved by the features of the independent claims. Further implementation forms are apparent from the dependent claims, the description and the figures.
The invention is based on the finding that by using the available microphones (two or more) of a mobile device, e.g. a mobile device 1 100 depicted in Fig. 1 1 or any other mobile device like a Smartphone or a Tablet PC for example, associated with an adapted user interface allows a flexible and adaptive multi-loudspeaker calibration of the audio rendering system comprising thereof. Such an audio rendering system improves the prior art calibration system by considering the position of the user which holds a mobile device used for the calibration process. In an implementation form, the audio rendering calibration system based on the mobile device comprises a mobile device with at least two microphones, synchronization means between the rendering system and the mobile device, e.g. WiFi, docking station, etc., rendering means for rendering of test or training signals, analysis means for analysis of the loudspeaker rendering system, e.g. position, frequency response, etc, and/or room characteristics, and compensation means for adaptive compensation of the audio rendering based on the analysis step performed by the analysis means.
By applying such audio rendering system where calibration is controlled by an adapted user interface evaluating information delivered by two or more microphones of the user's mobile device, the calibration process is significantly improved and the audio rendering is flexible and dynamic with respect to changing environments as will be presented in the following.
In order to describe the invention in detail, the following terms, abbreviations and notations will be used: audio
rendering: a reproduction technique capable of creating spatial sound fields in an extended area by means of loudspeakers or loudspeaker arrays, sweet
spot: listening position for optimal measurement of the audio signal as it is
actually received by the listener,
OS: operational system, App: application on a mobile device,
WiFi Wireless Fidelity according to IEEE 802.1 1 standard.
According to a first aspect, the invention relates to an audio rendering system for audio rendering room acoustics, comprising: a plurality of loudspeakers; a mobile device comprising at least two microphones, the mobile device being located at a room position; and a control device coupled to the plurality of loudspeakers and coupled to the mobile device, wherein the control device is configured to calibrate the audio rendering system based on the room position of the mobile device and on room positions of the
loudspeakers by evaluating information of the at least two microphones of the mobile device. The room position of the mobile device corresponds, for example, to a room position of a user using the mobile device.
The audio rendering system allows a dynamic calibration, in particular a calibration when the user moves or changes his listening position, based on the synchronization between the rendering system and the mobile device, synchronized by WiFi, docking station, etc., for example.
Moreover, the audio rendering system provides a control interface to the user and allows the user to control the calibration of the rendering system. Hence, the user can directly feedback on the performance of the calibration, modify the detected position of the loudspeaker and thus improve the performance of the audio rendering system.
In a first possible implementation form of the audio rendering system according to the first aspect, the control device is configured to start calibrating the audio rendering system by sending a training signal through the plurality of loudspeakers.
By sending a training signal through the loudspeakers, the calibration of the audio rendering is started at a predetermined point in time.
In a second possible implementation form of the audio rendering system according to the first implementation form of the first aspect, the mobile device is configured to record the training signal and to send the recorded training signal and/or information based thereupon to the control device.
When the mobile device records the training signal and sends the recorded training signal and/or information based thereupon to the control device, the control device is able to improve its audio rendering by exploiting data from a position inside the room. The room may be a closed room such as a theater, a concert hall or a small office room or it may be an open room such as an arena or a football stadium. When the recorded training signal is sent to the control device, complexity of the mobile device is low as processing is performed in the control device. When information based on the recorded training signal is sent to the control device, data being transmitted can be kept low, as pre-processing is performed in the mobile device and only key performance data is sent to the control device.
In a third possible implementation form of the audio rendering system according to the second implementation form of the first aspect, the information based on the recorded training signal comprises at least one of the following information: information on positions of the plurality of loudspeakers, information on a room characteristic, information on gains in signal paths between the plurality of loudspeakers and the at least two m rophones, information on delays in signal paths between the plurality of loudspeakers and the at least two microphones and information on transfer function in signal paths between the plurality of loudspeakers and the at least two microphones.
The more information the control device receives the better the accuracy of the audio rendering.
In a fourth possible implementation form of the audio rendering system according to the second implementation form or according to the third implementation form of thefirst aspect, the control device is configured to adapt the audio rendering of the audio rendering system based on the recorded training signal and/or information based thereupon received from the mobile device. Audio rendering is not limited to static environments, it performs well in dynamic environments by an adaptive rendering process based on the recorded training signal.
In a fifth possible implementation form of the audio rendering system according to the first aspect as such or according to any of the preceding implementation forms of the first aspect, the control device is located in a docking station of the mobile device.
The control device does not require a separate unit, it can be integrated in an existing unit such as a docking station. Implementation effort and costs can be kept low when an existing docking station is enhanced by control device functionality.
In a sixth possible implementation form of the audio rendering system according to the first aspect as such or according to any of the preceding implementation forms of the first aspect, each of the at least two microphones is located in the middle of a different edge of the mobile device.
When the microphones are located in the middle of different edges of the mobile device their spatial directivity is improved.
According to a second aspect, the invention relates to a portable electronic device, comprising: at least two microphones; and a processor configured to record a training signal received by the at least two microphones and to transmit the recorded training signal and/or information based thereupon via a transmit interface for calibration of an audio rendering system. The portable electronic device allows a dynamic calibration of an audio rendering system, in particular a calibration when the user moves or changes his listening position, based on the synchronization between the audio rendering system and the portable electronic device. The user is allowed to control the calibration of the audio rendering system.
Hence, the user can directly feedback on the performance of the calibration, modify the detected position of the loudspeaker and thus improve the performance of the audio rendering system.
In a first possible implementation form of the portable electronic device according to the second aspect, the processor comprises an analysis module configured for analyzing the recorded training signal to provide the information based on the recorded training signal.
When the processor comprises an analysis module for analyzing the recorded training signal, analyzed information based on the recorded training signal can be sent to the control device, thereby keeping the load of the interface between portable electronic
device and the control device low. Only key performance data found by the analysis module is sent to the control device.
In a second possible implementation form of the portable electronic device according to the second aspect as such or according to the first implementation form of the second aspect, the analysis module is configured to provide as information based on the recorded training signal at least one of the following information: information on a room
characteristic, information on room positions of the loudspeakers, information on gains in signal paths between the loudspeakers and the at least two microphones, information on delays in signal paths between the loudspeakers and the at least two microphones, and information on transfer function in signal paths between the loudspeakers and the at least two microphones.
The more information the control device receives the better the accuracy of the audio rendering.
In a third possible implementation form of the portable electronic device according to the second aspect as such or according to any of the preceding implementation forms of the second aspect, the portable electronic device comprises a synchronization circuit configured to synchronize the recording of the training signal and the transmission of the recorded training signal and/or the information based thereupon with a control device initiating the calibration of the audio rendering system.
Synchronizing the calibration process improves the accuracy of audio rendering.
In a fourth possible implementation form of the portable electronic device according to the second aspect as such or according to any of the preceding implementation forms of the second aspect, the portable electronic device comprises a graphical user interface configured for allowing a user to control the calibration of the audio rendering system by inputting information on a room characteristic used for adapting the audio rendering system.
The graphical user interface enables the direct interaction with the user who can indicate if a loudspeaker which is positioned on the front should be actually positioned on the rear. For instance, if the mobile device is equipped with only two microphones, the analysis
system can only discriminate the position of the loudspeaker according to one direction, e.g. left/right or front/back depending on the position of the microphones on the mobile device. The final set of loudspeaker positions and necessary rendering adaptation are then determined according to the recording and additional user information. The user, however, cannot provide the sufficient information on potential delay and equalization between channels. Recording is always necessary to achieve the optimal rendering.
The portable electronic device may comprise an App, i.e., an application tool on a mobile device with a graphic user interface, which can be installed in the mobile OS (operational system) directly. The user can directly get the feedback of the rendering system on the display, e.g. information on loudspeaker position, configuration of the room, etc. and control the adaptation of the rendering system without any difficulties.
In a fifth possible implementation form of the portable electronic device according to the second aspect as such or according to any of the preceding implementation forms of the second aspect, the portable electronic device further comprises earphones, wherein the processor is configured to record the training signal received by microphones integrated in the earphones for binaurally capturing the training signal at an ear canal of a user using the portable electronic device or to record the training signal by a combination of the at least two microphones and the microphones integrated in the earphones.
Thus, audio rendering can directly exploit information of the sweet spot. By a combined recording the audio rendering is improved and adapted to the listener's position. According to a third aspect, the invention relates to a method for calibrating an audio rendering system for audio rendering room acoustics, the method comprising: recording a training signal by at least two microphones of a mobile device being located in a room, the training signal being transmitted through a plurality of loudspeakers located in the room; transmitting the recorded training signal and/or information based thereupon to a control device configured for calibrating the audio rendering system.
This direct interaction with the user allows to always providing the rendering sweet spot to the listeners. It improves the multichannel audio rendering at the listening position for applications like Home Cinema, Home Theater, sound bar, docking station etc. based on an easy to use calibration tool based on mobile device as analysis/control tool.
In a first possible implementation form of the method according to the third aspect, the method further comprises: initiating the calibrating by the mobile device by transmitting an initiation signal to the control device.
By sending an initiation signal to the control device, the calibration of the audio rendering is acknowledged by the mobile device and thus started at a predetermined point in time.
The methods described herein may be implemented as software in a Digital Signal Processor (DSP), in a micro-controller or in any other side-processor or as hardware circuit within an application specific integrated circuit (ASIC).
The invention can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in combinations thereof.
BRIEF DESCRIPTION OF THE DRAWINGS
Further embodiments of the invention will be described with respect to the following figures, in which:
Fig. 1 shows a schematic diagram of a basic audio rendering system according to an implementation form; Fig. 2 shows a schematic diagram of an audio rendering system with the user being outside the system according to an implementation form;
Fig. 3 shows a schematic diagram of an audio rendering system with all loudspeakers being positioned in front of the user according to an implementation form;
Fig. 4 shows a schematic diagram of an audio rendering system with calibration based on a docking station according to an implementation form;
Fig. 5 shows a schematic diagram of a microphone arrangement on a Tablet PC according to an implementation form;
Fig. 6 shows a schematic diagram of an audio rendering system with calibration based on a docking station integrating a control device according to an implementation form; Fig. 7 shows a schematic diagram of an audio rendering system with calibration based on microphones installed in headphones of a listener according to an implementation form;
Fig. 8 shows a schematic diagram of a method for calibrating an audio rendering system according to an implementation form;
Fig. 9 shows a schematic diagram of a 5.1 multichannel loudspeaker system according to the recommendation ITU- BS 775-1 ;
Fig. 10 shows a schematic diagram of a non-standardized 5.1 multichannel loudspeaker system; and
Fig. 1 1 shows a schematic diagram of a microphone configuration of a conventional mobile phone. DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
There are two microphone configurations in current mobile phones. The first one corresponds to the configuration depicted in Fig. 1 1 but the second microphone 1 107 depicted in Fig. 1 1 is missing. It has only one single microphone 1 1 13 which is used for communication and any other mono sound pickup. This first configuration is based on omnidirectional microphone and cannot provide any stereo image.
The other configuration is illustrated in Fig. 1 1 and uses two omnidirectional microphones, the main microphone 1 1 13 and the auxiliary microphone 1 107. The main microphone 1 1 13 is used for the sound pick up, i.e. for communication application as well as simple audio/video recording. The auxiliary microphone 1 107 is used for noise cancellation and gain control. An omnidirectional microphone needs only one hole in the terminal as opposed to directional microphones. Indeed, the omnidirectional microphone offers a uniform directivity pattern in all the direction, equivalent to a sphere, and only one hole is then required for the microphone housing 1 1 15. The size of the hole and the actual
microphone housing 1 1 15 will affect the directivity of the complete system, but a single hole is sufficient for perfect sound pickup. Two omnidirectional microphones mounted with a spacing of several centimeters can also be used in order to obtain a stereo recording. Fig. 1 shows a schematic diagram of a basic audio rendering system 100 according to an implementation form.
A user or listener 109 sitting on a sofa holds a mobile device 105, e.g. a Smartphone or a tablet PC including at least two microphones 107. The mobile device 105 is connected to a control device 103, e.g. a set top box or an amplifier through a wireless or a wire line connection. The control device 103 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 100 comprises the following elements:
a plurality of loudspeakers 101 , e.g. a loudspeaker array or a loudspeaker system; a mobile device 105, e.g. a Smartphone or a tablet PC with multiple microphones
107, i.e., at least two omnidirectional or directional microphones;
open or standardized synchronization means, e.g. via WiFi, docking station, etc. between the audio rendering control device 103 and the mobile device 105;
analysis device of the audio rendering system and/or room characteristics; and control device 103 for the adaptive compensation of the audio rendering based on an analysis step of the analysis device.
In an implementation form, the analysis and control devices are implemented in separated devices. In an alternative implementation form, the analysis and control devices are implemented in the same device. The adaptation of the multi-loudspeaker audio rendering can be directly calculated in the analysis device if the test signal is known a priori by this device.
In an implementation form, for calibrating the audio rendering system 100, the user 109 of the mobile device 105 performs the following steps which interact with the control device 103:
starting the calibration procedure, e.g., with a mobile device application using a graphical user interface;
mobile device 105 sends indication to control device 103 of starting the procedure through the connection means, e.g. via WiFi, WLAN or docking station;
control device 103 starts the rendering of training signal through the loudspeaker system, i.e. through the multi-loudspeaker 101 shown in Fig. 1 ;
mobile device 105 records the training signal with multiple microphones 107; mobile device 105 analyzes the position of loudspeakers 101 and/or room characteristics;
a set of parameters, e.g. delay, gains, filters, etc. are extracted by the mobile device 105 to build a map of the loudspeaker system and to send these parameters to the control device 103. In an implementation form, at least positions of the loudspeakers 101 are transmitted by applying a specific protocol to exchange this information. In an alternative implementation form, the gains and delays are also transmitted. In an alternative implementation form, the mobile device 105 sends the recorded signals to the control device 103 and the analysis is done in the control device 103; and
control device 103 adapts the audio rendering based on the loudspeaker positions and/or room characteristic analysis.
In an implementation form, the mobile device 105 performs the computation of adaptation parameters such as delay, gains, filters, equalizer, etc. and sends these parameters to the control device 103. In an alternative implementation form, only the positions of the loudspeakers 101 are transmitted to the control device 103, the computation of the adaptation parameters are then performed in the rendering system, i.e., in the control device 103 in order to select the most appropriate processing, e.g., post processing of the audio channel signal, or adaptation in the coded domain.
Fig. 2 shows a schematic diagram of an audio rendering system 200 with the user 209 being outside the system 200 according to an implementation form.
A user or listener 209 sitting on a sofa outside the audio system holds a mobile device 205, e.g. a Smartphone or a tablet PC including at least two microphones 207. The mobile device 205 is connected to a control device 203, e.g. a set top box or an amplifier through a wireless or a wire line connection. The control device 203 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 200 comprises the following elements: a plurality of loudspeakers 201 , a mobile device 205 with at least two omnidirectional or directional microphones, a synchronization means
between the control device 203 and the mobile device 205, an analysis device of the audio rendering system and/or room characteristics and a control device 103 for the adaptive compensation of the audio rendering based on the analysis of the analysis device.
The calibration is performed analogously to the procedure described with respect to Fig. 1 . The audio rendering system 200 is able to calibrate the system if the user 209 is not located at the sweet spot of the audio rendering system even if the user 209 is sitting outside the audio system.
Fig. 3 shows a schematic diagram of an audio rendering system 300 with all loudspeakers being positioned in front of the user according to an implementation form.
A user or listener 309 sitting on a sofa behind a loudspeaker array 301 arranged in a line in front of the user 309. The user 309 holds a mobile device 305, e.g. a Smartphone or a tablet PC including at least two microphones 307. The mobile device 305 is connected to a control device 303, e.g. a set top box or an amplifier through a wireless or a wire line connection. The control device 303 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 300 comprises the following elements: a plurality of loudspeakers 301 arranged in line in front of the user 309, a mobile device 305 with at least two omnidirectional or directional microphones 307, synchronization means between the control device 303 and the mobile device 305, analysis device of the audio rendering system and/or room characteristics and a control device 303 for the adaptive compensation of the audio rendering based on the analysis of the analysis device.
The calibration is performed analogously to the procedure described with respect to Fig. 1 . The audio rendering system 300 is able to calibrate the system even if the
loudspeakers 301 are arranged in front of the user 309 and if there are no surround loudspeakers available in the audio system.
Fig. 4 shows a schematic diagram of an audio rendering system 400 with calibration based on a docking station 401 according to an implementation form.
A user or listener 409 is sitting on a sofa behind a docking station 401 comprising at least two loudspeakers. The user 409 holds a mobile device 405, e.g. a Smartphone or a tablet PC including at least two microphones 407. The mobile device 405 is connected to a control device 403, e.g. a set top box or an amplifier through a wireless or a wire line connection. The control device 403 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 400 comprises the following elements: loudspeakers integrated in a docking station 401 arranged in front of the user 409, a mobile device 405 with at least two omnidirectional or directional microphones 407, synchronization means between the control device 403 and the mobile device 405, analysis device of the audio rendering system and/or room characteristics and a control device 403 for the adaptive compensation of the audio rendering based on the analysis of the analysis device. Docking station 401 and control device 403 are separate units.
The calibration is performed analogously to the procedure described with respect to Fig. 1 . The audio rendering system 400 is able to calibrate the system even if there are only two loudspeakers integrated in the docking station 401 arranged in front of the user 409 and if there are no other front loudspeakers or surround loudspeakers available in the audio system. Fig. 5 shows a schematic diagram of a microphone arrangement on a Tablet PC 500 according to an implementation form.
The Tablet PC 500 comprises a number of four microphones 507 arranged in the middle of each edge of the tablet in order to better discriminate or distinguish the directions of the sounds. The tablet PC 500 corresponds to the mobile device described above with respect to Figures 1 to 4. The table PC 500 is adapted to perform the calibration analogously to the procedure described with respect to Fig. 1.
Fig. 6 shows a schematic diagram of an audio rendering system 600 with calibration based on a docking station integrating a control device according to an implementation form.
A user or listener 609 sitting on a sofa holds a mobile device 605, e.g. a Smartphone or a tablet PC including at least two microphones 607. The mobile device 605 is connected to a control device 603, e.g. a set top box or an amplifier through a wireless or a wire line
connection. The control device 603 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 600 comprises the following elements:
mobile device 605, e.g. a Smartphone or a tablet PC with multiple microphones 607, i.e., at least two omnidirectional or directional microphones;
docking station 601 with the possibility to connect to the mobile device 605, e.g., by a dock connector;
analysis device of the audio rendering system and/or room characteristics; and control device integrated in the docking station 601 for the adaptive compensation of the audio rendering based on an analysis step of the analysis device.
In an implementation form, for calibrating the audio rendering system 600, the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
Connecting the mobile device 605 on the docking station 601 , e.g., by the dock connector;
starting the calibration procedure, e.g., with a mobile device application using a graphical user interface;
mobile device 605 sends indication to the docking station 601 of starting the calibration procedure;
docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 provides the loudspeaker positions and/or room characteristics to the docking station 605 which adapts the rendering based on the loudspeaker positions and/or room characteristics analysis.
In an alternative implementation form, for calibrating the audio rendering system 600, the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
Connecting the mobile device 605 on the docking station 601 , e.g., by the dock connector;
starting the calibration procedure, e.g., with a mobile device application using a graphical user interface;
mobile device 605 sends indication to the docking station 601 of starting the calibration procedure;
docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 transmits the recorded signal to the docking station 601 and the analysis is done directly in the docking station 601 prior to the adaptation of the rendering based on the loudspeaker positions and/or room characteristics analysis of the control device in the docking station 601 .
In an alternative implementation form, for calibrating the audio rendering system 600, the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
Connecting the mobile device 605 on the docking station 601 , e.g., by the dock connector;
starting the calibration procedure by a timer integrated to the application to indicate how long the docking station must wait before starting to play the training sound;
mobile device 605 sends timing information to the docking station 601 to inform docking station 601 of starting the calibration procedure;
docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 provides the loudspeaker positions and/or room characteristics to the docking station 605 which adapts the rendering based on the loudspeaker positions and/or room characteristics analysis.
In an alternative implementation form, for calibrating the audio rendering system 600, the user 609 of the mobile device 605 performs the following steps which interact with the control device in the docking station 601 :
Connecting the mobile device 605 on the docking station 601 , e.g., by the dock connector;
starting the calibration procedure by a timer integrated to the application to indicate how long the docking station must wait before starting to play the training sound;
mobile device 605 sends timing information to the docking station 601 to inform docking station 601 of starting the calibration procedure;
docking station 601 starts the rendering of training signal through the loudspeaker system included in the docking station 601 ;
mobile device 605 records the training signal using the multiple microphones 607; mobile device 605 analyzes the position of loudspeaker in the docking station 601 and/or room characteristics;
user 609 puts the mobile device 605 on the docking station 601 and the mobile device 605 transmits the recorded signal to the docking station 601 and the analysis is done directly in the docking station 601 prior to the adaptation of the rendering based on the loudspeaker positions and/or room characteristics analysis of the control device in the docking station 601.
Fig. 7 shows a schematic diagram of an audio rendering system 700 with calibration based on microphones installed in headphones of a listener according to an
implementation form. A user or listener 709 is sitting on a sofa and holds a mobile device 705, e.g. a
Smartphone or a tablet PC including at least two microphones 707. The user 709 carries earphones with stereo microphones 713. The mobile device 705 is connected to a control device 703, e.g. a set top box or an amplifier through a wireless or a wire line connection. The control device 703 is performing the rendering adaptation, i.e., application of gains, delays, filters, etc. The audio rendering system 700 comprises the following elements: front loudspeakers 701 and surround loudspeakers 71 1 , a mobile device 705 with at least two omnidirectional or directional microphones 707, synchronization means between the control device 703 and the mobile device 705, analysis device of the audio rendering system and/or room characteristics and a control device 703 for the adaptive
compensation of the audio rendering based on the analysis of the analysis device.
The calibration is performed analogously to the procedure described with respect to Fig. 1 . The audio rendering system 700, however, is based on the use of stereo microphones 713 installed on the earphones or headphones and which are connected to the mobile device 705.
Compared to the previous implementation forms described with respect to Figures 1 to 6, this alternative scenario is based on binaural recording made through earphones or headphones with microphones 713 mounted on each side. Microphones 713 are directly integrated in the earphones/headphones capturing the audio signal directly at the ear canal. The user 709 connects the earphones/headphones to the mobile device 705 and the sound recording is done by this binaural microphone 713.
This implementation form is advantageously based on recording which is directly representative of human perception of the multi-loudspeaker rendering system. The rendering adaptation is directly based on the recording at the user's 709 ears which ensure that the adaptation, if done properly, will perfectly reconstruct the optimal multichannel signal at the ears of the listener 709.
In a further alternative implementation form, the analysis device combines the recording by the microphones 713 integrated to the earphones with the recording by the
microphones 707 of the mobile device 705. This combined recording provides a larger area of adaptation and limits the over tuning of the calibration algorithm to the user ears position. Fig. 8 shows a schematic diagram of a method 800 for calibrating an audio rendering system according to an implementation form.
The method 800 comprises recording 801 a training signal by at least two microphones of a mobile device being located in a room, the training signal being transmitted through a plurality of loudspeakers located in the room.
The method 800 comprises transmitting 803 the recorded training signal and/or information based thereupon to a control device configured for calibrating the audio rendering system.
In an implementation form, the method 800 comprises initiating the calibrating by the mobile device by transmitting an initiation signal to the control device.
From the foregoing, it will be apparent to those skilled in the art that a variety of methods, systems, computer programs on recording media, and the like, are provided. The present disclosure also supports a computer program product including computer executable code or computer executable instructions that, when executed, causes at least one computer to execute the performing and computing steps described herein.
The present disclosure also supports a system configured to execute the performing and computing steps described herein.
Many alternatives, modifications, and variations will be apparent to those skilled in the art in light of the above teachings. Of course, those skilled in the art readily recognize that there are numerous applications of the invention beyond those described herein. While the present inventions has been described with reference to one or more particular embodiments, those skilled in the art recognize that many changes may be made thereto without departing from the spirit and scope of the present invention. It is therefore to be understood that within the scope of the appended claims and their equivalents, the inventions may be practiced otherwise than as specifically described herein.
Claims
1 . Audio rendering system (100) for audio rendering room acoustics, comprising: a plurality of loudspeakers (101 ); a mobile device (105) comprising at least two microphones (107), the mobile device (105) being located at a room position ; and a control device (103) coupled to the plurality of loudspeakers (101 ) and coupled to the mobile device (105), wherein the control device (103) is configured to calibrate the audio rendering system (100) based on the room position of the mobile device (105) and based on room positions of the loudspeakers (101 ) by evaluating information of the at least two microphones (107) of the mobile device (105).
2. The audio rendering system (100) of claim 1 , wherein the control device (103) is configured to start calibrating the audio rendering system (100) by sending a training signal through the plurality of loudspeakers (101 ).
3. The audio rendering system (100) of claim 2, wherein the mobile device (105) is configured to record the training signal and to send the recorded training signal and/or information based thereupon to the control device (103).
4. The audio rendering system (100) of claim 3, wherein the information based on the recorded training signal comprises at least one of the following information: information on positions of the plurality of loudspeakers (101 ), information on a room characteristic, information on gains in signal paths between the plurality of loudspeakers (101 ) and the at least two microphones (107), information on delays in signal paths between the plurality of loudspeakers (101 ) and the at least two microphones (107), and information on transfer function in signal paths between the plurality of
loudspeakers (101 ) and the at least two microphones (107).
5. The audio rendering system (100) of claim 3 or claim 4, wherein the control device (103) is configured to adapt the audio rendering of the audio rendering system (100) based on the recorded training signal and/or information based thereupon received from the mobile device (105).
6. The audio rendering system (600) of one of the preceding claims, wherein the control device (103) is located in a docking station (601 ) of the mobile device (605).
7. The audio rendering system (100) of one of the preceding claims, wherein each of the at least two microphones (107) is located in the middle of a different edge of the mobile device (105).
8. A portable electronic device (500), comprising: at least two microphones (507); and a processor configured to record a training signal received by the at least two microphones (507) and to transmit the recorded training signal and/or information based thereupon via a transmit interface for calibration of an audio rendering system.
9. The portable electronic device (500) of claim 8, wherein the processor comprises an analysis module configured for analyzing the recorded training signal to provide the information based on the recorded training signal.
10. The portable electronic device (500) of claim 8 or claim 9, wherein the analysis module is configured to provide as information based on the recorded training signal at least one of the following information: information on a room characteristic, information on room positions of the loudspeakers, information on gains in signal paths between the loudspeakers and the at least two microphones (507), information on delays in signal paths between the loudspeakers and the at least two microphones (507), and information on transfer function in signal paths between the loudspeakers and the at least two microphones (507).
1 1. The portable electronic device (500) of one of claims 8 to 10, comprising a synchronization circuit configured to synchronize the recording of the training signal and the transmission of the recorded training signal and/or the information based thereupon with a control device initiating the calibration of the audio rendering system.
12. The portable electronic device (500) of one of claims 8 to 1 1 , comprising a graphical user interface configured for allowing a user to control the calibration of the audio rendering system by inputting information on a room characteristic used for adapting the audio rendering system.
13. The portable electronic device (500) of one of claims 8 to 12, further comprising earphones, wherein the processor is configured to record the training signal received by microphones (713) integrated in the earphones for binaurally capturing the training signal at an ear canal of a user using the portable electronic device (500) or to record the training signal by a combination of the at least two microphones (507) and the microphones (713) integrated in the earphones.
14. Method (800) for calibrating an audio rendering system for audio rendering room acoustics, the method comprising: recording (801 ) a training signal by at least two microphones of a mobile device being located in a room, the training signal being transmitted through a plurality of loudspeakers located in the room; transmitting (803) the recorded training signal and/or information based thereupon to a control device configured for calibrating the audio rendering system.
15. The method (800) of claim 14, further comprising: initiating the calibrating by the mobile device by transmitting an initiation signal to the control device.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2012/066775 WO2014032709A1 (en) | 2012-08-29 | 2012-08-29 | Audio rendering system |
EP12753717.3A EP2823650B1 (en) | 2012-08-29 | 2012-08-29 | Audio rendering system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2012/066775 WO2014032709A1 (en) | 2012-08-29 | 2012-08-29 | Audio rendering system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014032709A1 true WO2014032709A1 (en) | 2014-03-06 |
Family
ID=46785413
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2012/066775 WO2014032709A1 (en) | 2012-08-29 | 2012-08-29 | Audio rendering system |
Country Status (2)
Country | Link |
---|---|
EP (1) | EP2823650B1 (en) |
WO (1) | WO2014032709A1 (en) |
Cited By (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016004345A1 (en) * | 2014-07-03 | 2016-01-07 | Qualcomm Incorporated | Single-channel or multi-channel audio control interface |
EP2975861A1 (en) * | 2014-07-15 | 2016-01-20 | Sonavox Canada Inc. | Wireless control and calibration of audio system |
EP3018917A1 (en) * | 2014-11-06 | 2016-05-11 | Axis AB | Method and system for audio calibration of an audio device |
EP3057345A1 (en) * | 2015-02-16 | 2016-08-17 | Harman International Industries, Inc. | Mobile interface for loudspeaker optimization |
EP3094113A1 (en) * | 2015-05-14 | 2016-11-16 | Harman International Industries, Inc. | Techniques for autonomously calibrating an audio system |
WO2017049169A1 (en) * | 2015-09-17 | 2017-03-23 | Sonos, Inc. | Facilitating calibration of an audio playback device |
US9690271B2 (en) | 2012-06-28 | 2017-06-27 | Sonos, Inc. | Speaker calibration |
US9699555B2 (en) | 2012-06-28 | 2017-07-04 | Sonos, Inc. | Calibration of multiple playback devices |
US9706323B2 (en) | 2014-09-09 | 2017-07-11 | Sonos, Inc. | Playback device calibration |
US9743208B2 (en) | 2014-03-17 | 2017-08-22 | Sonos, Inc. | Playback device configuration based on proximity detection |
US9860670B1 (en) | 2016-07-15 | 2018-01-02 | Sonos, Inc. | Spectral correction using spatial calibration |
US9860662B2 (en) | 2016-04-01 | 2018-01-02 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US9864574B2 (en) | 2016-04-01 | 2018-01-09 | Sonos, Inc. | Playback device calibration based on representation spectral characteristics |
US9872119B2 (en) | 2014-03-17 | 2018-01-16 | Sonos, Inc. | Audio settings of multiple speakers in a playback device |
US9891881B2 (en) | 2014-09-09 | 2018-02-13 | Sonos, Inc. | Audio processing algorithm database |
RU2646337C1 (en) * | 2014-03-28 | 2018-03-02 | Самсунг Электроникс Ко., Лтд. | Method and device for rendering acoustic signal and machine-readable record media |
US9930470B2 (en) | 2011-12-29 | 2018-03-27 | Sonos, Inc. | Sound field calibration using listener localization |
US9936318B2 (en) | 2014-09-09 | 2018-04-03 | Sonos, Inc. | Playback device calibration |
US9952825B2 (en) | 2014-09-09 | 2018-04-24 | Sonos, Inc. | Audio processing algorithms |
US10003899B2 (en) | 2016-01-25 | 2018-06-19 | Sonos, Inc. | Calibration with particular locations |
US10045142B2 (en) | 2016-04-12 | 2018-08-07 | Sonos, Inc. | Calibration of audio playback devices |
US10063983B2 (en) | 2016-01-18 | 2018-08-28 | Sonos, Inc. | Calibration using multiple recording devices |
US10129679B2 (en) | 2015-07-28 | 2018-11-13 | Sonos, Inc. | Calibration error conditions |
US10127006B2 (en) | 2014-09-09 | 2018-11-13 | Sonos, Inc. | Facilitating calibration of an audio playback device |
US10129678B2 (en) | 2016-07-15 | 2018-11-13 | Sonos, Inc. | Spatial audio correction |
US10192563B2 (en) | 2014-03-26 | 2019-01-29 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for screen related audio object remapping |
US10200800B2 (en) | 2017-02-06 | 2019-02-05 | EVA Automation, Inc. | Acoustic characterization of an unknown microphone |
US10225656B1 (en) | 2018-01-17 | 2019-03-05 | Harman International Industries, Incorporated | Mobile speaker system for virtual reality environments |
US10284983B2 (en) | 2015-04-24 | 2019-05-07 | Sonos, Inc. | Playback device calibration user interfaces |
US10299061B1 (en) | 2018-08-28 | 2019-05-21 | Sonos, Inc. | Playback device calibration |
US10296282B2 (en) | 2012-06-28 | 2019-05-21 | Sonos, Inc. | Speaker calibration user interface |
CN110035350A (en) * | 2017-12-07 | 2019-07-19 | 哈曼国际工业有限公司 | Unmanned plane disposes speaker system |
US10372406B2 (en) | 2016-07-22 | 2019-08-06 | Sonos, Inc. | Calibration interface |
US10419864B2 (en) | 2015-09-17 | 2019-09-17 | Sonos, Inc. | Validation of audio calibration using multi-dimensional motion check |
US10446166B2 (en) | 2016-07-12 | 2019-10-15 | Dolby Laboratories Licensing Corporation | Assessment and adjustment of audio installation |
US10459684B2 (en) | 2016-08-05 | 2019-10-29 | Sonos, Inc. | Calibration of a playback device based on an estimated frequency response |
US10664224B2 (en) | 2015-04-24 | 2020-05-26 | Sonos, Inc. | Speaker calibration user interface |
US10734965B1 (en) | 2019-08-12 | 2020-08-04 | Sonos, Inc. | Audio calibration of a portable playback device |
US10837944B2 (en) | 2018-02-06 | 2020-11-17 | Harman International Industries, Incorporated | Resonator device for resonance mapping and sound production |
WO2021010884A1 (en) * | 2019-07-18 | 2021-01-21 | Dirac Research Ab | Intelligent audio control platform |
US11106423B2 (en) | 2016-01-25 | 2021-08-31 | Sonos, Inc. | Evaluating calibration of a playback device |
US11206484B2 (en) | 2018-08-28 | 2021-12-21 | Sonos, Inc. | Passive speaker authentication |
US11722821B2 (en) | 2016-02-19 | 2023-08-08 | Dolby Laboratories Licensing Corporation | Sound capture for mobile devices |
US11863952B2 (en) | 2016-02-19 | 2024-01-02 | Dolby Laboratories Licensing Corporation | Sound capture for mobile devices |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000261900A (en) | 1999-03-09 | 2000-09-22 | Sony Corp | Sound field correction method and acoustic device |
FR2850183A1 (en) | 2003-01-20 | 2004-07-23 | Remy Henri Denis Bruno | Acoustic field restoration controlling method, involves determining adaptation filter according to characteristic and directions of fixed restoration associated with entry signals for determining control signal of restoration unit |
WO2007017809A1 (en) * | 2005-08-05 | 2007-02-15 | Koninklijke Philips Electronics N.V. | A device for and a method of processing audio data |
US20100284544A1 (en) * | 2008-01-29 | 2010-11-11 | Korea Advanced Institute Of Science And Technology | Sound system, sound reproducing apparatus, sound reproducing method, monitor with speakers, mobile phone with speakers |
EP2378795A1 (en) | 2008-12-25 | 2011-10-19 | Pioneer Corporation | Sound field correction system |
WO2011139502A1 (en) * | 2010-05-06 | 2011-11-10 | Dolby Laboratories Licensing Corporation | Audio system equalization for portable media playback devices |
-
2012
- 2012-08-29 EP EP12753717.3A patent/EP2823650B1/en active Active
- 2012-08-29 WO PCT/EP2012/066775 patent/WO2014032709A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000261900A (en) | 1999-03-09 | 2000-09-22 | Sony Corp | Sound field correction method and acoustic device |
FR2850183A1 (en) | 2003-01-20 | 2004-07-23 | Remy Henri Denis Bruno | Acoustic field restoration controlling method, involves determining adaptation filter according to characteristic and directions of fixed restoration associated with entry signals for determining control signal of restoration unit |
WO2007017809A1 (en) * | 2005-08-05 | 2007-02-15 | Koninklijke Philips Electronics N.V. | A device for and a method of processing audio data |
US20100284544A1 (en) * | 2008-01-29 | 2010-11-11 | Korea Advanced Institute Of Science And Technology | Sound system, sound reproducing apparatus, sound reproducing method, monitor with speakers, mobile phone with speakers |
EP2378795A1 (en) | 2008-12-25 | 2011-10-19 | Pioneer Corporation | Sound field correction system |
WO2011139502A1 (en) * | 2010-05-06 | 2011-11-10 | Dolby Laboratories Licensing Corporation | Audio system equalization for portable media playback devices |
Cited By (161)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11197117B2 (en) | 2011-12-29 | 2021-12-07 | Sonos, Inc. | Media playback based on sensor data |
US11889290B2 (en) | 2011-12-29 | 2024-01-30 | Sonos, Inc. | Media playback based on sensor data |
US10455347B2 (en) | 2011-12-29 | 2019-10-22 | Sonos, Inc. | Playback based on number of listeners |
US10945089B2 (en) | 2011-12-29 | 2021-03-09 | Sonos, Inc. | Playback based on user settings |
US10986460B2 (en) | 2011-12-29 | 2021-04-20 | Sonos, Inc. | Grouping based on acoustic signals |
US11122382B2 (en) | 2011-12-29 | 2021-09-14 | Sonos, Inc. | Playback based on acoustic signals |
US11153706B1 (en) | 2011-12-29 | 2021-10-19 | Sonos, Inc. | Playback based on acoustic signals |
US11290838B2 (en) | 2011-12-29 | 2022-03-29 | Sonos, Inc. | Playback based on user presence detection |
US10334386B2 (en) | 2011-12-29 | 2019-06-25 | Sonos, Inc. | Playback based on wireless signal |
US11528578B2 (en) | 2011-12-29 | 2022-12-13 | Sonos, Inc. | Media playback based on sensor data |
US11825289B2 (en) | 2011-12-29 | 2023-11-21 | Sonos, Inc. | Media playback based on sensor data |
US11910181B2 (en) | 2011-12-29 | 2024-02-20 | Sonos, Inc | Media playback based on sensor data |
US11825290B2 (en) | 2011-12-29 | 2023-11-21 | Sonos, Inc. | Media playback based on sensor data |
US9930470B2 (en) | 2011-12-29 | 2018-03-27 | Sonos, Inc. | Sound field calibration using listener localization |
US11849299B2 (en) | 2011-12-29 | 2023-12-19 | Sonos, Inc. | Media playback based on sensor data |
US9913057B2 (en) | 2012-06-28 | 2018-03-06 | Sonos, Inc. | Concurrent multi-loudspeaker calibration with a single measurement |
US10045138B2 (en) | 2012-06-28 | 2018-08-07 | Sonos, Inc. | Hybrid test tone for space-averaged room audio calibration using a moving microphone |
US10412516B2 (en) | 2012-06-28 | 2019-09-10 | Sonos, Inc. | Calibration of playback devices |
US10674293B2 (en) | 2012-06-28 | 2020-06-02 | Sonos, Inc. | Concurrent multi-driver calibration |
US9788113B2 (en) | 2012-06-28 | 2017-10-10 | Sonos, Inc. | Calibration state variable |
US11064306B2 (en) | 2012-06-28 | 2021-07-13 | Sonos, Inc. | Calibration state variable |
US9699555B2 (en) | 2012-06-28 | 2017-07-04 | Sonos, Inc. | Calibration of multiple playback devices |
US9690271B2 (en) | 2012-06-28 | 2017-06-27 | Sonos, Inc. | Speaker calibration |
US10390159B2 (en) | 2012-06-28 | 2019-08-20 | Sonos, Inc. | Concurrent multi-loudspeaker calibration |
US10791405B2 (en) | 2012-06-28 | 2020-09-29 | Sonos, Inc. | Calibration indicator |
US11368803B2 (en) | 2012-06-28 | 2022-06-21 | Sonos, Inc. | Calibration of playback device(s) |
US11516606B2 (en) | 2012-06-28 | 2022-11-29 | Sonos, Inc. | Calibration interface |
US10296282B2 (en) | 2012-06-28 | 2019-05-21 | Sonos, Inc. | Speaker calibration user interface |
US11516608B2 (en) | 2012-06-28 | 2022-11-29 | Sonos, Inc. | Calibration state variable |
US10284984B2 (en) | 2012-06-28 | 2019-05-07 | Sonos, Inc. | Calibration state variable |
US9961463B2 (en) | 2012-06-28 | 2018-05-01 | Sonos, Inc. | Calibration indicator |
US10129674B2 (en) | 2012-06-28 | 2018-11-13 | Sonos, Inc. | Concurrent multi-loudspeaker calibration |
US11800305B2 (en) | 2012-06-28 | 2023-10-24 | Sonos, Inc. | Calibration interface |
US10045139B2 (en) | 2012-06-28 | 2018-08-07 | Sonos, Inc. | Calibration state variable |
US10129675B2 (en) | 2014-03-17 | 2018-11-13 | Sonos, Inc. | Audio settings of multiple speakers in a playback device |
US10412517B2 (en) | 2014-03-17 | 2019-09-10 | Sonos, Inc. | Calibration of playback device to target curve |
US10863295B2 (en) | 2014-03-17 | 2020-12-08 | Sonos, Inc. | Indoor/outdoor playback device calibration |
US10051399B2 (en) | 2014-03-17 | 2018-08-14 | Sonos, Inc. | Playback device configuration according to distortion threshold |
US9872119B2 (en) | 2014-03-17 | 2018-01-16 | Sonos, Inc. | Audio settings of multiple speakers in a playback device |
US10299055B2 (en) | 2014-03-17 | 2019-05-21 | Sonos, Inc. | Restoration of playback device configuration |
US11696081B2 (en) | 2014-03-17 | 2023-07-04 | Sonos, Inc. | Audio settings based on environment |
US10511924B2 (en) | 2014-03-17 | 2019-12-17 | Sonos, Inc. | Playback device with multiple sensors |
US10791407B2 (en) | 2014-03-17 | 2020-09-29 | Sonon, Inc. | Playback device configuration |
US11540073B2 (en) | 2014-03-17 | 2022-12-27 | Sonos, Inc. | Playback device self-calibration |
US9743208B2 (en) | 2014-03-17 | 2017-08-22 | Sonos, Inc. | Playback device configuration based on proximity detection |
RU2683380C2 (en) * | 2014-03-26 | 2019-03-28 | Фраунхофер-Гезелльшафт Цур Фердерунг Дер Ангевандтен Форшунг Е.Ф. | Device and method for repeated display of screen-related audio objects |
US11900955B2 (en) | 2014-03-26 | 2024-02-13 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for screen related audio object remapping |
US10854213B2 (en) | 2014-03-26 | 2020-12-01 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for screen related audio object remapping |
US10192563B2 (en) | 2014-03-26 | 2019-01-29 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for screen related audio object remapping |
US11527254B2 (en) | 2014-03-26 | 2022-12-13 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for screen related audio object remapping |
RU2646337C1 (en) * | 2014-03-28 | 2018-03-02 | Самсунг Электроникс Ко., Лтд. | Method and device for rendering acoustic signal and machine-readable record media |
US10149086B2 (en) | 2014-03-28 | 2018-12-04 | Samsung Electronics Co., Ltd. | Method and apparatus for rendering acoustic signal, and computer-readable recording medium |
US10382877B2 (en) | 2014-03-28 | 2019-08-13 | Samsung Electronics Co., Ltd. | Method and apparatus for rendering acoustic signal, and computer-readable recording medium |
US10687162B2 (en) | 2014-03-28 | 2020-06-16 | Samsung Electronics Co., Ltd. | Method and apparatus for rendering acoustic signal, and computer-readable recording medium |
US10073607B2 (en) | 2014-07-03 | 2018-09-11 | Qualcomm Incorporated | Single-channel or multi-channel audio control interface |
US10051364B2 (en) | 2014-07-03 | 2018-08-14 | Qualcomm Incorporated | Single channel or multi-channel audio control interface |
WO2016004345A1 (en) * | 2014-07-03 | 2016-01-07 | Qualcomm Incorporated | Single-channel or multi-channel audio control interface |
CN106664490A (en) * | 2014-07-03 | 2017-05-10 | 高通股份有限公司 | Single-channel or multi-channel audio control interface |
US9516444B2 (en) | 2014-07-15 | 2016-12-06 | Sonavox Canada Inc. | Wireless control and calibration of audio system |
EP2975861A1 (en) * | 2014-07-15 | 2016-01-20 | Sonavox Canada Inc. | Wireless control and calibration of audio system |
US9891881B2 (en) | 2014-09-09 | 2018-02-13 | Sonos, Inc. | Audio processing algorithm database |
US10127006B2 (en) | 2014-09-09 | 2018-11-13 | Sonos, Inc. | Facilitating calibration of an audio playback device |
US10271150B2 (en) | 2014-09-09 | 2019-04-23 | Sonos, Inc. | Playback device calibration |
US9936318B2 (en) | 2014-09-09 | 2018-04-03 | Sonos, Inc. | Playback device calibration |
US10701501B2 (en) | 2014-09-09 | 2020-06-30 | Sonos, Inc. | Playback device calibration |
US11625219B2 (en) | 2014-09-09 | 2023-04-11 | Sonos, Inc. | Audio processing algorithms |
US10154359B2 (en) | 2014-09-09 | 2018-12-11 | Sonos, Inc. | Playback device calibration |
US9706323B2 (en) | 2014-09-09 | 2017-07-11 | Sonos, Inc. | Playback device calibration |
US9952825B2 (en) | 2014-09-09 | 2018-04-24 | Sonos, Inc. | Audio processing algorithms |
US10599386B2 (en) | 2014-09-09 | 2020-03-24 | Sonos, Inc. | Audio processing algorithms |
US11029917B2 (en) | 2014-09-09 | 2021-06-08 | Sonos, Inc. | Audio processing algorithms |
US10127008B2 (en) | 2014-09-09 | 2018-11-13 | Sonos, Inc. | Audio processing algorithm database |
US9641948B2 (en) | 2014-11-06 | 2017-05-02 | Axis Ab | Method and system for audio calibration of an audio device |
EP3101920A1 (en) * | 2014-11-06 | 2016-12-07 | Axis AB | Method and peripheral device for providing a representation of how to alter a setting affecting audio reproduction of an audio device |
KR20160054404A (en) * | 2014-11-06 | 2016-05-16 | 엑시스 에이비 | Method and system for audio calibration of an audio device |
US9774968B2 (en) | 2014-11-06 | 2017-09-26 | Axis Ab | Method and system for audio calibration of an audio device |
TWI625061B (en) * | 2014-11-06 | 2018-05-21 | 安訊士有限公司 | Method and system for audio calibration of an audio device |
EP3018917A1 (en) * | 2014-11-06 | 2016-05-11 | Axis AB | Method and system for audio calibration of an audio device |
KR101694925B1 (en) | 2014-11-06 | 2017-01-11 | 엑시스 에이비 | Method and system for audio calibration of an audio device |
JP2016092836A (en) * | 2014-11-06 | 2016-05-23 | アクシス アーベー | Method and system for audio calibration of audio device |
CN105898663B (en) * | 2015-02-16 | 2020-03-03 | 哈曼国际工业有限公司 | Mobile interface for loudspeaker optimization |
CN105898663A (en) * | 2015-02-16 | 2016-08-24 | 哈曼国际工业有限公司 | Mobile Interface For Loudspeaker Optimization |
EP3057345A1 (en) * | 2015-02-16 | 2016-08-17 | Harman International Industries, Inc. | Mobile interface for loudspeaker optimization |
US10284983B2 (en) | 2015-04-24 | 2019-05-07 | Sonos, Inc. | Playback device calibration user interfaces |
US10664224B2 (en) | 2015-04-24 | 2020-05-26 | Sonos, Inc. | Speaker calibration user interface |
US10136234B2 (en) | 2015-05-14 | 2018-11-20 | Harman International Industries, Incorporated | Techniques for autonomously calibrating an audio system |
EP3094113A1 (en) * | 2015-05-14 | 2016-11-16 | Harman International Industries, Inc. | Techniques for autonomously calibrating an audio system |
US10462592B2 (en) | 2015-07-28 | 2019-10-29 | Sonos, Inc. | Calibration error conditions |
US10129679B2 (en) | 2015-07-28 | 2018-11-13 | Sonos, Inc. | Calibration error conditions |
JP2018533277A (en) * | 2015-09-17 | 2018-11-08 | ソノズ インコーポレイテッド | How to facilitate calibration of audio playback devices |
WO2017049169A1 (en) * | 2015-09-17 | 2017-03-23 | Sonos, Inc. | Facilitating calibration of an audio playback device |
JP7092829B2 (en) | 2015-09-17 | 2022-06-28 | ソノズ インコーポレイテッド | How to facilitate calibration of audio playback devices |
CN111314826A (en) * | 2015-09-17 | 2020-06-19 | 搜诺思公司 | Method performed by a computing device and corresponding computer readable medium and computing device |
US10585639B2 (en) | 2015-09-17 | 2020-03-10 | Sonos, Inc. | Facilitating calibration of an audio playback device |
JP2020195145A (en) * | 2015-09-17 | 2020-12-03 | ソノズ インコーポレイテッド | Facilitating calibration of audio playback device |
US11706579B2 (en) | 2015-09-17 | 2023-07-18 | Sonos, Inc. | Validation of audio calibration using multi-dimensional motion check |
US10419864B2 (en) | 2015-09-17 | 2019-09-17 | Sonos, Inc. | Validation of audio calibration using multi-dimensional motion check |
US11197112B2 (en) | 2015-09-17 | 2021-12-07 | Sonos, Inc. | Validation of audio calibration using multi-dimensional motion check |
CN111314826B (en) * | 2015-09-17 | 2021-05-14 | 搜诺思公司 | Method performed by a computing device and corresponding computer readable medium and computing device |
US11099808B2 (en) | 2015-09-17 | 2021-08-24 | Sonos, Inc. | Facilitating calibration of an audio playback device |
EP3531714A3 (en) * | 2015-09-17 | 2019-10-16 | Sonos Inc. | Facilitating calibration of an audio playback device |
US11803350B2 (en) | 2015-09-17 | 2023-10-31 | Sonos, Inc. | Facilitating calibration of an audio playback device |
US11432089B2 (en) | 2016-01-18 | 2022-08-30 | Sonos, Inc. | Calibration using multiple recording devices |
US10063983B2 (en) | 2016-01-18 | 2018-08-28 | Sonos, Inc. | Calibration using multiple recording devices |
US10405117B2 (en) | 2016-01-18 | 2019-09-03 | Sonos, Inc. | Calibration using multiple recording devices |
US10841719B2 (en) | 2016-01-18 | 2020-11-17 | Sonos, Inc. | Calibration using multiple recording devices |
US11800306B2 (en) | 2016-01-18 | 2023-10-24 | Sonos, Inc. | Calibration using multiple recording devices |
US10003899B2 (en) | 2016-01-25 | 2018-06-19 | Sonos, Inc. | Calibration with particular locations |
US10390161B2 (en) | 2016-01-25 | 2019-08-20 | Sonos, Inc. | Calibration based on audio content type |
US11006232B2 (en) | 2016-01-25 | 2021-05-11 | Sonos, Inc. | Calibration based on audio content |
US10735879B2 (en) | 2016-01-25 | 2020-08-04 | Sonos, Inc. | Calibration based on grouping |
US11516612B2 (en) | 2016-01-25 | 2022-11-29 | Sonos, Inc. | Calibration based on audio content |
US11184726B2 (en) | 2016-01-25 | 2021-11-23 | Sonos, Inc. | Calibration using listener locations |
US11106423B2 (en) | 2016-01-25 | 2021-08-31 | Sonos, Inc. | Evaluating calibration of a playback device |
US11863952B2 (en) | 2016-02-19 | 2024-01-02 | Dolby Laboratories Licensing Corporation | Sound capture for mobile devices |
US11722821B2 (en) | 2016-02-19 | 2023-08-08 | Dolby Laboratories Licensing Corporation | Sound capture for mobile devices |
US11736877B2 (en) | 2016-04-01 | 2023-08-22 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US10405116B2 (en) | 2016-04-01 | 2019-09-03 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US10880664B2 (en) | 2016-04-01 | 2020-12-29 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US10884698B2 (en) | 2016-04-01 | 2021-01-05 | Sonos, Inc. | Playback device calibration based on representative spectral characteristics |
US9864574B2 (en) | 2016-04-01 | 2018-01-09 | Sonos, Inc. | Playback device calibration based on representation spectral characteristics |
US10402154B2 (en) | 2016-04-01 | 2019-09-03 | Sonos, Inc. | Playback device calibration based on representative spectral characteristics |
US11212629B2 (en) | 2016-04-01 | 2021-12-28 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US9860662B2 (en) | 2016-04-01 | 2018-01-02 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
US11379179B2 (en) | 2016-04-01 | 2022-07-05 | Sonos, Inc. | Playback device calibration based on representative spectral characteristics |
US10750304B2 (en) | 2016-04-12 | 2020-08-18 | Sonos, Inc. | Calibration of audio playback devices |
US11889276B2 (en) | 2016-04-12 | 2024-01-30 | Sonos, Inc. | Calibration of audio playback devices |
US10299054B2 (en) | 2016-04-12 | 2019-05-21 | Sonos, Inc. | Calibration of audio playback devices |
US11218827B2 (en) | 2016-04-12 | 2022-01-04 | Sonos, Inc. | Calibration of audio playback devices |
US10045142B2 (en) | 2016-04-12 | 2018-08-07 | Sonos, Inc. | Calibration of audio playback devices |
US10446166B2 (en) | 2016-07-12 | 2019-10-15 | Dolby Laboratories Licensing Corporation | Assessment and adjustment of audio installation |
US11736878B2 (en) | 2016-07-15 | 2023-08-22 | Sonos, Inc. | Spatial audio correction |
US9860670B1 (en) | 2016-07-15 | 2018-01-02 | Sonos, Inc. | Spectral correction using spatial calibration |
US10750303B2 (en) | 2016-07-15 | 2020-08-18 | Sonos, Inc. | Spatial audio correction |
US10129678B2 (en) | 2016-07-15 | 2018-11-13 | Sonos, Inc. | Spatial audio correction |
US10448194B2 (en) | 2016-07-15 | 2019-10-15 | Sonos, Inc. | Spectral correction using spatial calibration |
US11337017B2 (en) | 2016-07-15 | 2022-05-17 | Sonos, Inc. | Spatial audio correction |
US10853022B2 (en) | 2016-07-22 | 2020-12-01 | Sonos, Inc. | Calibration interface |
US11531514B2 (en) | 2016-07-22 | 2022-12-20 | Sonos, Inc. | Calibration assistance |
US10372406B2 (en) | 2016-07-22 | 2019-08-06 | Sonos, Inc. | Calibration interface |
US11237792B2 (en) | 2016-07-22 | 2022-02-01 | Sonos, Inc. | Calibration assistance |
US10853027B2 (en) | 2016-08-05 | 2020-12-01 | Sonos, Inc. | Calibration of a playback device based on an estimated frequency response |
US10459684B2 (en) | 2016-08-05 | 2019-10-29 | Sonos, Inc. | Calibration of a playback device based on an estimated frequency response |
US11698770B2 (en) | 2016-08-05 | 2023-07-11 | Sonos, Inc. | Calibration of a playback device based on an estimated frequency response |
US10200800B2 (en) | 2017-02-06 | 2019-02-05 | EVA Automation, Inc. | Acoustic characterization of an unknown microphone |
CN110035350A (en) * | 2017-12-07 | 2019-07-19 | 哈曼国际工业有限公司 | Unmanned plane disposes speaker system |
US10377486B2 (en) | 2017-12-07 | 2019-08-13 | Harman International Industries, Incorporated | Drone deployed speaker system |
US11084583B2 (en) | 2017-12-07 | 2021-08-10 | Harman International Industries, Incorporated | Drone deployed speaker system |
CN110035350B (en) * | 2017-12-07 | 2022-07-05 | 哈曼国际工业有限公司 | Unmanned aerial vehicle deploys speaker system |
US10225656B1 (en) | 2018-01-17 | 2019-03-05 | Harman International Industries, Incorporated | Mobile speaker system for virtual reality environments |
US10837944B2 (en) | 2018-02-06 | 2020-11-17 | Harman International Industries, Incorporated | Resonator device for resonance mapping and sound production |
US10582326B1 (en) | 2018-08-28 | 2020-03-03 | Sonos, Inc. | Playback device calibration |
US10299061B1 (en) | 2018-08-28 | 2019-05-21 | Sonos, Inc. | Playback device calibration |
US10848892B2 (en) | 2018-08-28 | 2020-11-24 | Sonos, Inc. | Playback device calibration |
US11350233B2 (en) | 2018-08-28 | 2022-05-31 | Sonos, Inc. | Playback device calibration |
US11877139B2 (en) | 2018-08-28 | 2024-01-16 | Sonos, Inc. | Playback device calibration |
US11206484B2 (en) | 2018-08-28 | 2021-12-21 | Sonos, Inc. | Passive speaker authentication |
WO2021010884A1 (en) * | 2019-07-18 | 2021-01-21 | Dirac Research Ab | Intelligent audio control platform |
US11728780B2 (en) | 2019-08-12 | 2023-08-15 | Sonos, Inc. | Audio calibration of a portable playback device |
US11374547B2 (en) | 2019-08-12 | 2022-06-28 | Sonos, Inc. | Audio calibration of a portable playback device |
US10734965B1 (en) | 2019-08-12 | 2020-08-04 | Sonos, Inc. | Audio calibration of a portable playback device |
Also Published As
Publication number | Publication date |
---|---|
EP2823650A1 (en) | 2015-01-14 |
EP2823650B1 (en) | 2020-07-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2823650B1 (en) | Audio rendering system | |
US9877133B2 (en) | Sound collection and reproduction system, sound collection and reproduction apparatus, sound collection and reproduction method, sound collection and reproduction program, sound collection system, and reproduction system | |
EP3011763B1 (en) | Method for generating a surround sound field, apparatus and computer program product thereof. | |
US8477970B2 (en) | Systems, methods, and apparatus for controlling sounds in a three-dimensional listening environment | |
US9609141B2 (en) | Loudspeaker localization with a microphone array | |
CN101129088B (en) | Portable device with enhanced stereo image | |
US20110188662A1 (en) | Method of rendering binaural stereo in a hearing aid system and a hearing aid system | |
US10306358B2 (en) | Sound system | |
EP2816823B1 (en) | Audio system and audio apparatus and channel mapping method thereof | |
JP2007135094A (en) | Sound field correcting apparatus | |
US10708691B2 (en) | Dynamic equalization in a directional speaker array | |
US20190394567A1 (en) | Dynamically Adapting Sound Based on Background Sound | |
CN101489173B (en) | Signal processing apparatus, signal processing method | |
US20190391783A1 (en) | Sound Adaptation Based on Content and Context | |
US20190394603A1 (en) | Dynamic Cross-Talk Cancellation | |
US9294861B2 (en) | Audio signal processing device | |
US20190394598A1 (en) | Self-Configuring Speakers | |
US10531221B1 (en) | Automatic room filling | |
US10511906B1 (en) | Dynamically adapting sound based on environmental characterization | |
US20190246230A1 (en) | Virtual localization of sound | |
US10440473B1 (en) | Automatic de-baffling | |
CN111510847B (en) | Micro loudspeaker array, in-vehicle sound field control method and device and storage device | |
KR102609084B1 (en) | Electronic apparatus, method for controlling thereof and recording media thereof | |
US10484809B1 (en) | Closed-loop adaptation of 3D sound | |
US20190394570A1 (en) | Volume Normalization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12753717 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012753717 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |