WO2012164444A1 - An audio system and method of operating therefor - Google Patents

An audio system and method of operating therefor Download PDF

Info

Publication number
WO2012164444A1
WO2012164444A1 PCT/IB2012/052580 IB2012052580W WO2012164444A1 WO 2012164444 A1 WO2012164444 A1 WO 2012164444A1 IB 2012052580 W IB2012052580 W IB 2012052580W WO 2012164444 A1 WO2012164444 A1 WO 2012164444A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
channel
spatial
render
channels
Prior art date
Application number
PCT/IB2012/052580
Other languages
French (fr)
Inventor
Frédéric ROSKAM
Sylvain Jean CHOISEL
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Publication of WO2012164444A1 publication Critical patent/WO2012164444A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/40Visual indication of stereophonic sound image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2205/00Details of stereophonic arrangements covered by H04R5/00 but not provided for in any of its subgroups
    • H04R2205/024Positioning of loudspeaker enclosures for spatial sound reproduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • H04S3/002Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/301Automatic calibration of stereophonic sound system, e.g. with test microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation

Definitions

  • the invention relates to an audio system and a method of operation therefor, and in particular, but not exclusively, to generation for drive signals for loudspeakers of spatial surround sound system.
  • a 5.1 home cinema system provides spatial sound via five loudspeakers being positioned with one speaker directly in front of the listening position (the center channel), one speaker to the front left of the listening position, one speaker to the front right of the listening position, one speaker to the rear left of the listening position, and one speaker to the rear right of the listening position.
  • a non-spatial low frequency speaker is provided.
  • Such conventional systems are based the reproduction of audio signals at specific nominal positions relative to the listening position.
  • One speaker is typically provided for each audio channel and therefore speakers must be positioned at positions corresponding to the predetermined or nominal positions for the system.
  • the requirement for loudspeakers to be in specific positions is often found to be highly undesirable to users and is particularly inconvenient in the consumer market.
  • the loudspeakers are still required to positioned in approximately the nominal configuration. Indeed, typically the compensation algorithm will fail if the actual loudspeaker positions deviate too much from the assumed nominal positions/angles.
  • an improved audio system would be advantageous and in particular a system allowing increased flexibility, facilitated operation, increased freedom in sound transducer placement, improved perceived audio quality, improved audio quality, an improved spatial user experience and/or improved performance would be advantageous.
  • the Invention seeks to preferably mitigate, alleviate or eliminate one or more of the above mentioned disadvantages singly or in any combination.
  • an audio system for generating drive signals for a plurality of sound render channels, each sound render channel providing a drive signal for a sound transducer;
  • the audio system comprising: a user interface for receiving a user input graphical representation of positions of the sound transducers; a processor for determining sound render positions for the sound transducers in response to the user input graphical representation; an input for receiving a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position; a signal linker for determining a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions, an association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel; and a driver for generating drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.
  • the invention may provide an advantageous audio system for providing spatial audio.
  • a more flexible system may often be provided with an increased freedom and flexibility in the positioning of the physical sound transducers.
  • a low complexity, practical and easy-to-use user interface may allow the user to easily and intuitively provide relevant information for the system to configure the sound reproduction to the specific configuration.
  • a user of a home cinema system may more freely and flexibly position speakers.
  • the speaker positions may then be indicated to the system by the user inputting a graphical representation of the speaker position with the system subsequently selecting and adjusting the signals sent to the individual speakers such that the spatial sound presentation is adapted to the specific transducer positions.
  • the signal linker may for each spatial input channel select a subset of the sound render channels with the signal of the spatial input channel then being used to generate the drive signals for the selected sound render channels.
  • the right front input channel may be linked with one or more sound render channels having positions towards the right and front of the (nominal) listening position.
  • a graphical representation allows a particularly advantageous user experience as it provides an intuitive means for indicating sound source positions.
  • the interface to the user may be simple, easy to use, and highly intuitive.
  • the approach may allow an improved configuration and adaptation of the system since the graphical representation may provide information that e.g. is not available from simple measurements by a test-microphone at a listening position.
  • the graphical representation may indicate a listening position which can be adjusted by the user.
  • the signal linker may further determine the set of associations in response to the listening position.
  • the signal linker is arranged to associate a first spatial channel with a first sound render channel if the first sound render channel has a sound render position being a closest sound render position to a nominal position of the first spatial channel.
  • This may provide an improved spatial experience in many scenarios, and may in particular in many scenarios allow a perception of the first spatial channel originating from a position close to the nominal position for the first spatial channel.
  • the approach may further in many scenarios provide a low complexity implementation and/or operation.
  • the set of associations may associate more than one sound render channel with one spatial channel.
  • the driver is arranged to generate drive signals for at least two sound render channels from an input signal of a first spatial channel to provide a virtual sound source position for the first spatial channel closer to a nominal position of the first spatial channel than sound render positions of the at least two sound render channels.
  • This may allow improved flexibility and may in many scenarios provide an improved audio experience.
  • it may allow the sound rendered from the first channel to be perceived to originate from approximately the nominal position without requiring any physical sound transducers to be positioned at or necessarily close to that position.
  • the set of associations may associate more than one spatial channel with one sound render channel.
  • the approach may relax the relation between the number and assumed/nominal positions of spatial channels and the actual number and positions of sound transducers used for the audio rendering.
  • the driver is arranged to adjust at least one of an audio level and a delay of at least one of the drive signals in response to the sound render positions.
  • the audio system may provide an improved adaptation of the reproduced sound to the specific sound transducer setup. Specifically, the variations in the characteristics from the sound propagation from sound transducers to a listener as a function of the relative positions may be at least partially compensated. As a result an improved spatial experience may be provided.
  • the user interface comprises: a display for displaying the user input graphical representation with icons representing positions of the sound transducers; an input interface for moving the icons on the display in response to a user input.
  • This may provide a particularly advantageous user interface for inputting sound transducer setup information.
  • the input interface is operable to receive a user selection of an icon representing a position of a sound transducer; and wherein the audio system further comprises an audio generator for providing a test drive signal to the sound render channel corresponding to the selected icon.
  • the audio system further comprises an audio generator for providing a test drive signal to the sound render channel corresponding to the selected icon.
  • a test drive signal may be generated for the corresponding sound render channel resulting in the corresponding sound transducer radiating sound.
  • all other sound transducers may be kept silent (or the test signal may be a recognizable signal, such as a single tone).
  • the process may provide an efficient feedback to a user and substantially facilitate the user input.
  • the feature may in particular allow a very flexible and user friendly system allowing the user to freely and randomly connect or couple the different sound transducers (specifically loudspeakers) to the different output channels of the multi-channel system.
  • wires from loudspeakers may randomly be connected to the outputs from an amplifier without any consideration of any matching between the loudspeakers and outputs.
  • Any loudspeaker may be connected to any output.
  • the audio generator is arranged to adjust a volume of the test drive signal in response to a distance between a nominal listening position and the sound render position of the sound render channel corresponding to the selected icon.
  • This may further facilitate the user input and/or may improve the accuracy of the graphical representation. In particular, it may allow improved distance information to be provided.
  • the audio system further comprises: a microphone input for receiving a microphone signal; and a calibrator for setting at least one of a delay, volume and sound render direction for at least one of the drive signals in response to the microphone signal.
  • the calibrator may perform a setup or configuration process based on the graphical representation. This may in particular allow a closer adaptation to specific conditions of the audio environment.
  • a close interworking between adaptation based on user input and on (e.g. automatic) calibration based on a microphone signal may provide improved performance in many scenarios.
  • the microphone based calibration may be used to fine tune parameters having approximate values determined from the graphical representation.
  • the audio system further comprises an initializer for initializing a calibration by the calibrator in response to the sound render positions.
  • the user interface is part of a portable device operable to be remote from the driver.
  • the portable device may for example be a remote control or a mobile phone.
  • the user input graphical representation further comprises an indication of a characteristic of a listening environment, and the driver is arranged to generate the drive signals in response to the indication.
  • the characteristics may for example include positions of walls, furniture, etc.
  • the user input graphical representation comprises an indication of a listening position
  • the driver is arranged to generate the drive signals in response to the listening position.
  • This may provide a particularly advantageous system, and in particular may provide a practical, feasible, user friendly, and high performance system which can quickly and easily adapt and/or optimize the sound reproduction at different positions/areas.
  • the driver may be arranged to adapt a direction of sound radiation for at least one sound render channel in response to the characteristic.
  • a method of generating drive signals for a plurality of sound render channels each sound render channel providing a drive signal for a sound transducer; the method comprising: receiving a user input graphical representation of positions of the sound transducers; determining sound render positions for the sound transducers in response to the user input graphical
  • a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position;
  • Figures 1-3 illustrate examples of loudspeaker setups for spatial audio provision
  • FIG. 4 illustrates an example of an audio system in accordance with some embodiments of the invention
  • Figure 5 illustrates an example of a graphical representation of an audio system in accordance with some embodiments of the invention
  • Figure 6 illustrates an example of a portable device of an audio system in accordance with some embodiments of the invention
  • Figure 7 illustrates an example of elements of a portable device of an audio system in accordance with some embodiments of the invention.
  • Figure 8 illustrates an example of elements of a driver circuit of an audio system in accordance with some embodiments of the invention.
  • Fig. 1 illustrates an example of a nominal setup for a five channel surround sound system.
  • the loudspeakers are assumed to be positioned around a listening position 101 with a speaker directly in front of the listening position 101 (the center speaker 103), a speaker to the front left of the listening position (the front left speaker 105), a speaker to the front right of the listening position (the front right speaker 107), a speaker to the rear left of the listening position (the left surround speaker 109), and a speaker to the rear right of the listening position (the right surround speaker 111).
  • the spatial audio signal is generated to provide the desired spatial experience when the loudspeakers are positioned in accordance with the nominal setup relative to the listening position. Accordingly, users are required to position their speakers at specific locations relative to the listening position in order to achieve the optimum spatial experience. However, in practice this is inconvenient and in many cases impossible or at least unacceptable to do. Therefore, typical surround sound setups only approximate the nominal setup resulting in a degraded spatial experience.
  • an audio system which is arranged to provide a spatial experience yet to provide additional freedom in the positioning of loudspeakers and/or a facilitated loudspeaker set-up. Indeed, the system adapts to specific loudspeaker setups, and seeks to provide an approach where a user may position speakers relatively freely with the system then adapting its operation to the specific speaker setup.
  • the approach seeks to adapt the audio system to the specific speaker setup.
  • the system may provide a very advantageous spatial experience in many scenarios. For example, a user may freely position speakers to surround the listening position but without being at predetermined or nominal positions. The audio system then adapts to the speaker positions to provide a spatial experience that still relatively closely corresponds to that which would have been provided by a direct rendering of the received multi-channel signal from loudspeakers at the nominal positions. Examples of possible loudspeaker setups are illustrated in Figs. 2 and 3.
  • each output channel is statically and fixedly linked to a an input spatial channel so that the corresponding speaker must be placed at the corresponding nominal position
  • the system described in the following implements a dynamic linking between the spatial input channels and the output channels.
  • each output channel is not associated with a specific position and is not predetermined to correspond to a specific input channel.
  • the audio system merely amplifies each input signal to generate an output signal that directly corresponds to the input spatial channel, and thus is for a speaker that must be positioned at the nominal position to provide the desired spatial effect.
  • sound transducers and specifically loudspeakers may be positioned by the user in the audio environment without consideration of which input channel is to be rendered by the speaker.
  • the sound transducer may specifically be a loudspeaker, and the following description will for clarity and brevity assume that each sound transducer is a loudspeaker (e.g. comprising a plurality of drivers).
  • the user may position loudspeakers without considering the link to the input channels.
  • the audio system may then proceed to adapt the processing such that the implemented speaker setup provides a spatial experience corresponding to that which would be achieved for a conventional setup with speakers at the nominal positions.
  • the audio system allows a large variation in the actual speaker positions with the system adapting itself to reflect the specific speaker setup.
  • Fig. 4 illustrates an example of an audio system in accordance with some embodiments of the invention.
  • the audio system comprises a receiver 401 which receives a spatial multichannel signal.
  • the multichannel signal comprises a plurality of spatial channels, each of which is associated with a nominal sound source position.
  • the input signal is thus a signal where each channel is intended to be rendered from a specific nominal position relative to the listening position.
  • the input signal is a five channel surround sound system with a nominal speaker setup as illustrated in Fig. 1.
  • the input signal comprises a center channel, a front left channel, a front right channel, a surround left channel and a surround right channel.
  • the input signal may also comprise non- spatial signals, such as a Low Frequency Effects (LFE) channel.
  • LFE Low Frequency Effects
  • the receiver 401 demultiplexes the received multichannel signal to provide the individual spatial channel signals.
  • the input signal may be an encoded signal and the receiver 401 may include a decoder for generating the signals of the individual spatial channels.
  • the multichannel signal may be provided as a downmixed signal wherein the spatial channels are downmixed to a lower number of channels (e.g. to a mono or stereo signal).
  • the receiver 401 may include functionality for upmixing, or equivalent ly the multichannel signal may be considered to be received by the receiver 401 from an internal or external upmixing unit.
  • the audio system generates drive signals for a number of sound render channels.
  • Each sound render channel provides one drive signal for a sound transducer which in the specific example is a loudspeaker.
  • each sound render channel (output channel) corresponds to a specific sound source position in the listening environment, namely to the position of the loudspeaker driven by the drive signal.
  • the audio system generates the loudspeaker drive signals from the spatial input signals.
  • the link/ association between the input channels and loudspeakers (and loudspeaker positions) is not predetermined or fixed but is dynamically updated and determined for the individual loudspeaker setup.
  • the receiver 401 is coupled to a driver 403 which generates a drive signal for each sound render channel from the input signals.
  • the driver 403 receives the signals of all the spatial input signals and in response generates the drive signals for each loudspeaker.
  • Fig. 4 illustrates an amplifier 405 and a loudspeaker 407 for each sound render channel, but it will be considered that these elements may be considered external to or as part of the audio system.
  • the amplifiers 405 amplify the signal of each sound render channel to the appropriate levels for the loudspeakers 407.
  • the loudspeakers 407 are coupled to the audio system via physical wires, and in this example the amplifiers 405 may be included in a suitable device together with the receiver 401 and the driver 403.
  • the device may for example be a home cinema amplifier. However, in the example, the device does not have any
  • the device merely provides a number of outputs that may be connected to speakers at any position.
  • the user needs not consider which speaker output to use for a specific speaker but can rather use any of the outputs for any of the speakers.
  • the user can simple connect the speaker wires freely to the different speaker outputs. This simplifies the setting up of e.g. a home cinema system which for many users in especially the consumer market is considered to otherwise be complex and cumbersome.
  • the signal may e.g. be fed wirelessly to the speakers and the amplifiers 405 may be implemented together with the speakers 407 rather than in the driver unit.
  • the connections between a centralized audio device and the individual speakers may be implemented over a shared wireless network.
  • the loudspeakers may register with the centralized audio device and be assigned a network address. Each loudspeaker may be treated in exactly the same way and without any considerations of positions when the address is assigned. This may allow for a very simple, flexible and efficient set up process.
  • the audio system does not have any predetermined knowledge and assumption about positions of speakers. Rather, it is initially only known that a number of speakers are connected to the system but there is no spatial assumption or information associated with the individual speaker outputs (and thus with the sound render channels).
  • the system is arranged to dynamically adapt and select the links between the input channels and the sound render channels.
  • the driver 403 generates the drive signals for the render channels from input signals of the spatial channels based on a set of associations between the input spatial channels and the sound render channels.
  • an association may indicate that a first render channel is suitable for providing the perception of a sound source originating from in front of the listening position, and the drive signal for that sound render channel will therefore be generated from the input center channel.
  • Another association may indicate that a second sound render channel is suitable for providing the perception of a sound source originating from the right and front of the listening position, and the drive signal for that sound render channel will therefore be generated from the front right channel.
  • associations will be provided for all input channels but in some embodiments or scenarios, one or more input channels may not have any associations and therefore may not be rendered.
  • the determination of suitable associations is based on a graphical representation of the loudspeaker setup provided by a user.
  • the audio system of Fig. 4 comprises a user interface 409 which receives a user input graphical representation of the positions of the sound transducers.
  • the user interface 409 in the example of Fig. 4 comprises a touch screen wherein a graphical representation of the listening environment is displayed.
  • the graphical representation is a simple and stylized representation of the environment with a graphical element (icon) representing the listening position and a graphical element (icon) for each sound render channel.
  • a graphical representation is illustrated in Fig. 5.
  • the user may in this example simply move the icons representing the loudspeakers to positions that reflect the positions of the loudspeakers in the environment.
  • the user can move the graphical elements representing the speakers (in the example of Fig. 5 represented by dark circles) to correspond to the physical location of the speakers relative to the user.
  • the top of the screen corresponds to the position directly in front of the user/listening position.
  • the graphical representation is presented on a display which is a touch screen display. Accordingly, the user can simply touch the icons representing the loudspeaker positions and move them to the desired position in the graphical representation.
  • a very simple and intuitive, user interface is provided.
  • the user interface is coupled to a position processor 411 which receives data describing the graphical representation.
  • the position processor 411 proceeds to determine sound render positions for the sound transducers in response to the graphical representation.
  • the position processor 411 determines position values for the loudspeakers based on the graphical representation.
  • the position values may for example determine a distance and angle from the listening position to the individual speakers.
  • the distances may either be absolute or relative values.
  • the positions may be determined as coordinates in a two dimensional representation of the listening environment. In a simple embodiment, the positions of the loudspeakers in the listening environment may simply be determined by scaling of the relative distances between the icons in the graphical representation.
  • the position processor 411 is coupled to a linker 413 which proceeds to determine the set of associations between the input spatial channels and the output sound render channels based on the positions determined for the loudspeakers and on the nominal sound source positions.
  • the linker 413 generates a number of associations where each association represents a signal path link from the corresponding spatial channel to the corresponding sound render channel.
  • the association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel.
  • An association between a spatial channel and a sound render channel may specifically be indicative of a weight of the spatial channel in the drive signal for the sound render channel. A zero weight corresponds to no association, i.e. to a scenario where the sound render channel contains no contribution from the spatial channel.
  • the set of associations are fed to the driver 403 which proceeds to generate drive signals for the output sound render channels based on the set of the associations.
  • the linker 413 may generate one association for each input spatial channel.
  • one loudspeaker is used to render each input spatial channel.
  • the linker may select the associations between loudspeakers and spatial channels such that the speaker indicated by the graphical representation to be closest to the nominal position for a given spatial channel is associated with the sound render channel for that speaker.
  • the driver 403 may simply correspond to a switch matrix that can switch the input signals of the spatial signals to the best suited output sound render channel.
  • the central audio device may be provided with simple speaker outputs that are not labeled but rather are merely provided as identical and nonspecific speaker outputs.
  • the user can then place the speakers and connect them to the device using any speaker output for any speaker.
  • the user then moves the graphical elements representing the speakers on the touch screen display to represent their physical location relative to the user.
  • the audio system in response proceeds to switch the input channels to the appropriate speaker outputs to provide a suitable spatial experience.
  • the user interface may in many scenarios be provided as a portable device which can be operated remotely from the central speaker drive unit.
  • the user interface may be provided as an application on a mobile phone or by a remote control.
  • Fig. 6 shows an example where the user interface is provided in a mobile phone and Fig. 7 illustrates elements of a remote portable device for providing the user interface.
  • the portable device comprises a display 701 which is driven by a display controller 703 which presents the graphical representation on the display.
  • the display 701 is in the example a touch screen which is coupled to an input 705 that can detect and interpret the touches on the screen.
  • the portable device comprises a user interface controller 707 which receives the signal from the input 705 and provides data describing the graphical
  • the user interface 707 When a user operates the screen, this is detected by the user interface controller 707 which accordingly adjusts the graphical representation.
  • the user interface 707 is further coupled to a transceiver 709 which communicates with the central audio device.
  • the communication may for example be via a wireless communication, such as for example using a BluetoothTM connection.
  • the transceiver 709 transmits data describing the graphical representation to the central audio device which proceeds to determine the transducer positions and the associations.
  • the exact functionality distribution between the central audio unit and the portable device may depend on the preferences and requirements of the individual embodiment.
  • the position processor 411 and possibly the linker 413 may be implemented in the portable device or in the central audio device.
  • the routing between input channels (the spatial channels) and the output channels (the sound render channels) is determined based on the user input graphical representation. This may substantially facilitate the setting up by non- expert users and may provide a much improved user experience.
  • the user input graphical representation may also be used to adapt signal processing characteristics of the system.
  • an audio level and/or delay for the drive signals may be set in response to the sound render positions determined from the graphical interface.
  • the delay and sound level may be increased for an increasing distance from the loudspeakers to the listening position.
  • the system may on the basis of the graphical representation compensate for different propagation delays and attenuations thereby providing an improved audio experience.
  • each spatial channel may be rendered in a single sound render channel and each sound render channel may render only one spatial channel.
  • a spatial channel may be rendered in a plurality of channels and/or a sound render channel may comprise contributions from a plurality of channels. Further, the number of sound render channels may be different from the number of spatial channels.
  • Each association between a spatial channel and a sound render channel indicates that a signal path exists from the spatial channel to the sound render channel.
  • the set of associations generated by the linker may comprise a plurality of associations for a spatial channel and/or a plurality of associations for a sound render channel.
  • the set of associations may associate more than one sound render channel with one spatial channel and/or may associate more than one spatial channel with one sound render channel.
  • the system may be arranged to render a spatial channel through a plurality of sound transducers such that the combined effect results in a perceived sound source of the spatial channel to be closer to the nominal position of the spatial channel than of the positions of the loudspeakers rendering the sound.
  • the system uses the information of the loudspeaker positions to generate virtual sound positions in between the loudspeakers.
  • the linker 413 may for each spatial channel determine the two sound transducers that are closest to the nominal position on each side.
  • the signal of the spatial channel may then be rendered by these two loudspeakers using amplitude panning to provide a perception of the sound source being in between the two loudspeakers. For example, if the nominal position is in the middle of the two loudspeaker positions, the gain for the two drive signals are equal and if the nominal position is closer to one of the loudspeakers, the gain for this loudspeaker will be set higher than for the other loudspeaker.
  • the overall gain may be set dependent on the distance from the listening position to the two loudspeaker relative to the distance to the nominal position and/or relative to the distance to loudspeakers used for other channels.
  • each spatial channel is thus rendered by two sound transducers and each sound transducer may render more than one spatial channel.
  • each drive signal may be generated as a weighted combination of the input spatial channels, i.e. each sound render channel may possibly comprise contributions from all spatial channels.
  • each sound render channel may possibly comprise contributions from all spatial channels.
  • the output drive signals may be determined as:
  • x n represents input spatial channel n
  • y m represents output sound render channel m
  • Ck,i represents the association between input spatial channel k and output sound render channel. If no association exists the value of the corresponding matrix coefficient is zero.
  • the output drive signals may be determined as:
  • the output drive signals may thus be generated from a gain matrix which provides associations between input spatial channels and output sound render channels.
  • a large number of the coefficients of the gain matrix will be zero corresponding to there being no association between the corresponding input spatial channel and output sound render channel.
  • the signal paths may also include other compensations or signal processing such as e.g. frequency compensations (e.g. filtering).
  • the signal paths may include a delay which specifically may compensate for differences in the distance between the sound transducers and the listening position.
  • the driver 403 may for each spatial channel provided the functionality illustrated in Fig. 8.
  • a signal path with a variable gain and delay may be provided to each of the sound render channels.
  • the values of the gain and delay may be determined fully or partially based on the graphical representation. If no association is provided between a spatial channel and a sound render channel, the corresponding gain will be zero. It will be appreciated that the delays will typically be the same for all spatial channels for a given sound render channel, and that a single delay may accordingly be applied to the drive signal generated by combining the individual contributions from the different spatial channels.
  • the system may accordingly provide a very flexible setup.
  • a user may simply connect (e.g. directly, or using a wireless or network based connections) as many speakers as he would like to use.
  • the speakers may be positioned freely but may preferably be situated to surround the listening position.
  • the user then uses his portable device to quickly and easily generate a graphical representation of the speaker setup, e.g. simply by sliding some icons to appropriate positions on a touch screen.
  • the system automatically adapts to provide an appropriate spatial experience, and may specifically proceed to generate virtual sound sources from positions corresponding to (or approximating) the nominal positions of the received spatial multi-channel signal.
  • the user may freely select how many speakers to use and may position these freely.
  • Increasing the number of speakers used, and providing a more even distribution of the loudspeakers around the listening position, may typically increase the accuracy of the virtual sound source positioning and the resulting spatial experience. However, in some embodiments it may be more attractive to reduce cost and complexity, and therefore the number of loudspeakers may be kept low. Thus, the user has a high degree of flexibility in both how many and where to position loudspeakers, and the same audio system may be used as it can easily adapt to the individual configuration.
  • the system may be arranged to generate a sound from a sound transducer in response to a selection of the corresponding icon on the display. For example, when the use touches an icon corresponding to a sound source position on the display, this icon may e.g. change color to indicate that it has been selected.
  • an audio generator may provide a test drive signal on the corresponding sound render channel resulting in the loudspeaker connected to the output of that sound render channel rendering the corresponding sound.
  • the test drive signal may typically represent a distinct and detectable sound such as a pure tone or a white noise signal.
  • a distinct and detectable sound such as a pure tone or a white noise signal.
  • the user will immediately hear a characteristic and distinct sound coming from the corresponding loudspeaker. He may then easily slide or move the icon to a position corresponding to the position of the loudspeaker radiating the sound.
  • No predetermined information or correlation between loudspeaker positions and the sound render channels is required, and specifically any output can be used with any loudspeaker.
  • the generated test signal and thus the rendered test audio, may have a volume that is adjusted in response to the distance from the nominal listening position to the position of the icon for the sound transducer position on the display. Specifically, the volume may be increased as the icon is moved further away and decreased when it is moved closer. This may assist the user in positioning the icons to reflect the actual sound positions, and may compensate for differences in the distances and thus in attenuation of the sound from the different loudspeakers.
  • the processing of the audio system may be setup based only on the graphical representation.
  • the system may further be arranged to perform an automatic or semi-automatic calibration based on a microphone signal.
  • the system may comprise a microphone input which is coupled to a microphone that can be positioned at the listening position.
  • a calibrator may then receive the microphone signal and adjust a characteristic of the signal processing dependent on the microphone signal. Specifically, a delay, volume and/or sound render direction of the drive signals may be adjusted dependent on the microphone signal.
  • a test generator may generate a test signal which is fed to one of the sound transducers.
  • the time difference between the generated test signal and the received signal by the microphone may be used to determine the propagation delay from the sound transducer to the listening position. This may be performed for all sound render channels and the appropriate relative delays may be applied to ensure that sound is received substantially simultaneously.
  • the approach may combine the approach of calibrating/adapting the system based on the graphical representation with a microphone based calibration/adaptation.
  • An improved performance may be achieved by combining the two approaches.
  • the calibration based on the graphical representation may provide a coarse
  • the sound render positions determined from the graphical representation may be used to initialize the calibration process based on the microphone signal.
  • the graphical representation may be used to determine initial estimates for delay and volume as well as to setup the virtual sound source positioning.
  • the calibration based on the microphone signal may then proceed based on these initial values.
  • Calibration processes may often determine local optimizations and by initializing the calibration with approximate initial values, the chance of the determined local optimum is also the global optimum is also substantially increased.
  • the graphical representation can be used to determine or adapt to characteristics that are typically not addressed by a microphone calibration. For example, the angle of the sound reaching the listening position may be estimated based on the graphical representation. Further, the virtual sound source positioning may be determined based on the graphical representation with the subsequent microphone calibration being used to calibrate characteristics of the virtual sound source.
  • the user input graphical representation may comprise an indication of a characteristic of a listening environment. This additional information may then be used to adapt the signal processing of the driver 403.
  • the graphical representation may be enhanced to allow different listening positions to be indicated, such that the drive signal generation may compensate for differences in the listening position.
  • the user may be able to indicate the presence of walls or other obstacles in the listening environment. This may specifically indicate sound reflections.
  • the graphical representation can be used to estimate e.g. a direction in which sound should be radiated from a sound transducer in order to reach the listening position from a reflected direction corresponding to a desired virtual sound source direction. For example, a wall to the rear of the listening position may be indicated on the graphical representation. In response, sound may be radiated towards this wall to provide reflected sound reaching the listening position from behind.
  • an embodiment of the invention may be physically, functionally and logically implemented in any suitable way. Indeed the functionality may be implemented in a single unit, in a plurality of units or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units, circuits and processors.

Abstract

An audio system generates drive signals for sound render channels driving a sound transducer. A user interface (409) receives a user input graphical representation of positions of the sound transducers and a processor (411) determines sound render positions for the sound transducers in response to the user input graphical representation. An input (401) receives a spatial multichannel signal comprising a plurality of spatial channels associated with a nominal sound source position. A signal linker (413) determines a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions where an association between a spatial channel and a sound render channel indicates that the sound render channel is to be used for rendering the spatial channel. A driver (403) then generates drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.

Description

An audio system and method of operating therefor
FIELD OF THE INVENTION
The invention relates to an audio system and a method of operation therefor, and in particular, but not exclusively, to generation for drive signals for loudspeakers of spatial surround sound system.
BACKGROUND OF THE INVENTION
Spatial sound reproduction beyond simple stereo has become commonplace through applications such as home cinema systems. Typically such systems use loudspeakers positioned at specific spatial positions relative to a listening position. For example, a 5.1 home cinema system provides spatial sound via five loudspeakers being positioned with one speaker directly in front of the listening position (the center channel), one speaker to the front left of the listening position, one speaker to the front right of the listening position, one speaker to the rear left of the listening position, and one speaker to the rear right of the listening position. In addition, a non-spatial low frequency speaker is provided.
Such conventional systems are based the reproduction of audio signals at specific nominal positions relative to the listening position. One speaker is typically provided for each audio channel and therefore speakers must be positioned at positions corresponding to the predetermined or nominal positions for the system. However, the requirement for loudspeakers to be in specific positions is often found to be highly undesirable to users and is particularly inconvenient in the consumer market.
Although users are expected to position speakers approximately at nominal positions relative to the listening position, deviations will occur in real use cases in consumers' homes. In particular, the angle and distance to the loudspeakers from the listening position varies from one home cinema setup to another. In order to address such variations, automatic optimization procedures have been developed. Typically, such procedures are based on positioning a microphone at the listening position and radiating test signals from the loudspeakers. The test signals are received by the microphone and the system adapts the processing of the audio channels dependent on the received test signals. For example, the volume for each individual channel may be adjusted based on the measured test signal for the corresponding loudspeakers.
However, although such compensation may improve the spatial listening experience and may provide some additional compensation for variations in the speaker setup, the loudspeakers are still required to positioned in approximately the nominal configuration. Indeed, typically the compensation algorithm will fail if the actual loudspeaker positions deviate too much from the assumed nominal positions/angles.
Hence, an improved audio system would be advantageous and in particular a system allowing increased flexibility, facilitated operation, increased freedom in sound transducer placement, improved perceived audio quality, improved audio quality, an improved spatial user experience and/or improved performance would be advantageous.
SUMMARY OF THE INVENTION
Accordingly, the Invention seeks to preferably mitigate, alleviate or eliminate one or more of the above mentioned disadvantages singly or in any combination.
According to an aspect of the invention there is provided an audio system for generating drive signals for a plurality of sound render channels, each sound render channel providing a drive signal for a sound transducer; the audio system comprising: a user interface for receiving a user input graphical representation of positions of the sound transducers; a processor for determining sound render positions for the sound transducers in response to the user input graphical representation; an input for receiving a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position; a signal linker for determining a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions, an association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel; and a driver for generating drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.
The invention may provide an advantageous audio system for providing spatial audio. A more flexible system may often be provided with an increased freedom and flexibility in the positioning of the physical sound transducers. A low complexity, practical and easy-to-use user interface may allow the user to easily and intuitively provide relevant information for the system to configure the sound reproduction to the specific configuration. For example, a user of a home cinema system may more freely and flexibly position speakers. The speaker positions may then be indicated to the system by the user inputting a graphical representation of the speaker position with the system subsequently selecting and adjusting the signals sent to the individual speakers such that the spatial sound presentation is adapted to the specific transducer positions.
The signal linker may for each spatial input channel select a subset of the sound render channels with the signal of the spatial input channel then being used to generate the drive signals for the selected sound render channels. For example, the right front input channel may be linked with one or more sound render channels having positions towards the right and front of the (nominal) listening position.
The use of a graphical representation allows a particularly advantageous user experience as it provides an intuitive means for indicating sound source positions. Thus, the interface to the user may be simple, easy to use, and highly intuitive.
The approach may allow an improved configuration and adaptation of the system since the graphical representation may provide information that e.g. is not available from simple measurements by a test-microphone at a listening position.
In some embodiments, the graphical representation may indicate a listening position which can be adjusted by the user. In such embodiments, the signal linker may further determine the set of associations in response to the listening position.
In accordance with an optional feature of the invention, the signal linker is arranged to associate a first spatial channel with a first sound render channel if the first sound render channel has a sound render position being a closest sound render position to a nominal position of the first spatial channel.
This may provide an improved spatial experience in many scenarios, and may in particular in many scenarios allow a perception of the first spatial channel originating from a position close to the nominal position for the first spatial channel. The approach may further in many scenarios provide a low complexity implementation and/or operation.
In accordance with an optional feature of the invention, the set of associations may associate more than one sound render channel with one spatial channel.
This may allow improved flexibility and may in many scenarios provide an improved audio experience. The approach may relax the relation between the number and assumed/nominal positions of spatial channels and the actual number and positions of sound transducers used for the audio rendering. A single sound transducer may thus be used by a plurality of input channels, e.g. to generate virtual sound positions for the spatial channels. In accordance with an optional feature of the invention, the driver is arranged to generate drive signals for at least two sound render channels from an input signal of a first spatial channel to provide a virtual sound source position for the first spatial channel closer to a nominal position of the first spatial channel than sound render positions of the at least two sound render channels.
This may allow improved flexibility and may in many scenarios provide an improved audio experience. In particular, it may allow the sound rendered from the first channel to be perceived to originate from approximately the nominal position without requiring any physical sound transducers to be positioned at or necessarily close to that position.
In accordance with an optional feature of the invention, the set of associations may associate more than one spatial channel with one sound render channel.
This may allow improved flexibility and may in many scenarios provide an improved audio experience. The approach may relax the relation between the number and assumed/nominal positions of spatial channels and the actual number and positions of sound transducers used for the audio rendering.
In accordance with an optional feature of the invention, the driver is arranged to adjust at least one of an audio level and a delay of at least one of the drive signals in response to the sound render positions.
The audio system may provide an improved adaptation of the reproduced sound to the specific sound transducer setup. Specifically, the variations in the characteristics from the sound propagation from sound transducers to a listener as a function of the relative positions may be at least partially compensated. As a result an improved spatial experience may be provided.
In accordance with an optional feature of the invention, the user interface comprises: a display for displaying the user input graphical representation with icons representing positions of the sound transducers; an input interface for moving the icons on the display in response to a user input.
This may provide a particularly advantageous user interface for inputting sound transducer setup information.
In accordance with an optional feature of the invention, the input interface is operable to receive a user selection of an icon representing a position of a sound transducer; and wherein the audio system further comprises an audio generator for providing a test drive signal to the sound render channel corresponding to the selected icon. This may facilitate operation and/or may allow a more flexible setup. For example, it may mitigate or avoid the requirement for predetermined associations between sound transducers and nominal positions and/or between the sound transducers and icons on the user interface. The approach may allow facilitated identification of corresponding icons and sound transducers.
For example, when a user touches an icon representing a sound transducer/sound transducer position on a touch screen display, a test drive signal may be generated for the corresponding sound render channel resulting in the corresponding sound transducer radiating sound. At the same time all other sound transducers may be kept silent (or the test signal may be a recognizable signal, such as a single tone). The process may provide an efficient feedback to a user and substantially facilitate the user input.
The feature may in particular allow a very flexible and user friendly system allowing the user to freely and randomly connect or couple the different sound transducers (specifically loudspeakers) to the different output channels of the multi-channel system. As a specific example, wires from loudspeakers may randomly be connected to the outputs from an amplifier without any consideration of any matching between the loudspeakers and outputs. Any loudspeaker may be connected to any output.
In accordance with an optional feature of the invention, the audio generator is arranged to adjust a volume of the test drive signal in response to a distance between a nominal listening position and the sound render position of the sound render channel corresponding to the selected icon.
This may further facilitate the user input and/or may improve the accuracy of the graphical representation. In particular, it may allow improved distance information to be provided.
In accordance with an optional feature of the invention, the audio system further comprises: a microphone input for receiving a microphone signal; and a calibrator for setting at least one of a delay, volume and sound render direction for at least one of the drive signals in response to the microphone signal.
This may provide improved audio quality and/or performance in many environments. The calibrator may perform a setup or configuration process based on the graphical representation. This may in particular allow a closer adaptation to specific conditions of the audio environment.
A close interworking between adaptation based on user input and on (e.g. automatic) calibration based on a microphone signal may provide improved performance in many scenarios. For example, the microphone based calibration may be used to fine tune parameters having approximate values determined from the graphical representation.
In accordance with an optional feature of the invention, the audio system further comprises an initializer for initializing a calibration by the calibrator in response to the sound render positions.
This may provide improved audio quality and/or performance in many environments. For example, many automatic calibration/ configuration processes are based on finding local optimizations for the relevant parameters. By initializing such a calibration or configuration based on the graphical representation, the probability that the nearest local optimization is also a global optimization can be substantially increased. In addition, a much faster calibration can often be achieved.
In accordance with an optional feature of the invention, the user interface is part of a portable device operable to be remote from the driver.
This may provide a particularly advantageous user experience. The portable device may for example be a remote control or a mobile phone.
In accordance with an optional feature of the invention, the user input graphical representation further comprises an indication of a characteristic of a listening environment, and the driver is arranged to generate the drive signals in response to the indication.
This may provide improved quality of the audio or spatial experience. The characteristics may for example include positions of walls, furniture, etc.
In accordance with an optional feature of the invention, the user input graphical representation comprises an indication of a listening position, and the driver is arranged to generate the drive signals in response to the listening position.
This may provide a particularly advantageous system, and in particular may provide a practical, feasible, user friendly, and high performance system which can quickly and easily adapt and/or optimize the sound reproduction at different positions/areas.
In some embodiments, the driver may be arranged to adapt a direction of sound radiation for at least one sound render channel in response to the characteristic.
This may provide increased flexibility in sound transducer setup and/or may provide an improved spatial experience from the specific positions. E.g. the approach may be used to provide a sound reaching the listening position from a direction at which no sound transducer has been positioned. According to an aspect of the invention there is provided a method of generating drive signals for a plurality of sound render channels, each sound render channel providing a drive signal for a sound transducer; the method comprising: receiving a user input graphical representation of positions of the sound transducers; determining sound render positions for the sound transducers in response to the user input graphical
representation; receiving a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position;
determining a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions, an association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel; and generating drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.
These and other aspects, features and advantages of the invention will be apparent from and elucidated with reference to the embodiment(s) described hereinafter.
BRIEF DESCRIPTION OF THE DRAWINGS
Embodiments of the invention will be described, by way of example only, with reference to the drawings, in which
Figures 1-3 illustrate examples of loudspeaker setups for spatial audio provision;
Figure 4 illustrates an example of an audio system in accordance with some embodiments of the invention;
Figure 5 illustrates an example of a graphical representation of an audio system in accordance with some embodiments of the invention;
Figure 6 illustrates an example of a portable device of an audio system in accordance with some embodiments of the invention;
Figure 7 illustrates an example of elements of a portable device of an audio system in accordance with some embodiments of the invention; and
Figure 8 illustrates an example of elements of a driver circuit of an audio system in accordance with some embodiments of the invention.
DETAILED DESCRIPTION OF SOME EMBODIMENTS OF THE INVENTION The following description focuses on embodiments of the invention applicable to a surround sound home cinema system, and in particular to a five channel surround sound home cinema system. However, it will be appreciated that the invention is not limited to this application but may be applied to many other spatial audio systems.
Home cinema systems provide a spatial experience using a plurality of loudspeakers positioned at or close to nominal positions. Thus, a spatial multi-channel signal is provided with a signal for each of the nominal positions of a loudspeaker. Fig. 1 illustrates an example of a nominal setup for a five channel surround sound system.
In the example, the loudspeakers are assumed to be positioned around a listening position 101 with a speaker directly in front of the listening position 101 (the center speaker 103), a speaker to the front left of the listening position (the front left speaker 105), a speaker to the front right of the listening position (the front right speaker 107), a speaker to the rear left of the listening position (the left surround speaker 109), and a speaker to the rear right of the listening position (the right surround speaker 111).
The spatial audio signal is generated to provide the desired spatial experience when the loudspeakers are positioned in accordance with the nominal setup relative to the listening position. Accordingly, users are required to position their speakers at specific locations relative to the listening position in order to achieve the optimum spatial experience. However, in practice this is inconvenient and in many cases impossible or at least unacceptable to do. Therefore, typical surround sound setups only approximate the nominal setup resulting in a degraded spatial experience.
Indeed, the restrictions on speaker positioning is considered one of the main disadvantages of spatial audio systems. Indeed, users would prefer to be able to position speakers at positions reflecting the specific preferences and requirements in the individual environment.
In the following, an audio system will be described which is arranged to provide a spatial experience yet to provide additional freedom in the positioning of loudspeakers and/or a facilitated loudspeaker set-up. Indeed, the system adapts to specific loudspeaker setups, and seeks to provide an approach where a user may position speakers relatively freely with the system then adapting its operation to the specific speaker setup.
Thus, rather than requiring the positioning of the speakers to be adapted to the spatial audio, the approach seeks to adapt the audio system to the specific speaker setup.
The system may provide a very advantageous spatial experience in many scenarios. For example, a user may freely position speakers to surround the listening position but without being at predetermined or nominal positions. The audio system then adapts to the speaker positions to provide a spatial experience that still relatively closely corresponds to that which would have been provided by a direct rendering of the received multi-channel signal from loudspeakers at the nominal positions. Examples of possible loudspeaker setups are illustrated in Figs. 2 and 3.
In contrast to conventional systems where each output channel is statically and fixedly linked to a an input spatial channel so that the corresponding speaker must be placed at the corresponding nominal position, the system described in the following implements a dynamic linking between the spatial input channels and the output channels. Thus each output channel is not associated with a specific position and is not predetermined to correspond to a specific input channel. This is in direct contrast to conventional approaches wherein the audio system merely amplifies each input signal to generate an output signal that directly corresponds to the input spatial channel, and thus is for a speaker that must be positioned at the nominal position to provide the desired spatial effect.
In the system described in the following, sound transducers and specifically loudspeakers may be positioned by the user in the audio environment without consideration of which input channel is to be rendered by the speaker. The sound transducer may specifically be a loudspeaker, and the following description will for clarity and brevity assume that each sound transducer is a loudspeaker (e.g. comprising a plurality of drivers). In the system, the user may position loudspeakers without considering the link to the input channels. The audio system may then proceed to adapt the processing such that the implemented speaker setup provides a spatial experience corresponding to that which would be achieved for a conventional setup with speakers at the nominal positions. Thus, rather than require a close correlation between the actual speaker positions and the nominal positions, the audio system allows a large variation in the actual speaker positions with the system adapting itself to reflect the specific speaker setup.
Fig. 4 illustrates an example of an audio system in accordance with some embodiments of the invention.
The audio system comprises a receiver 401 which receives a spatial multichannel signal. The multichannel signal comprises a plurality of spatial channels, each of which is associated with a nominal sound source position. The input signal is thus a signal where each channel is intended to be rendered from a specific nominal position relative to the listening position. In the specific example, the input signal is a five channel surround sound system with a nominal speaker setup as illustrated in Fig. 1. Thus, the input signal comprises a center channel, a front left channel, a front right channel, a surround left channel and a surround right channel. It will be appreciated that the input signal may also comprise non- spatial signals, such as a Low Frequency Effects (LFE) channel.
In the example, the receiver 401 demultiplexes the received multichannel signal to provide the individual spatial channel signals. In some embodiments, the input signal may be an encoded signal and the receiver 401 may include a decoder for generating the signals of the individual spatial channels. In some embodiments, the multichannel signal may be provided as a downmixed signal wherein the spatial channels are downmixed to a lower number of channels (e.g. to a mono or stereo signal). In such embodiments, the receiver 401 may include functionality for upmixing, or equivalent ly the multichannel signal may be considered to be received by the receiver 401 from an internal or external upmixing unit.
The audio system generates drive signals for a number of sound render channels. Each sound render channel provides one drive signal for a sound transducer which in the specific example is a loudspeaker. Hence, each sound render channel (output channel) corresponds to a specific sound source position in the listening environment, namely to the position of the loudspeaker driven by the drive signal. Thus, the audio system generates the loudspeaker drive signals from the spatial input signals. However, in the system, the link/ association between the input channels and loudspeakers (and loudspeaker positions) is not predetermined or fixed but is dynamically updated and determined for the individual loudspeaker setup.
In the system of Fig. 4, the receiver 401 is coupled to a driver 403 which generates a drive signal for each sound render channel from the input signals. Thus, the driver 403 receives the signals of all the spatial input signals and in response generates the drive signals for each loudspeaker. Fig. 4 illustrates an amplifier 405 and a loudspeaker 407 for each sound render channel, but it will be considered that these elements may be considered external to or as part of the audio system. The amplifiers 405 amplify the signal of each sound render channel to the appropriate levels for the loudspeakers 407.
In the specific example, the loudspeakers 407 are coupled to the audio system via physical wires, and in this example the amplifiers 405 may be included in a suitable device together with the receiver 401 and the driver 403. The device may for example be a home cinema amplifier. However, in the example, the device does not have any
predetermined relationship between the sound render channel outputs (i.e. speaker outputs) and assumed positions of the corresponding loudspeakers. Rather, the device merely provides a number of outputs that may be connected to speakers at any position. Thus, the user needs not consider which speaker output to use for a specific speaker but can rather use any of the outputs for any of the speakers. Thus, the user can simple connect the speaker wires freely to the different speaker outputs. This simplifies the setting up of e.g. a home cinema system which for many users in especially the consumer market is considered to otherwise be complex and cumbersome.
In other embodiments, the signal may e.g. be fed wirelessly to the speakers and the amplifiers 405 may be implemented together with the speakers 407 rather than in the driver unit. For example, the connections between a centralized audio device and the individual speakers may be implemented over a shared wireless network. The loudspeakers may register with the centralized audio device and be assigned a network address. Each loudspeaker may be treated in exactly the same way and without any considerations of positions when the address is assigned. This may allow for a very simple, flexible and efficient set up process.
Thus, in the system of Fig. 4, the audio system does not have any predetermined knowledge and assumption about positions of speakers. Rather, it is initially only known that a number of speakers are connected to the system but there is no spatial assumption or information associated with the individual speaker outputs (and thus with the sound render channels).
Accordingly it is not possible for the system to use a predetermined allocation of input channels to sound render channels. For example, it is not known which output (sound render channel) is connected to a speaker which is positioned to best provide a perception of a sound source directly in front of the listening position. It is therefore not possible to determine which output (sound render channel) should be fed the input signal received for the center channel.
Instead, the system is arranged to dynamically adapt and select the links between the input channels and the sound render channels. In particular, the driver 403 generates the drive signals for the render channels from input signals of the spatial channels based on a set of associations between the input spatial channels and the sound render channels.
For example, an association may indicate that a first render channel is suitable for providing the perception of a sound source originating from in front of the listening position, and the drive signal for that sound render channel will therefore be generated from the input center channel. Another association may indicate that a second sound render channel is suitable for providing the perception of a sound source originating from the right and front of the listening position, and the drive signal for that sound render channel will therefore be generated from the front right channel. Typically associations will be provided for all input channels but in some embodiments or scenarios, one or more input channels may not have any associations and therefore may not be rendered.
In the system, the determination of suitable associations is based on a graphical representation of the loudspeaker setup provided by a user.
Specifically, the audio system of Fig. 4 comprises a user interface 409 which receives a user input graphical representation of the positions of the sound transducers.
The user interface 409 in the example of Fig. 4 comprises a touch screen wherein a graphical representation of the listening environment is displayed. In the example, the graphical representation is a simple and stylized representation of the environment with a graphical element (icon) representing the listening position and a graphical element (icon) for each sound render channel. An example of such a graphical representation is illustrated in Fig. 5.
The user may in this example simply move the icons representing the loudspeakers to positions that reflect the positions of the loudspeakers in the environment. Thus, the user can move the graphical elements representing the speakers (in the example of Fig. 5 represented by dark circles) to correspond to the physical location of the speakers relative to the user. In the example, the top of the screen corresponds to the position directly in front of the user/listening position.
In the example, the graphical representation is presented on a display which is a touch screen display. Accordingly, the user can simply touch the icons representing the loudspeaker positions and move them to the desired position in the graphical representation. Thus, a very simple and intuitive, user interface is provided.
The user interface is coupled to a position processor 411 which receives data describing the graphical representation. The position processor 411 proceeds to determine sound render positions for the sound transducers in response to the graphical representation. Thus, the position processor 411 determines position values for the loudspeakers based on the graphical representation. The position values may for example determine a distance and angle from the listening position to the individual speakers. The distances may either be absolute or relative values. Additionally or alternatively, the positions may be determined as coordinates in a two dimensional representation of the listening environment. In a simple embodiment, the positions of the loudspeakers in the listening environment may simply be determined by scaling of the relative distances between the icons in the graphical representation.
The position processor 411 is coupled to a linker 413 which proceeds to determine the set of associations between the input spatial channels and the output sound render channels based on the positions determined for the loudspeakers and on the nominal sound source positions. The linker 413 generates a number of associations where each association represents a signal path link from the corresponding spatial channel to the corresponding sound render channel. The association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel. An association between a spatial channel and a sound render channel may specifically be indicative of a weight of the spatial channel in the drive signal for the sound render channel. A zero weight corresponds to no association, i.e. to a scenario where the sound render channel contains no contribution from the spatial channel.
The set of associations are fed to the driver 403 which proceeds to generate drive signals for the output sound render channels based on the set of the associations.
As a low complexity example, the linker 413 may generate one association for each input spatial channel. In this example, one loudspeaker is used to render each input spatial channel. The linker may select the associations between loudspeakers and spatial channels such that the speaker indicated by the graphical representation to be closest to the nominal position for a given spatial channel is associated with the sound render channel for that speaker.
In such an example, the driver 403 may simply correspond to a switch matrix that can switch the input signals of the spatial signals to the best suited output sound render channel.
In such a system, the central audio device may be provided with simple speaker outputs that are not labeled but rather are merely provided as identical and nonspecific speaker outputs. The user can then place the speakers and connect them to the device using any speaker output for any speaker. The user then moves the graphical elements representing the speakers on the touch screen display to represent their physical location relative to the user. The audio system in response proceeds to switch the input channels to the appropriate speaker outputs to provide a suitable spatial experience. The user interface may in many scenarios be provided as a portable device which can be operated remotely from the central speaker drive unit. For example, the user interface may be provided as an application on a mobile phone or by a remote control.
Fig. 6 shows an example where the user interface is provided in a mobile phone and Fig. 7 illustrates elements of a remote portable device for providing the user interface. The portable device comprises a display 701 which is driven by a display controller 703 which presents the graphical representation on the display. The display 701 is in the example a touch screen which is coupled to an input 705 that can detect and interpret the touches on the screen. The portable device comprises a user interface controller 707 which receives the signal from the input 705 and provides data describing the graphical
representation to the display controller 703. When a user operates the screen, this is detected by the user interface controller 707 which accordingly adjusts the graphical representation. The user interface 707 is further coupled to a transceiver 709 which communicates with the central audio device. The communication may for example be via a wireless communication, such as for example using a Bluetooth™ connection. The transceiver 709 transmits data describing the graphical representation to the central audio device which proceeds to determine the transducer positions and the associations.
It will be appreciated that the exact functionality distribution between the central audio unit and the portable device may depend on the preferences and requirements of the individual embodiment. For example, the position processor 411 and possibly the linker 413 may be implemented in the portable device or in the central audio device.
In the previous example, the routing between input channels (the spatial channels) and the output channels (the sound render channels) is determined based on the user input graphical representation. This may substantially facilitate the setting up by non- expert users and may provide a much improved user experience.
However, in addition, the user input graphical representation may also be used to adapt signal processing characteristics of the system. In particular, an audio level and/or delay for the drive signals may be set in response to the sound render positions determined from the graphical interface. In particular, the delay and sound level may be increased for an increasing distance from the loudspeakers to the listening position. Thus, the system may on the basis of the graphical representation compensate for different propagation delays and attenuations thereby providing an improved audio experience.
In some embodiments, each spatial channel may be rendered in a single sound render channel and each sound render channel may render only one spatial channel. However, in other embodiments, a spatial channel may be rendered in a plurality of channels and/or a sound render channel may comprise contributions from a plurality of channels. Further, the number of sound render channels may be different from the number of spatial channels.
Each association between a spatial channel and a sound render channel indicates that a signal path exists from the spatial channel to the sound render channel. The set of associations generated by the linker may comprise a plurality of associations for a spatial channel and/or a plurality of associations for a sound render channel. Thus, in some embodiments, the set of associations may associate more than one sound render channel with one spatial channel and/or may associate more than one spatial channel with one sound render channel.
Indeed, the system may be arranged to render a spatial channel through a plurality of sound transducers such that the combined effect results in a perceived sound source of the spatial channel to be closer to the nominal position of the spatial channel than of the positions of the loudspeakers rendering the sound. Thus, the system uses the information of the loudspeaker positions to generate virtual sound positions in between the loudspeakers.
As a low complexity example, the linker 413 may for each spatial channel determine the two sound transducers that are closest to the nominal position on each side. The signal of the spatial channel may then be rendered by these two loudspeakers using amplitude panning to provide a perception of the sound source being in between the two loudspeakers. For example, if the nominal position is in the middle of the two loudspeaker positions, the gain for the two drive signals are equal and if the nominal position is closer to one of the loudspeakers, the gain for this loudspeaker will be set higher than for the other loudspeaker. Furthermore, the overall gain may be set dependent on the distance from the listening position to the two loudspeaker relative to the distance to the nominal position and/or relative to the distance to loudspeakers used for other channels.
In such an embodiment, each spatial channel is thus rendered by two sound transducers and each sound transducer may render more than one spatial channel.
In a more general implementation, each drive signal may be generated as a weighted combination of the input spatial channels, i.e. each sound render channel may possibly comprise contributions from all spatial channels. For example, for five spatial channels and four sound transducers (and thus four sound render channels), the output drive signals may be determined as:
Figure imgf000017_0001
where xn represents input spatial channel n, ym represents output sound render channel m, and Ck,i represents the association between input spatial channel k and output sound render channel. If no association exists the value of the corresponding matrix coefficient is zero.
For five spatial channels and six sound transducers, the output drive signals may be determined as:
Figure imgf000017_0002
In this example, the output drive signals may thus be generated from a gain matrix which provides associations between input spatial channels and output sound render channels. Typically, a large number of the coefficients of the gain matrix will be zero corresponding to there being no association between the corresponding input spatial channel and output sound render channel.
In some embodiments, the signal paths may also include other compensations or signal processing such as e.g. frequency compensations (e.g. filtering). In particular, the signal paths may include a delay which specifically may compensate for differences in the distance between the sound transducers and the listening position.
As an example, the driver 403 may for each spatial channel provided the functionality illustrated in Fig. 8. Thus, a signal path with a variable gain and delay may be provided to each of the sound render channels. The values of the gain and delay may be determined fully or partially based on the graphical representation. If no association is provided between a spatial channel and a sound render channel, the corresponding gain will be zero. It will be appreciated that the delays will typically be the same for all spatial channels for a given sound render channel, and that a single delay may accordingly be applied to the drive signal generated by combining the individual contributions from the different spatial channels.
The system may accordingly provide a very flexible setup. A user may simply connect (e.g. directly, or using a wireless or network based connections) as many speakers as he would like to use. The speakers may be positioned freely but may preferably be situated to surround the listening position. The user then uses his portable device to quickly and easily generate a graphical representation of the speaker setup, e.g. simply by sliding some icons to appropriate positions on a touch screen. In response, the system automatically adapts to provide an appropriate spatial experience, and may specifically proceed to generate virtual sound sources from positions corresponding to (or approximating) the nominal positions of the received spatial multi-channel signal. The user may freely select how many speakers to use and may position these freely. Increasing the number of speakers used, and providing a more even distribution of the loudspeakers around the listening position, may typically increase the accuracy of the virtual sound source positioning and the resulting spatial experience. However, in some embodiments it may be more attractive to reduce cost and complexity, and therefore the number of loudspeakers may be kept low. Thus, the user has a high degree of flexibility in both how many and where to position loudspeakers, and the same audio system may be used as it can easily adapt to the individual configuration.
In some embodiments, the system may be arranged to generate a sound from a sound transducer in response to a selection of the corresponding icon on the display. For example, when the use touches an icon corresponding to a sound source position on the display, this icon may e.g. change color to indicate that it has been selected. In response to the selection, an audio generator may provide a test drive signal on the corresponding sound render channel resulting in the loudspeaker connected to the output of that sound render channel rendering the corresponding sound.
The test drive signal may typically represent a distinct and detectable sound such as a pure tone or a white noise signal. Thus, when the icon is selected, the user will immediately hear a characteristic and distinct sound coming from the corresponding loudspeaker. He may then easily slide or move the icon to a position corresponding to the position of the loudspeaker radiating the sound. Thus a low complexity and intuitive approach is achieved for the user allowing him to easily identify which speaker is associated with which icon. No predetermined information or correlation between loudspeaker positions and the sound render channels is required, and specifically any output can be used with any loudspeaker.
In some embodiments, the generated test signal, and thus the rendered test audio, may have a volume that is adjusted in response to the distance from the nominal listening position to the position of the icon for the sound transducer position on the display. Specifically, the volume may be increased as the icon is moved further away and decreased when it is moved closer. This may assist the user in positioning the icons to reflect the actual sound positions, and may compensate for differences in the distances and thus in attenuation of the sound from the different loudspeakers.
In some embodiments, the processing of the audio system may be setup based only on the graphical representation. However, in many embodiments, the system may further be arranged to perform an automatic or semi-automatic calibration based on a microphone signal.
Specifically, the system may comprise a microphone input which is coupled to a microphone that can be positioned at the listening position. A calibrator may then receive the microphone signal and adjust a characteristic of the signal processing dependent on the microphone signal. Specifically, a delay, volume and/or sound render direction of the drive signals may be adjusted dependent on the microphone signal.
For example, a test generator may generate a test signal which is fed to one of the sound transducers. The time difference between the generated test signal and the received signal by the microphone may be used to determine the propagation delay from the sound transducer to the listening position. This may be performed for all sound render channels and the appropriate relative delays may be applied to ensure that sound is received substantially simultaneously.
Thus, the approach may combine the approach of calibrating/adapting the system based on the graphical representation with a microphone based calibration/adaptation. An improved performance may be achieved by combining the two approaches. For example, the calibration based on the graphical representation may provide a coarse
adaptation/calibration with some parameters being fine-tuned using the microphone measurement.
Specifically, the sound render positions determined from the graphical representation may be used to initialize the calibration process based on the microphone signal. For example, the graphical representation may be used to determine initial estimates for delay and volume as well as to setup the virtual sound source positioning. The calibration based on the microphone signal may then proceed based on these initial values. Calibration processes may often determine local optimizations and by initializing the calibration with approximate initial values, the chance of the determined local optimum is also the global optimum is also substantially increased.
Furthermore, the graphical representation can be used to determine or adapt to characteristics that are typically not addressed by a microphone calibration. For example, the angle of the sound reaching the listening position may be estimated based on the graphical representation. Further, the virtual sound source positioning may be determined based on the graphical representation with the subsequent microphone calibration being used to calibrate characteristics of the virtual sound source.
In some embodiments, the user input graphical representation may comprise an indication of a characteristic of a listening environment. This additional information may then be used to adapt the signal processing of the driver 403.
For example, the graphical representation may be enhanced to allow different listening positions to be indicated, such that the drive signal generation may compensate for differences in the listening position.
As another example, the user may be able to indicate the presence of walls or other obstacles in the listening environment. This may specifically indicate sound reflections. Thus, the graphical representation can be used to estimate e.g. a direction in which sound should be radiated from a sound transducer in order to reach the listening position from a reflected direction corresponding to a desired virtual sound source direction. For example, a wall to the rear of the listening position may be indicated on the graphical representation. In response, sound may be radiated towards this wall to provide reflected sound reaching the listening position from behind.
It will be appreciated that the above description for clarity has described embodiments of the invention with reference to different functional circuits, units and processors. However, it will be apparent that any suitable distribution of functionality between different functional circuits, units or processors may be used without detracting from the invention. For example, functionality illustrated to be performed by separate processors or controllers may be performed by the same processor or controllers. Hence, references to specific functional units or circuits are only to be seen as references to suitable means for providing the described functionality rather than indicative of a strict logical or physical structure or organization. The invention can be implemented in any suitable form including hardware, software, firmware or any combination of these. The invention may optionally be
implemented at least partly as computer software running on one or more data processors and/or digital signal processors. The elements and components of an embodiment of the invention may be physically, functionally and logically implemented in any suitable way. Indeed the functionality may be implemented in a single unit, in a plurality of units or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units, circuits and processors.
Although the present invention has been described in connection with some embodiments, it is not intended to be limited to the specific form set forth herein. Rather, the scope of the present invention is limited only by the accompanying claims. Additionally, although a feature may appear to be described in connection with particular embodiments, one skilled in the art would recognize that various features of the described embodiments may be combined in accordance with the invention. In the claims, the term comprising does not exclude the presence of other elements or steps.
Furthermore, although individually listed, a plurality of means, elements, circuits or method steps may be implemented by e.g. a single circuit, unit or processor.
Additionally, although individual features may be included in different claims, these may possibly be advantageously combined, and the inclusion in different claims does not imply that a combination of features is not feasible and/or advantageous. Also the inclusion of a feature in one category of claims does not imply a limitation to this category but rather indicates that the feature is equally applicable to other claim categories as appropriate.
Furthermore, the order of features in the claims do not imply any specific order in which the features must be worked and in particular the order of individual steps in a method claim does not imply that the steps must be performed in this order. Rather, the steps may be performed in any suitable order. In addition, singular references do not exclude a plurality. Thus references to "a", "an", "first", "second" etc. do not preclude a plurality. Reference signs in the claims are provided merely as a clarifying example shall not be construed as limiting the scope of the claims in any way.

Claims

CLAIMS:
1. An audio system for generating drive signals for a plurality of sound render channels, each sound render channel providing a drive signal for a sound transducer; the audio system comprising:
a user interface (409) for receiving a user input graphical representation of positions of the sound transducers;
a processor (411) for determining sound render positions for the sound transducers in response to the user input graphical representation;
an input (401) for receiving a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position;
a signal linker (413) for determining a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions, an association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel; and
a driver (403) for generating drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.
2. The audio system of claim 1 wherein the signal linker (413) is arranged to associate a first spatial channel with a first sound render channel if the first sound render channel has a sound render position being a closest sound render position to a nominal position of the first spatial channel.
3. The audio system of claim 1 wherein the set of associations may associate more than one sound render channel with one spatial channel.
4. The audio system of claim 1 wherein the driver (403) is arranged to generate drive signals for at least two sound render channels from an input signal of a first spatial channel to provide a virtual sound source position for the first spatial channel closer to a nominal position of the first spatial channel than sound render positions of the at least two sound render channels.
5. The audio system of claim 1 wherein the set of associations may associate more than one spatial channel with one sound render channel.
6. The audio system of claim 1 wherein the driver (403) is arranged to adjust at least one of an audio level and a delay of at least one of the drive signals in response to the sound render positions.
7. The audio system of claim 1 wherein the user interface (409) comprises:
a display for displaying the user input graphical representation with icons representing positions of the sound transducers;
an input interface for moving the icons on the display in response to a user input.
8. The audio system of claim 7 wherein the input interface is operable to receive a user selection of an icon representing a position of a sound transducer; and wherein the audio system further comprises an audio generator for providing a test drive signal to the sound render channel corresponding to the selected icon.
9. The audio system of claim 8 wherein the audio generator is arranged to adjust a volume of the test drive signal in response to a distance between a nominal listening position and the sound render position of the sound render channel corresponding to the selected icon.
10. The audio system of claim 1 further comprising:
a microphone input for receiving a microphone signal; and
a calibrator for setting at least one of a delay, volume and sound render direction for at least one of the drive signals in response to the microphone signal.
11. The audio system of claim 10 further comprising an initializer for initializing a calibration by the calibrator in response to the sound render positions.
12. The audio system of claim 1 wherein the user interface (409) is part of a portable device operable to be remote from the driver.
13. The audio system of claim 1 wherein the user input graphical representation further comprises an indication of a characteristic of a listening environment, and the driver
(403) is arranged to generate the drive signals in response to the indication.
14. The audio system of claim 1 wherein the user input graphical representation comprises an indication of a characteristic of a listening position, and the driver (403) is arranged to generate the drive signals in response to the listening position.
15. A method of generating drive signals for a plurality of sound render channels, each sound render channel providing a drive signal for a sound transducer; the method comprising:
receiving a user input graphical representation of positions of the sound transducers;
determining sound render positions for the sound transducers in response to the user input graphical representation;
receiving a spatial multichannel signal comprising a plurality of spatial channels, each spatial channel being associated with a nominal sound source position;
determining a set of associations between the spatial channels and the sound render channels in response to the sound render positions and the nominal sound source positions, an association between a spatial channel and a sound render channel indicating that the sound render channel is to be used for rendering the spatial channel; and
generating drive signals for the sound render channels from input signals of the spatial channels in response to the set of associations.
PCT/IB2012/052580 2011-06-01 2012-05-23 An audio system and method of operating therefor WO2012164444A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP11168342.1 2011-06-01
EP11168342 2011-06-01

Publications (1)

Publication Number Publication Date
WO2012164444A1 true WO2012164444A1 (en) 2012-12-06

Family

ID=46246115

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2012/052580 WO2012164444A1 (en) 2011-06-01 2012-05-23 An audio system and method of operating therefor

Country Status (1)

Country Link
WO (1) WO2012164444A1 (en)

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150208187A1 (en) * 2014-01-17 2015-07-23 Sony Corporation Distributed wireless speaker system
US9288597B2 (en) 2014-01-20 2016-03-15 Sony Corporation Distributed wireless speaker system with automatic configuration determination when new speakers are added
EP2996354A1 (en) * 2014-09-11 2016-03-16 Genelec OY Loudspeaker control
EP3024253A1 (en) * 2014-11-21 2016-05-25 Harman Becker Automotive Systems GmbH Audio system and method
US9369801B2 (en) 2014-01-24 2016-06-14 Sony Corporation Wireless speaker system with noise cancelation
US9426551B2 (en) 2014-01-24 2016-08-23 Sony Corporation Distributed wireless speaker system with light show
CN106488363A (en) * 2016-09-29 2017-03-08 Tcl通力电子(惠州)有限公司 Sound channel distribution method and device of audio output system
US9693169B1 (en) 2016-03-16 2017-06-27 Sony Corporation Ultrasonic speaker assembly with ultrasonic room mapping
US9693168B1 (en) 2016-02-08 2017-06-27 Sony Corporation Ultrasonic speaker assembly for audio spatial effect
US9699579B2 (en) 2014-03-06 2017-07-04 Sony Corporation Networked speaker system with follow me
US9756444B2 (en) 2013-03-28 2017-09-05 Dolby Laboratories Licensing Corporation Rendering audio using speakers organized as a mesh of arbitrary N-gons
US9794724B1 (en) 2016-07-20 2017-10-17 Sony Corporation Ultrasonic speaker assembly using variable carrier frequency to establish third dimension sound locating
US9826332B2 (en) 2016-02-09 2017-11-21 Sony Corporation Centralized wireless speaker system
US9826330B2 (en) 2016-03-14 2017-11-21 Sony Corporation Gimbal-mounted linear ultrasonic speaker assembly
US9854362B1 (en) 2016-10-20 2017-12-26 Sony Corporation Networked speaker system with LED-based wireless communication and object detection
US9866986B2 (en) 2014-01-24 2018-01-09 Sony Corporation Audio speaker system with virtual music performance
US9924286B1 (en) 2016-10-20 2018-03-20 Sony Corporation Networked speaker system with LED-based wireless communication and personal identifier
WO2018072214A1 (en) * 2016-10-21 2018-04-26 向裴 Mixed reality audio system
US10075791B2 (en) 2016-10-20 2018-09-11 Sony Corporation Networked speaker system with LED-based wireless communication and room mapping
US10582330B2 (en) 2013-05-16 2020-03-03 Koninklijke Philips N.V. Audio processing apparatus and method therefor
US10623859B1 (en) 2018-10-23 2020-04-14 Sony Corporation Networked speaker system with combined power over Ethernet and audio delivery
CN111886881A (en) * 2018-04-24 2020-11-03 索尼公司 Signal processing device, sound channel setting method, program, and speaker system
CN116634349A (en) * 2023-07-21 2023-08-22 深圳隆苹科技有限公司 Audio output system capable of automatically distributing sound channels and use method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6741273B1 (en) * 1999-08-04 2004-05-25 Mitsubishi Electric Research Laboratories Inc Video camera controlled surround sound
FR2850183A1 (en) * 2003-01-20 2004-07-23 Remy Henri Denis Bruno Acoustic field restoration controlling method, involves determining adaptation filter according to characteristic and directions of fixed restoration associated with entry signals for determining control signal of restoration unit
US6798889B1 (en) * 1999-11-12 2004-09-28 Creative Technology Ltd. Method and apparatus for multi-channel sound system calibration
US20040264704A1 (en) * 2003-06-13 2004-12-30 Camille Huin Graphical user interface for determining speaker spatialization parameters
US20050220309A1 (en) * 2004-03-30 2005-10-06 Mikiko Hirata Sound reproduction apparatus, sound reproduction system, sound reproduction method and control program, and information recording medium for recording the program
WO2006054270A1 (en) * 2004-11-22 2006-05-26 Bang & Olufsen A/S A method and apparatus for multichannel upmixing and downmixing
WO2006131894A2 (en) * 2005-06-09 2006-12-14 Koninklijke Philips Electronics N.V. A method of and system for automatically identifying the functional positions of the loudspeakers of an audio-visual system
WO2008046530A2 (en) * 2006-10-16 2008-04-24 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for multi -channel parameter transformation
WO2010140088A1 (en) * 2009-06-03 2010-12-09 Koninklijke Philips Electronics N.V. Estimation of loudspeaker positions
US20110091055A1 (en) * 2009-10-19 2011-04-21 Broadcom Corporation Loudspeaker localization techniques

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6741273B1 (en) * 1999-08-04 2004-05-25 Mitsubishi Electric Research Laboratories Inc Video camera controlled surround sound
US6798889B1 (en) * 1999-11-12 2004-09-28 Creative Technology Ltd. Method and apparatus for multi-channel sound system calibration
FR2850183A1 (en) * 2003-01-20 2004-07-23 Remy Henri Denis Bruno Acoustic field restoration controlling method, involves determining adaptation filter according to characteristic and directions of fixed restoration associated with entry signals for determining control signal of restoration unit
US20040264704A1 (en) * 2003-06-13 2004-12-30 Camille Huin Graphical user interface for determining speaker spatialization parameters
US20050220309A1 (en) * 2004-03-30 2005-10-06 Mikiko Hirata Sound reproduction apparatus, sound reproduction system, sound reproduction method and control program, and information recording medium for recording the program
WO2006054270A1 (en) * 2004-11-22 2006-05-26 Bang & Olufsen A/S A method and apparatus for multichannel upmixing and downmixing
WO2006131894A2 (en) * 2005-06-09 2006-12-14 Koninklijke Philips Electronics N.V. A method of and system for automatically identifying the functional positions of the loudspeakers of an audio-visual system
WO2008046530A2 (en) * 2006-10-16 2008-04-24 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for multi -channel parameter transformation
WO2010140088A1 (en) * 2009-06-03 2010-12-09 Koninklijke Philips Electronics N.V. Estimation of loudspeaker positions
US20110091055A1 (en) * 2009-10-19 2011-04-21 Broadcom Corporation Loudspeaker localization techniques

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9756444B2 (en) 2013-03-28 2017-09-05 Dolby Laboratories Licensing Corporation Rendering audio using speakers organized as a mesh of arbitrary N-gons
US11503424B2 (en) 2013-05-16 2022-11-15 Koninklijke Philips N.V. Audio processing apparatus and method therefor
US10582330B2 (en) 2013-05-16 2020-03-03 Koninklijke Philips N.V. Audio processing apparatus and method therefor
US11197120B2 (en) 2013-05-16 2021-12-07 Koninklijke Philips N.V. Audio processing apparatus and method therefor
EP2997742B1 (en) * 2013-05-16 2022-09-28 Koninklijke Philips N.V. An audio processing apparatus and method therefor
US11743673B2 (en) 2013-05-16 2023-08-29 Koninklijke Philips N.V. Audio processing apparatus and method therefor
US20150208187A1 (en) * 2014-01-17 2015-07-23 Sony Corporation Distributed wireless speaker system
US9560449B2 (en) * 2014-01-17 2017-01-31 Sony Corporation Distributed wireless speaker system
US9288597B2 (en) 2014-01-20 2016-03-15 Sony Corporation Distributed wireless speaker system with automatic configuration determination when new speakers are added
US9866986B2 (en) 2014-01-24 2018-01-09 Sony Corporation Audio speaker system with virtual music performance
US9369801B2 (en) 2014-01-24 2016-06-14 Sony Corporation Wireless speaker system with noise cancelation
US9426551B2 (en) 2014-01-24 2016-08-23 Sony Corporation Distributed wireless speaker system with light show
US9699579B2 (en) 2014-03-06 2017-07-04 Sony Corporation Networked speaker system with follow me
CN105430576B (en) * 2014-09-11 2019-06-04 珍尼雷克公司 Device and method for loudspeaker control
JP2016059047A (en) * 2014-09-11 2016-04-21 ゲネレク オーワイ Control for loudspeaker
CN105430576A (en) * 2014-09-11 2016-03-23 珍尼雷克公司 Loudspeaker control
EP2996354A1 (en) * 2014-09-11 2016-03-16 Genelec OY Loudspeaker control
US9686626B2 (en) 2014-11-21 2017-06-20 Harman Becker Automotive Systems Gmbh Audio system and method
CN105635906A (en) * 2014-11-21 2016-06-01 哈曼贝克自动系统股份有限公司 Audio system and method
US20160150346A1 (en) * 2014-11-21 2016-05-26 Harman Becker Automotive Systems Gmbh Audio system and method
EP3024253A1 (en) * 2014-11-21 2016-05-25 Harman Becker Automotive Systems GmbH Audio system and method
US9693168B1 (en) 2016-02-08 2017-06-27 Sony Corporation Ultrasonic speaker assembly for audio spatial effect
US9826332B2 (en) 2016-02-09 2017-11-21 Sony Corporation Centralized wireless speaker system
US9826330B2 (en) 2016-03-14 2017-11-21 Sony Corporation Gimbal-mounted linear ultrasonic speaker assembly
US9693169B1 (en) 2016-03-16 2017-06-27 Sony Corporation Ultrasonic speaker assembly with ultrasonic room mapping
US9794724B1 (en) 2016-07-20 2017-10-17 Sony Corporation Ultrasonic speaker assembly using variable carrier frequency to establish third dimension sound locating
CN106488363A (en) * 2016-09-29 2017-03-08 Tcl通力电子(惠州)有限公司 Sound channel distribution method and device of audio output system
CN106488363B (en) * 2016-09-29 2020-09-22 Tcl通力电子(惠州)有限公司 Sound channel distribution method and device of audio output system
US10075791B2 (en) 2016-10-20 2018-09-11 Sony Corporation Networked speaker system with LED-based wireless communication and room mapping
US9924286B1 (en) 2016-10-20 2018-03-20 Sony Corporation Networked speaker system with LED-based wireless communication and personal identifier
US9854362B1 (en) 2016-10-20 2017-12-26 Sony Corporation Networked speaker system with LED-based wireless communication and object detection
WO2018072214A1 (en) * 2016-10-21 2018-04-26 向裴 Mixed reality audio system
CN111886881B (en) * 2018-04-24 2022-11-25 索尼公司 Signal processing device, sound channel setting method, recording medium, and speaker system
US11356789B2 (en) 2018-04-24 2022-06-07 Sony Corporation Signal processing device, channel setting method, and speaker system
EP3787318A4 (en) * 2018-04-24 2021-06-30 Sony Corporation Signal processing device, channel setting method, program and speaker system
CN111886881A (en) * 2018-04-24 2020-11-03 索尼公司 Signal processing device, sound channel setting method, program, and speaker system
US10623859B1 (en) 2018-10-23 2020-04-14 Sony Corporation Networked speaker system with combined power over Ethernet and audio delivery
CN116634349A (en) * 2023-07-21 2023-08-22 深圳隆苹科技有限公司 Audio output system capable of automatically distributing sound channels and use method
CN116634349B (en) * 2023-07-21 2023-11-10 深圳隆苹科技有限公司 Audio output system capable of automatically distributing sound channels and use method

Similar Documents

Publication Publication Date Title
WO2012164444A1 (en) An audio system and method of operating therefor
US20200366994A1 (en) Automatic discovery and localization of speaker locations in surround sound systems
KR101925708B1 (en) Distributed wireless speaker system
JP5992409B2 (en) System and method for sound reproduction
EP3092824B1 (en) Calibration of virtual height speakers using programmable portable devices
JP6258587B2 (en) Speaker system and operation method thereof
JP4449998B2 (en) Array speaker device
US9955262B2 (en) Device and method for driving a sound system and sound system
JP2016509429A (en) Audio apparatus and method therefor
AU2001239516B2 (en) System and method for optimization of three-dimensional audio
JP2013535894A5 (en)
AU2001239516A1 (en) System and method for optimization of three-dimensional audio
JPH09233593A (en) Method for controlling performance of speaker
WO2007135581A2 (en) A device for and a method of processing audio data
US7676049B2 (en) Reconfigurable audio-video surround sound receiver (AVR) and method
JP4450764B2 (en) Speaker device
US20070104331A1 (en) Multi-channel audio system and method for generating virtual speaker sound
WO2013064943A1 (en) Spatial sound rendering system and method
US20200084537A1 (en) Automatically movable speaker to track listener or optimize sound performance
JP2011188287A (en) Audiovisual apparatus
CN109716795B (en) Networked microphone device, method thereof and media playback system
WO2024073401A2 (en) Home theatre audio playback with multichannel satellite playback devices
RU2575883C2 (en) Acoustic system and operation method thereof
KR20220120587A (en) System and method for virtual sound effect using invisible speaker

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12726867

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12726867

Country of ref document: EP

Kind code of ref document: A1