B&O Tech: Active Room Compensation – Some Details, Part 1

#40 in a series of articles about the technology behind Bang & Olufsen loudspeakers

 

Once-upon-a-time, I wrote a posting explaining why a loudspeaker’s biggest enemy is your listening room. The basic problem is as follows:

  • when you listen to a loudspeaker in a room, roughly 99% or more of what you hear is sound that goes outwards from the loudspeaker, fills up the room, bounces around and then gets to you. Only about 1% or less of what you hear actually comes directly to you from the loudspeaker…
  • rooms have early reflections that mess up the tone colour and the imaging of the loudspeakers
  • rooms have room modes that make the bass response extremely uneven, and very different from place to place in the room
  • rooms have reverberation that, generally, makes some frequency bands sound louder because they last longer in time

So, to borrow a phrase from South Park’s Mr. Mackey, “rooms are bad, m’kay?

For this posting, we’re going to focus only on the third issue there: that of “room modes” and a couple of ways to deal with them. To begin with, let’s talk about what a room mode is.

The short version

The PowerPoint-single-slide-of-bullet-points version of an explanation of room compensation is as follows:

  • A room mode is a resonant frequency where the room “sings along” with the sound of the loudspeaker, making that pitch (or note) sound louder than others.
  • This resonance can be measured using a microphone
  • To reduce the effect of this problem, the loudspeaker can either
    • reduce the amount of energy it emits at that particular frequency
    • actively absorb energy at that frequency
  • This means that one detrimental aspect of the room’s acoustical behaviour on the sound of the loudspeaker will be reduced

 

That’s it. However, of course, the real story is a little more nuanced (or maybe just more complicated) than that. So, if you’re interested in knowing a little more, read on!

 

 

Resonance

If you put a kid on a swing and push him (not too hard… we don’t want anyone to get hurt here…) and stand back, you’ll watch him swing back and forth, decreasing a little in distance from the resting position each time, until eventually, he’ll come to a stop. Now, do it again, and pay attention to his movement. He moves forwards, away from the resting position until he reaches a high point, then stops, comes backwards past the resting position to a high point, stops, and then moves forwards again. We’ll call that total return trip from going forwards at the resting position to being back again, going forwards at the resting position, one cycle. The number of times the kid does this for each period of time (let’s say, in this case, each minute) is called the frequency – how frequently he’s repeating the motion. The maximum distance he moves from the resting position (to the highest point in the cycle) is called the amplitude.

Now, if you pay really close attention to the movement of the kid on the swing, you might notice that, even though the amplitude of the cycle decreases over time, the frequency doesn’t change.

Another example of this kind of motion is shown in the video in this posting. However, that video is not for the faint-of-heart. Watch it at your own risk…

Okay, if you watched that video and you’re still awake, let’s move on.

There are a couple of important concepts to glean from this discussion so far.

  • If you have a system that can resonate (like a spring and a mass, or a kid on a swing), it will “want to” resonate at a fundamental frequency.
  • You can trigger that system to resonate at its fundamental frequency by injecting energy into the system (like lifting the mass, or pushing the kid)

Another example of a system that “wants” to resonate is a string that’s fixed on both ends – like a guitar string. It has a fundamental frequency that is determined by the string’s mass and tension. The cool thing about a string is that it also resonates at multiples of that fundamental frequency (better known as harmonics or overtones – which almost mean the same thing, but not quite – but the difference is irrelevant here). So, if you have a guitar string that’s tuned to 100 Hz (an abbreviation for “Hertz” which is the word we use to mean “cycles per second”) then it will also resonate at harmonics with frequencies of 200 Hz, 300 Hz, 400 Hz, and so on. If you inject energy into the string (by plucking it with your finger, for example), then the string’s total vibration will include resonances at all those frequencies, on up to ∞ Hz (in theory… not in practice…). You’ll have a resonance at each individual harmonic, each with its own amplitude (how much the string is vibrating or moving away from its resting position – the higher the amplitude, the louder it is) and the total sum of all of these results in the shape (and sound) of the string.

To see this in action, check out this video.

At this point, you’re probably wondering “what does this have to do with my room?” We’re getting there… I promise.

One last example of a system that resonates is the air in a pipe (like an organ pipe, for example). If you could shrink yourself down to the size of a molecule and get inside an organ pipe, you’d see that you’re looking down a long tube that’s capped at both ends (we won’t talk about the other kind of pipe that’s open on one end… let’s pretend those don’t exist…). If you face one end of the pipe and push the air molecule next to you towards it, it will push the one in front of it, which will push the one in front of it, and so on, until the shoving match ends at the cap at the end of the pipe. (See this page for a good animation of the story so far…) That last molecule can’t push the “wall” so it bounces back, which winds up in a return shoving match (or pressure wave…) that will eventually push against you, and you’ll push the molecule behind you, which keeps repeating until the wavefront gets to the cap at the opposite end of the pipe, which reflects, and sends wavefront back to you again.

Now, what happens if, while all of that is happening, you are pushing repeatedly? Every time the wave bounces back at you, you push it again in the same direction that the particles want to move in (this is exactly the same as pushing a kid on a swing at exactly the right time to make him swing higher and higher – you time the push so that he’s moving in the same direction as you’re pushing).

Take a look at the top part of the animation below.

Temp Caption

Let’s say that you’re the red molecule and you’re in a pipe. You push and pull the two adjacent particles (one in front, one behind) exactly in sync with the wave that’s reflecting off the two ends of the pipe, and you’re helping to inject energy into the fundamental resonance of the air column inside the pipe. The result is something called a “standing wave”. it’s called this because it looks like the wave is “standing still” and just changing in instantaneous amplitude over time, but the reality is that this is just an illusion. What is actually happening is that the right-going wave is aligned with the left-going wave to perfectly cancel each other out at all times in the middle of the pipe, and to result in big changes in pressure at the ends of the pipe. Check out the third animation on this page for a good explanation of this summing or “superposition” of waves resulting in a standing wave.

Just like a string, a pipe will also resonate at harmonics of the fundamental, in multiples of the frequency of the lowest resonance. So, if the pipe resonates at 150 Hz, then you will also have resonances at 300 Hz, 450 Hz, and so on… I’ll show some animations with examples of these later in the posting.

Room Modes

The nice thing (at least it’s nice in that it helps our understanding – it’s not actually nice in real life…) is that a room behaves just like a very big, very wide, organ pipe. So, for example, when I sit in a listening room that is 5 m wide by 6 m long, some of the resonances that I hear “singing along” with every sound source in the room are exactly the same frequencies as the ones that would come out of organ pipes 5 m long and 6 m long.

Take a look at that animation above once again. You can think of the red dot as the loudspeaker, pushing and pulling the adjacent air particles in time. You sit at the listening position at the black dot – the adjacent air particles push and pull your eardrum in and out of your head. The pressure of the sound wave (the only thing we’re going to worry about in this posting) can be thought of as a measure of how close two adjacent air particles are (shown by the three vertical lines in the top part, and represented by the entire bottom part of the animation).

So, as the loudspeaker woofer (for example) moves in and out of the enclosure on a kick drum hit, it injects energy into the room at many different frequencies. If the frequency of the resonance in the room is one of the frequencies in the woofer’s signal, then that note will sound much louder than the other notes, making for a very uneven bass. This is because the room “wants” to resonate (or “sing along”) at that particular frequency, so a little energy coming into it will give a large result (just like a series of small, but well-timed, pushes of a kid on a swing can build up over time to result in the kid moving back and forth much more than you’re pushing).

This will happen not only at the fundamental frequency of the resonance, but its harmonics. The second harmonic is shown below.

animation_mode_02

The third harmonic looks like the animation below.

animation_mode_03

And the fourth harmonic looks like the animation below.

animation_mode_04

We won’t go any higher than this.

There are some things to notice in these animations.

  1. The air particles at the walls on either end of the room experience the maximum pressure change. This is because one of the particles cannot push the wall, so it receives the full effect of the pressure.
  2. There are places in the room where the pressure does not change over time. For example, in the case of Mode #1 in the first animation, the point at the exact middle of the room doesn’t have a change in pressure. This doesn’t mean that the air particles are not moving – they are. However, the ones in the middle of the room are staying the same distance from each other – so the pressure doesn’t change over time.

If you are sitting at a point in the room where there is no change in pressure caused by the mode (say, for example, you’re the black dot in Mode #2, above), then you will not be able to hear the room mode. It’s still happening around you – but at your exact position, since there is no change in pressure over time, it doesn’t exist. However, this doesn’t mean that you won’t hear anything at that particular frequency – you will – it’s just that the sound you hear (from the loudspeaker) doesn’t have the room mode singing along with it at your location.

Similarly, if the loudspeaker is placed at a location where there is no change in pressure when the room mode is ringing (for example, the location of the red circle in Mode #3), then the loudspeaker will not trigger this mode to ring. Energy will come out of the loudspeaker at that frequency, and you’ll be able to hear it – but the room mode will not “sing along” with the signal, since the loudspeaker can’t inject energy into it. (Note that this is statement is based on at least two assumptions: (1) that the loudspeaker is a pressure source, and (2) that it is infinitely small. At least one of these assumptions is incorrect.)

This means that, looking at the four animations above, and assuming that the red dot is the loudspeaker and the black dot is the listener:

  • We need to fix Mode #1, since the loudspeaker and the listener are “coupled” to it (meaning that they are not sitting on a point where there is no pressure change in the mode).
  • We do not need to fix Mode #2 since, although the loudspeaker is triggering the mode to ring, the listener cannot hear it. (However, if the loudspeaker was placed somewhere else, this mode could be a problem.)
  • We do not need to fix Mode #3 since the loudspeaker cannot trigger the mode, so no one hears it (although the mode will occur for other sound sources in the room).  (Also, if the listening position was placed somewhere else, this mode could be a problem.)
  • We need to fix Mode #4.
A simulated example of the total result of two modes, the fundamental at 100 Hz, and the 4th harmonic at 400 Hz (see the above text to explain why 200 Hz and 300 Hz are missing...)
Figure 5: A simulated, over-simplified example of the total effect on the magnitude response of two modes, the fundamental at 100 Hz, and the 4th harmonic at 400 Hz (see the above text to explain why 200 Hz and 300 Hz are missing…)

How can we fix this?

As I said in the PowerPoint version above, we have two basic strategies for dealing with room modes.

The first is a “symptom attenuation” approach: the room mode makes some frequencies louder than others, so we’ll just reduce the level of those frequencies by the same amount. If the room boosts 100 Hz by 6 dB, then we’ll put a dip of -6 dB at 100 Hz in the signal sent to the loudspeaker. This doesn’t actually correct the problem – it just covers it up. It’s a bit like taking a pain killer because you’ve broken your leg. You are no longer in pain, but the problem still exists… The trick here is to measure the response of the loudspeaker at the listening position (this is easy) and find out which frequencies appear to be boosted by the room. We then make an equalisation which is that same response turned “upside down” so a peak in the measured response becomes a dip in the loudspeaker’s processing.

As I said, this covers up the problem but it doesn’t solve it. The thing to remember is that a room mode appears to be louder because the room moves the energy in time. For example, if you pluck a guitar string, you “activate it” for a very short period of time. The resonance of the string takes that energy and extends it in time so you hear a note that goes on for seconds after the pluck. A room mode does the same thing – it “rings” in time, smearing the energy coming out of your loudspeakers (say, a kick drum hit) and extending it like a guitar string ringing. If you send less energy out of the loudspeaker at the same frequency of the room mode, then the total result at the listening position will be quieter, but it’s still ringing in time. (So, instead of plucking the guitar string and then stopping it, we’re just plucking with less force.)

The second method is a little more complicated, but works better. Since we have control of the loudspeaker, what we can do is to send the sound we want out of it – but we’ll add an extra signal that cancels the room mode by working against it. If we go back to the original description where we were a little air particle in a pipe, we send the signal out, and when it comes back, reflected off the cap, we’ll send a signal out of the loudspeaker that “sucks it back in”. This way, the loudspeaker acts as an active absorber. (This is a little like the way a person on a trampoline can stop jumping by using his/her legs to absorb the push-back from the springs. See this video at 0:53, for example.)

The great thing is that, if you’re smart and a little lucky, both of these approaches are the same thing. If we consider the room mode to have the same characteristics as a “minimum phase” peaking filter (which, in isolation, it does), then, if we can measure it correctly, we can implement its exact reciprocal. This means that the magnitude response of the filter we insert into the loudspeaker’s processing will be the exact opposite of the room’s response. In addition to this, the phase response of the filter is, at any frequency, opposite in polarity to the phase response of the room mode (in isolation). This means that both approaches are rolled into one solution.

The same two modes shown in the previous figure, decomposed into the two individual resonances.
Figure 6: The same two modes shown in Figure 5, decomposed into the two individual resonances.

 

The correction filter we would need to cancel the effects of the two modes shown in the previous figure. Note that the phase responses are polarity-inverted copies of the modes' phase responses.
Figure 7: The correction filter we would need to cancel the effects of the two modes shown in Figure 6. Note that the phase responses are polarity-inverted copies of the modes’ phase responses. If we take the signals shown in Figure 6 and process them with a filter with the characteristics shown in Figure 7, the result would be a perfectly flat signal, as if nothing had happened to it.

 

Of course, this is very theoretical. However, in practice the concept holds true enough that it results in a noticeable improvement in the sound of a loudspeaker in the listening position, assuming that you are careful about the measurements you make around that position. If you choose to make a room correction filter that applies to the whole room, and not just one listening position, then you should measure the room in many different and varied locations. This will give you a “general” compensation that is better, on average, for everyone in the room – although it might make one specific location (like, say, the sweet spot) worse. If, instead, you just measure one location by placing the microphone in a tight area around the sweet spot, for example, then the compensation filter will be optimised for that location, but it will probably make the rest of the room worse as a result.

Active Room Compensation gives you this option to customise the room compensation filters as they best suit your needs. If you’re the kind of person with one chair and no friends, then you measure around that chair and you never leave it while you’re listening. If you’re the kind of person who has lots of friends, or who moves around while you listen, then you create a compensation filter that fixes the entire room, on average.

Of course, in a perfect world, you would be able to make both of these measurements in advance and save each one as a preset – then you would be able to select the one that best suits the situation when the time arises… Or, you could combine them and build a filter that corrects the room, with emphasis on the sweet spot…

P.S. As I mentioned briefly in this posting, Active Room Compensation has one additional feature – multichannel processing. That will be explained later in Part 2.

B&O Tech: The Heat of the Moment

#39 in a series of articles about the technology behind Bang & Olufsen loudspeakers

 

“If you saw a heat wave, would you wave back?”

– Steven Wright

 

Part 1: The Myth(?)

Once upon a time, back in the “olden days” (as my kids call them), before I worked at Bang & Olufsen, I did a Master’s degree at McGill University as a tonmeister (hence the domain name…). In practice, this means that I was trained to be someone who knew as much about things like room acoustics, microphone techniques, recording practices and electronics as I did about music. (Of course, if I knew nothing about music, then this equation falls apart…) In practice, what it meant was that I spent a LOT of time in recording studios (literally more time than I remember) doing recordings or just listening to them.

Now, let’s think about the design of a typical “good” recording studio control room. Usually, you have a mixing console in the middle of the room, and at least one pair of small studio monitor loudspeakers sitting on the meter bridge. Also, up on the front wall, there’s a large pair of “soffit mounted” loudspeakers built into it. Figure 1, below, shows an example of this.

 

Figure 1. A typical recording studio. Notice that there are a pair of small loudspeakers (usually called "nearfield monitors") sitting above the mixing console, and a pair of large "soffit-mounted" loudspeakers on the walls.
Figure 1. A typical recording studio. Notice that there are a pair of small loudspeakers (usually called “nearfield monitors”) sitting above the mixing console, and a pair of large “soffit-mounted” loudspeakers on the walls.

 

One story I heard back in the “olden days” was that, if you’re building a recording studio, you should always keep the amplifiers for your big soffit-mounted loudspeakers away from (or at least above) the loudspeakers themselves. You should never NEVER put the amps directly under the loudspeakers. The story went that the reason for this was that hot air rising from the amplifiers would waft past the tweeters, modulating (changing) the speed of sound of the air between the tweeters and the listening position. And, since this modulation was random, then this heat would sometimes cause the right channel to arrive at the listening position earlier than the left, sometimes later. The result (according to the story) would be a constant shifting of the phantom image, moving left and right instead of staying exactly in the middle where it should be. The analogy/metaphor/explanation was that the image location would “shimmer” in the same way that there appears to be water on hot asphalt in the distance on a summer day, or the way you can see heat rising from a hot surface like a barbecue.

 

Fast-forward about 25 years, and I’m sitting in a meeting at B&O, discussing mechanical design options for high-end loudspeakers. Someone in the room says “one possibility is to put the radiators for the amplifiers on the face of the loudspeaker, just under the midranges and tweeters” and I say “you can’t do that – you’ll screw up the phantom imaging”. Heads turn to look at me, the expressions on the faces obviously say “what on earth are you talking about?” So, I tell the story that I heard back in the olden days…

Quick sidebar: if you don’t already know, almost all Bang & Olufsen loudspeakers are active loudspeakers, which means that the amplifiers and power supply are inside the loudspeaker enclosure. Amplifiers and power supplies are never 100% efficient. In addition to this, because we use DSP to control the loudspeaker, we can push our loudspeaker drivers to their limits without melting their voice coils – this also results in a lot of heat. All of this means that our loudspeakers need a lot of thought into their thermal design – basically, we need to make sure that the heat can get out, which is why we need radiators like the one shown in the photo on the right (which shows the bottom of the back of a BeoLab 20. The vertical stripes are a metal radiator attached to the electronics inside.). We now return you to your regularly scheduled story…

Some of the engineers in the room say “that’s crazy” – and others (the ones I like) say “let’s find out…” So, using the modern tools for finding-things-out, we started by typing things like “loudspeaker amplifier heat phantom image” into Google. This resulted in no information – or at least no useful information. So this meant that we had to do an experiment.

Part 2: The pilot experiment

Okay, so the question is: does heat rising in front of a tweeter have an effect on the sound that it radiates? And, if so, what is/are the effect(s) and what kind of magnitude of effect are we talking about? The easiest way to find this out is to exaggerate the effect. So, we got a tweeter and a hot plate (which lives in the acoustic department exclusively for warming risalamande at the yearly Julefrokost (the ubiquitous Danish “Christmas lunch” literally translated, although it can happen as early as November, it rarely starts at lunchtime, and typically extends well into the evening…) and we set them up as shown in Figure 2.

 

Figure 2: The setup for the first experiment showing the tweeter and the hot plate on the Crane in the Cube.
Figure 2: The setup for the first experiment showing the tweeter and the hot plate on the Crane in the Cube. The BeoLab 6000 in the background was not part of the experiment.

 

We put a microphone in front of the tweeter at a distance far enough away that the microphone wasn’t getting heated by the hot plate (about a metre or so…), and we sent a high frequency tone (20 kHz) to the tweeter from a sine wave generator. Then we simply sent the original signal and the output of the microphone to an oscilloscope so that we could look at them in real time.

 

Figure 3. The original signal and the output of the tweeter at the microphone location when the hot plate was turned off.
Figure 3. The original signal and the output of the tweeter at the microphone location when the hot plate was turned off. (Note: Exposure time: 30 seconds)

 

Figure 3 shows the result when the hot plate was turned off. The original signal and the microphone signal are both clean, as you can see there.

 

Figure 4. The original signal and the output of the tweeter at the microphone location when the hot plate was set to "medium".
Figure 4. The original signal and the output of the tweeter at the microphone location when the hot plate was set to “medium”. (Note: Exposure time: 30 seconds)

 

Figure 4 shows the same thing, but the hot plate is now set to “medium”. As you can see there, the output from the tweeter is much more “blurry”. This is because it is moving on the screen over the 30 seconds of the photo exposure. Note that it is moving in both the X-axis (meaning the level was changing over time) and the Y-axis (meaning that the time of travel from the loudspeaker to the microphone was changing over time).

 

Figure 5. The original signal and the output of the tweeter at the microphone location when the hot plate was at maximum.
Figure 5. The original signal and the output of the tweeter at the microphone location when the hot plate was at maximum. (Note: Exposure time: 30 seconds)

 

Figure 5 shows the same thing, but the hot plate is now set to “maximum” (hot enough to boil water or to burn risalamande). As you can see there, the output from the tweeter is even more “blurry”for the same reason. It’s also fairly easy to see in that photo that the output from the tweeter is modulating in both axes.

One thing to be clear of here is that we are looking at the effect of modulating the temperature of the air in front of the tweeter. We’re not changing the temperature of the tweeter itself (at least, not with the hot plate – the tweeter heats itself, but that’s another story) or the microphone.

 

Part 3: The “real” experiment

Okay, so we had some proof that the story that I heard back in the recording studio might be true. At the very least, we found out that it would be easy to measure the effect of heat rising in front of a tweeter. Now we had to get a little more useful data by analysing the signal from the microphone a little better. Looking at the signal on an oscilloscope is good for intuition, but it’s hard to graph the results.

So, this time we changed the setup a little. We used a little studio monitor loudspeaker (with some minor alterations – we taped over the ports to ensure that we weren’t modulating the air around the loudspeaker as a result of airflow through the port, just in case… for the lower frequencies… Since we were only playing sine tones and we’re not really interested in the absolute magnitude response of the loudspeaker, the lack of a port didn’t matter too much to us…) and we recorded the microphone output instead of just looking at the signal on a ‘scope. (We were also careful not to move during the recordings, partly to reduce background noise, but also to avoid creating air currents … just in case…) The total setup (although it’s still pretty simple) is shown in Figure 6.

 

Figure 6. A small studio monitor with sealed ports and the hotplate with a microphone.
Figure 6. A small studio monitor with sealed ports and the hotplate with a microphone. The gray panels are an acoustically absorptive foam called Basotect. We use that a lot inside loudspeaker cabinets, so we have plenty lying around…

We also tried measuring the air temperature in front of the loudspeaker, but this was difficult because it was fluctuating so much and so rapidly. So, instead, we put a power meter on the mains feed to the hot plate to get a better idea of how much constant power we were using.

Figure 7. The power monitor on the hotplate.
Figure 7. The power monitor on the hotplate.

 

 

The results

Our analysis was to see how much the phase and the magnitude were changing over time with different frequencies and different settings of the hot plate temperature. This was done by analysing the recordings of the microphone signal for a 30-second period. The results turned out to be different for different frequencies – and surprisingly large in some cases…

Some details for the geeks-amongst-you:

  • The recordings were done with a 192 kHz sampling rate so that we had plenty of bandwidth to play with for the next step.
  • We took each 30-second recording and band-limited it to a frequency band 2 octaves wide using second-order high pass and low pass filters, with the tone being the centre frequency (geometrically). (In other words, for example, the 16 kHz recording was band-limited from 8 kHz to 32 kHz). The reason for this was to reduce the influence of background noise in the recordings for the level analysis.
  • The phase analysis was done based on the output of a single FFT bin which provided its own noise immunity.
  • The recordings were done at 5 power consumption settings of the hot plate: 0 W, 75 W, 150 W, 300 W and 700 W. These roughly corresponded to air temperatures in front of the tweeter 25º C, 28º C, 30º C, 34º C, and 44º C. I say “roughly” because the air temperature in front of the tweeter was varying quickly and widely, so this is just a ballpark measurement to get a rough idea of the effect of the hot plate. Due to this uncertainty, the plots below are relative to the power consumption of the hot plate rather than the air temperature.

 

Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 250 Hz tone.
Figure 8a: Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 250 Hz tone.

 

Change in phase of the signal at the microphone over time for different hot plate temperatures. 250 Hz tone.
Figure 8b: Change in phase of the signal at the microphone over time for different hot plate temperatures. 250 Hz tone.

As you can see above in Figures 8a and 8b, there was almost no effect on a 250 Hz tone, even when we pushed the hot plate up to 700 Watts. There is a very small change in magnitude, however this might be caused by background noise.

Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 1 kHz tone.
Figure 9a: Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 1 kHz tone.
Change in phase of the signal at the microphone over time for different hot plate temperatures. 1 kHz tone.
Figure 9b: Change in phase of the signal at the microphone over time for different hot plate temperatures. 1 kHz tone.

As you can see above in Figures 9a and 9b, at 1 kHz the effect is not getting much more noticeable.

Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 4 kHz tone.
Figure 10a: Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 4 kHz tone.
Change in phase of the signal at the microphone over time for different hot plate temperatures. 4 kHz tone.
Figure 10b: Change in phase of the signal at the microphone over time for different hot plate temperatures. 4 kHz tone.

Figures 10a and 10b, show the modulation for 4 kHz. Now we’re starting to see some effects…

Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 16 kHz tone.
Figure 11a: Change in magnitude of the signal at the microphone over time for different hot plate temperatures. 16 kHz tone.

 

Change in phase of the signal at the microphone over time for different hot plate temperatures. 16 kHz tone.
Figure 11b: Change in phase of the signal at the microphone over time for different hot plate temperatures. 16 kHz tone.

Figures 11a and 11b, show the modulation for 16 kHz . You can see there that, in a worst-case scenario, we’re getting magnitude changes on the order of +/- 3 dB or so (which would mean, in a stereo system, a worst-case difference between the two loudspeakers’ signals of about 6 dB), and phase changes on the order of about +/- 40 degrees (so a worst-case difference between the two loudspeakers’ signals of about 80 degrees).

If we then look at the peak-to-peak variation in the above graphs and plot that as a function of hot plate power consumption for the different frequencies, we get the results shown in Figures 12 and 13.

Total result of the relationship between the peak-to-peak change in magnitude vs hot plate power consumption for different frequencies.
Figure 12: Total result of the relationship between the peak-to-peak change in magnitude vs hot plate power consumption for different frequencies.

 

Total result of the relationship between the peak-to-peak change in phase vs hot plate power consumption for different frequencies.
Figure 13: Total result of the relationship between the peak-to-peak change in phase vs hot plate power consumption for different frequencies.

 

So, the conclusion from the measurements was that the effect was considerable at high frequencies and high temperatures, but might be significant at lower values as well… the question was to ask “what is significant?”

Luckily, for the phase modulation part of the question, we had a partial answer from an experiment that we did a long time ago. In a paper called “Audibility of Phase Response Differences in a Stereo Playback System. Part 2: Narrow-Band Stimuli in Headphones and Loudspeakers” (Presented at the 125th AES Convention,  Paper Number ) , Sylvain Choisel and I reported that subjects were able to detect phase response differences in loudspeakers with a threshold of approximately 50 degrees phase difference, regardless of frequency – although we only tested up to 8 kHz… Apart from the 16 kHz measurements in this experiment, none of the frequencies we measured came close to a worst-case difference of 50 degrees – so we’re probably safe on that one. (Although geeks might be interested to see the below for a little more discussion about this issue looking at it from a perspective of time instead of phase.)

The magnitude response differences may be a similar story. Many studies have been done to find out what differences in level between loudspeakers result in changes in phantom image placement – for example, see this site for a good summary. As can be seen in that graph, a 1.5 dB difference (chosen because that’s the worse-case value for the 1 kHz curve in the plot in Figure 12) in level between the two loudspeakers (measured at the listening position) will result in something like a 15% to 20% shift in phantom image location away from centre (that’s 15% to 20% of the distance to one of the loudspeakers from centre) which will probably be noticeable if you’re paying attention, although it might sound more like a “fuzzy” centre image instead of a moving one… A 6 dB difference (roughly the worst-case value for the 16 kHz curve) can result in an approximately 60%  shift in phantom image location. Note, however that we might be jumping to (incorrect) conclusions here, since some of the data in the plots on that PDF file I linked to above are from listening tests with full-range signals. Whether the values will be valid for narrow-band signals, especially at 16 kHz is unknown…

 

Part 4: Phase vs. Time

Take a look at the side view of our setup, shown below in Figure 14.

 

Figure X: Side view with a simplified view of what is happening.
Figure 14: Side view with a simplified view of what is happening.

 

This is a simplified version of what is happening in the experiment. The sound wave travels outwards from the tweeter, through the air. However, that air has one area where the temperature is significantly different than the rest of the room.

The speed of sound in air can be calculated as follows:

c = 331.3 + (0.6 * T) m/s

where T is the air temperature in degrees Celsius.

So, for example, if the temperature of the air increases by 10ºC, then the speed of sound increases by 6 m/s. If the temperature increases by 50ºC, then the speed of sound increases by 30 m/s. This will change the time of arrival of the sound at the microphone – the hotter the air, the earlier the sound arrives.

However, this change is a change in time. This converts to a change in phase in the sine wave that we were measuring – but you have to remember that a given change in time is equivalent to a change in phase that increases with frequency. For example, a 1 ms delay change is equivalent to a 45º phase shift at 125 Hz, but a 90º phase shift at 250 Hz, a 180º phase shift at 500 Hz, and a 360º phase shift at 1 kHz, and so on…

So, another way to look at this problem is to consider the change in interchannel delay instead of interchannel phase. If you take a look at this site again, you can see that our perception of the phantom image location in a stereo system is very sensitive to changes in interchannel delay. For example, according to Simonsen, an interchannel delay difference of only 200 µs (that’s 0.0002 seconds!) will result in a 30% change in phantom image location. In other words, if you have a “correct” stereo configuration where your loudspeakers are placed +/- 30º from centre front, then delaying one loudspeaker by 200 µs will move the centre phantom image to 9º off-centre (or 30% of the way to the earlier loudspeaker).

Let’s simplify, just to get an idea of what we’re talking about in our experimental results:

Case 1: If the microphone is 3 m from the tweeter and the air temperature is 25º C for all 3 m of travel, then the speed of sound for the entire trip is 346.3000 m/s and the total travel time is therefore 8.6630 ms.

Case 2: If the microphone is 3 m from the tweeter and the air temperature for the first 50 cm of travel is 44º C and the remaining 2.5 m of travel is 25º C, then the speed of sound for the first 50 cm is 357.7 m/s and for the remaining 2.5 m it’s 346.3 m/s. This means you have a total travel time of 8.6170 ms.

The difference between Case 1 and Case 2 is 8.6630 – 8.6170 = 0.046 ms or 46 µs, which, according to Simonsen’s results would correspond to something like a 5% shift in centre image location. Audible? Perhaps, but just barely… However, if you’re checking my math, you’ll note that a 46 µs change in time is equivalent to a 264º change in phase at 16 kHz, which was much larger than the roughly 85º we measured in the analysis above (see Figure 13)… So, my simplified example here seems to be exaggerating the effect, which means that a guess (even a calculated guess) of a 46 µs interchannel latency difference is probably an overestimate…

So, we might be able to conclude that the effect of this with respect to change in time of arrival are also inaudible, according to the data…

 

Part 5: Why is the level changing over time?

You may notice that I’ve avoided an obvious question throughout the above: “Why does the magnitude of the signal change?” If you look at this site, for example, you’ll see that although there is a change in sound absorption with change in air temperature, it is a very small effect. So, why do we measure such a large change in magnitude (or level) correlated with the temperature of the hot plate? There are two possible explanations for this.

The first was already discussed. At low temperature settings, the variation in level is probably simply an artefact of the background noise. All of these measurements were done in the Cube at B&O – which is normally used for a different kind of measurement which is more immune to background noises. This means that the Cube itself doesn’t have to be quiet, so it isn’t soundproofed from the outside world. Extraneous noises such as wind, rain, or someone walking by pushing a cart in the hallway, are easily audible if you’re sitting in the Cube. However, for the measurements in this experiment, those background noises will infect the results.

The second explanation is not related to background noise. As you can see in Figure 12, at higher frequencies, there is a definite increase in level variations caused by increasing temperature of the hot plate. However, the effect is too big to be attributed to absorption changes related to temperature changes. So, what is it? Our best guess is that it’s the result of the changes in thermal currents (in other words, the hot air rising) in front of the tweeter resulting in thermal gradients (differences of temperature in space). This causes differences in speed of sound in space causing refraction of the wavefront. The higher the frequency, the more the tweeter is “beaming” so the effect is more evident than it is at low frequencies where the loudspeaker is more omnidirectional. (Spinning a flashlight generates a more obvious than spinning a naked light bulb.) As the temperature changes in time, the sound beam from the tweeter moves towards and away from the microphone input, which looks like a change in level.

 

Figure 15:
Figure 15: A simplified diagram showing our guess as to why we see a change in magnitude at the microphone.

 

If we had used a cluster of microphones instead, we would very probably have seen the “beam” moving to another microphone as it went quiet at the main mic… Maybe we’ll do that experiment someday and call it “research” – but for now we have the answer we were looking for, so we’ll leave that one for another day…

 

The conclusion

So, after all that, our end conclusion was, for an active high-end loudspeaker, we should not put radiators directly below the midranges and tweeters – it’s better to put them behind the loudspeaker like we do already in the BeoLab 5 or BeoLab 20, for example. Or perhaps above the tweeter could also be an option – although that might be a challenge for the visual designers…

There’s no guarantee that the artefacts will be audible (in fact, they almost certainly won’t if you’re not paying attention to the phantom image location, or if you aren’t sitting in the sweet spot… but if those were the case, we probably wouldn’t be talking about a high-end loudspeaker…) but the artefacts are certainly measurable – and that’s enough for us to go looking for a better place to put the amplifier radiators.

And now, I guess this also means that when someone types “loudspeaker amplifier heat phantom image” into Google, they’ll get an answer…

 

B&O Tech: Beam Width Control – A Primer

#38 in a series of articles about the technology behind Bang & Olufsen loudspeakers

In the last posting, I talked about a little experiment we did that made us realise that control of a loudspeaker’s directivity (or more simply “beam width”) would be a useful parameter in the hands of a listener. And, in a previous article, I talked a little about why that is probably true. This week, let’s get our hands dirty and talk about how Beam Width Control can be accomplished.

Part 1: What is sound?

Okay – we’re really getting back to basics, but it won’t take long. I promise. In order to get to where we’re headed, we have to have a fairly intuitive grasp of what sound is. At a basic level, sound is a change in air pressure over time. Today’s atmospheric (or barometric) pressure is about 100 kiloPascals (give or take). That pressure can be thought of as a measure of the average distances between the air molecules in the room you’re sitting in. It is also important to note that (unless you’re very unlucky and very uncomfortable) the pressure inside your head is roughly the same as the pressure outside your head (otherwise, you’d probably be talking about going to visit a doctor…). The important thing about this is that this means that the pressure that is exerted on the outside of your eardrum (by the air molecules sitting next to it in your ear canal) is almost identical to the pressure that it exerted on the inside of your eardrum (the part that’s inside your head). Since the two sides are being pressed with the same force, then the eardrum stays where it should.

However, if we momentarily change the air pressure outside your head, then you eardrum will be pushed inwards (if the pressure is higher on the outside of your head) or pulled outwards (if the pressure is lower). This small, momentary change in pressure can be caused by many things, but one of those things is, for example, a woofer getting pushed out of its cabinet (thus making a “high pressure” where the air molecules are squeezed closer together than normal) or pulled into its cabinet (thus making a “low pressure” and pulling the air molecules apart).

So, the woofer pushes out of the cabinet, squeezes the air molecules together which push your eardrum into your head which does a bunch of things that result in you hearing a sound like a kick drum.

Part 2: Frequency and wavelength

Take a loudspeaker into your back yard and play a tone with the same pitch as a “Concert A” (the note you use to tune a violin – or, at least, the note violinists use to start tuning a violin…). This note has a frequency of 440 Hz, which means that your loudspeaker will be moving from its resting position, outwards, back, inwards, and back to the resting position (a full cycle of the tone) 440 times every second. This also means that your eardrum is getting pushed inwards and and pulled outwards 440 times every second.

Now, turn off the loudspeaker and walk 344 m away. When you get there, turn on the tone again. You’ll notice that you don’t hear it right away because sound travels pretty slowly in air. In fact, it will take almost exactly one second from the time you turn on the loudspeaker to the time the sound reaches you, since the speed of sound is 344 m/s (you’re 344 m away, so it takes one second for the sound to get to you).

If, at the instant you start to hear the tone, you were able to freeze time – or at least freeze the air molecules between you and the loudspeaker, you’d see something interesting (at least, I think it’s interesting). The instant you hear the start of the tone (the first cycle), the 441st cycle is starting to be generated by the loudspeaker (because it’s producing 440 cycles per second, and it’s been one second since it started playing the tone). This means that, in your frozen-time-world, there are 440 high and low pressure zones “floating” in the 344 m of air between you and the loudspeaker. Given this information, we can calculate how long the length of one wave is (better known as the “wavelength” of the tone). It’s 344 m divided by 440 cycles per second = 78.2 cm.

So, this means that there are 78.2 cm between each high pressure “bump” in the air. This also means that there is half that distance (39.1 cm) between adjacent high and low pressure zones in the air.

Figure 1: High pressure (red) and low pressure (blue) zones radiating outwards from a sound source (the black dot)
Figure 1: High pressure (red) and low pressure (blue) zones radiating outwards from a sound source (the black dot). Note that the scale goes from -1 to 1.

 

Part 3: Interference

Now let’s take a case where we have two woofers. We’ll put them side by side and we’ll send the same 440 Hz tone to them. So, when one woofer goes out, the other one does as well. This means that each one creates a high pressure in front of it. If you’re exactly the same distance from the two woofers, then those two high pressures add together and you get an even higher pressure at the listening position (assuming that there are no reflecting surfaces nearby). This is called constructive interference. This happens when you have two loudspeaker drivers that are either the same distance from the listening position as in Figure 2 or in exactly the same location (which is not possible), as in Figure 3.

 

 

Figure 2: Two sound sources playing the same signal located at the same distance from the listening position.
Figure 2: Two sound sources playing the same signal located at the same distance from the listening position.

 

Figure 2: Two sound sources located in the same position (this is not possible). Note that the pattern of radiation is the same as in Figure 1, it's just twice as much pressure.
Figure 3: Two sound sources located in the same position (this is not possible). Note that the pattern of radiation is the same as in Figure 1, it’s just twice as much pressure – the scale now goes from -2 to 2 because the pressures from the two sources add perfectly in all locations.

 

If we change the wiring to the woofers so that they are in opposite polarity, then something different happens. Now, when one woofer goes outwards, the other goes inwards by the same amount. This means that one creates a high pressure in front of it while the other creates a low pressure of the same magnitude. If you’re exactly the same distance from the two woofers, then those two pressures (one high and one low) arrive at the same time and add together, so you get a perfect cancellation at the listening position (assuming that there are no reflecting surfaces nearby). This is called destructive interference. This is shown in Figure 4 and 5.

 

igure 2: Two sound sources playing the same signal located at the same distance from the listening position.
Figure 4: Two sound sources playing the same signal in opposite polarity located at the same distance from the listening position.

 

 

Figure 4. Two sound sources playing the same signal in opposite polarity. Notice that they cancel each other "on axis" to the pair of sources (to the right of the black dots in the figure).
Figure 4. Two sound sources playing the same signal in opposite polarity. Notice that they cancel each other “on axis” to the pair of sources (to the right of the black dots in the figure). The cancellation is seen as a greenish-colour which corresponds to a maximum pressure of 0, or no sound…

 

However, an important thing about what I just said was that you are the same distance from the two woofers. What if you’re not? Let’s say, for example, that you’ve changed your angle to the “front” of the loudspeaker (which has two loudspeaker drivers). This means that you’re closer to one driver than the other. The result is that, although the two drivers are playing the same signal at the same time, your difference in distance causes their individual signals to have different phases at the listening position – possibly even a completely opposite polarity – as is shown in Figure 5.

 

igure 2: Two sound sources playing the same signal located at the same distance from the listening position.
Figure 5: Two sound sources playing the same signal located at different distances from the listening position. Notice that, although the signals leave the drivers in the same polarity, they are opposite in polarity at the listening position because the difference in distance results in a difference in time of arrival at the listening position. The more distant driver is farther away by one half of a wavelength in this case.

 

If you consider Figure 2 and Figure 5 together (they’re really two different views of the same situation, since both loudspeakers are playing the same signal simultaneously) and you include all other listening positions, then you get Figure 6. As you can see there, there is no sound (again, indicated by the greenish colour which means no change in pressure) at the angle shown in Figure 5.

Figure 3: Two sound sources separated by a distance of one wavelength of the tone they are producing. Note that the two sources add perfectly at all distances directly "in front" (to the right, in the figure) because any position on that line is the same distance from each source.
Figure 6: Two sound sources separated by a distance of one wavelength of the tone they are producing. Note that the two sources add perfectly at all distances directly “in front” (to the right, in the figure) because any position on that line is the same distance from each source.

 

So, Figure 6 shows that, for a position directly in front of the sound sources (to the right of the black dots, in the figure), the result is identical to that in Figure 3 – the two sources add together perfectly if you’re the same distance from them. However, if you start to move away from that line, so that you’re closer to one sound source than the other, then, at some angle, you will get a high pressure from the upper sound source at the same time as you get the previous low pressure from the lower source (because it’s farther away from you, and therefore a bit late in time…) This high+low = nothing, and you hear no sound.

 

Another, slightly more nuanced way to do this is to not just change the polarity of one of the sound sources, but to alter its phase instead. This is a little like applying a delay to the signal sent to the driver, but it’s a delay that is specific to the frequency that we’re considering. An example of this is shown in Figure 7.

Figure 5. Two sound sources playing the same signal with slightly different phases (delay). Notice that they cancel each other in a different direction than the previous cases.
Figure 7. Two sound sources playing the same signal with slightly different phases (another word for “delay”, sort of…). Notice that they cancel each other in a different direction than the previous cases.

This means that, by changing the phase instead of the polarity of the drivers, I can “steer” the direction of the beam and the “null” (the area where the sound is cancelled out)

It should be noted that, at a different separation of the drivers (more accurately stated: at a different relationship between the wavelength of the signal and the distance between the drivers) , the behaviour would be different. For example, look at Figure 8 which shows the same situation as Figure 6 – the only difference is that the two sound sources are half the distance apart (but the wavelength of the signal has remained the same).

Figure 6: Two sound sources separated by a distance of one wavelength of the tone they are producing. Note that the two sources add perfectly at all distances directly "in front" (to the right, in the figure) because any position on that line is the same distance from each source.
Figure 8: Two sound sources separated by a distance of one half wavelength of the tone they are producing. Compare this to Figure 6.

Part 4: Putting it together

So, at a very basic level (so far) we can determine which direction sound is radiated in based on a relationship between

  1. the wavelength of the signal it is producing (which is a function of its frequency)
  2. the distance between the sound sources (the loudspeaker drivers), and
  3. the polarity and/or the phase of the signals sent to the drivers.

All of the examples shown above assume that the two sound sources (the loudspeaker drivers) are playing at the same level, which is not necessarily the case – we can choose to play a given frequency at whatever level we want using a rather simple filter. So, by reducing the level of the “interfering” driver, we can choose how much the directivity of the radiated sound is affected.

In addition to this, all of the examples shown above assume that the sound sources are omnidirectional – that they radiate sound equally at all frequencies – which is not true in real life. In real life, a loudspeaker driver (like a tweeter or a woofer) has some natural directivity pattern. This pattern is different at different frequencies and is influenced by the physical shapes of the driver and the enclosure it’s mounted on.

So, let’s start putting this together.

If I take a loudspeaker driver – say, a midrange unit – and I mount it on the front of a loudspeaker enclosure and measure its directivity pattern  (the geeky term for the width of its sound beam) across its frequency range and beyond I’ll see something like this:

Figure 9: The horizontal directivity (or "beam width") of a midrange driver mounted on the front of a loudspeaker enclosure.
Figure 9: The horizontal directivity (or “beam width”) of a midrange driver mounted on the front of a loudspeaker enclosure. We don’t care about its behaviour in the very low end very high end (outside the red lines), since those frequency bands are taken care of by woofers and tweeters.

As you can see in that plot, in its low frequency range (the red line on the left), the midrange driver is radiating sound with a wide beam – not completely omnidirectional, but very wide. In its high frequency region (around the right-hand red line) (but still not high enough to be re-routed to the tweeter) the midrange is “beaming” – meaning that the beam is getting too narrow. Our goal is to somehow make the beam widths at these frequency bands (and the ones in between) more alike. So, we want to reduce the width of the beam in the low frequencies and increase the width of the beam in the high frequencies. How can we do this? We’ll use more midrange drivers!

Let’s take two more midrange drivers (one on either side of the “front” or “main” one) and use those to control the beam width. In the lower frequencies, we’ll send a signal to the two side drivers that cancel the energy radiated to the sides of the loudspeaker – this reduces the width of the beam compared to using the front midrange by itself. At higher frequencies, we’ll send a signal to the two side drivers that add to the signal from the front driver to make the width of the beam wider. At a frequency in the middle, we don’t have to do anything, since the width of the beam from the front driver is okay by itself, so at that frequency, we don’t send anything to the adjacent drivers.

“Okay”, I hear you cry, “that’s fine for the beam width looking from the point of view of the front driver, but what happens as I come around towards the rear of the loudspeaker?” Excellent question! As you come around the rear of the loudspeaker, you won’t get much contribution from the front midrange, so the closest “side” midrange driver is the “main” driver in that direction. And, as we saw in the previous paragraph, the signal coming out of that driver is pretty strange (because we intentionally made it strange using the filters – it’s in opposite polarity in its low end, it doesn’t produce anything in its midrange, and it’s quiet, but has a “correct” polarity in its high end). So, we’ll need to put in more midrange drivers to compensate again and clean up the signal heading towards the rear of the loudspeaker. (What we’re doing here is basically the same as we did before, we’re using the “rear” drivers to ensure that all frequencies heading in one rearward direction are equally loud – so we make the “rear” drivers do whatever they have to do at whatever frequency bands they have to do it in to make this true.)

“Okay”, I hear you cry, “that’s fine for the beam width looking from the point of view of the rear of the loudspeaker, but what happens as we go back towards the front of the loudspeaker? Won’t the signals from the rear-facing drivers screw up the signal going forwards?” Excellent question! The answer is “yes”. So, this means that after the signal from the rear-facing midrange drivers is applied (which compensate for the side-facing midrange drivers which compensate for the front facing midrange driver) then we have to go back to the front and find out the effects of the compensation of the compensation on the “original” signal and compensate for that – so we start in a loop where we are compensating for the compensation for the for the compensation for the compensation for the… Well, you get the idea…

The total result is a careful balancing act where some things are known or given:

  • the frequency range of the music being played through a given loudspeaker driver (this is limited by the crossover)
  • the natural directivity patterns of the loudspeaker drivers on the loudspeaker enclosure within that frequency range
  • the locations of the loudspeaker drivers in three-dimensional space
  • the orientations of the loudspeaker drivers in three-dimensional space (in other words, which way they’re pointing)

Note that these last two have been calculated and optimised based on a combination of the natural directivity patterns of the loudspeaker drivers and the desired beam widths we want to make available.

As a result each individual loudspeaker driver gets its own customised filter that controls

  • the level of the signal at any given frequency
  • the phase (which includes delay and polarity – sort of…) of the signal at any given frequency

 

Figure X: A conceptual illustration showing the different relationships between the various loudspeaker drivers for two different frequencies.
Figure 10: A conceptual illustration showing the different relationships between the various (in this illustration, seven) loudspeaker drivers for two different frequencies. The sum of all loudspeaker drivers results in the same beam width and the same output level on-axis at both frequencies.

By controlling the individual output levels and phases of each loudspeaker driver at each frequency it produces, we can change the overall level of the combined signals from all the loudspeaker drivers in a given direction. If we want to be loud at 0º (on-axis) and 20 dB quieter at 90º (to the side), we just apply the filters to all of the drivers to make that happen. If we want the loudspeaker to be only 10 dB down at 90º, then this just means changing to a different set of filters. This can be done independently at different frequencies – with the end goal to make all frequencies behave more similarly, as I talked about in this posting and this posting.

Also, since the filters are merely settings of the DSP (the digital signal processing “brain” of the loudspeaker), we can change the beam width of the loudspeaker merely by loading different filters – one for each loudspeaker driver in the system.

The end result is a loudspeaker that can play many different roles, as was shown in the different plots in this posting. In one mode, the beam width can be set to “narrow” for situations where you have one chair, and no friends and you want the ultimate “sweet spot” experience for an hour or two. In another mode, the beam width can be set to “wide”, resulting in a directivity that is very much like an improved version of the wide dispersion of the BeoLab 5. In yet another mode, the beam width can be set to “omni”, sending sound in all directions, making a kind of “party mode”.

 

Part 5: The “side” effect: Beam Direction Control

In order to be able to have a Beam Width Control, a loudspeaker must have a number of identical loudspeaker drivers. A collection of tweeters, midranges and woofers, some on the front, some on the sides and some to the rear of the loudspeaker. In addition, each of these drivers must have its own amplifier and DSP channel so that it is independently controllable by the system.

This means that, instead of using the drivers to merely control the width of the beam in the front of the loudspeaker, they can also be used somewhat differently. We could, instead, decide to “rotate” the sound, so that the “main” tweeter and midrange are the ones on the side (or even the rear) of the loudspeaker instead of the ones on the front. This means that, if you’re sitting in a location that is not directly in front of the loudspeakers (say, to the sides, or even behind), it could be possible to improve the sound quality by “rotating” the “on-axis” direction of the loudspeaker towards your listening position. This by-product of Beam Width Control is called Beam Direction Control.

P.S. Apologies for the pun…

 

 

For more information on Beam Width Control:

What is “Beam Width Control”?

Shark Fins and the birth of Beam Width Control

B&O Tech: Shark Fins and the birth of Beam Width Control

#37 in a series of articles about the technology behind Bang & Olufsen loudspeakers

In the last posting, I described a little experiment that we did where we could easily be accused of asking a stupid question and getting an obvious answer. This posting is a little different, although I’ll also describe a little experiment where we asked a much-less-stupid question and got an answer we weren’t looking for… However, before I describe the experiment itself, we’ll have to start with a little background information.

Let’s say that we wanted to build a “simple” two-way loudspeaker with a tweeter and a woofer in a box like the one shown in Figure 1.

Figure 1. A simple two-way loudspeaker
Figure 1. A simple two-way loudspeaker.

Since we’re Bang & Olufsen, it is implied that this will be a fully active, DSP-based loudspeaker. Let’s also, for the sake of simplicity, say that we used loudspeaker drivers and other components with absolutely no linear distortion artefacts (I can dream, can’t I?). This means that we can apply any kind of filter we like to each of the two loudspeaker drivers to result in whatever response we want (in a given direction in a “free field” (a space with no reflecting surfaces, extending to infinity in all directions) – for more information on this, please read this posting and/or this posting).

Now, it’s fairly reasonable to assume that one portion of a good starting point for a loudspeaker’s acoustic response is to have a flat magnitude response when measured on-axis in a free field. This means that, if you have the loudspeaker in a space that has no reflections, and you are directly in front of it, no frequency will be louder than any other (within some limits, of course…). In fact, this is the way a lot of studio monitor loudspeakers are tuned (which makes sense, since many studio control rooms don’t have a lot of reflections as was discussed here).

The problem is that, if you do make a loudspeaker that is flat-on-axis, you’ll probably have problems in its power response (for a description of what a loudspeaker’s “power response” is, see here). This is dependent on a lot of things like the crossover frequency, the sizes and shapes of the drivers, the phase responses of the crossover filters, the shape of the cabinet, and other things. However, if we were to simplify, we could say that a two-way loudspeaker (say, with a 4th order Linkwitz-Riley crossover, just to pick one type…) that is flat on-axis, will have a dip in its power response at the crossover frequency. This is because, although the two drivers (the tweeter and the woofer) add together nicely in one location (on-axis, directly in front of the loudspeaker) they do not add together nicely anywhere else (because the distances to the two drivers from the listening position are probably not matched, particularly when you go vertical…).

So, one basic problem with building a “simple” two-way loudspeaker is that you have to choose to either have a flat magnitude response on-axis, or a smooth power response (without a dip) – but you can’t have both. (If you want to really dig into this, I’d recommend starting with J. Vanderkooy & S.P. Lipshitz, “Power Response of Loudspeakers with Noncoincident Drivers — The Influence of Crossover Design,” J. Audio Eng. Soc., Vol. 34, No. 4, pp. 236-244 (Apr. 1986).)

This basic problem raised a question in the head of one of our acoustical engineers, Gert Munch. He started to wonder how we could build a loudspeaker that could have a flat magnitude response on-axis and still have a smooth power response that didn’t suffer from a dip at the crossover. One possible solution is to build a loudspeaker with the desired on-axis response, and then somehow create an additional portion of the loudspeaker that could “fill up” the dip in the power response by sending energy into the room without it affecting the on-axis response.

One possible way to do this is to use an extra loudspeaker with a “dipole” characteristic – a two-sided loudspeaker where the same signal is sent to both sides, but in opposite polarity. This means that, when one side of the loudspeaker produces a high pressure, the other side produces a low. When you sit on one side of the loudspeaker or the other, then you hear a signal. However, if you sit “on edge” to the loudspeaker, you hear nothing, since the two sides of the loudspeaker cancel each other out (in theory).

So, Gert’s idea was to add a two-way dipole loudspeaker on the top of a normal two-way loudspeaker and to just use the dipole (which became known as a “shark fin” – shown in Figure 2)  to add energy around the crossover frequency of the “normal” loudspeaker. Since the edge of the dipole was facing forwards, the theory was that none of its sound would have an effect on the on-axis response of the two-way loudspeaker below it.

Figure 2. A two-way loudspeaker with a side-facing dipole loudspeaker on top.
Figure 2. A two-way loudspeaker with a side-facing dipole loudspeaker on top. Note that the other side of the dipole looks identical to the one you see.

 

So, the question to answer was:

“Which sounds better:

  • a loudspeaker tuned to be flat on-axis, but with no correction for its power response
  • a loudspeaker that was tuned to have a smooth power response (in other words, with a “bump” around the crossover)
  • a loudspeaker that was tuned to be flat on-axis and had a power response correction provided by the dipole

?”

Since each of the 6 loudspeaker drivers in our model loudspeaker was independently controllable, this was an easy comparison to do on-the-fly, so we tried it out.

Also, just to be certain, we tried two different orientations of the dipole, both of which had the “null” – the edge of the loudspeaker – facing the listening position. The first orientation is shown in Figures 2 and 3. The second orientation is shown in Figure 4.

 

Figure 3. The dipoles in the first orientation.
Figure 3. The dipoles in the first orientation.

 

Figure 3. The dipoles in a second orientation.
Figure 4. The dipoles in a second orientation.

 

So, a bunch of us sat in the listening room for a week or so, listening to the various possible tunings of the loudspeakers, in various positions in the room, with various types of music. We each had an opinion about the whether we liked one configuration more or less than another. However, one thing that was noticeable was that there was an effect not only on the timbral impression of the loudspeakers (the tone colour) – but also on the spatial presentation of the pair. As we switched between the different tunings, phantom images changed in width and distance, and envelopment (when it existed in the recording) also changed.

This was particularly noticeable when the loudspeakers were closer to a reflecting surface, as is shown in Figure 5.

Figure 5. A listening configuration where one loudspeaker was intentionally placed close to a reflecting surface.
Figure 5. A listening configuration where one loudspeaker was intentionally placed close to a reflecting surface.

 

The end result was a bunch of different conclusions:

  1. The three different configurations of the loudspeaker (listed above) sounded different. (this was not a big surprise)
  2. A two-way loudspeaker tuned to have a smooth power response and a loudspeaker with a flat on-axis response with a shark fin smoothing the power response sounded different – but mostly spatially.
  3. Switching between those two loudspeaker configurations listed in Point #2 basically meant that we were switching between two different loudspeakers with the same power response, but different directivities (a fancy word meaning “directional behaviours”).
  4. The dipole loudspeaker’s output was audible at the listening position because of the sidewall reflections in the case where the dipole was on top.
  5. Switching the dipole on and off had a noticeable effect on the phantom image distance and width… When the dipole was turned on, the sidewall reflections were obviously louder, and the phantom images moved to a distance from the listening position roughly the same as the distance to the loudspeakers themselves. At the same time, the phantom image width increased and became “cloudy” or “fuzzy” and less precise.

 

All of those little conclusions could be folded into two big ones:

  1. The original idea of using a dipole to fill in the power response of the loudspeaker doesn’t work well enough in all situations because we don’t know where the reflective surfaces around the loudspeaker will be in any given configuration.
  2. Being able to control the directivity of the loudspeaker (specifically, how much energy is coming from the sidewall reflections) is a very interesting control parameter that we should look into…

 

So, this is where the concept of the shark fin died, but where the idea of “Beam Width Control” began.

Well… more accurately: it was the beginning of our thinking that being able to change the beam width (or the “directivity”, if you’re a geek) of the loudspeaker would be a useful “handle” in the hands of our customers. The idea that was born was something along the lines of the following: “If a customer could change the beam width of a loudspeaker, then (s)he could change it from being a narrow-directivity loudspeaker for people with one chair and no friends to a wide-directivity loudspeaker for people with a sofa and some friends to an omni-directivity loudspeaker for people throwing a party – all just by the flick of a switch…”

 

For more information on Beam Width Control:

What is “Beam Width Control”?

Beam Width Control – A Primer

B&O Tech: It’s lonely at the top

#36 in a series of articles about the technology behind Bang & Olufsen loudspeakers

 

“In all affairs it’s a healthy thing now and then to hang a question mark on the things you have long taken for granted.”

-Bertrand Russell

 

If you were to get your hands on Harry Potter’s invisibility cloak and you were to walk into the acoustics department at Bang & Olufsen and eavesdrop on conversations, you’d sometimes be amazed at some of the questions we ask each other. That’s particularly true when we’re working on a new concept, because, if the concept is new, then it’s also new for us. The good thing is that all of my colleagues (and I) are ready and willing at any time to ask what might be considered to be a stupid (or at least a very basic) question.

One of those questions that we asked each other recently seemed like a basic one – why do we always put the tweeter on the top? It seems like there are very few loudspeakers that don’t do this (of course, there are exceptions – take the BeoLab Penta, for example, which has the tweeter in the middle). However, more often than not, when we (and most other people) make a loudspeaker, we put the loudspeaker drivers in ascending order of frequency – woofers on the bottom, tweeters on the top. Maybe this is because woofers are heavier, and if you stand a BeoLab 5 on its head, it will fall over – but that’s not really the question we were asking…

The REAL question we were asking ourselves at the time was: if we were to build a multi-way loudspeaker – let’s say a 3-way, with a woofer, a midrange and a tweeter, and if the crossovers were such that the bulk of the “interesting” information (say, the vocal range) was coming from the midrange, then why would we not put the midrange at ear height and put the tweeter between it and the woofer? For example, imagine we made BeoLab 20 without an Acoustic Lens on top, would it be better to arrange the drivers like the version on the left of Figure 1 or the version on the right? Which one would sound better?

 

Figure 1: Two versions of a BeoLab 20-like loudspeaker with different driver arrangements.
Figure 1: Two versions of a BeoLab 20-like loudspeaker with different driver arrangements.

 

After answering that question, there’s a second question would follow closely behind: how close together do the drivers with adjacent frequency bands (i.e. the woofer and the midrange or the tweeter and the midrange) have to be in order for them to sound like one thing? Of course, these two questions are inter-related. If your midrange and tweeter are so far apart that they sound like different sound sources, then you would probably be more interested in where the voice was coming from than where the cymbals were coming from…

Of course, step one in answering the second question could be to calculate/simulate the response of the loudspeaker, based on distance between drivers, the crossover frequency (and therefore the wavelengths of the frequency band we’re interested in), the slopes of the crossover filters, and so on. It would also be pretty easy to make a prototype model out of MDF, put the loudspeaker drivers in there, do the vertical directivity measurements of the system in the Cube, and see how well the theory matches reality.

However, the question we were really interested in was “but how would it sound?” – just to get a rough idea before going any further. And I have to stress here that we were really talking about getting a rough idea. What I’m about to describe to you is a little undertaking that we put together in a day – just to find out what would happen. This was not a big, scientifically-valid experiment using a large number of subjects and intensive statistics to evaluate the results. It was a couple of guys having a chat over coffee one morning when one of them said “I wonder what would happen if we put the midrange on top…” and then, off they went to a listening room to find out.

One thing we have learned in doing both “quick-n-dirty” listening comparisons and “real, scientifically valid listening tests” is that the placement of a loudspeaker in a room, has a huge effect on how the loudspeaker sounds. So, when we’re comparing two loudspeakers, we try to put them as close together as possible. So, we tried different versions of this. In the first, we took two pairs of loudspeakers, and put the left loudspeakers in each pair side-by-side, with one pair upside down and the other right-side up, as shown in Figure 2.

 

Figure X: Another arrangement of the loudspeakers used in the first part of the experiment.
Figure 2: One arrangement of the loudspeakers used in the first part of the experiment. “Pair A” is in black and “Pair B” is in red.

 

We then switched between the right-side up pair and the upside down pair, listening for a change in vertical position of the image. Note that we tried two arrangements of this – one where both right-side up loudspeakers were to the left of the upside-down loudspeakers. The other where the “right-side up” loudspeakers were the “outside” pair, as shown in Figure 3.

Figure X: Another arrangement of the loudspeakers used in the first part of the experiment.
Figure 3: Another arrangement of the loudspeakers used in the first part of the experiment.

 

There are advantages and disadvantages of both of these arrangements – but in both cases, there is a lateral shift in the stereo image. When switching between pairs, either you get a left-right shift in image, or a change in width… It turned out that this change was more distracting than the vertical arrangement of the drivers, so we changed configuration to the one shown in Figure 4, below.

Figure X: Another arrangement of the loudspeakers used in the first part of the experiment.
Figure 4: Another arrangement of the loudspeakers used in the first part of the experiment. Note that, in this version, we used the loudspeaker pairs differently. “Pair A” used the red loudspeakers as tweeters and the black loudspeakers as woofers using an external crossover. “Pair B” used the red loudspeakers as woofers and the black loudspeakers as tweeters.

 

Now, instead of switching between loudspeakers, we pretended that one of them was a tweeter and other was a mid-woofer, and switched which was which, on the fly. Our “virtual” crossover was close-ish to the real crossover in the loudspeakers (our crossover was at 3.5 kHz, if you’re curious), so you could say that we were sort-of changing between using the upper tweeter + the lower woofer and using the upper woofer + lower tweeter, the “acoustical centres” of which are roughly at the same height. (remember – this was a quick-n-dirty experiment…)

 

Figure X: A photo of the loudspeakers used in the first part of the experiment.
Figure 5: A photo of one configuration of the  loudspeakers used in the first part of the experiment. Note that the BeoLab 9 and the BeoPlay V1 were not part of the experiment – we just didn’t bother to clean up before getting busy…

 

 

Figure X: A photo of the loudspeakers used in the first part of the experiment.
Figure 6: A photo of one configuration of the loudspeakers used in the first part of the experiment.

 

After having listened to these three configurations of loudspeakers, we decided that the vertical arrangement of the drivers was not important with the vertical separation we were using.

This brought us to the second part of the question… If the tweeter and the midrange were further apart, would we have a preference? So, we kept our virtual crossover, using one loudspeaker as the “tweeter” and the other as the “mid-woofer”, and we moved the loudspeakers further apart, one example of which is shown in Figure 7. (One thing to note here is that when I say “further apart” I’m talking about the separation in the vertical angles of the sources – not necessarily their actual distance from each other. For example, if the loudspeakers were 1 m apart vertically, and you were level with one of the loudspeakers, but the listening position was a kilometre away, then the vertical angular separation (0.057 degrees) would be much smaller than if you were 1 m away (45 degrees)…)

 

Figure X: A photo of the loudspeakers used in the second part of the experiment.
Figure 7: A photo of the loudspeakers used in the second part of the experiment. This was one (extreme) example of a vertical separation.

The answer we arrived at at the end was that, when the vertical separation of the drivers gets extreme (perhaps I should use the word “silly” instead), we preferred the configuration where the “mid-woofer” was at ear-height. However, this was basically a choice of which version we disliked less (“preference” is a loaded word…). When the drivers get far enough apart, of course, they no longer “hang together” as a single device without some extra signal processing tricks.

So, we went to lunch having made the decision that, as long as the tweeters and the midranges are close enough to each other vertically, we really didn’t have a preference as to which was on top, so, if anyone asked, we would let the visual designers decide. (Remember that “close enough” is not only determined by this little experiment – it is also determined by the wavelengths of the crossover frequencies and whether or not there are more drivers to consider. For example, in the example in Figure 1, it might be that we don’t care about the relative placement of the tweeter and midrange in isolation – but perhaps putting the tweeter between the midrange and woofer will make them too far apart to have a nice vertical directivity behaviour across the lower crossover…)

 

Addendum

This isn’t the first time we asked ourselves such a question. Although I was not yet working at B&O at the time, my colleagues tell me that, back in the days when they were developing the BeoLab 3500 and the BeoLab 7-1 (both of which are “stereo” loudspeakers – essentially two multi-way loudspeakers in a single device) , they questioned the driver arrangement as well. Should the tweeters or the midranges / woofers be on the outside? You can see in the photos of the 3500 on this page that they decided to put the lower-frequency bands wider because the overall impression with stereo signals was wider than if the tweeters were on the outside.