Category Archives: Audio

Skywalker’s Michael Semanick: Mixing SFX for Star Wars: The Last Jedi

By Jennifer Walden

Oscar-winning re-recording mixer Michael Semanick from Skywalker Sound mixed the sound effects, Foley and backgrounds on Star Wars: The Last Jedi, which has earned an Oscar nomination for Sound Mixing.

Technically, this is not Semanick’s first experience with the Star Wars franchise — he’s credited as an additional mixer on Rogue One — but on The Last Jedi he was a key figure in fine-tuning the film’s soundtrack. He worked alongside re-recording mixers Ren Klyce and David Parker, and with director Rian Johnson, to craft a soundtrack that was bold and dynamic.

Recently, Semanick shared his story of what went into mixing the sound effects on The Last Jedi. He mixed at Skywalker in Nicasio, California, on the Kurosawa Stage. (And come back next week for our interview with Ren Klyce about mixing the music and sounds.)

You have all of these amazing elements — Skywalker’s effects, John Williams’s score and the dialogue. How do you bring clarity to what could potentially be a chaotic soundtrack?
Yes, there are a lot of elements that come in, and you have to balance these things. It’s easy on a film like this to get bombastic and assault the audience, but that’s one of the things that Rian didn’t want to do. He wanted to create dynamics in the track and to get really quiet so that when it does get loud it’s not overly loud.

So when creating that I have to look at all of the elements coming in and see what we’re trying to do in each specific scene. I ask myself, “What’s this scene about? What’s this storyline? What’s the music doing here? Is that the thread that takes us to the next scene or to the next place? What are the sound effects? Do we need to hear these background sounds, or do we need just the hard effects?”

Essentially, it’s me trying to figure out how many frequencies are available and how much dialogue has to come through so the audience doesn’t lose the thread of the story. It’s about deciding when it’s right to feature the sound effects or take the score down to feature a big explosion and then bring the score back up.

It’s always a balancing act, and it’s easy to get overwhelmed and throw it all in there. I might need a line of dialogue to come through, so the backgrounds go. I don’t want to distract the audience. There is so much happening visually in the film that you can’t put sound on everything. Otherwise, the audience wouldn’t know what to focus on. At least that’s my approach to it.

How did you work with the director?
As we mixed the film with Rian, we found what types of sounds defined the film and what types of moments defined the film in terms of sound. For example, by the time you reach the scene when Vice Admiral Holdo (Laura Dern) jumps to hyperspace into the First Order’s fleet, everything goes really quiet. The sound there doesn’t go completely out — it feels like it goes out, but there’s sound. As soon as the music peaks, I bring in a low space tone. Well, if there was a tone in space, I imagine that is what it would sound like. So there is sound constantly through that scene, but the quietness goes on for a long time.

One of the great things about that scene was that it was always designed that way. While I noted how great that scene was, I didn’t really get it until I saw it with an audience. They became the soundtrack, reacting with gasps. I was at a screening in Seattle, and when we hit that scene and you could hear that the people were just stunned, and one guy in the audience went, “Yeah!”

There are other areas in the film where we go extremely quiet or take the sound out completely. For example, when Rey (Daisy Ridley) and Kylo Ren (Adam Driver) first force-connect, the sound goes out completely… you only hear a little bit of their breathing. There’s one time when the force connection catches them off guard — when Kylo had just gotten done working out and Rey was walking somewhere — we took the sound completely out while she was still moving.

Rian loved it because when we were working on that scene we were trying to get something different. We used to have sound there, all the way through the scene. Then Rian said, “What happens if you just start taking some of the sounds out?” So, I started pulling sounds out and sure enough, when I got the sound all the way out — no music, no sounds, no backgrounds, no nothing — Rian was like, “That’s it! That just draws you in.” And it does. It pulls you into their moment. They’re pulled together even though they don’t want to be. Then we slowly brought it back in with their breathing, a little echo and a little footstep here or there. Having those types of dynamics worked into the film helped the scene at the end.

Rian shot and cut the picture so we could have these moments of quiet. It was already set up, visually and story-wise, to allow that to happen. When Rey goes into the mirror cave, it’s so quiet. You hear all the footsteps and the reverbs and reflections in there. The film lent itself to that.

What was the trickiest scene to mix in terms of the effects?
The moment Kylo Ren and Rey touch hands via the force connection. That was a real challenge. They’re together in the force connection, but they weren’t together physically. We were cutting back and forth from her place to Kylo Ren’s place. We were hearing her campfire and her rain. It was a very delicate balance between that and the music. We could have had the rain really loud and the music blasting, but Rian wanted the rain and fire to peel away as their hands were getting closer. It was so quiet and when they did touch there was just a bit of a low-end thump. Having a big sound there just didn’t have the intimacy that the scene demanded. It can be so hard to get the balance right to where the audience is feeling the same thing as the characters. The audience is going, “No, oh no.” You know what’s going to come, but we wanted to add that extra tension to it sonically. For me, that was one of the hardest scenes to get.

What about the action scenes?
They are tough because they take time to mix. You have to decide what you want to play. For example, when the ships are exploding as they’re trying to get away before Holdo rams her ship into the First Order’s, you have all of that stuff falling from the ceiling. We had to pick our moments. There’s all of this fire in the background and TIE fighters flying around, and you can’t hear them all or it will be a jumbled mess. I can mix those scenes pretty well because I just follow the story point. We need to hear this to go with that. We have to have a sound of falling down, so let’s put that in.

Is there a scene you had fun with?
The fight in Snoke’s (Andy Serkis) room, between Rey and Kylo Ren. That was really fun because it was like wham-bam, and you have the lightsaber flying around. In those moments, like when Rey throws the lightsaber, we drop the sound out for a split second so when Kylo turns it on it’s even more powerful.

That scene was the most fun, but the trickiest one was that force-touch scene. We went over it a hundred different ways, to just get it to feel like we were with them. For me, if the sound calls too much attention to itself, it’s pulling you out of the story, and that’s bad mixing. I wanted the audience to lean in and feel those hands about to connect. When you take the sound out and the music out, then it’s just two hands coming together slowly. It was about finding that balance to make the audience feel like they’re in that moment, in that little hut, and they’re about to touch and see into each other’s souls, so to speak. That was a challenge, but it was fun because when you get it, and you see the audience react, everyone feels good about that scene. I feel like I did something right.

What was one audio tool that you couldn’t live without on this mix?
For me, it was the AMS Neve DFC Gemini console. All the sounds came into that. The console was like an instrument that I played. I could bring any sound in from any direction, and I could EQ it and manipulate it. I could put reverb on it. I could give the director what he wanted. My editors were cutting the sound, but I had to have that console to EQ and balance the sounds. Sometimes it was about EQing frequencies out to make a sound fit better with other sounds. You have to find room for the sounds.

I could move around on it very quickly. I had Rian sitting behind me saying, “What if you roll back and adjust this or try that.” I could ease those faders up and down and hit it just right. I know how to use it so well that I could hear stuff ahead of what I was doing.

The Neve DFC was invaluable. I could take all the different sound formats and sample rates and it all came through the console, and in one place. It could blend all those sources together; it’s a mixing bowl. It brought all the sounds together so they could all talk to each other. Then I manipulated them and sent them out and that was the soundtrack — all driven by the director, of course.

Can you talk about working with the sound editor?
The editors are my right-hand people. They can shift things and move things and give me another sound. Maybe I need one with more mid-range because the one in there isn’t quite reading. We had a lot of that. Trying to get those explosions to work and to come through John Williams’ score, sometimes we needed something with more low-end and more thump or more crack. There was a handoff in some scenes.

On The Last Jedi, I had sound effects editor Jon Borland with me on the stage. Bonnie Wild had started the project and had prepped a lot of the sounds for several reels — her and Jon and Ren Klyce, who oversaw the whole thing. But Jon was my go-to person on the stage. He did a great job. It was a bit of a daunting task, but Jon is young and wants to learn and gave it everything he had. I love that.

What format was the main mix?
Everything was done in Atmos natively, then we downmixed to 7.1 and 5.1 and all the other formats. We were very diligent about having the downmixed versions match the Atmos mix the best that they could.

Any final thoughts you’d like to share?
I’m so glad that Rian chose me to be part of the mix. This film was a lot of fun and a real collaborative effort. Rian is the one who really set that tone. He wanted to hear our ideas and see what we could do. He wasn’t sold on one thing. If something wasn’t working, he would try things out until it did. It was literally sorting out frequencies and getting transitions to work just right. Rian was collaborative, and that creates a room of collaboration. We wanted a great track for the audience to enjoy… a track that went with Rian’s picture.

Super Bowl: Sound Lounge’s audio post for Pepsi, NFL and more

By Jennifer Walden

Super Bowl Sunday is an unofficial national holiday in this country, with almost as much excitement for the commercials that air as for the actual game. And regardless of which teams are playing, New York’s advertising and post communities find themselves celebrating, because they know the work they are providing will be seen by millions and talked about repeatedly in offices and on social media. To land a Super Bowl ad is a pretty big deal, and audio post facility Sound Lounge has landed seven!

Tom Jucarone

In this story, president/mixer/sound designer Tom Jucarone, mixer/sound designer Rob DiFondi and mixer/sound designer Glen Landrum share details on how they helped to craft the Super Bowl ads for Pepsi, E*Trade, the NFL and more.

Pepsi This is the Pepsi via Pepsi’s in-house creative team
This spot looks at different Pepsi products through the ages and features different pop-culture icons — like Cindy Crawford — who have endorsed Pepsi over the years. The montage-style ad is narrated by Jimmy Fallon.

Sonically, what’s unique about this spot?
Jucarone: What’s unique about this spot is the voiceover — it’s Jimmy Fallon. Sound-wise, the spot was about him and the music more than anything else. The sound effects were playing a very secondary role.

Pepsi had a really interesting vision of how they wanted Jimmy to sound. We spent a lot of time making his voice work well against the music. The Pepsi team wanted Fallon’s voice to have a fullness yet still be bright enough to cut through the heavy-duty music track. They wanted his voice to sound big and full, but without losing the personality.

What tools helped?
I used a few plug-ins on his voice. Obviously, there was some EQ but I also used one plug-in called MaxxBass by Waves, which is a bass enhancement plug-in. With that, I was able to manipulate where on the low-end I could affect his voice with more fullness. Then we added a touch of reverb to make it a bit bigger. For that, I used Audio Ease’s Altiverb but it’s very slight.

Persil Game-Time Stain-Time via DDB New York
In this spot, there’s a time-out during the big game and an announcer on TV taps on the television’s glass — from the inside. He points out a guacamole stain on one viewer’s shirt, then comes through the TV and offers up a jug of laundry detergent. The man’s shirt flies off, goes into the washer and comes out perfect. Suddenly, the shirt is back on the viewer’s body and the announcer returns to inside the TV.

Sonically, what’s unique about this spot?
Jucarone: It’s an interesting spot because it’s so totally different from what you’d expect to see during the Super Bowl. It’s this fun, little quirky spot. This guy comes out of the TV and turns all these people onto this product that cleans their clothes. There was no music, just a few magical sound effects. It’s a dialogue-driven spot, so the main task there was to clean up the dialogue and make it clear.

iZotope is my go-to tool for dialogue clean up. I love that program. There are so many different ways to attack the clean up. I’m working on a spot now that has dialogue that is basically not savable, but I think I can save it with iZotope. It’s a great tool — one of the best ones to have. I used RX 6 a lot on the Persil spot, particularly for this one guy who whispers, “What is going on?” The room tone was pretty heavy on that line, and it was one of the funniest lines, so we really wanted that one to be clear.

The approach to all these spots was to find out what unique sonic pieces are important to the story, and those are the ones you want to highlight. Back before the CALM Act, everyone was trying to make their commercial louder than everybody else’s. Now that we have that regulation, we’re a bit more open to making a spot more cinematic. We have a greater opportunity for storytelling.

E*Trade This is Getting Old via MullenLowe
In this spot, a collection of senior citizens sing about still being in the workforce — “I’m eighty-five, and I want to go home.” It’s set to the music of Harry Belafonte’s song “Day-O.” From lifeguard to club DJ, their careers are interesting, sure, but they really want nothing more than to retire.

Sonically, what’s unique about this spot?
Jucarone: That spot was difficult because of all the different voices involved, all the different singers. The agency worked with mixer Rob DiFondi and me on this one. Rob did the final mix.

The spot has a music track with solo and group performances. They had recorded the performers at a recording studio and then brought those tracks to us as a playlist of roughly 20 different versions. There were multiple people with multiple different versions, and the challenge was going through all of those to find the most unique and funniest voices for each person. So that took some time. Then, we had to match all of those voices so they sounded similar in tone. We had to re-mix each voice as we found it and used it because it wasn’t already processed. Then we had to also craft the group.
I worked with the agency to get the solo performances finalized and then Rob, the other mixer on it, took over and created the group performances. He had to combine all of these singular voices to make it sound like they were all singing together in a group, which was pretty difficult. It turned out to be a very complex session. We had multiple versions because they wanted to have choices after the fact.

What tools did you use?
There were a couple of different reverbs that really helped on this spot. We used the Waves Renaissance Reverb, and Avid’s Reverb One. We used a fun analog modeling EQ called Waves V-EQ4, which is modeled after a Neve 1081 console EQ. We wanted the individual voices to sound like they were singing together, one after another.

Any particular challenge?
DiFondi: My big job was the background chorus. We had to make a group of eight elderly background singers sound much larger. The problem there was layering the same eight people four times doesn’t net you the same as having 32 individuals. So what I did was treat each track separately. I varied the timing of each layer and I put each one in a separate room using different reverb settings and in the end that gave us the sound of a much larger chorus though we had only eight people.

Super Bowl NFL Celebrations to Come via Grey New York
NY Giants players Eli Manning and Odell Beckham, Jr. re-enact the famous last dance from Dirty Dancing, including the legendary lift at the end. The spot starts out realistic with on-camera dialogue for Eli and Odell during a team practice, but then it transitions into more of a music video as the players get wrapped up in the dance.

Sonically, what’s unique about this spot?
Landrum: There was lots of mastering on the main music track to get it to pop and be loud on-air. I used the iZotope Neutron for my music track mastering. I love that plug-in and have been using it and learning more about it. It has great multi-band compression, and the exciter is a cool addition to really finesse frequencies.

I think the most interesting part of the process was working with the director and agency creatives and producers to edit the music to match the storyboard they had before the actual shoot. We cut a few versions of varying lengths to give some flexibility. They used the music edits on-set so the guys could dance to it. I thought this was so smart because they would know what’s working and what isn’t while on-set and could adjust accordingly. I know they had a short shoot day so this had to help.

Everything worked out perfectly. I think they edited in less than a week (editor Geoff Hounsell from Arcade Edit, NY) and we mixed in a day or less. The creatives and producers involved with this spot and the NFL account are an awesome group. They make decisions and get it done and the result was amazing. Also, our expert team of producers here made the process smooth as silk during the stressful Super Bowl time.


Jennifer Walden is a New Jersey-based audio engineer and writer.  Follow her on Twitter @AudioJeney.

Cinna 1.2

Oscar Watch: The Shape (and sound) of Water

Post production sound mixers Christian Cooke and Brad Zoern, who are nominated (with production mixer Glen Gauthier) for their work on Fox’s The Shape of Water, have sat side-by-side at mixing consoles for nearly a decade. The frequent collaborators, who handle mixing duties at Deluxe Toronto, faced an unusual assignment given that the film’s two lead characters never utter a single word of actual dialogue. In The Shape of Water, which has been nominated for 13 Academy Awards, Elisa (Sally Hawkins) is mute and the creature she falls in love with makes undefined sounds. This creative choice placed more than the usual amount of importance on the rest of the soundscape to support the story.

L-R: Nathan Robitaille, J. Miles Dale, Brad Zoern, director Guillermo del Toro, Christian Cooke, Nelson Ferreira, Filip Hosek, Cam McLauchlin, video editor Sidney Wolinsky, Rob Hegedus, Doug Wilkinson.

Cooke, who focused on dialogue and music, and Zoern, who worked with effects, backgrounds and Foley, knew from the start that their work would need to fit into the unique and delicate tone that infused the performances and visuals. Their work began, as always, with pre-dubs followed by three temp mixes of five days each, which allowed for discussion and input from director Guillermo del Toro. It was at the premixes that the mixers got a feel for del Toro’s conception for the film’s soundtrack. “We were more literal at first with some of the sounds,” says Zoern. “He had ideas about blending effects and music. By the time we started on the five-week-long mix, we had a very clear idea about what he was looking for.”

The final mix took place in one of Deluxe Toronto’s five stages, which have identical acoustic qualities and the same Avid Pro Tools-based Harrison MP4D/Avid S6 hybrid console, JBL M2 speakers and Crown amps.

The mixers worked to shape sonic moments that do more than represent “reality,” but create mood and tension. This includes key moments such as the sound of a car’s windshield wipers that build in volume until they take over the track in the form of a metronome-like beat underlining the tension of the moment. One pivotal scene finds Richard Strickland (Michael Shannon) paying a visit to Zelda Fuller (Octavia Spencer). As Strickland speaks, Zelda’s husband Brewster (Martin Roach) watches television. “It was an actual mono track from a real show,” Cooke explains. “It starts out sounding roomy and distant as it would really have sounded. As the scene progresses, it expands, getting more prominent and spreading out around the speakers [for the 5.1 version]. By the end of the scene, the audio from the TV has become something totally different from what it started the scene as and then we melded that seamlessly into Alexandre Desplat’s score.”

Beyond the aesthetic work of building a sound mix, particularly one so fluid and expressionistic, post production mixers must also collaborate on a large number of technical decisions during the mix to ensure the elements have the right amount of emotional punch without calling attention to themselves. Individual sounds, even specific frequencies, vie for audience attention and the mixers orchestrate and layer them.

“It’s raining outside when they come into the room,” Zoern notes about the above scene. “We want to initially hear the sound of the rain to have a context for the scene. You never just want dialogue coming out of nowhere; it needs to live in a space. But then we pull that back to focus on the dialogue, and then the [augmented] audio from the TV gains prominence. During the final mix, Chris and I are always working together, side by side, to meld the hundreds of sounds the editors have built in a way that reflects the story and mood of the film.”

“We’re like an old married couple,” Cooke jokes. “We finish each other’s sentences. But it’s very helpful to have that kind of shorthand in this job. We’re blending so many pieces together and if people notice what we’ve done, we haven’t done our jobs.”


Super Bowl: Heard City’s audio post for Tide, Bud and more

By Jennifer Walden

New York audio post house Heard City put their collaborative workflow design to work on the Super Bowl ad campaign for Tide. Philip Loeb, partner/president of Heard City, reports that their facility is set up so that several sound artists can work on the same project simultaneously.

Loeb also helped to mix and sound design many of the other Super Bowl ads that came to Heard City, including ads for Budweiser, Pizza Hut, Blacture, Tourism Australia and the NFL.

Here, Loeb and mixer/sound designer Michael Vitacco discuss the approach and the tools that their team used on these standout Super Bowl spots.

Philip Loeb

Tide’s It’s a Tide Ad campaign via Saatchi & Saatchi New York
Is every Super Bowl ad really a Tide ad in disguise? A string of commercials touting products from beer to diamonds, and even a local ad for insurance, are interrupted by David Harbour (of Stranger Things fame). He declares that those ads are actually just Tide commercials, as everyone is wearing such clean clothes.

Sonically, what’s unique about this spot?
Loeb: These spots, four in total, involved sound design and mixing, as well as ADR. One of our mixers, Evan Mangiamele, conducted an ADR session with David Harbour, who was in Hawaii, and we integrated that into the commercial. In addition, we recorded a handful of different characters for the lead-ins for each of the different vignettes because we were treating each of those as different commercials. We had to be mindful of a male voiceover starting one and then a female voiceover starting another so that they were staggered.

There was one vignette for Old Spice, and since the ads were for P&G, we did get the Old Spice pneumonic and we did try something different at the end — with one version featuring the character singing the pneumonic and one of him whistling it. There were many different variations and we just wanted, in the end, to get part of the pneumonic into the joke at the end.

The challenge with the Tide campaign, in particular, was to make each of these vignettes feel like it was a different commercial and to treat each one as such. There’s an overall mix level that goes into that but we wanted certain ones to have a little bit more dynamic range than the others. For example, there is a cola vignette that’s set on a beach with people taking a selfie. David interrupts them by saying, “No, it’s a Tide ad.”

For that spot, we had to record a voiceover that was very loud and energetic to go along with a loud and energetic music track. That vignette cuts into the “personal digital assistant” (think Amazon’s Alexa) spot. We had to be very mindful of these ads flowing into each other while making it clear to the viewer that these were different commercials with different products, not one linear ad. Each commercial required its own voiceover, its own sound design, its own music track, and its own tone.

One vignette was about car insurance featuring a mechanic in a white shirt under a car. That spot isn’t letterbox like the others; it’s 4:3 because it’s supposed to be a local ad. We made that vignette sound more like a local ad; it’s a little over-compressed, a little over-equalized and a little videotape sounding. The music is mixed a little low. We wanted it to sound like the dialogue is really up front so as to get the message across, like a local advertisement.

What’s your workflow like?
Loeb: At Heard City, our workflow is unique in that we can have multiple mixers working on the same project simultaneously. This collaborative process makes our work much more efficient, and that was our original intent when we opened the company six years ago. The model came to us by watching the way that the bigger VFX companies work. Each artist takes a different piece of the project and then all of the work is combined at the end.

We did that on the Tide campaign, and there was no other way we could have done it due to the schedule. Also, we believe this workflow provides a much better product. One sound artist can be working specifically on the sound design while another can be mixing. So as I was working on mixing, Evan was flying in his sound design to me. It was a lot of fun working on it like that.

What tools helped you to create the sound?
One plug-in we’re finding to be very helpful is the iZotope Neutron. We put that on the master bus and we have found many settings that work very well on broadcast projects. It’s a very flexible tool.

Vitacco: The Neutron has been incredibly helpful overall in balancing out the mix. There are some very helpful custom settings that have helped to create a dynamic mix for air.

Tourism Australia Dundee via Droga5 New York
Danny McBride and Chris Hemsworth star in this movie-trailer-turned-tourism-ad for Australia. It starts out as a movie trailer for a new addition to the Crocodile Dundee film franchise — well, rather, a spoof of it. There’s epic music featuring a didgeridoo and title cards introducing the actors and setting up the premise for the “film.” Then there’s talk of miles of beaches and fine wine and dining. It all seems a bit fishy, but finally Danny McBride confirms that this is, in fact, actually a tourism ad.

Sonically, what’s unique about this spot?
Vitacco: In this case, we were creating a fake movie trailer that’s a misdirect for the audience, so we aimed to create sound design that was both in the vein of being big and epic and also authentic to the location of the “film.”

One of the things that movie trailers often draw upon is a consistent mnemonic to drive home a message. So I helped to sound design a consistent mnemonic for each of the title cards that come up.

For this I used some Native Instruments toolkits, like “Rise & Hit” and “Gravity,” and Tonsturm’s Whoosh software to supplement some existing sound design to create that consistent and branded mnemonic.

In addition, we wanted to create an authentic sonic palette for the Australian outback where a lot of the footage was shot. I had to be very aware of the species of animals and insects that were around. I drew upon sound effects that were specifically from Australia. All sound effects were authentic to that entire continent.

Another factor that came into play was that anytime you are dealing with a spot that has a lot of soundbites, especially ones recorded outside, there tends to be a lot of noise reduction taking place. I didn’t have to hit it too hard because everything was recorded very well. For cleanup, I used the iZotope RX 6 — both the RX Connect and the RX Denoiser. I relied on that heavily, as well as the Waves WNS plug-in, just to make sure that things were crisp and clear. That allowed me the flexibility to add my own ambient sound and have more control over the mix.

Michael Vitacco

In RX, I really like to use the Denoiser instead of the Dialogue Denoiser tool when possible. I’ll pull out the handles of the production sound and grab a long sample of noise. Then I’ll use the Denoiser because I find that works better than the Dialogue Denoiser.

Budweiser Stand By You via David Miami
The phone rings in the middle of the night. A man gets out of bed, prepares to leave and kisses his wife good-bye. His car radio announces that a natural disaster is affecting thousands of families who are in desperate need of aid. The man arrives at a Budweiser factory and helps to organize the production of canned water instead of beer.

Sonically, what’s unique about this spot?
Loeb: For this spot, I did a preliminary mix where I handled the effects, the dialogue and the music. We set the preliminary tone for that as to how we were going to play the effects throughout it.

The spot starts with a husband and wife asleep in bed and they’re awakened by a phone call. Our sound focused on the dialogue and effects upfront, and also the song. I worked on this with another fantastic mixer here at Heard City, Elizabeth McClanahan, who comes from a music background. She put her ears to the track and did an amazing job of remixing the stems.

On the master track in the Pro Tools session, she used iZotope’s Neutron, as well as the FabFilter Pro-L limiter, which helps to contain the mix. One of the tricks on a dynamic mix like that — which starts off with that quiet moment in the morning and then builds with the music in the end — is to keep it within the restrictions of the CALM Act and other specifications that stipulate dynamic range and not just average loudness. We had to be mindful of how we were treating those quiet portions and the lower portions so that we still had some dynamic range but we weren’t out of spec.


Jennifer Walden is a New Jersey-based audio engineer and writer. Follow her on Twitter @AudioJeney.


Genelec intros The Ones series of near-field monitors

Genelec is now offering point source monitoring with The Ones series, featuring the established 8351 three-way monitor along with the new 8341 and 8331. These small three-way coaxial monitors are housed in enclosures no larger than a traditional two-way Genelec 8040 or 8030. Their coaxial driver design provides accurate imaging and improved sound quality with clear accuracy, both on- and off-axis, vertically as well as horizontally. Also, there are no visible woofers.

Like the 8351, both the 8341 and 8331 can be orientated horizontally or vertically using an adjustable IsoPod base for isolation. But while the 8341 and 8331 both echo the 8351 in form and function, the new models have been entirely reengineered and feature ultra-compact dimensions: 13.78 in. x 9.33 in. x 9.57 in. [350 mm x 237 mm x 243 mm] for the 8341, and 11.77 in. x 7.44 in. x 8.70 in. [299 mm x 189 mm x 212 mm] for the 8331.

Innovations include a motor assembly that sees both the midrange and the tweeter share the same compact magnet system, reducing size and weight with no reduction in response. The midrange coaxial driver cone is now composed of concentric sections, optimizing midrange linearity — as does the DCW, which covers the entire front face of the enclosure. Despite the size of the 8341 and 8331, each unit incorporates three stages of dedicated Class D amplification.

Short-term maximum output capacity is 110 dB SPL for the 8341 (at 1 m) and 104 dB SPL for the 8331 (at 1 m), with accuracy better than ±1.5 dB, and respective frequency responses start at 45 Hz and 38 Hz (-6 dB) and extend beyond 40 kHz both for the analog and digital inputs.

The coaxial design allows for ultra-near-field listening, creating a dramatic improvement in the direct sound-to-reverberant sound ratio and further reducing the room’s influence while monitoring. The listening distance may be as short as 16 inches, with no loss of precision.

The Ones were recently used by Richard Chycki for his latest project, a 5.1 mix of The Tragically Hip – A National Celebration.


Capturing Foley for Epix’s Berlin Station

Now in its second season on Epix, the drama series Berlin Station centers on undercover agents, diplomats and whistleblowers inhabiting a shadow world inside the German capital.

Leslie Bloome

Working under the direction of series supervising sound editor Ruy Garcia, Westchester, New York-based Foley studio Alchemy Post Sound is providing Berlin Station with cinematic sound. Practical effects, like the clatter of weapons and clinking glass, are recorded on the facility’s main Foley stage. Certain environmental effects are captured on location at sites whose ambience is like the show’s settings. Interior footsteps, meanwhile, are recorded in the facility’s new “live” room, a 1,300-square-foot space with natural reverb that’s used to replicate the environment of rooms with concrete, linoleum and tile floors.

Garcia wants a soundtrack with a lot of detail and depth of field,” explains lead Foley artist and Alchemy Post founder Leslie Bloome. “So, it’s important to perform sounds in the proper perspective. Our entire team of editors, engineers and Foley artists need to be on point regarding the location and depth of field of sounds we’re recording. Our aim is to make every setting feel like a real place.”

A frequent task for the Foley team is to come up with sounds for high-tech cameras, surveillance equipment and other spy gadgetry. Foley artist Joanna Fang notes that sophisticated wall safes appear in several episodes, each one featuring differing combinations of electronic, latch and door sounds. She adds that in one episode a character has a microchip concealed in his suit jacket and the Foley team needed to invent the muffled crunch the chip makes when the man is frisked. “It’s one of those little ‘non-sounds’ that Foley specializes in,” she says. “Most people take it for granted, but it helps tell the story.”

The team is also called on to create Foley effects associated with specific exterior and interior locations. This can include everything from seedy safe houses and bars to modern office suites and upscale hotel rooms. When possible, Alchemy prefers to record such effects on location at sites closely resembling those pictured on-screen. Bloome says that recording things like creaky wood floors on location results in effects that sound more real. “The natural ambiance allows us to grab the essence of the moment,” he explains, “and keep viewers engaged with the scene.”

Footsteps are another regular Foley task. Fang points out that there is a lot of cat-and-mouse action with one character following another or being pursued, and the patter of footsteps adds to the tension. “The footsteps are kind of tough,” she says. “Many of the characters are either diplomats or spies and they all wear hard soled shoes. It’s hard to build contrast, so we end up creating a hierarchy, dark powerful heels for strong characters, lighter shoes for secondary roles.”

For interior footsteps, large theatrical curtains are used to adjust the ambiance in the live stage to fit the scene. “If it’s an office or a small room in a house, we draw the curtains to cut the room in half; if it’s a hotel lobby, we open them up,” Fang explains. “It’s amazing. We’re not only creating depth and contrast by using different types of shoes and walking surfaces, we’re doing it by adjusting the size of the recording space.”

Alchemy edits their Foley in-house and delivers pre-mixed and synced Foley that can be dropped right into the final mix seamlessly. “The things we’re doing with location Foley and perspective mixing are really cool,” says Foley editor and mixer Nicholas Seaman. “But it also means the responsibility for getting the sound right falls squarely on our shoulders. There is no ‘fix in the mix.’ From our point of view, the Foley should be able to stand on its own. You should be able to watch a scene and understand what’s going on without hearing a single line of dialogue.”

The studio used Neumann U87 and KMR81 microphones, a Millennia mic-pre and Apogee converter, all recorded into Avid Pro Tools on a C24 console. In addition to recording a lot of guns, Alchemy also borrowed a Doomsday prep kit for some of the sounds.

The challenge to deliver sound effects that can stand up to that level of scrutiny keeps the Foley team on its toes. “It’s a fascinating show,” says Fang. “One moment, we’re inside the station with the usual office sounds and in the next edit, we’re in the field in the middle of a machine gun battle. From one episode to the next, we never know what’s going to be thrown at us.”


Review: Blackmagic’s DaVinci Resolve 14 for editing

By Brady Betzel

Resolve 14 has really stepped up Blackmagic’s NLE game with many great new updates over the past few months. While I typically refer to Resolve as a high-end color correction and finishing tools, this review will focus on the Editing tab.

Over the last two years, Resolve has grown from a high-end color correction and finishing app to include a fully-capable nonlinear editor, media organizer and audio editing tool. Fairlight is not currently at the same level as Avid Pro Tools, but it is still capable, and with a price of free or at most $299 you can’t lose. For this review, I am using the $299 version, which has a few perks — higher than UHD resolutions; higher than 60 frames per second timelines; the all-important spatial and/or temporal noise reduction; many plugins like the new face tracker; multi-user collaboration; and much more. The free version will work with resolutions up to UHD at up to 60fps and still gives you access to all of the powerful base tools like Fairlight and the mighty color correction tool set.

Disclaimer: While I really will try and focus on the Editing tab, I can’t make any promises I won’t wander.

Digging In
My favorite updates to Resolve 14’s Editing tab revolve around collaboration and conforming functions, but I even appreciate some smaller updates like responsiveness while trimming and video scopes on the edit page. And don’t forget the audio waveforms being visible on the source monitor!

With these new additions, among others, I really do think that Resolve is also becoming a workable nonlinear editor much like industry standards such as Avid Media Composer, Adobe Premiere Pro and Apple Final Cut Pro X. You can work from ingest to output all within one app. When connected to a collaborative project there is now bin-locking, sharing bins and even a chat window.

Multicam works as expected with up to 16 cameras in one split view. I couldn’t figure out how to watch all of the angles in the source monitor while playing down the sequence in the record monitor, so I did a live switch (something I love to do in Media Composer). I also couldn’t figure out how to adjust the multi-cam after it had been created, because say, for instance, audio was one frame out of sync or I needed to add another angle later on. But the multicam worked and did its job by allowing me to sync by in point, out point, timecode, sound or marker. In addition, you can make the multicam a different frame rate than your timeline, which is handy.

[Editor’s Note: Blackmagic says: “There are a few ways to do that. You can right click on the multicam clip and select ‘open in timeline.’ Or you can pause over any segment of a multicam clip, click on a different angle and swap out the shots. Most importantly, you get into multicam edit mode by clicking on the drop down menu on the lower left hand corner of the source viewer and selecting Multicam mode.”]

Another addition is the Position Lock located in the middle right, above the timeline. The Position Lock keeps all of your clips locked in time in your timeline. What is really interesting about this is that it still allows you to trim and apply other effects to clips while locking the position of your clips in place. This is extremely handy when doing conforms and online passes of effects when you don’t want timing and position of clips to change. It’s a great safety net. There are some more fancy additions like re-time curves directly editable in the timeline. But what I would really love is a comprehensive overhaul of the Title Tool that would allow for direct manipulation of the text on top of the video. It would be nice to have a shortcut to use the title as a matte for other footage for some quick and fancy titling effects, but maybe that is what Fusion is for? The title tool works fine and will now give you nice crisp text even when blown up. The bezier curves really come in handy here to make animations ease in and out nicely.

If you start and finish within Resolve 14, your experience will most likely be pretty smooth. For anyone coming from another NLE — like Media Composer or Premiere — there are a few things you will have to get used to, but overall it feels like the interface designers of Resolve 14 kept the interface familiar for those “older” editors, yet also packed it with interesting features to keep the “YouTube” editors’ interest piqued. As someone who’s partial to Media Composer, I really like that you can choose between frame view in the timeline and clips-only view, leaving out thumbnails and waveform views in the timeline.

I noticed a little bit of a lag when editing with the thumbnail frames turned on. I also saw recently that Dave Dugdale on YouTube found an interesting solution to the possible bug. Essentially, one of the thumbnail views of the timeline was a little slower at re-drawing when zooming into a close view in a sequence Regardless, I like to work without thumbnails, and that view seemed to work fluidly for me.

After working for about 12 minutes I realized I hadn’t saved my work and Resolve didn’t auto-saved. This is when I remembered hearing about the new feature “Live Save.” It’s a little tricky to find, but the Live Save feature lives under the DaVinci Resolve Menu > User > Auto Save and is off by default — I really think this should be changed. Turn this fuction on and your Resolve project will continually save, which in turn saves you from unnecessary conniptions when your project crashes and you try to find the spot that was last saved.

Coming from another NLE, the hardest thing for me to get used to in a new app was the keyboard layouts and shortcuts. Typically, trimming works similar to other apps and overwriting; ripple edits, dissolves and other edit functions don’t change, but the placement of their shortcuts does. In Resolve 14, you can access the keyboard shortcut commands in the same spot as the Live Save, but under the Keyboard Mapping menu under User. From here you can get grounded quickly by choosing a preset that is similar to your NLE of choice — Premiere, FCP X, Media Composer — or Resolve’s default keyboard layout, which isn’t terrible. If this could be updated to how apps like Premiere and/or Avid have their keyboard layouts designed, it would be a lot easier to navigate. Meaning there is usually a physical representation of a keyboard that allows you to drag your shortcuts to and from it realtime.

Right now, Resolve’s keyboard mapper is text-based and a little cumbersome. Overall, Resolve’s keyboard shortcuts (when in the editing tab) are pretty standard, but it would do you well to read and go through basic moves like trimming, trimming the heads and tails of clips or even just trimming by plus or minus and the total frames you want to trim.

Something else I discovered when trimming was when you go into actual “trim mode,” it isn’t like other NLEs where you can immediately start trimming. I had to click on the trim point with my mouse or pen, then I could use keyboard shortcuts to trim. This is possibly a bug, but what I would really love to happen is when you enter “trim mode,” you would see trimming icons at the A and B sides of the nearest clips on the selected tracks. This would allow you to immediately trim using keyboard shortcuts without any mouse clicks. In my mind, the more mouse clicks I have to use to accomplish a task means time wasted. This leads to having less time to spend on “important” stuff like story, audio, color, etc. When time equals money, every mouse click means money out of my pocket. [Note from Blackmagic: “In our trim tools you can also enter trim mode by hitting T on the keyboard. We did not put in specific trim tool icons on purpose because we have an all-in-one content sensitive trim tool that changes based on where you place the cursor. And if you prefer trimming with realtime playback, hit W for dynamic trim mode, and then click on the cut you want to trim with before hitting JKL to play the trim.”]

I have always treated Resolve as another app in my post workflow — I wasn’t able to use it all the way from start to finish. So in homage to the old way of working, a.k.a. “a round trip workflow,” I wanted to send a Media Composer sequence to Resolve by way of a linked AAF, then conform the media clips and work from there. I had a few objectives, but the main one was to make sure my clips and titles came over. Next was to see if any third-party effects would translate into Resolve from Media Composer and, finally, I wanted to conform an “updated” AAF to the original sequence using Resolve’s new “Compare with Current Timeline” command.

This was a standard 1080p, 23.98 sequence (transcoded to one mezzanine DNx175x codec with 12 frame handles) with plenty of slates, titles, clips, speed ramps, Boris Continuum Complete and Sapphire Effect. Right off the bat all of the clip-based media came over fine and in its correct time and place in the timeline. Unfortunately, the titles did not come over and were offline — none of them were recognized as titles so they couldn’t be edited. Dissolves came over correctly, however none of the third-party BCC or Sapphire effects came across. I didn’t really expect the third-party effects to come over, but at some point, in order to be a proper conforming application, Resolve will need to figure out a way to translate those when sending sequences from an NLE to Resolve. This is more of a grand wish, but in order to be a force in the all-in-one app for the post finishing circle, this is a puzzle that will need to be solved.

Otherwise, for those who want to use alternative nonlinear editing systems, they will have to continue using their NLE as the editor, Resolve as a color-only solution, and the NLE as their finisher. And from what I can tell Blackmagic wants Resolve to be your last stop in the post pipeline. Obviously, if you start your edit in Resolve and use third-party OpenFX (OFX) like BCC or Sapphire, you shouldn’t have any problems.

Last on my list was to test the new Compare with Current Timeline command. In order for this option to pop up when you right click, you must be in the Media tab with the sequence you want to compare to the one loaded. You then need to find the sequence you want to compare from, right click on it and click Compare with Current Timeline. Once you click the sequences you want to compare, a new window will pop up with the option to view the Diff Index. The Diff Index is a text-based list of each new edit next to the timeline that visually compares your edits between the two sequences. This visual representation of the edits between the sequences is where you will apply those changes. There are marks identifying what has changed, and if you want to apply those changes you must right click and hit Apply Changes. My suggestion is to duplicate your sequence before you apply changes (actually you should be constantly duplicating your sequence as a backup as a general rule). The Compare with Current Timeline function is pretty incredible. I tested it using an AAF I had created in Media Composer and compared it against an AAF made from the same sequence but with some “creative” changes and trimmed clips — essentially a locked sequence that suddenly became unlocked while in Online/Color and needed to reflect the latest changes from the offline edit.

I wasn’t able to test Resolve 14 in a shared-project environment, so I couldn’t test a simultaneous update coming from another editor. But this can come in really handy for anyone who has to describe any changes made to a particular sequence or for that pesky online editor that needs to conform a new edit while not losing all their work.

I can’t wait to see the potential of this update, especially if we can get Resolve to recognize third-party effects from other NLEs. Now don’t get me wrong, I’m not oblivious to the fact that asking Resolve engineers to figure out how to recognize third-party effects in an AAF workflow is a pie-in-the-sky scenario. If it was easy it probably would have already been done. But it is a vital feature if Blackmagic wants Resolve to be looked at like a Flame or Media Composer but with a high-end coloring solution and audio finishing solution. While I’m at it, I can’t help but think that Resolve may eventually include Fusion as another tab maybe as a paid add-on, which would help to close that circle to being an all-in-one post production solution.

Summing Up
In the end, Resolve 14 has all the makings of becoming someone’s choice as a sole post workflow solution. Blackmagic has really stepped up to the plate and made a workable and fully functional NLE. And, oh yeah not to mention it is one of the top color correction tools being used in the world.

I did this review of the editing tab using Blackmagic Design’s DaVinci Resolve 14.2. Find the latest version here. And check out our other Resolve review — this one from a color and finishing perspective.


Brady Betzel is an Emmy-nominated online editor at Margarita Mix in Hollywood, working on Life Below Zero and Cutthroat Kitchen. You can email Brady at bradybetzel@gmail.com. Follow him on Twitter @allbetzroff.


Coco’s sound story — music, guitars and bones

By Jennifer Walden

Pixar’s animated Coco is a celebration of music, family and death. In the film, a young Mexican boy named Miguel (Anthony Gonzalez) dreams of being a musician just like his great-grandfather, even though his family is dead-set against it. On the evening of Día de los Muertos (the Mexican holiday called Day of the Dead), Miguel breaks into the tomb of legendary musician Ernesto de la Cruz (Benjamin Bratt) and tries to steal his guitar. The attempted theft transforms Miguel into a spirit, and as he flees the tomb he meets his deceased ancestors in the cemetery.

Together they travel to the Land of the Dead where Miguel discovers that in order to return to life he must have the blessing of his family. The matriarch, great-grandmother Mamá Imelda (Alanna Ubach) gives her blessing with one stipulation, that Miguel can never be a musician. Feeling as though he cannot live without music, Miguel decides to seek out the blessing of his musician great-grandfather.

Music is intrinsically tied to the film’s story, and therefore to the film’s soundtrack. Ernesto de la Cruz’s guitar is like another character in the film. The Skywalker Sound team handled all the physical guitar effects, from subtle to destructive. Although they didn’t handle any of the music, they covered everything from fret handling and body thumps to string breaks and smashing sounds. “There was a lot of interaction between music and effects, and a fine balance between them, given that the guitar played two roles,” says supervising sound editor/sound designer/re-recording mixer Christopher Boyes, who was just nominated for a CAS award for his mixing work on Coco. His Skywalker team on the film included co-supervising sound editor J.R. Grubbs, sound effects editors Justin Doyle and Jack Whittaker, and sound design assistant Lucas Miller.

Boyes bought a beautiful guitar from a pawn shop in Petaluma near their Northern California location, and he and his assistant Miller spent a day recording string sounds and handling sounds. “Lucas said that one of the editors wanted us to cut the guitar strings,” says Boyes. “I was reluctant to cut the strings on this beautiful guitar, but we finally decided to do it to get the twang sound effects. Then Lucas said that we needed to go outside and smash the guitar. This was not an inexpensive guitar. I told him there was no way we were going to smash this guitar, and we didn’t! That was not a sound we were going to create by smashing the actual guitar! But we did give it a couple of solid hits just to get a nice rhythmic sound.”

To capture the true essence of Día de los Muertos in Mexico, Boyes and Grubbs sent effects recordists Daniel Boyes, Scott Guitteau, and John Fasal to Oaxaca to get field recordings of the real 2016 Día de los Muertos celebrations. “These recordings were essential to us and director Lee Unkrich, as well as to Pixar, for documenting and honoring the holiday. As such, the recordings formed the backbone of the ambience depicted in the track. I think this was a crucial element of our journey,” says Boyes.

Just as the celebration sound of Día de los Muertos was important, so too was the sound of Miguel’s town. The team needed to provide a realistic sense of a small Mexican town to contrast with the phantasmagorical Land of the Dead, and the recordings that were captured in Mexico were a key building block for that environment. Co-supervising sound editor Grubbs says, “Those recordings were invaluable when we began to lay the background tracks for locations like the plaza, the family compound, the workshop, and the cemetery. They allowed us to create a truly rich and authentic ambiance for Miguel’s home town.”

Bone Collecting
Another prominent set of sounds in Coco are the bones. Boyes notes that director Unkrich had specific guidelines for how the bones should sound. Characters like Héctor (Gael García Bernal), who are stuck in the Land of the Dead and are being forgotten by those still alive, needed to have more rattle-y sounding bones, as if the skeleton could come apart easily. “Héctor’s life is about to dissipate away, just as we saw with his friend Chicharrón [Edward James Olmos] on the docks, so their skeletal structure is looser. Héctor’s bones demonstrated that right from the get-go,” he explains.

In contrast, if someone is well remembered, such as de la Cruz, then the skeletal structure should sound tight. “In Miguel’s family, Papá Julio [Alfonso Arau] comically bursts apart many times, but he goes back together as a pretty solid structure,” explains Boyes. “Lee [Unkrich] wanted to dig into that dynamic first of all, to have that be part of the fabric that tells the story. Certain characters are going to be loose because nobody remembers them and they’re being forgotten.”

Creating the bone sounds was the biggest challenge for Boyes as a sound designer. Unkrich wanted to hear the complexity of the bones, from the clatter and movement down to the detail of cartilage. “I was really nervous about the bones challenge because it’s a sound that’s not easily embedded into a track without calling attention to itself, especially if it’s not done well,” admits Boyes.

Boyes started his bone sound collection by recording a mobile he built using different elements, like real bones, wooden dowels, little stone chips and other things that would clatter and rattle. Then one day Boyes stumbled onto an interesting bone sound while making a coconut smoothie. “I cracked an egg into the smoothie and threw the eggshell into the empty coconut hull and it made a cool sound. So I played with that. Then I was hitting the coconut on concrete, and from all of those sources I created a library of bone sounds.” Foley also contributed to the bone sounds, particularly for the literal, physical movements, like walking.

According to Grubbs, the bone sounds were designed and edited by the Skywalker team and then presented to the directors over several playbacks. The final sound of the skeletons is a product of many design passes, which were carefully edited in conjunction with the Foley bone recordings and sometimes used in combination with the Foley.

L-R: J.R. Grubbs and Chris Boyes

Because the film is so musical, the bone tracks needed to have a sense of rhythm and timing. To hit moments in a musical way, Boyes loaded bone sounds and other elements into Native Instruments’ Kontakt and played them via a MIDI keyboard. “One place for the bones that was really fun was when Héctor went into the security office at the train station,” says Boyes.

Héctor comes apart and his fingers do a little tap dance. That kind of stuff really lent to the playfulness of his character and it demonstrated the looseness of his skeletal structure.”

From a sound perspective, Boyes feels that Coco is a great example of how movies should be made. During editorial, he and Grubbs took numerous trips to Pixar to sit down with the directors and the picture department. For several months before the final mix, they played sequences for Unkrich that they wanted to get direction on. “We would play long sections of just sound effects, and Lee — being such a student of filmmaking and being an animator — is quite comfortable with diving down into the nitty-gritty of just simple elements. It was really a collaborative and healthy experience. We wanted to create the track that Lee wanted and wanted to make sure that he knew what we were up to. He was giving us direction the whole way.”

The Mix
Boyes mixed alongside re-recording mixer Michael Semanick (music/dialogue) on Skywalker’s Kurosawa Stage. They mixed in native Dolby Atmos on a DFC console. While Boyes mixed, effects editor Doyle handled last-minute sound effects needs on the stage, and Grubbs ran the logistics of the show. Grubbs notes that although he and Boyes have worked together for a long time this was the first time they’ve shared a supervising credit.

“J.R. [Grubbs] and I have been working together for probably 30 years now.” Says Boyes. “He always helped to run the show in a very supervisory way, so I just felt it was time he started getting credit for that. He’s really kept us on track, and I’m super grateful to him.”

One helpful audio tool for Boyes during the mix was the Valhalla Room reverb, which he used on Miguel’s footsteps inside de la Cruz’s tomb. “Normally, I don’t use plug-ins at all when I’m mixing. I’m a traditional mixer who likes to use a console and TC Electronic’s TC 6000 and the Leixcon 480 reverb as outboard gear. But in this one case, the Valhalla Room plug-in had a preset that really gave me a feeling of the stone tomb.”

Unkrich allowed Semanick and Boyes to have a first pass at the soundtrack to get it to a place they felt was playable, and then he took part in the final mix process with them. “I just love Lee’s respect for us; he gives us time to get the soundtrack into shape. Then, he sat there with us for 9 to 10 hours a day, going back and forth, frame by frame at times and section by section. Lee could hear everything, and he was able to give us definitive direction throughout. The mix was achieved by and directed by Lee, every frame. I love that collaboration because we’re here to bring his vision and Pixar’s vision to the screen. And the best way to do that is to do it in the collaborative way that we did,” concludes Boyes.


Jennifer Walden is a New Jersey-based audio engineer and writer.


The 54th annual CAS Award nominees

The Cinema Audio Society announced the nominees for the 54th Annual CAS Awards for Outstanding Achievement in Sound Mixing. There are seven creative categories for 2017, and the Outstanding Product nominations were revealed as well.

Here are this year’s nominees:

Baby Driver

Motion Picture – Live Action

Baby Driver

Production Mixer – Mary H. Ellis, CAS

Re-recording Mixer – Julian Slater, CAS

Re-recording Mixer – Tim Cavagin

Scoring Mixer – Gareth Cousins, CAS

ADR Mixer – Mark Appleby

Foley Mixer – Glen Gathard

Dunkirk

Production Mixer – Mark Weingarten, CAS

Re-recording Mixer – Gregg Landaker

Re-recording Mixer – Gary Rizzo, CAS

Scoring Mixer – Alan Meyerson, CAS

ADR Mixer – Thomas J. O’Connell

Foley Mixer – Scott Curtis

Star Wars: The Last Jedi

Production Mixer – Stuart Wilson, CAS

Re-recording Mixer – David Parker

Re-recording Mixer – Michael Semanick

Re-recording Mixer – Ren Klyce

Scoring Mixer – Shawn Murphy

ADR Mixer – Doc Kane, CAS

Foley Mixer – Frank Rinella

The Shape of Water

Production Mixer – Glen Gauthier

Re-recording Mixer – Christian T. Cooke, CAS

Re-recording Mixer – Brad Zoern, CAS

Scoring Mixer – Peter Cobbin

ADR Mixer – Chris Navarro, CAS

Foley Mixer – Peter Persaud, CAS

Wonder Woman

Production Mixer – Chris Munro, CAS

Re-recording Mixer – Chris Burdon

Re-recording Mixer – Gilbert Lake, CAS

Scoring Mixer – Alan Meyerson, CAS

ADR Mixer – Nick Kray

Foley Mixer – Glen Gathard

 

Motion Picture Animated

The Lego Batman Movie

Cars 3

Original Dialogue Mixer – Doc Kane, CAS

Re-recording Mixer – Tom Meyers

Re-recording Mixer – Michael Semanick

Re-recording Mixer – Nathan Nance

Scoring Mixer – David Boucher

Foley Mixer – Blake Collins

Coco

Original Dialogue Mixer – Vince Caro

Re-recording Mixer – Christopher Boyes

Re-recording Mixer – Michael Semanick

Scoring Mixer – Joel Iwataki

Foley Mixer – Blake Collins

Despicable Me 3

Original Dialogue Mixer – Carlos Sotolongo

Re-recording Mixer – Randy Thom, CAS

Re-recording Mixer – Tim Nielson

Re-recording Mixer – Brandon Proctor

Scoring Mixer – Greg Hayes

Foley Mixer – Scott Curtis

Ferdinand

Original Dialogue Mixer – Bill Higley, CAS

Re-recording Mixer – Randy Thom, CAS

Re-recording Mixer – Lora Hirschberg

Re-recording Mixer – Leff Lefferts

Scoring Mixer – Shawn Murphy

Foley Mixer – Scott Curtis

The Lego Batman Movie

Original Dialogue Mixer – Jason Oliver

Re-recording Mixer – Michael Semanick

Re-recording Mixer – Gregg Landaker

Re-recording Mixer – Wayne Pashley

Scoring Mixer – Stephen Lipson

Foley Mixer – Lisa Simpson

 

Motion Picture – Documentary

An Inconvenient Sequel: Truth to Power

Production Mixer – Gabriel Monts

Re-recording Mixer – Kent Sparling

Re-recording Mixer – Gary Rizzo, CAS

Re-recording Mixer – Zach Martin

Scoring Mixer – Jeff Beal

Foley Mixer – Jason Butler

Long Strange Trip

Eric Clapton: Life in 12 Bars

Re-recording Mixer – Tim Cavagin

Re-recording Mixer – William Miller

ADR Mixer – Adam Mendez, CAS

Gaga: Five Feet Two

Re-recording Mixer – Jonathan Wales, CAS

Re-recording Mixer – Jason Dotts

Jane

Production Mixer – Lee Smith

Re-recording Mixer – David E. Fluhr, CAS

Re-recording Mixer – Warren Shaw

Scoring Mixer – Derek Lee

ADR Mixer – Chris Navarro, CAS

Foley Mixer – Ryan Maguire

Long Strange Trip

Production Mixer – David Silberberg

Re-recording Mixer – Bob Chefalas

Re-recording Mixer – Jacob Ribicoff

 

Television Movie Or Mini-Series

Big Little Lies: “You Get What You Need”

Production Mixer – Brendan Beebe, CAS

Re-recording Mixer – Gavin Fernandes, CAS

Re-recording Mixer – Louis Gignac

Black Mirror: “USS Callister”

Production Mixer – John Rodda, CAS

Re-recording Mixer – Tim Cavagin

Fargo

Re-recording Mixer – Dafydd Archard

Re-recording Mixer – Will Miller

ADR Mixer – Nick Baldock

Foley Mixer – Sophia Hardman

Fargo: ”The Narrow Escape Problem”

Production Mixer – Michael Playfair, CAS

Re-recording Mixer – Kirk Lynds, CAS

Re-recording Mixer – Martin Lee

Scoring Mixer – Michael Perfitt

Sherlock: “The Lying Detective”

Production Mixer –John Mooney, CAS

Re-recording Mixer – Howard Bargroff

Scoring Mixer – Nick Wollage

ADR Mixer – Peter Gleaves, CAS

Foley Mixer – Jamie Talbutt

Twin Peaks: “Gotta Light?”

Production Mixer – Douglas Axtell

Re-recording Mixer –Dean Hurley

Re-recording Mixer – Ron Eng

 

Television Series – 1-Hour

Better Call Saul: “Lantern”

Production Mixer – Phillip W. Palmer, CAS

Re-recording Mixer – Larry B. Benjamin, CAS

Re-recording Mixer – Kevin Valentine

ADR Mixer – Matt Hovland

Foley Mixer – David Michael Torres, CAS

Game of Thrones: “Beyond the Wall”

Game of Thrones

Production Mixer – Ronan Hill, CAS

Production Mixer – Richard Dyer, CAS

Re-recording Mixer – Onnalee Blank, CAS

Re-recording Mixer – Mathew Waters, CAS

Foley Mixer – Brett Voss, CAS

Stranger Things: “The Mind Flayer”

Production Mixer – Michael P. Clark, CAS

Re-recording Mixer – Joe Barnett

Re-recording Mixer – Adam Jenkins

ADR Mixer – Bill Higley, CAS

Foley Mixer – Anthony Zeller, CAS

The Crown: “Misadventure”

Production Mixer – Chris Ashworth

Re-recording Mixer – Lee Walpole

Re-recording Mixer – Stuart Hilliker

Re-recording Mixer – Martin Jensen

ADR Mixer – Rory de Carteret

Foley Mixer – Philip Clements

The Handmaid’s Tale: “Offred”

Production Mixer – John J. Thomson, CAS

Re-recording Mixer – Lou Solakofski

Re-recording Mixer – Joe Morrow

Foley Mixer – Don White

 

Television Series – 1/2 Hour

Ballers: “Yay Area”

Production Mixer – Scott Harber, CAS

Re-recording Mixer – Richard Weingart, CAS

Re-recording Mixer – Michael Colomby, CAS

Re-recording Mixer – Mitch Dorf

Black-ish: “Juneteenth, The Musical”

Production Mixer – Tom N. Stasinis, CAS

Re-recording Mixer – Peter J. Nusbaum, CAS

Re-recording Mixer – Whitney Purple

Modern Family: “Lake Life”

Production Mixer – Stephen A. Tibbo, CAS

Re-recording Mixer – Dean Okrand, CAS

Re-recording Mixer – Brian R. Harman, CAS

Silicon Valley: “Hooli-Con”

Production Mixer – Benjamin A. Patrick, CAS

Re-recording Mixer – Elmo Ponsdomenech

Re-recording Mixer – Todd Beckett

Veep: “Omaha”

Production Mixer – William MacPherson, CAS

Re-recording Mixer – John W. Cook II, CAS

Re-recording Mixer – Bill Freesh, CAS

 

Television Non-Fiction, Variety Or Music Series Or Specials

American Experience: “The Great War – Part 3”

Production Mixer – John Jenkins

Re-Recording Mixer – Ken Hahn

Anthony Bourdain: Parts Unknown: “Oman”

Re-Recording Mixer – Benny Mouthon, CAS

Anthony Bourdain: Parts Unknown

Deadliest Catch: “Last Damn Arctic Storm”

Re-Recording Mixer – John Warrin

Rolling Stone: “Stories from the Edge”

Production Mixer – David Hocs

Production Mixer – Tom Tierney

Re-Recording Mixer – Tom Fleischman, CAS

Who Killed Tupac?: “Murder in Vegas”

Production Mixer – Steve Birchmeier

Re-Recording Mixer – John Reese

 

Nominations For Outstanding Product – Production

DPA – DPA Slim

Lectrosonics – Duet Digital Wireless Monitor System

Sonosax – SX-R4+

Sound Devices – Mix Pre- 10T Recorder

Zaxcom – ZMT3-Phantom

 

Nominations For Outstanding Product – Post Production

Dolby – Dolby Atmos Content Creation Tools

FabFilter – Pro Q2 Equalizer

Exponential Audio – R4 Reverb

iZotope – RX 6 Advanced

Todd-AO – Absentia DX

The Awards will be presented at a ceremony on February 24 at the Omni Los Angeles Hotel at California Plaza. This year’s CAS Career Achievement Award will be presented to re-recording mixer Anna Behlmer, the CAS Filmmaker Award will be given to Joe Wright and the Edward J. Greene Award for the Advancement of Sound will be presented to Tomlinson Holman, CAS. The Student Recognition Award winner will also be named and will receive a cash prize.

Main Photo: Wonder Woman

Dynasty composer Paul Leonard-Morgan

By Randi Altman

Scottish-born composer Paul Leonard-Morgan, who owns a BAFTA award and Emmy nomination, has a resume that is as eclectic as it is long. He has worked on television (Limitless), films (The Numbers Station) and games (Dawn of War III). He has also produced music for artists such as No Doubt (Push and Shove).

In addition to the Wormwood miniseries for Netflix, one of Leonard-Morgan’s most recent projects is the soundtrack for The CW’s reboot of the show Dynasty. We recently reached out to him to talk about the show, the way he works and what’s next.

L-R: Dynasty showrunner Sallie Patrick, Paul Leonard-Morgan and director Brad Sieberling with various musicians.

The name Dynasty comes with certain expectations and history. Did you use the original as an inspiration or borrow bits from the original as an homage?
I remember watching Dynasty as a child, but other than the main theme I couldn’t begin to tell you what the music was like, other than it was pretty orchestral — Bill Conti is such a phenomenal composer. So right from the outset our showrunner Sallie Patrick and director Brad Sieberling and I wanted to do a title sequence with a modernized version of the iconic theme. People don’t tend to do title sequences these days, so it was very cool of The CW to let us do it.

We got a bunch of players into Capitol Studios and overlaid the orchestra onto my beats and synths. I brought in an old friend and Grammy-winning producer Troy Nokaan to pump up the beats a bit. And, of course, there was Tom (Hooten), principal trumpet player with the LA Philharmonic. For me, this is what the whole series’ ethos is about — tying the old to the new. Recording these players in the iconic Capitol Studios, where people like Sinatra recorded… we got such a vintage vibe going on. But then we added modern beats and synths – that’s what the whole score has become. Adding a cool ‘80s twist to modern sounds and orchestra. But other than the titles, the rest of the score does its own thing.

Can you talk about what the show’s producers wanted for the score? Did you have a lot of input?
We had detailed discussions at the start about what we wanted to achieve. Everything to do with the ‘80s is so trendy now — from fashion to music, but there’s a fine line between adding ’80s elements to give the music a nice edge, and creating an ’80s pastiche, which sounds dated.

I produce a lot of bands, so I started taking some of those beats and then adding in lots of analog synths. And then our scoring sessions added an orchestra. I was really keen to use a string section, as I felt that Dynasty is so iconic, giving it a small section would add that touch of class to it. The beats — the clicks, claps and kicks — are what gives the Fallon character her swagger — the synths give it the pace, and the orchestra gives it the cinematic quality. I was keen to find a sound that would become instantly recognizable as that Dynasty sound.

How would you describe your score? 
Unique!

Can you walk us through your process? How do you begin? What inspires you? 
I start by watching the episode with the director, editor and writer and then have a spotting session. We work out where the music should come in and out, but even that is open to interpretation, as sometimes their vision might be different from mine. They might imagine short musical cues, where I’m envisaging longer, shaped pieces.

For example, there’s a piece in the episode I’ve just finished (110) that lasts the entire part 4. Obviously, it’s not full-on drums the whole time, but doing cues like that give it some real shape and add to the visuals filmic qualities. After the spotting sessions, I go away and start writing. After a while, you get a feel for what’s working and what’s not — when to leave the dialogue alone and when to try and help it. We’re all pretty keen on not making the music too emotionally leading in this series. We want to let the acting do that, instead of sign-posting every happy/sad moment. When everyone’s happy, we’ll start orchestrating the music, get the parts ready, and then go off to Capitol, or another studio, to record the real players.

The schedule is pretty crazy — I have a week to score each episode. So while we’re recording the real players, the dub is in its final day. As we finish mixing each cue, we then start sending them over the Internet to the dub stage, where they quickly lay them in and balance the levels with dialogue and FX. They’re lucky that I don’t get the chance to go and sit in the dub much, as we’re literally mixing to the last second!

What tools do you use to create a score?
I use MOTU’s Digital Performer to write, produce and pre-mix, then everything gets transferred to Avid ProTools for the main recording session and final mix. Obviously, I have a million samples and lots of original analog synths.

You work in many different parts of the music world — TV, films and games. Do you have a preference? How are those hats different, or are they not very different at all?
It sounds like a cop-out, but I really don’t have a preference. I like working in different fields, as I always feel that brings a freshness and different take to the next project, consciously and sub-consciously. For example, I was scoring a series of plays for The National Theatre in London a few years ago — at the same time I was scoring the film Walking With Dinosaurs in LA and the game Battlefield Hardline — and that theatre score was so different from many things I’d done before. But it led to me working with the incredible filmmaker Errol Morris for his film The B Side, and subsequently his new Netflix series Wormwood.

Dynasty came more from my work with bands. I like working in different genres, as it keeps pushing me out of my comfort zone, which I feel is really important as an artist.

You are building a new studio. Can you talk about that?
It’s been a process! Two weeks to go! Before I moved to LA with my family, I had just completed building my studio in Glasgow, Scotland. Then we moved over here, as I was living on planes between the UK and the US. This was about three years ago. I’ve been renting a studio, but finally the time came to buy a house and it’s got a huge guesthouse in the backyard (2,000 square feet), so I decided to get it properly treated.

We pulled down most of the inside and spent the last six months soundproofing and giving it the proper acoustic treatments, etc. But it’s insane, as I’ve hardly been out of my studio in Santa Monica while the build process has been going on, so the contractors have been FaceTiming me to show me how the progress is going, Trying to make decisions after a week of 20-hour days is hard.

I was keen to move to a place that had birds and nature. Coming from Scotland I like my space, which is not the easiest thing to find in LA. I insisted on having tons of windows in the studios for daylight to pour in — something that is great for me, but awful acoustically, so the acoustic guys spent weeks designing it so the glass wouldn’t affect the sound! But it’s looking fantastic, and I’ll have the ability to record up to 20 players in there. The irony is, having moved to what I thought was a pretty quiet neighborhood, I have a mega-famous hip-hop artist right next to me. His soundproofing had better be as good as mine!

What’s next for you project-wise?
Other than the rest of the season on Dynasty (we’re not even halfway there yet!), I’m working on a game score for the next year and a half, and have a new film starting in the New Year. I’ll also be working with my team on The Grand Tour, Amazon’s big series. Errol Morris’ Wormwood was recently released on Netflix — that’s been a life highlight for me!