In music, the journey from intention to sound is a dance of knobs and ears. An artist tweaks filters, envelopes, and waveforms, chasing a target that only exists in memory—the moment when a synth sounds exactly like what the brain hears. Today’s researchers are teaching computers to join that dance, not by replacing the musician but by building a feedback loop that nudges a synthesizer toward a target sound. The question they ask is deceptively simple: is there one universal way to measure how close two sounds are?
From the University of Alberta, a team led by Amir Salimi with Abram Hindle and Osmar R. Zaïane set out to test whether a single loss function can reliably guide a wide range of synthesis methods. Their answer is a nuanced, surprisingly human one: no. The best way to compare sounds depends on the instrument you’re tuning, the sonic goal you’re chasing, and the specific musical tool you’re using. It’s a reminder that even in the age of clever algorithms, sound design remains a creative negotiation between machine and maker.
A map of the sound-matching landscape
Sound-matching is the art of guiding a digital instrument to reproduce a target sound. It’s not about mimicking every knob turn; it’s about converging toward a sonic fingerprint that captures the essence of the target. When the process is automated, you’re playing the same back-and-forth loop humans use in the studio: listen, adjust, listen again, adjust again—until the target feels within reach.
To formalize this loop, you need four moving parts. First, a differentiable synthesizer g that takes a set of parameters theta and outputs audio x. Second, a target sound t you want to imitate. Third, a representation function phi that turns sounds into a form engineers can compare. And fourth, a loss L that measures how far the synthesized sound is from the target in that representation space. The elegant twist in differentiable sound-matching is that you can tighten the loop with gradient-based optimization, nudging theta in the direction that reduces L.
The authors push this idea with a deliberate twist: they explore how the choice of loss and the choice of synthesis interact. If L is a hammer, the shape of the nail is the synthesis method g. They argue that, historically, people have tested losses in narrow settings with a single or very similar synths, then claimed a winner. But in the wild world of actual sound design, the same loss that shines on one machine can falter on another. The study deliberately builds four differentiable synthesizers—representing subtractive, additive, and modulation-based approaches—and pairs each with four different losses. The goal isn’t to crown a universal champion, but to map how the landscape changes when the instrument changes.
Four losses, four synths, a forest of trials
The experiment centers on four differentiable loss functions: two that live in the spectrum domain and two that are more time- or envelope-focused. Three losses hailing from tradition—L1_Spec and SIMSE_Spec, both spectrogram-based, plus JTFS, which uses a joint time-frequency representation—sit alongside a fourth, the DTW_Envelope loss, which borrows dynamic time warping to align envelopes over time. Each loss is applied to one of four differentiable synthesizer programs that span common design philosophies: subtractive, additive, and two AM/FM-like modulations. The four synths are BP-Noise (a bare-bones subtractive path using a band-limited filter on noise), Add-SineSaw (an additive blend of a sine and a sawtooth), Noise-AM (noise modulated by a low-frequency oscillator), and SineSaw-AM (a blended carrier that has amplitude modulation behaving like an LFO for a carved timbre).
That setup is not an academic curiosity. It’s a deliberate attempt to test the core intuition: does a single metric truly do the work across the diverse ways people sculpt sound? The team ran 300 randomized sound-matching trials for every pairing of loss and synthesizer, with a maximum of 200 gradient steps per trial. They assessed output similarity in three ways: an automatic assessment using P-Loss (the distance between parameter vectors) and MSS (a spectrogram-based distance), plus a blind listening test where two authors scored final outputs on a five-point scale. The alignment between automatic scores and human judgments turned out to be surprisingly strong, though not perfect—an encouraging sign that these measures capture something meaningful about sonic similarity, but not a replacement for listening with human ears.
What emerged is a clear pattern: the performance of a loss function is not universal. For the BP-Noise synthesizer, spectrogram-based losses reigned, with SIMSE_Spec and L1_Spec often near the top in automatic and human rankings. For Add-SineSaw, JTFS consistently rose to the top across all evaluation methods, suggesting that a richer time-frequency representation can better capture the subtleties of additive timbres. Noise-AM and SineSaw-AM produced yet different winners, with DTW_Envelope frequently outshining traditional spectral losses in those setups. The big takeaway is not that one loss is inherently superior, but that the fit between loss and synth is a defining constraint on success.
Beyond the headline results, the analysis included a thorough post hoc look at the statistics. The researchers used non-parametric tests to see whether differences in performance were meaningful across losses for each synth, and they found many cases where the ranking of losses differed by program. In other words, the same loss can be excellent with one synth and mediocre with another, reinforcing the central claim: synthesis choice matters as much as, if not more than, the loss choice.
Implications: what this changes about designing sound with AI
The most striking implication is perhaps the simplest: there is no one-size-fits-all loss function for differentiable, iterative sound matching. If you want to replicate a target sound on a particular machine, you might pick one loss; switch the synth, and you should consider a different one. That undermines the dream of a universal, plug-and-play metric that dethrones human judgment. Instead, the research nudges us toward a more nuanced, instrument-aware framework for metric design.
But the story isn’t a cautionary tale about the limits of automation. It’s a celebration of creativity in the metric space. The study finds that two losses not always preferred in prior literature—DTW_Envelope and SIMSE_Spec—delivered strong results in several configurations. This suggests there is meaningful room for experimentation in how we quantify similarity, not just in how we model the synthesizer. If the goal is to empower sound designers to explore more expressive timbres and more faithful or more interpretive reproductions, expanding the toolbox matters as much as refining the models themselves.
Another meaningful thread is the emphasis on diverse synthesis methods. Previous work often stuck to a narrow slice of DSP, which can yield a narrow view of what works. By testing across subtractive, additive, and modulation-based paths, the researchers reveal a broader ecology of interactions between loss landscapes and parameter spaces. The practical upshot is clear: a developer building a sound-matching tool should anticipate that the best learning objective will depend on the instrument it’s interfacing with, and it should be designed to accommodate multiple synthesis styles rather than optimize for a single archetype.
There is a larger cultural point tucked into these findings. Sound design is a creative practice at its core, and even when we lean on gradient descent and differentiable DSP, the human ear, memory, and taste remain essential. The work invites future researchers to imagine loss functions that adapt to a designer’s evolving goals, perhaps through feedback loops that learn from human preferences or through reinforcement learning that navigates the gradient landscape with a more exploratory, artist-friendly mindset. It also hints at practical directions for the near term: toolmakers should provide practitioners with a palette of losses and a choice of synthesis engines, and encourage empirical testing across real-world sound targets rather than rely on a single benchmark scenario.
From a developer’s perspective, the takeaway is pragmatic: ship a toolkit, not a single hammer. If you want to help musicians sculpt sounds with AI, you’ll want to support multiple loss representations and a family of differentiable synths, plus straightforward ways to test them with actual ears. The era of chasing a single state-of-the-art metric is giving way to an era of instrument-aware, human-centered design where the metric itself can be as expressive as the sounds it helps create.
In the end, the study’s core message resonates beyond the lab: sound is memory, emotion, and craft. The way we measure and tune it should reflect that depth, not pretend the world is a flat field where any tool will do. If there is a single lesson, it is this: let the instrument speak. The metric should listen—and adapt.
Institution and authors: the work was conducted by researchers at the University of Alberta, with Amir Salimi as the lead author and collaborators Abram Hindle and Osmar R. Zaïane. Their efforts illuminate a future where the artistry of sound design and the precision of machine learning work in concert, not in competition.