Problem with Bolt-On MMMs
MTA and MMM were built for fundamentally different jobs. If your measurement provider started in MTA, here's what might be happening inside your model.
Linnea Zielinski · 10 min read
If your MMM provider started in MTA, here's what might be happening with your data
Think about what happens when a deep-sea fishing boat gets retrofitted for river tours. The crew knows water. They know boats. But the vessel was designed for open ocean conditions, and no amount of repainting changes the fact that the hull, the engine size, and the navigation systems were all built for a fundamentally different environment. It works, sort of. But every trip comes with tradeoffs the brochure doesn't mention.
Something similar happens when a marketing measurement platform built around multi-touch attribution (MTA) adds a marketing mix model (MMM) on top. The team understands measurement. They understand data. But the architecture underneath, the assumptions baked into how data flows, how channels get evaluated, and what the model is trained to look for, was built for a different job. And that shapes everything.
For brands making real budget decisions on the back of MMM outputs, knowing where those outputs came from matters.
Key takeaways
- MTA and MMM are built on fundamentally different assumptions about data. MTA tracks individual users across touchpoints; MMM works from aggregated signals and doesn't require individual tracking at all.
- When an MMM is layered on top of MTA infrastructure, it can inherit a bias toward observable digital signals, the very signals an MMM was designed to exist independently of.
- MTA-origin models often treat channels as independent contributors to revenue, which creates structural blind spots around how upper-funnel and lower-funnel activity actually interact.
- This architecture can systematically undervalue upper-funnel investment because MTA is, by nature, a bottom-of-funnel tool focused on what's measurable near the point of conversion.
- Cross-channel halo effects—the spillover revenue that awareness spend generates in branded search, organic traffic, direct traffic, and retail—are often missed entirely when a model is structured around per-channel credit assignment.
- Saturation assumptions inherited from MTA can push brands toward pulling back spend earlier than necessary, leaving real scaling opportunity on the table.
- The right question to ask isn't just "does my provider offer an MMM?" It's "what was the model actually built to do, and does that match what I need it to measure?"
MTA and MMM are built for different jobs
It's easy to treat MTA and MMM as two versions of the same thing: both are measurement approaches, both deal with marketing channels, both try to connect spend to revenue. But the questions they were designed to answer are genuinely different, and so are the assumptions they run on. (We also have a deep dive into MMM vs MTA if you want to read further.)
What MTA was designed to do
Multi-touch attribution was built to answer a user-level question: across the touchpoints a specific customer encountered before converting, how should we distribute credit? It traces individual journeys—a click on a Meta ad, a branded search, a direct visit—and assigns weight to each one based on whatever rule the model applies (last touch, linear, time-decay, and so on).
That architecture depends heavily on being able to observe individual users. Cookies, pixels, device IDs, and identity graphs are the infrastructure MTA runs on. When those signals are clean and complete, MTA can tell you a lot about the path to conversion. When they're degraded—which is increasingly the reality as privacy restrictions tighten—the model's accuracy degrades with them.
What MMM was designed to do
Marketing mix modeling takes a completely different approach. Instead of tracking individuals, MMM works at the aggregate level: it looks at how changes in spend across channels and campaigns correspond to changes in total revenue over time. It doesn't need to follow a single user anywhere. That's not a limitation. It's the point of this technology. MMM was specifically designed to answer portfolio-level questions without relying on individual tracking, which is why it holds up better in a privacy-constrained world.
The questions an MMM is built to answer are also broader. Not just "which touchpoint before this conversion should get credit," but "how does running more YouTube spend affect our branded search volume three weeks from now?" or "what's the real contribution of this awareness campaign after accounting for everything else going on in the business?" Those are fundamentally different questions, and they require fundamentally different modeling assumptions.
Why the architecture underneath matters
Understanding that these tools were built for different purposes isn't just an intellectual exercise. The assumptions a platform was founded on get baked into how the model is structured, what data it prioritizes, and how it interprets the signals it sees. When an MMM layer gets built on top of MTA infrastructure, those inherited assumptions don't disappear.
The pull toward observable digital signals
MTA providers have years of data pipelines, integrations, and organizational muscle built around user-level digital signals. That orientation shapes how data gets ingested, cleaned, and weighted. When an MMM is built on top of that foundation, it can carry a pull toward the same observable digital signals—platform-reported clicks, session data, last-touch conversions—that MTA depended on to function.
The problem is that an MMM's value is precisely that it doesn't need to rely on those signals. A well-specified MMM can measure the revenue contribution of a TV campaign, an out-of-home placement, or an awareness-only digital campaign that generated almost no direct clicks but meaningfully lifted branded search volume in the weeks that followed. If the model's underlying orientation is toward observable digital events, that broader picture gets distorted.
Channel independence and why it breaks things
MTA models assign credit channel by channel. Each touchpoint in a customer's journey gets its own credit allocation. The model is designed to treat channels as discrete contributors to a conversion event.
The issue is that real marketing systems don't work that way. Upper-funnel spend changes how lower-funnel channels perform. A heavy Meta prospecting push in October means your retargeting campaigns in November are reaching warmer audiences, so their conversion rates go up. A YouTube campaign that ran for six weeks has been slowly building brand familiarity, which means your paid search click-through rates start improving as more people recognize the brand name when they see the ad. These aren't independent contributions, they're interactions, and an MMM that treats channels as isolated from one another will get the math wrong on all of them.
When an MMM inherits the channel-independence framing of MTA, it can produce attributions that look internally consistent but systematically misrepresent how marketing is actually driving revenue. You end up with a model that's confident and stable-looking but pointing in the wrong direction.
Funnel blindness and what gets undervalued
MTA is, structurally, a bottom-of-funnel tool. Its inputs are the events closest to conversion, including clicks, sessions, and direct purchases. Upper-funnel activity shows up in MTA primarily when it produces a trackable touchpoint, which is often not how awareness spend works. Someone who saw your CTV ad and later searched your brand name directly probably doesn't appear in your MTA model as a CTV-influenced conversion. They appear as a branded search conversion, or a direct visit.
An MMM built on MTA infrastructure often carries this same funnel orientation. The model's inductive "muscle memory" points toward what's measurable near the point of conversion. Upper-funnel investment—prospecting, awareness, reach-based campaigns—tends to look less effective than it actually is because the model isn't structured to trace the longer, less linear path between an awareness impression and a downstream sale.
This isn't a small discrepancy. Brands that consistently undervalue upper-funnel spend based on their measurement data end up cutting the channels that are quietly feeding their entire conversion funnel. The damage tends to be slow and hard to attribute to the measurement platform, but it's real.
What this looks like in practice
These structural issues show up in the outputs that marketing teams actually use to make decisions.
Attribution that shifts under small changes
A well-specified MMM produces stable attribution when the underlying data is stable. If you extend a date range, rerun the model with slightly different configurations, or look at a different time window, the story shouldn't change dramatically. Channel contributions might shift a bit at the margins, but the core picture should hold.
One sign that a model may be structurally misspecified is when outputs are unusually sensitive to small changes in inputs. If swapping a date range shifts your Meta contribution by 30%, or if adding a few weeks of data reshuffles your channel rankings, that's not just noise. It's often a sign that the model is resolving an ambiguity in the data based on its own internal assumptions rather than on signal in the data itself. Those assumptions, in an MTA-origin model, may be optimized for a different kind of measurement problem than the one you're actually trying to solve.
Saturation calls that push you toward underspending
MTA click data tends to show diminishing returns as a channel gets more exposure to the same audience pool…because it does. More impressions to an audience that's already seen your ad several times produces fewer new clicks. That's a real phenomenon in the digital click environment. But it's not the same thing as true spend saturation at the campaign or channel level.
An MMM that inherits this framing can apply overly aggressive saturation assumptions, telling brands to pull back spend earlier than the data actually warrants. If your model is consistently recommending budget cuts across your highest-spend channels, it's worth asking whether those recommendations are coming from a genuine read on diminishing returns or from an assumption about saturation that was calibrated for click-based data rather than for revenue-level modeling.
Cross-channel halo effects going unmeasured
One of the most consequential things a strong MMM does is trace the full revenue contribution of a campaign, including the revenue it generates indirectly. When a prospecting campaign on Meta runs successfully, some of its impact shows up as direct conversions (people who clicked and purchased). But more of its impact often shows up elsewhere: in a lift in branded search volume, in an uptick in direct traffic, in stronger organic search performance, and sometimes in Amazon sales that have no trackable connection to the original ad.
These spillover effects, what Prescient calls marketing halo effects, are only visible to a model that's looking at the full revenue picture simultaneously rather than assigning credit channel by channel. A model structured around per-channel attribution tends to miss them, which means the campaigns generating the most halo revenue consistently look undervalued in the attribution output. Brands running on that data end up making scaling decisions against a picture that's materially incomplete.
Questions worth asking your provider
Before trusting budget decisions to any measurement platform, it's worth understanding what the model was actually built to do. That doesn't require a deep technical conversation; a few pointed questions can surface a lot. Ask:
- How the model handles upper-funnel and lower-funnel channels differently.
- What happens to attribution when a brand runs a heavy awareness push and conversion performance lifts in the weeks that follow (where does that revenue get credited?)
- Whether channel contributions are modeled as independent of each other or whether the model accounts for how spending in one channel changes the performance of another.
The answers to those questions tell you a lot about the model's underlying architecture. A platform that struggles to answer them, or defaults quickly to "trust the output," may not have a strong answer because the model genuinely isn't structured to handle those dynamics. That's worth knowing before you make your next media budget call.
Where Prescient comes in
Prescient's MMM was built from the ground up as a marketing mix model, not adapted from another methodology or an open-source framework. That means its core architecture is designed to reflect how marketing systems actually behave: channels interact, upper-funnel spend shapes lower-funnel performance, and the same dollar can produce very different results depending on context and timing. Our MMM models these dynamics directly rather than treating them as noise, which is why it can surface the halo effects that other models miss and produce stable attribution even across complex, multi-channel spending environments.
For brands that want to see the full picture of what their marketing spend is doing, that foundation matters. If you'd like to see how Prescient's approach compares to what you're working with now, book a demo.
FAQs
Can an MTA provider build a good MMM?
It's possible, but it requires building the MMM as a genuinely separate product with its own modeling architecture rather than extending existing MTA infrastructure. The challenge is organizational and technical: years of investment in user-level tracking, pixel-based data pipelines, and click-oriented measurement tend to shape how a team thinks about measurement problems. Building an MMM that doesn't carry any of that orientation requires deliberately starting from different assumptions, which is a harder lift than it sounds and not something every vendor does.
What's the difference between MTA and MMM data inputs?
MTA relies primarily on user-level digital signals: cookie data, device IDs, click events, and session-level behavior tracked across touchpoints. MMM works from aggregated inputs: total spend by channel and campaign, total revenue, and external factors like promotions, seasonality, and macroeconomic conditions. Because MMM doesn't depend on individual tracking, it's not affected by ad blockers, cookie deprecation, or iOS privacy changes in the same way MTA is.
Why does it matter if my MMM treats channels as independent?
Because they aren't. Upper-funnel channels like awareness and prospecting influence how lower-funnel channels like retargeting and branded search perform. When a model treats each channel's contribution as independent—meaning it doesn't account for those interactions—it will systematically misattribute revenue. Channels that primarily drive downstream lift in other channels will consistently look undervalued, and channels that benefit from that lift will look like stronger standalone performers than they actually are. Over time, budget decisions made on that data pull spend away from where it's actually needed.
Should I stop using MTA entirely?
That depends on your measurement stack and what questions you're trying to answer. MTA can provide useful user-journey context for digital channels where tracking is still reliable. The issue isn't that MTA has no value, it's that MTA and MMM answer different questions, and using an MMM that was built on MTA infrastructure can give you something that looks like an MMM but behaves more like a sophisticated MTA. If you're making full-funnel budget decisions, you need a model that was actually designed for that job.
See the data behind articles like this
Get a custom analysis of your media mix
Prescient AI shows you exactly which channels drive revenue — so you can stop guessing and start optimizing.
Book a demoKeep reading
View all
Problems with Vibe-Coded MMMs
Read article
Prescient Branded Search Attribution FAQs
Read article
Most incrementality tests are not rigorous randomized controlled trials—here’s why
Read article
YouTube MMM benchmarks
Read article
AppLovin MMM benchmarks – part II
Read article
AppLovin MMM benchmarks – part I
Read article