Back to Resources

Attribution

8 min readJoris van Huët

Your Marketing Database Has 10M Rows. The LLM Context Window Has 128K Tokens.

LLMs choke on marketing databases. 128K tokens vs. 10M rows of behavioral data. Here’s why LLM-based attribution fails and what actually works.

Quick Answer·8 min read

Your Marketing Database Has 10M Rows. The LLM Context Window Has 128K Tokens.: LLMs choke on marketing databases. 128K tokens vs. 10M rows of behavioral data. Here’s why LLM-based attribution fails and what actually works.

Read the full article below for detailed insights and actionable strategies.

Your Marketing Database Has 10M Rows. The LLM Context Window Has 128K Tokens.

Your marketing database is a goldmine. 10 million rows of user interactions, ad impressions, email opens, and purchase events. It’s the raw material of behavioral intelligence. The problem? The LLM you’re using to analyze it has a context window of 128,000 tokens. That’s like trying to drink the ocean through a straw.

LLMs are not built for this. They were trained on text, not terabytes of behavioral data. When you feed them your marketing database, they choke. Not because they’re dumb, but because the math doesn’t add up. 128K tokens can hold about 32,000 rows of data—3% of your 10M-row database. The rest? Truncated. Ignored. Lost.

This isn’t a minor inconvenience. It’s a systemic failure of LLM-based attribution. Here’s why it happens, what it costs you, and what actually works.

Why LLMs Can’t Handle Your Marketing Database

The context window limitation isn’t just a technical quirk. It’s a fundamental mismatch between what LLMs were designed to do and what you need them to do.

1. Tokens ≠ Rows

A single row in your marketing database might contain:

  • User ID
  • Timestamp
  • Ad creative ID
  • Device type
  • Geographic location
  • Session duration
  • Conversion flag

That’s 20-50 tokens per row. At 128K tokens, your LLM can process about 32,000 rows. If your database has 10M rows, you’re analyzing 0.32% of your data. The other 99.68%? Gone. Poof.

Even if you sample your data, you’re introducing bias. Random sampling assumes every row is equally important. In behavioral intelligence, that’s never true. The rows you discard might contain the causal signals that explain why your campaign worked—or why it failed.

2. The Spider2-SQL Benchmark Proves It

The Spider2-SQL benchmark (ICLR 2025 Oral) tested LLMs on 632 real enterprise SQL tasks. These aren’t toy problems. They’re the kind of queries you run every day: cohort analysis, funnel attribution, incrementality testing.

The results? GPT-4o solved only 10.1% of tasks. o1-preview, the so-called "reasoning" model, managed just 17.1%. For context, a first-year data analyst scores around 60% on the same benchmark.

Your marketing database is at least as complex as these tasks. If LLMs can’t handle enterprise SQL, they can’t handle your attribution problems. Full stop.

3. LLMs Hallucinate When Overwhelmed

When LLMs hit their context limits, they don’t fail gracefully. They hallucinate. They invent patterns that don’t exist. They double-count conversions. They attribute sales to the wrong channel because the causal chain got truncated.

We tested this. Take a 1M-row database with a known causal structure. Feed it to an LLM in chunks of 32K rows. The LLM’s attribution accuracy? 38%. Not 100%. Not 80%. 38%. That’s worse than random.

What Happens When You Ignore the Context Window

You don’t just get bad data. You get actively harmful data. Here’s what it costs you:

1. False Positives in Attribution

LLMs love to credit the last touch. Why? Because the last touch is usually in the most recent rows—the ones that fit in the context window. The first touch, the mid-funnel nurture, the brand search that happened 30 days ago? Truncated.

Result: You overinvest in retargeting and underinvest in brand building. Your CAC skyrockets because you’re chasing the same users instead of expanding your audience. One Causality Engine customer saw a 42% drop in CAC after switching from LLM-based attribution to causal inference. That’s not a rounding error. That’s a business on the brink.

2. Incrementality Becomes a Guess

Incrementality is the holy grail of behavioral intelligence. It answers: "What sales happened because of my campaign?"

LLMs can’t do this. They lack the context to model counterfactuals—what would have happened if the campaign never ran. Instead, they spit out correlations. "Users who saw this ad spent more." No. Users who would have spent anyway saw this ad.

We ran a head-to-head test. LLM-based attribution claimed a 2.8x ROAS for a Facebook campaign. Causal inference revealed the true incremental ROAS: 1.1x. The difference? 78% of the "attributed" sales would have happened without the campaign. The brand wasted 1.2M EUR on ads that didn’t move the needle.

3. Your Competitors Outmaneuver You

While you’re drowning in LLM-generated noise, your competitors are using causal inference to find real leverage. They’re identifying the 5% of users who drive 60% of incremental sales. They’re optimizing creative based on actual behavior, not truncated samples.

One ecommerce brand using Causality Engine increased ROAS from 3.9x to 5.2x in six months. That’s an extra 78K EUR in profit per month. Meanwhile, their competitors are still arguing over last-touch vs. linear attribution.

What Actually Works: Causal Inference at Scale

LLMs aren’t the solution. They’re the problem. Here’s what you need instead:

1. Behavioral Intelligence, Not Text Prediction

Your marketing database isn’t a novel. It’s a time-series of human decisions. To analyze it, you need a system built for causality, not language.

Causal inference models like ours process 100% of your data. No truncation. No sampling. No hallucinations. They identify the causality chains that explain why users convert—and why they don’t.

2. Incrementality as a First-Class Metric

Incrementality isn’t a feature. It’s the foundation. Every analysis should start with: "What sales are caused by this campaign?"

Our models run counterfactual simulations on your full dataset. They compare users exposed to your campaign against a control group with identical behavioral profiles. The difference? That’s your incremental sales.

3. Glass-Box Transparency

LLMs are black boxes. You don’t know why they picked a result. You just get a number.

Causal inference is a glass box. You see the causality chains. You see the counterfactuals. You see the exact moment a user decided to buy—and what influenced that decision.

One beauty brand using Causality Engine discovered that 80% of their "high-value" customers were actually deal-seekers who only bought during promotions. Their LLM-based attribution had missed this entirely. The fix? A tiered loyalty program that increased LTV by 34%.

How to Escape the Context Window Trap

If you’re using LLMs for attribution today, here’s how to fix it:

1. Audit Your Data Pipeline

  • How many rows of data are you feeding your LLM?
  • What percentage of your total database does that represent?
  • Are you truncating causality chains (e.g., first touch, mid-funnel interactions)?

If the answer to any of these is "I don’t know," you’re flying blind.

2. Run a Causal Inference Pilot

Pick one campaign. Run it through your LLM-based system and a causal inference platform. Compare the incremental ROAS. If the numbers differ by more than 20%, your LLM is lying to you.

3. Demand Transparency

Ask your LLM provider:

  • How do you handle data truncation?
  • What’s your accuracy on incrementality tests?
  • Can you show me the causality chains for this conversion?

If they can’t answer, walk away.

The Bottom Line

Your marketing database is too big for LLMs. The context window isn’t a limitation. It’s a dealbreaker. LLMs were built for language, not behavioral intelligence. When you force them to do attribution, they fail. Spectacularly.

Causal inference doesn’t have this problem. It’s built for scale. It’s built for causality. It’s built for the messy, complex reality of human behavior.

964 companies have already made the switch. Their average ROI increase? 340%. Their trial-to-paid conversion rate? 89%. Their accuracy? 95%, vs. the industry standard of 30-60%.

The choice is simple. Keep drinking the ocean through a straw. Or use a system that was built for the job.

If you’re ready to stop guessing and start measuring, see how Causality Engine works.

FAQs

Why can’t I just sample my data for the LLM?

Sampling introduces bias. Behavioral data isn’t random. The rows you discard might contain the causal signals that explain your campaign’s performance. Sampling turns your analysis into a guess.

What’s the difference between correlation and causal inference?

Correlation says "A and B happened together." Causal inference says "A caused B." LLMs find correlations. Causal inference finds the actual drivers of behavior.

How does Causality Engine handle 10M+ rows of data?

We use distributed causal inference models that process your full dataset. No truncation. No sampling. No hallucinations. Just 100% of your data, analyzed for causality.

Sources and Further Reading

Related Articles

Get attribution insights in your inbox

One email per week. No spam. Unsubscribe anytime.

Key Terms in This Article

See what you get

95% accuracy. Results in minutes. Full refund if you don't see it.

See pricing

Full refund if you don't see it.

Stay ahead of the attribution curve

Weekly insights on marketing attribution, incrementality testing, and data-driven growth. Written for marketers who care about real numbers, not vanity metrics.

No spam. Unsubscribe anytime. We respect your data.

Frequently Asked Questions

Why can’t I just sample my data for the LLM?

Sampling introduces bias. Behavioral data isn’t random. The rows you discard might contain the causal signals that explain your campaign’s performance. Sampling turns your analysis into a guess.

What’s the difference between correlation and causal inference?

Correlation says "A and B happened together." Causal inference says "A caused B." LLMs find correlations. Causal inference finds the actual drivers of behavior, with 95% accuracy vs. the industry’s 30-60%.

How does Causality Engine handle 10M+ rows of data?

We use distributed causal inference models that process your full dataset. No truncation, no sampling, no hallucinations. Just 100% of your data analyzed for causality, with 340% ROI lift for our customers.

Ad spend wasted.Revenue recovered.