Metric Deep Dive · 1 of 3

Latent Brand Association

What ChatGPT, Claude, and Gemini already believe about your brand — before they ever touch the web.


What is Latent Brand Association?

Every AI model — ChatGPT, Claude, Gemini — spent months of training time reading billions of documents from across the web. News articles. Wikipedia. Forum posts. Product reviews. Research papers. Company pages. Subreddits. Court filings. Recipe blogs. Everything.

Out of all that reading, each model came out with opinions. Or no opinion. Or the wrong opinion. About your brand, about your category, about your competitors.

Those opinions live inside the model's neural weights. Nobody put them there on purpose. They formed on their own, because when a model sees "Nike" appear next to "athletic" ten million times in its training data, the connection sticks. When it sees a smaller brand appear once or twice in a random forum post, the connection is weak — or nonexistent.

Latent Brand Association is the measurement of those baked-in beliefs. What the model associates with your brand. How confident those associations are. Whether they're accurate, positive, stale, or entirely made up.

Quick analogy

Think of LBA as the "first impression" the model has when it hears your brand name. Before it does any research. Before it Googles anything. What comes to mind?

For Nike, what comes to mind is athletic footwear, the Swoosh, "Just Do It," superstar athlete endorsements. Strong, consistent, positive associations.

For a newly launched startup with your exact same name, what comes to mind might be… nothing. Or worse, it might come back confidently describing the wrong company.

This is different from Top of Mind, which measures whether the model recalls your brand at all when asked "what are the best X for Y?" LBA asks a different question: given that the model knows about your brand, what does it believe about you?

A brand can be top-of-mind but carry negative associations. A brand can have great associations but be too obscure to surface unprompted. These are different problems with different fixes, which is why we measure them separately.


Why LBA Matters More Than a Single AI "Rank"

Anyone can open ChatGPT, type "what's the best CRM?", and check whether your brand shows up. Do that five times and you'll get five different answers. Some will include you, some won't. That single number is almost useless on its own.

LBA is different. We're not looking at one response. We're looking at the shape of everything the model knows about you. That shape doesn't jump around between queries. It's the deep, underlying pattern — formed during training, stable enough to actually measure.

Because LBA is baked into model weights, three things follow:

  • It changes slowly. A model "remembers" your brand as it existed when it was last trained — often 12 to 24 months ago. You could have completely repositioned, launched new products, or doubled your market share since then. The model doesn't know.
  • It's hard to fake. You can't sprint your way to good LBA by running a content campaign this week. You can move it, but only over quarters and years of authoritative coverage.
  • It's the foundation for everything else. If the model has strong, positive associations with your brand, your retrieval visibility, your citation rates, and your recall all benefit. If the model has negative or hollow associations, even perfect content will run uphill.
The geological layer

We call LBA the geological layer of AI visibility. It shifts on the scale of training cycles, not weeks. That makes it the hardest metric to change — but also the most structural. Brands with strong LBA have an advantage their competitors can't replicate quickly.


How We Measure LBA

Measuring what's baked into a model's weights takes more than one clever question. Here's how we do it.

Step 1: Five Probe Prompts, Recall Only

We ask each model the same five questions about your brand, with web search explicitly disabled:

  1. "What is [your brand] known for?"
  2. "What are [your brand]'s main strengths and weaknesses?"
  3. "Who should use [your brand]? Who should avoid it?"
  4. "How does [your brand] compare to its main competitors?"
  5. "What do people typically complain about with [your brand]?"

These are designed to probe the model's training-baked beliefs from different angles. Positive, negative, comparative, behavioral. A brand with strong LBA produces rich, specific, consistent answers to all five. A brand with weak LBA produces vague, generic, or inconsistent answers.

Step 2: Five Iterations Per Prompt

Every prompt runs five times.

Why? Because LLMs are probabilistic. The same question asked twice produces two different responses. We learned this the hard way testing Nike — the most recognizable brand slogan in advertising history, "Just Do It", appeared in only one out of three early test runs. A single query would have missed it.

The Nike surprise

We assumed Nike's core associations would be identical across every run. They weren't. The category ("athletic footwear") was 100% stable. But iconic differentiators oscillated: "Just Do It" appeared in 1 of 3 runs. "Air Max" in 1 of 3. "Premium pricing" in 1 of 3.

If we'd only run one iteration, we'd have produced a Nike profile missing the brand's most famous slogan. Multiple iterations aren't just statistical rigor — they're the difference between seeing the full picture and seeing a coin flip.

Step 3: A Control Prompt

Alongside the brand probes, we run a control prompt with your brand name swapped for a generic category. "What is a typical CRM known for?" rather than "What is [YourBrand] known for?"

Why? Because some models fake recall by riffing on your brand name. When we tested a small, niche brand, all five runs returned polished, plausible-sounding associations — but every one of them was just a generic description of its category. The model wasn't actually remembering that brand. It was reading the words in the name and improvising.

By comparing the brand-specific answers to the category-generic answers, we can tell the difference between "the model knows your brand" and "the model is faking it by describing your category." Two very different realities that produce similar-looking responses.

Step 4: The Familiarity Probe

Alongside the five descriptive probes and the control, we ask the model one meta-question: "On a scale of 1-10, rate your familiarity with this brand. 1 = I've never heard of it. 10 = deep, confident knowledge of specific features, pricing, and competitor positioning."

Why? Because the descriptive probes don't catch a specific failure mode: when the model is asked "what is [obscure brand] known for?", it will confidently hallucinate plausible-sounding features. We tested this with a completely fabricated brand name — the model dutifully produced polished descriptions of rank tracking, backlink analysis, and site audits, as if the brand really existed. The descriptive probes can't tell confident-and-correct from confident-and-wrong.

The familiarity probe asks the model to introspect. It's specifically trained to calibrate uncertainty when asked directly. Real brands consistently rate 6-10; obscure-but-real brands rate 4-6; fabricated or unknown brands rate 1-3. We run it five times and average — consistency itself is informative, since wildly varying self-ratings are their own signal.

The fake-brand test

To validate the familiarity probe, we fabricated a brand called "SEO Frog Bot" — a plausible-sounding SEO tool that doesn't exist. The descriptive probes generated rich, confident answers about its supposed features. The familiarity probe returned 1 on every single iteration. The gap between those two signals is exactly what the final score captures.

Step 5: Extract, Classify, Cluster

For every response, a second cheap model extracts individual associations and tags each one:

  • Polarity: positive, neutral, or negative
  • Freshness: current or outdated
  • Factuality: true, false (hallucination), or unverifiable
  • Meta-uncertainty: is the model itself flagging that it doesn't know? ("Generic placeholder brand name," "low brand recognition")

Then we cluster associations that mean the same thing but are worded differently. "Expired domain finder," "Expired domains marketplace," and "Finding expired domains" are the same association. Raw string-matching would treat them as three unrelated items; clustering treats them as one.

Step 6: Score Across Three Models

Everything above runs for ChatGPT, Claude, and Gemini — independently. Each model has different training data, different cutoff dates, different biases. A brand can score well on ChatGPT and be invisible to Gemini, or vice versa. We report a per-model breakdown plus a combined score, so you can see exactly where your weak spots are.

What it costs to measure

Per brand: 5 descriptive probes + 1 control + 1 familiarity probe, each run 5 times across 3 models = 105 primary calls, plus roughly 25 extraction calls (extraction runs only on the descriptive probes, not familiarity or the integer-only probe, and only once per brand-level probe), plus a clustering pass. Total cost on cost-effective models like GPT-5-mini, Gemini Flash, and Claude Haiku is around 5–7 cents per brand, per analysis run.


The Six LBA Patterns We've Found

We've run this methodology against real brands ranging from Nike and Ahrefs to obscure B2B tools and a completely made-up brand we invented for testing. The results sort into six recognizable patterns. Yours almost certainly falls into one of these.

Pattern 1 · Product-Strong
Expected score: 70–90

Example: Ahrefs (ahrefs.com)

When we ran our methodology against Ahrefs, every single run produced specific product names the model had learned by heart:

  • Site Explorer — 4 of 5 runs (product name)
  • Content Explorer — 3 of 5 runs (product name)
  • Ahrefs Webmaster Tools — 3 of 5 runs, with "(free)" correctly noted twice
  • Keyword Difficulty metric — named by name
  • Link Intersect tool — named by name

This is the strongest LBA pattern there is. The model doesn't just know what category Ahrefs competes in — it knows their specific products, pricing tiers, reputation signals ("trusted by SEO professionals"), and their free vs paid split. An Ahrefs user has a structural advantage: when someone asks ChatGPT "what's the best SEO platform?", Ahrefs gets named with specific product features while competitors get generic descriptions.

What to do if you're in this pattern
  • Don't relax. Product-Strong is earned, but it's not permanent. New product launches, rebrands, or acquisitions might not make it into the next training cycle unless you actively seed coverage.
  • For every new product: push for independent reviews, comparison articles, and user-generated coverage on high-authority domains within 12 months of launch.
  • Monitor by model version. When a new model drops, re-test. A competitor that invested heavily in the last 12 months of authoritative coverage might start eroding your lead.
  • Protect the named products. If a flagship product gets renamed, the old name is what the model still knows. Maintain both names in content for a full training cycle after any rename.
Pattern 2 · Iconic but Oscillating
Expected score: 65–80

Example: Nike (nike.com)

Testing Nike gave us a surprising result. The category-level associations were rock solid — every single run mentioned athletic footwear, sports apparel, athlete endorsements, global marketing. Category ownership was perfect.

But look at what oscillated:

  • "Just Do It slogan" — appeared in only 1 of 3 early runs (33%)
  • "Air technology (Air Max)" — appeared in only 1 run
  • "Premium pricing and brand prestige" — appeared in only 1 run

Even Nike — whose slogan is arguably the most famous in advertising history — has a roughly one-in-three chance of surfacing "Just Do It" on any given query. The category is locked in, but the differentiators aren't.

This pattern is common for large, well-known brands with many products, slogans, and campaigns. The category association is unbreakable. The specific attributes are stable-ish but not universal.

What to do if you're in this pattern
  • Category dominance is fine — don't try to fix it. Your structural position is already strong.
  • Identify which differentiators oscillate. For Nike, it's slogans and specific technologies. For your brand, it might be a signature feature, a founder's name, or a campaign tagline.
  • Concentrate content weight around those oscillating differentiators. If "Air Technology" is your weak link, push for articles where "Air Technology" is in the headline on domains that training pipelines crawl (Wikipedia, major publications, review sites).
  • Don't bury your differentiators inside long-form prose. Short, frequently-repeated associations win more training-data weight than long, complex ones.
Pattern 3 · Integration-Strong
Expected score: 55–75

Example: DomCop (domcop.com)

DomCop is a niche expired-domain marketplace. Not a household name by any stretch. But the model knows it surprisingly well — and specifically, it knows DomCop's partnerships.

Across five test runs, the model consistently mentioned:

  • "Integration with Moz and Majestic" (data partnerships)
  • "Integrates GoDaddy and NameJet auctions" (auction feed partners)
  • "Filters for DA, PA, TF, CF" (specific SEO metrics, some from partners)
  • "Bulk domain analysis" (core capability)

This pattern shows up for brands whose value proposition is defined by who they integrate with. The model learned DomCop ≈ "the thing that aggregates expired domain auctions with Moz data." Strong association, but structurally dependent on those named partners.

What to do if you're in this pattern
  • Lean into your "X for Y" framing. Your strongest LBA hook is the integration relationship — emphasize it in press, SEO, and product marketing.
  • Be aware of dependency risk. If a partnership changes or an integration breaks, your LBA weakens on the next training cycle. Contingency: build standalone brand content that doesn't depend on partner names.
  • Add product-level associations. Beyond "integrates with Moz," you want named features: "DomCop Watchlist," "DomCop Auction Tracker." Whatever your version of this is, push it into the same authoritative-coverage pipeline.
  • Diversify your partner mentions. If Moz ever loses relevance, you don't want to be left as "the thing that integrates with a tool nobody uses." Multiple partners = more resilient integration-strong LBA.
Pattern 4 · Category-Known, Product-Unknown
Expected score: 25–45

Example: Medicube (medicube.us)

Every one of our five test runs for Medicube started with the exact same phrase: "Korean skincare brand." Perfect 5-for-5 category match. Every run also included "acne treatment products." Pore care, dermatologist-developed formulations, and cica ingredients showed up across runs.

But not once did the model name a specific Medicube product. No Age-R devices. No Zero Pore Pads. No Red Kollagen. The model knows exactly what Medicube does and who it's for — but not the specific products that distinguish them from other acne-focused K-beauty brands.

This is probably the most common LBA pattern for mid-sized brands. You've built enough category awareness that the model knows your positioning. You haven't built enough product awareness for the model to cite specific offerings.

What to do if you're in this pattern
  • Your job: turn category awareness into product awareness. The diagnosis is clear and so is the fix.
  • Push product-specific content into authoritative sources. Independent reviews of named products. "Best of" lists. Comparison articles. User guides.
  • Use structured data. Product schema on your own site helps future crawlers extract product names reliably. This matters more than it sounds — it's free and compounding.
  • Wikipedia matters more than people think. If your flagship product qualifies for a Wikipedia page with citations, the odds of future training cycles capturing the product name jump dramatically.
  • Press coverage with product names in headlines. Headlines are weighted heavily in training data. "Medicube's Age-R Device Reviewed" beats "We Tested Medicube's Skincare" by a large margin.
  • Patience. This pattern takes 12–24 months to convert fully, tied to major model training cycles. Start now, stay consistent.
Pattern 5 · Entity Collision
Expected score: 10–30

Example: Domino (domino.com)

This one surprised us. We tested "Domino" with the domain domino.com — which is actually the Condé Nast interior-design magazine. The model's five runs split cleanly in half:

  • Runs 1, 2, 3 → Domino's Pizza. Pizza delivery. Domino's Tracker. Hand-tossed crust. Sides like chicken and breadsticks.
  • Runs 4, 5 → Domino magazine. Home design. Interior decorating. Room makeovers. Celebrity home features.

Zero overlap between the two groups. Each group was internally coherent. The model picked an interpretation and stuck with it — but which interpretation won flipped unpredictably between runs.

This is name collision: your brand shares a name with a more famous entity, and the model's default interpretation is the famous one. If you're Domino magazine, roughly 60% of generic "Domino" queries are getting answered about pizza.

What to do if you're in this pattern
  • Recognize this is partially outside your control. You can't un-famous Domino's Pizza. You can't defeat a household name on its own turf. Set realistic expectations.
  • Always use a distinguishing descriptor in your own content. "Domino magazine," "Domino Home Design." Never just "Domino" in standalone contexts.
  • Teach your customers to disambiguate. If AI agents are part of the buyer journey, the way you're queried matters. "Domino magazine interior design tips" will reliably find you. "Domino tips" won't.
  • Consider adding a consistent descriptor suffix. Think of it like a rebrand-adjacent move — anywhere your brand name appears in content you control, include the descriptor. Over time, the model learns "Domino + descriptor" as a distinct entity.
  • Accept that LBA is structurally capped when you share a name with a more famous brand. Focus on stable LLM Authority in retrieval mode, which is less affected by namesake confusion because the domain context clarifies the entity.
Pattern 6 · Unknown or Confabulated
Expected score: 0–10

Example: AcmeWidgetsXYZ (acmewidgetsxyz.com) — a completely made-up brand we created for testing

When we asked the model about a fictional brand that doesn't exist anywhere in training data, the model did one of two things. Both are informative.

In 3 of 5 runs, the model honestly flagged uncertainty: "Generic placeholder brand name," "Novelty or demo website," "Low brand recognition," "Unclear company provenance." That's good behavior — the model refusing to hallucinate.

In 2 of 5 runs, the model confidently invented a full backstory: "Replacement parts and components," "DIY repair and hobby supplies," "Fast shipping and fulfillment," "Volume discounts for bulk orders," "Compatibility with popular brands." Pure fabrication, delivered with total confidence. If we'd run only one test and landed on one of those runs, we'd have an entirely fake LBA profile for a brand that doesn't exist.

This is the pattern you'll see for genuinely new or obscure brands: startups pre-launch, brands that rebranded under a new name in the last year, regional brands without international coverage, or brands with name changes. You're getting either honest "I don't know" responses or confident hallucinations, and there's no way to predict which on any given run.

What to do if you're in this pattern
  • Don't panic. This is the normal starting position for any new brand. Everyone was here once.
  • The only real fix is time plus authoritative coverage. Future training cycles will pick you up if you build a legitimate footprint. There are no shortcuts.
  • Quick wins for coverage: Wikipedia (if you qualify), Crunchbase, AngelList, industry-specific listings, founder interviews on podcasts with transcripts, LinkedIn company pages.
  • Medium-term: press coverage, podcast guest appearances, interviews with the founder on high-authority domains. Aim for 20–30 citations across reputable sources in your first 12 months.
  • Long-term expectation: measurable LBA progress takes 12–24 months, tied to major model training cycles. Plan for patience.
  • Lean on retrieval in the meantime. Your LLM Authority Score in retrieval mode doesn't require the model to already know you. Publish authoritative, well-cited content, and web search will surface you even while training-data recall is still catching up.
  • Monitor for hallucinations carefully. When a model invents facts about your brand, those confabulations can spread through screenshot-based marketing content before the model even knows who you are. Flag and counter them explicitly.

How Your LBA Score Is Calculated

For the curious: your LBA score combines three signals from the probes above, each scored from 0 to 1. They're multiplied together, scaled to 100, and then calibrated against empirical top-tier brands so the category leaders land in the low 90s instead of an unreachable 100.

Signal What it measures
Association quality The polarity and factuality of everything the model associates with your brand. Positive associations help. Negative associations hurt. Hallucinations (factually false claims) are penalized hardest.
Meta-uncertainty factor How often the model itself flagged uncertainty ("generic placeholder," "low brand recognition"). Lots of uncertainty flags = the model doesn't really know you.
Familiarity The model's self-rated confidence about your brand, on a 1-10 scale, averaged over five iterations. This is the sharpest discriminator: real well-known brands score 8-10, niche-but-real brands 5-7, hallucinated or unknown brands 1-3.

These three numbers get multiplied together. Multiplication matters: any single signal near zero drives the final score near zero, even if the others are strong. A brand can't have meaningfully "good LBA" if the model is inventing fake products for it (quality collapses), hedging every answer (meta-factor collapses), or admitting it's never heard of the brand (familiarity collapses).

After the multiplication, we apply one calibration step: a 10-point confidence floor is subtracted and the result is scaled up by ~1.9×, then clamped to [0, 100]. The result: a brand the model has literally never heard of — where familiarity lands at 1/10 — drops to 0. Category leaders who truthfully can't do much more land in the low 90s.

Score ranges by pattern
  • Product-Strong (Ahrefs-type): 85–95 — the model genuinely knows your specific products and differentiators
  • Iconic but Oscillating (Nike-type): 70–85 — great category awareness, differentiators drift between runs
  • Integration-Strong (DomCop-type): 55–75 — strong but narrow associations, famous within a niche
  • Category-Known, Product-Unknown (Medicube-type): 25–45 — model recognizes the category you're in but not your specific product
  • Entity Collision (Domino-type): 10–30 — model keeps talking about a more famous namesake
  • Unknown / Confabulated (made-up-brand-type): 0–10 — familiarity bottoms out and pulls the score to the floor

We report each of the three sub-signals alongside your final score. That way, when a score is low, you can see exactly which factor is holding it down — which tells you what to work on.


The Universal Playbook for Improving LBA

Regardless of your current pattern, the levers that move LBA are the same. They're just slow, and they compound.

1. Authoritative coverage

Articles about your brand on domains that model training pipelines actually crawl. Not every site carries equal weight. Wikipedia, major publications, industry-specific media, established subreddits, academic papers, and Reddit threads (yes, really — they show up in training data more than people realize) are worth far more than a thousand syndicated press releases.

2. Structured brand identity

Consistent name, category, and descriptor across every property. "Medicube, the Korean dermatologist-developed skincare brand" repeated across 100 sources moves LBA more than one standalone profile on the New York Times. Training data rewards repetition with variation, not one-off prestige.

3. Product-level naming

If the model knows your category but not your products, the fix is product-level content density. Get specific product names into independent reviews, comparison articles, user discussions. Named products in headlines are worth more than named products in body text.

4. Time and training cycles

LBA updates on the scale of the next major model release. ChatGPT retrains roughly every 12–18 months. Claude and Gemini are similar. Anything you start today shows up in the model 12–24 months from now — which means you need to start early and stay consistent, not rush at the finish line.

5. Monitor oscillation

If your associations are unstable across runs, that's a signal you don't have enough training-data weight yet. Stability itself is a goal. Watch for differentiators that appear in 2 or 3 of 5 runs — those are the ones that need more reinforcement.

6. Disambiguate aggressively if you share a name

If you're in the entity collision pattern, a consistent descriptor ("Domino magazine," not "Domino") across every property you control is worth more than any individual piece of content. You're not trying to beat the famous namesake — you're trying to become a distinct entity the model can learn separately.


What Comes Next

LBA is the foundation metric. But it's only one of three. Once you know what the model already believes about you, the next questions are:

  • How often does the model actually bring you up in real queries? (LLM Authority Score)
  • Does the model recall you at all without a web search? (Top of Mind)

Each metric diagnoses a different layer of the AI visibility stack. LBA is the deepest, the hardest to move, but also the most structural. When your LBA is strong, every other metric gets easier. When it's weak, you can still earn visibility through retrieval-based Authority while you patiently build LBA over training cycles.

← Back to the overview of all three metrics