Insights
BERTNLPsemantic SEO

BERT (Language Model): Why Google's 2018 Update Still Determines Whether Your Content Ranks or Gets Buried

Bidirectional Encoder Representations from Transformers (BERT) is a language model introduced in October 2018 by researchers at Google. BERT learns to represent text as a sequence of vectors using self-supervised learning, which allows Google Search to evaluate whether a page genuinely ...

Dendro SEO 13 min read

Topic Ecosystem

Amber nodes = topics covered in this article — hover to explore connections

Bidirectional Encoder Representations from Transformers (BERT) is a language model introduced in October 2018 by researchers at Google. BERT learns to represent text as a sequence of vectors using self-supervised learning, which allows Google Search to evaluate whether a page genuinely answers a query rather than just matching keywords. Google deployed BERT across Google Search in October 2019, and BERT now evaluates content on every search query processed by Google.

What Is BERT and Why Did Google Build It?

BERT is a language model that Google built to understand the meaning of words based on surrounding context, not word count. Google released BERT in October 2018 to fix a specific failure: Google Search was misreading queries because Google’s previous systems processed words in isolation rather than as connected meaning.

The One-Sentence Definition Your Team Actually Needs

BERT — Bidirectional Encoder Representations from Transformers — is a language model that reads every word in a sentence in relation to every other word, simultaneously, so that Google can determine what a search query actually means rather than just matching keywords.

BERT’s core attributes:

  • Type: Language model based on transformer architecture
  • Developer: Google Research
  • Release date: October 2018 (research paper); October 2019 (deployed in Google Search)
  • Function: Generates a sequence of vectors that represents the meaning of text in context
  • Learning method: Self-supervised learning on large text corpora
  • Deployment scope: Applied to Google Search queries and page content globally

What Google Was Trying to Fix When Google Released BERT in 2018

Before BERT, Google Search used systems that read search queries from left to right or right to left — one direction at a time. A query like “can you get medicine for someone at a pharmacy” produced results about pharmacies in general, because Google’s systems weighted individual keywords rather than the relationship between words like “for someone.”

Google built BERT to process all words in a query simultaneously — reading in both directions at once — so that Google Search could interpret the full meaning of a phrase, not just the highest-frequency terms.

Why ‘Reading Like a Human’ Is More Than a Metaphor

BERT does not read like a human. BERT performs a measurable technical operation: BERT maps every word in a sentence to a position in a high-dimensional space — a mathematical map where words used in similar contexts appear near each other — so that Google can determine whether a page genuinely answers a query. Google then uses those positions — the sequence of vectors — to determine whether a page genuinely answers a query.

The business implication is direct: content written to insert keywords at a set frequency sends a different signal to BERT than content written to answer a question completely.

How Did Google Use BERT to Change What ‘Relevance’ Means?

Google used BERT to replace keyword frequency as the primary signal of relevance. After BERT’s deployment, Google evaluated content by analyzing contextual word relationships and sentence structure. Content that answered a question completely ranked above content that repeated a keyword more often.

Before BERT: Why Keyword Frequency Used to Be Enough

Before BERT, Google Search ranked pages primarily by measuring keyword density — the number of times a target keyword appeared relative to total word count. An SEO agency that inserted a keyword phrase 15 times into a 500-word page produced a page that Google Search ranked competitively, regardless of whether that page actually answered the reader’s question.

Google Search measured content relevance in the pre-BERT era by keyword presence, not meaning.

After BERT: How Context and Sentence Structure Changed the Scoring

BERT applies natural language processing to evaluate every sentence on a page as part of an interconnected structure. BERT does not score a keyword in isolation. BERT scores the relationship between a keyword and the words that surround the keyword.

A sentence like “Our product helps businesses grow” provides no contextual word relationships that BERT can use to determine what the product does, who the business serves, or what “grow” means in that context. A sentence like “Our inventory management software reduces stock shortfall incidents by 30% for retail businesses with 3 to 20 locations” gives BERT a complete, scorable unit of meaning.

Search query interpretation under BERT rewards specificity and penalizes vagueness.

The Business Result: Rankings Shifted and Traffic Moved

Semrush documented ranking volatility across multiple industries in the weeks following BERT’s October 2019 deployment in Google Search. Pages with high keyword density and low substantive content dropped in organic search visibility. Pages with complete, clearly written answers to specific questions gained positions.

The traffic redistribution was not random. BERT moved organic traffic from vague pages toward pages that demonstrated natural language understanding of the reader’s question.

What Happened to Sites When BERT Rolled Out in October 2019?

Sites with keyword-stuffed or thin content lost organic search visibility within weeks of BERT’s October 2019 deployment. Sites with clearly written, question-answering content gained rankings. The pattern held across industries and search categories, not just in one niche or content type.

The Sites That Dropped and What the Sites Had in Common

Analysis of post-BERT ranking shifts by Moz and Semrush identified 4 measurable characteristics shared by sites that lost ground:

  1. High keyword density: Pages repeated target phrases at a rate that exceeded natural sentence construction.
  2. Thin content: Pages contained fewer than 500 words on topics that require substantive explanation to answer completely.
  3. Vague sentence structure: Pages used generic statements without named subjects, specific outcomes, or defined entities.
  4. Low content quality signals: Pages lacked internal structure — no clear subheadings, no lists, no answers to the specific questions implied by the page’s target query.

Moz confirmed in algorithm update analysis that content quality signals — not domain authority alone — drove the ranking shifts in October 2019.

The Sites That Gained Ground and What the Sites Did Differently

Sites that gained organic search positions after BERT’s deployment had written content structured around answering a specific question completely. Sites that gained organic search positions used natural sentence construction, named entities with defined attributes, and page structures that made the answer to the target query extractable within the first 2 to 3 sentences.

Organic search visibility increased for pages where BERT could identify a clear question and a clear answer without parsing vague or keyword-dense text.

Why This Was Not a One-Time Update — BERT Is Now the Baseline

BERT is not an algorithm update that Google applied once and then moved past. Google integrated BERT into Google Search’s core infrastructure in 2019. Every search query Google processes today passes through BERT. Every piece of content Google evaluates today is evaluated against BERT’s standard of natural language understanding.

Any content investment made today — blog posts, landing pages, service pages — will be evaluated by BERT. Content that would have failed BERT’s evaluation in 2019 will fail BERT’s evaluation today.

What Does BERT Actually Read When BERT Evaluates Your Content?

BERT reads sentence structure, word relationships, and whether a page answers a specific question with complete, contextually coherent language. BERT does not count keyword frequency. BERT scores the semantic meaning of every sentence in relation to the query it is trying to answer.

How BERT Handles Ambiguous Words — and Why Your Page Title Matters More Than Ever

BERT uses bidirectional attention to resolve ambiguous words. The word “bank” carries different meanings depending on the words that surround “bank.” BERT reads the full sentence — in both directions from “bank” — to assign the correct meaning before evaluating whether the page is relevant to a query.

BERT uses page titles, H1 headings, and the first 100 words to interpret every subsequent sentence on the page. A page title that uses a keyword without specifying what the page actually explains forces BERT to work harder to determine relevance. According to Backlinko’s search ranking research, pages that rank in position 1 versus position 5 receive more than four times the click-through rate — a gap that begins with whether BERT can resolve a page’s meaning immediately from its title and opening content.

Why Long, Vague Paragraphs Confuse the Model and Cost Rankings

BERT performs passage-level understanding — BERT evaluates individual paragraphs and sections as discrete units of meaning, not just the page as a whole. A 200-word paragraph that uses abstract language and no named entities provides BERT with a low-density signal: BERT extracts few scorable meaning units from the paragraph.

BERT’s transformer architecture scores text by attention-weighted word positions — vague paragraphs produce low-confidence scores that reduce the probability of ranking for a target query, directly lowering organic traffic potential.

The business result: vague paragraphs reduce ranking position even when the page targets the correct keyword.

Clear Writing Produces Both Higher Rankings and Higher Conversion Rates

Clear writing performs two functions simultaneously. Clear writing gives BERT the semantic meaning BERT needs to rank a page. Clear writing gives a human reader the specific answer the reader came to the page to find.

A page that ranks because BERT identified clear, specific answers also converts at a higher rate because the reader receives a clear, specific answer. The content quality standard that BERT requires is the same standard that produces a page a reader trusts enough to contact a business.

What Is the Business Risk of Ignoring How Google Reads Content Today?

The business risk is measurable: content built on keyword insertion rather than complete answers will underperform in organic search, produce fewer leads per published page, and waste budget on production that BERT has already discounted. The risk compounds with every new page published under the same flawed approach.

Signs Your Current Content Was Written for the Pre-BERT Era

A marketing director can identify pre-BERT content using 5 observable signals:

  1. Keyword repetition without variation: The same phrase appears in nearly every paragraph in the same form.
  2. No entity definitions: The page references products, services, or concepts without defining what those entities are or what the entities do.
  3. Generic page titles: Titles use category words (“SEO Services,” “Marketing Tips”) without specifying the question the page answers.
  4. Low specificity: Claims use adjectives (“leading,” “comprehensive,” “best”) without numeric or factual support.
  5. No structured answers: The page does not contain a direct answer to the question implied by its target keyword within the first 100 words.

What a BERT-Resistant Content Problem Costs in Leads and Traffic

A site with 50 pre-BERT pages — pages built on keyword density rather than complete answers — carries 50 ranking liabilities. Each page competes for organic traffic against pages that meet BERT’s content quality standard. Each page that ranks below position 5 in Google Search produces fewer than 7% of available clicks, according to Backlinko’s click-through rate research.

50 underperforming pages at reduced click-through rates represent a quantifiable traffic gap. That traffic gap translates directly into a lead volume gap.

The Question to Ask Your Agency or Writer Before Approving Any Content Brief

Content briefs that omit a defined target question produce pages BERT cannot score as relevant. Before approving a content brief, ask: “What specific question does this page answer, and where on the page does the complete answer appear?”

A brief that cannot answer that question will produce a page that BERT cannot score as relevant. A page that BERT cannot score as relevant will not rank. A page that does not rank will not generate organic leads.

How Does Entity-Focused Content Perform Better Under BERT?

Content structured around named entities — defined subjects with stated attributes — performs better under BERT because BERT is built to identify entities, resolve their meaning, and match entity-based content to entity-based queries. Vague content provides no entities for BERT to identify.

Why Entity-Focused Content Aligns with How BERT Processes Search Queries

BERT processes search queries by identifying the entity the searcher is asking about and the relationship the searcher wants to understand. A query like “what does inventory management software do for small retailers” contains 2 entities — “inventory management software” and “small retailers” — and 1 relationship — “does for.”

Entity-first content states the entity, defines the entity’s attributes, and describes the entity’s relationships to other named entities. Language models like BERT score entity-first content more precisely than content that references concepts without naming or defining them.

Topical authority — the depth and breadth of entity coverage across a site — signals to BERT that a site is a reliable source for a given subject area. Sites with high topical authority rank for more queries in their subject area, increasing total organic traffic volume without requiring additional domain authority.

What ‘Writing for Humans’ Looks Like in Practice for a B2B Site

Writing for humans, in the context of BERT performance, means 4 specific practices:

  1. Name every entity on first reference with a complete definition — never use a pronoun where a named entity belongs.
  2. Answer the section question in the first sentence — do not build to the answer through background explanation.
  3. Use specific numbers instead of general adjectives — “reduces processing time by 40%” instead of “significantly faster.”
  4. Structure each paragraph around one claim — one subject, one predicate, one object, one idea.

A B2B site that applies these 4 practices produces content that BERT can evaluate precisely and that human readers find immediately useful. DendroSEO is an SEO agency that applies entity-first content architecture and natural language writing standards to build content programs that meet BERT’s evaluation criteria — without producing keyword-stuffed volume that BERT discounts.

How Content Architecture Decisions Upstream Affect BERT Performance Downstream

Content architecture — the decisions made before a word is written — determines whether BERT can identify a clear topical structure across a site. A site with 30 pages that each answer a distinct, clearly defined question on a single topic sends BERT a coherent topical signal. A site with 30 pages that each repeat the same 5 keywords in different orders sends BERT a redundant, low-value signal.

Structured content, where each page occupies a defined role within a topical cluster, allows BERT to evaluate the site’s semantic relevance at both the page level and the site level. Content architecture decisions made at the brief stage determine BERT performance at the ranking stage.

What Should Marketing Leaders Know Before Approving Any Content Investment?

Marketing leaders who control content budgets need these 7 reference points before approving new content investment. Content built on keyword density rather than complete answers will underperform under BERT’s evaluation criteria, waste production budget, and reduce organic lead volume with every page published under the same flawed approach.

  1. BERT is a language model — BERT is a defined technical system introduced by Google in October 2018 that evaluates content meaning, not keyword frequency. BERT is not a trend. BERT is Google Search’s current standard.

  2. BERT has been the baseline since October 2019 — every page published since October 2019 has been evaluated by BERT. Every page published today will be evaluated by BERT.

  3. Keyword stuffing actively costs rankings — content built on keyword density rather than complete answers receives lower content quality signals from BERT and ranks below content that answers questions clearly.

  4. Thin content is a measurable liability — pages with low word count, vague sentences, and no entity definitions perform below their potential organic traffic ceiling under BERT’s evaluation criteria.

  5. Clear writing and ranking performance are the same objective — content that BERT scores as relevant is also content that human readers find useful. The content quality standard required for organic growth is the same standard required for lead generation.

  6. Entity-first content structure aligns with BERT’s processing model — content that names entities, defines entity attributes, and describes entity relationships gives BERT precise meaning signals and ranks more consistently.

  7. Content architecture decisions determine BERT outcomes — the question a page answers, the structure of that answer, and the page’s relationship to other pages on the site are all decisions made before writing begins. Natural language understanding starts with a content plan, not a keyword list.

BERT (Bidirectional Encoder Representations from Transformers) is classified as a DefinedTerm in structured data. BERT is a language model developed by Google Research, introduced in October 2018, and deployed in Google Search in October 2019. BERT applies transformer architecture and self-supervised learning to evaluate the semantic meaning of text as a sequence of vectors.

Ready to build?

Let's look at your architecture.

Apply to Work Together →