AI Summary
Mike King at iPullRank, who literally introduced the concept of chunking to the SEO space, makes the core point bluntly: AI engines retrieve and rank passages, not pages. If you understand how chunking actually works inside a RAG pipeline, you can structure content that gets retrieved disproportionately. If you don’t, you’re writing for an algorithm that doesn’t exist anymore.
What chunking is (and what it is not)
Chunking is the process AI systems use to break long content into smaller passages before embedding and retrieval. King clarifies: chunking happens inside the RAG system, not on your page. You don’t “chunk” your content – the system chunks it. What you control is how easy you make that chunking.
Common chunking strategies in production RAG systems:
- Fixed-size chunking. Naive but common. Splits at every N tokens regardless of meaning. Loses context at boundaries.
- Recursive character chunking. Splits on paragraph, then sentence, then word boundaries. The default in LangChain.
- Semantic chunking. Uses embeddings to detect topic shifts and splits there. Higher quality, more compute.
- Late chunking. Embeds the full document first, then derives chunks from the contextualized embeddings. Highest quality, newest.
On-page changes that improve chunk extraction
- One idea per paragraph. If a paragraph covers two distinct ideas, the chunker will split them across two chunks and lose context. Tighten paragraphs to one main claim each.
- Self-contained sections. Each H2 should be readable in isolation. Define key terms inside the section even if you defined them earlier – the chunker may extract this section without the earlier one.
- Lead with the answer. The first sentence of each section should answer the implicit question of the heading. This is what semantic chunkers anchor on.
- Limit cross-section pronouns. “This means…” referring to the previous section makes the chunk un-extractable. Repeat the antecedent.
- Use lists and tables for enumerable content. Chunkers preserve structural boundaries. Lists and tables become atomic, citable units.
The Ahrefs counter-argument and what it gets right
Ahrefs argues that “chunk optimization” is overrated, and they’re partially correct. Two things they get right:
- You should not contort your writing for hypothetical chunk boundaries. Good editorial writing is already chunk-friendly.
- Adding artificial micro-headings to “create chunks” hurts readability without helping retrieval.
What they understate: the difference between writing for human readers and writing for human readers AND retrieval is real but small. The structural changes that help chunking – tight paragraphs, self-contained sections, lead-with-the-answer – are the same changes editors have demanded for a century. Do them well and you optimize for both.
Frequently Asked Questions
Should I split my long-form articles into smaller posts to help AI retrieval?
What's the ideal paragraph length for AI retrieval?
Do I need to add explicit chunk boundaries (like horizontal rules)?
Want this implemented for your brand?
I help growth-stage companies own their category in AI search. Book a strategy call.