Skip to main content
Methodology Guide

Keyword clustering with neuro-symbolic methods: structured data meets language models

Pure embedding-based clustering groups keywords by surface similarity. Neuro-symbolic approaches add structured taxonomies, label grammars, and synonym layers so every keyword lands in the right topic - even when the words look nothing alike.

Trusted by leading e-commerce brands

Visual ComfortTwinklBigjigs ToysDewaeleDiscountMugsDependsRVshareKleinanzeigen

Why one-keyword-per-page fails for e-commerce

The old playbook was simple: find a keyword, build a page, repeat. For a site with 500 categories that might mean 500 pages, each targeting a single phrase. The problem is that search engines no longer evaluate pages by isolated keyword matches. Google groups queries by topic, not keyword, and a single category page can rank for dozens of related queries if the topic coverage is strong.

When you build pages keyword by keyword, you get cannibalization: three nearly identical pages competing for overlapping queries, none of them comprehensive enough to win. Topic clustering solves this by grouping related keywords first, then building one authoritative page per topic.

The question is how you group those keywords. The approach you choose determines whether your clusters are accurate or full of false matches.

The problem with pure ML clustering

Embedding-based clustering sounds elegant in theory, but in practice it hits several hard limits that matter for e-commerce.

Surface similarity traps

Embeddings place "leather sofa" and "leather cleaner" close together because they share a dominant token. A shopper searching "leather sofa" and one searching "leather cleaner" have completely different purchase intent, but a vector-only approach merges them into one cluster.

Result: Category pages that mix products with cleaning supplies, confusing both shoppers and crawlers.

Synonym blindness

"Couch" and "sofa" mean the same thing, but their embeddings are not identical. "Trainers" and "sneakers" diverge even further because one is British English and the other American. Without an explicit synonym layer, these end up in separate clusters - splitting demand you should be capturing on one page.

Result: Two thin pages competing with each other instead of one strong page capturing both audiences.

No awareness of catalog structure

Embeddings know nothing about your product taxonomy. They cannot distinguish "dining table" from "coffee table" if both appear in similar linguistic contexts. Your site already has a category hierarchy that encodes exactly this distinction - but a pure ML pipeline ignores it.

Result: Clusters that cut across your existing navigation, creating content that does not fit your site architecture.

Abbreviation and modifier confusion

"AC unit" and "air conditioning unit" are the same product. "4K TV" and "ultra-high-definition television" target the same buyer. Without a grammar layer that recognizes abbreviations and modifier equivalences, ML clusters fragment this demand.

Result: Keyword lists that look complete but silently miss 20-40% of real search demand.

What is neuro-symbolic keyword clustering?

Neuro-symbolic AI combines two complementary strengths: the pattern recognition of neural networks (the "neuro" part) with the logical precision of structured knowledge systems (the "symbolic" part). For keyword clustering, this means pairing language-model embeddings with explicit taxonomies, label grammars, synonym dictionaries, and rules.

The neural component handles fuzzy language understanding - recognizing that "mid-century modern desk" and "MCM writing table" share meaning even though they share almost no tokens. The symbolic component enforces structure: a taxonomy that says desks belong under "Office Furniture," not "Kitchen," and a grammar that maps "MCM" to "mid-century modern."

This is the methodology behind Topic Sieve. Instead of treating keyword grouping as a single-step vector similarity problem, it runs keywords through multiple layers of structured and learned intelligence.

Five layers of neuro-symbolic clustering

Each layer adds a different kind of intelligence. Together they catch what any single method misses.

1

Topic taxonomy

A hierarchical tree of topics that mirrors how your products are organized. "Outdoor Furniture > Patio Dining > Dining Sets" is a branch. Every keyword gets mapped to a leaf node in this tree. Keywords that land on the same node (or sibling nodes) belong to the same topic cluster.

What it catches: Structural relationships that embeddings miss. "Patio dining set" and "outdoor table and chairs" both map to the same taxonomy node even though their embedding vectors may diverge.
2

Labels

Labels are reusable tags that describe attributes: "waterproof," "kids," "under $50," "king size." A single keyword might carry multiple labels. Labels let you create sub-clusters within a topic - for example, splitting "waterproof hiking boots" from "leather hiking boots" even though both sit under the same taxonomy node.

What it catches: Attribute-level distinctions that determine whether two keywords belong on the same page or separate pages. A furniture site needs different category pages for "outdoor dining sets" and "indoor dining sets" even though both are dining sets.
3

Grammars

Grammars are pattern rules that normalize how keywords are expressed. They handle word order ("shoes running" to "running shoes"), compound words ("bookcase" to "book case"), and modifier positions. Grammars reduce the surface variation in your keyword list before clustering begins, so the clustering step sees cleaner input.

What it catches: Syntactic variations that inflate keyword counts without representing genuine demand differences. Without grammars, you might build separate pages for "blue velvet sofa" and "velvet sofa blue" - both of which are the same query in different word order.
4

Synonym and abbreviation layers

An explicit dictionary that maps equivalent terms: couch = sofa, trainers = sneakers, AC = air conditioning, 4K = ultra-high-definition. This layer is separate from the neural embeddings because synonyms need to be deterministic - "couch" should always resolve to the same canonical term, regardless of the surrounding context.

What it catches: Regional vocabulary differences, brand-specific abbreviations, and industry shorthand that language models handle inconsistently. This is especially important for product data enrichment pipelines where product titles use abbreviations that shoppers spell out.
5

Language model verification

After the symbolic layers have done their work, an LLM reviews borderline cases: keywords that the taxonomy placed in one cluster but the embeddings suggest belong in another. The LLM acts as a referee, using broader world knowledge to break ties. This step is deliberately last - the structured layers handle 90% of decisions deterministically, and the LLM only weighs in on the ambiguous remainder.

What it catches: Edge cases where rigid rules would be wrong. For example, a taxonomy might place "standing desk converter" under "Desks" and a grammar might normalize it to "desk converter standing," but an LLM recognizes it is an accessory, not a desk, and moves it to the right cluster.

Implicit demand: the keywords your tools never show you

Standard keyword research tools report search volume for exact phrases. They cannot tell you about implicit demand - the queries people would search for if they knew the vocabulary, or the variations a tool has not yet indexed.

Neuro-symbolic clustering exposes implicit demand by working from your product data outward. If your catalog contains "ergonomic mesh office chair with lumbar support," the system generates candidate keywords from product attributes (mesh, ergonomic, lumbar support) and then validates them against search data. This surfaces queries your competitors target but that never appeared in a keyword tool export.

The Demand Without Supply report in Similar AI shows exactly these gaps: topics with proven search demand where your site has no matching page.

Explicit vs implicit demand

Explicit demand (what tools show)
"ergonomic office chair" - 12,100/mo
"mesh office chair" - 6,600/mo
"lumbar support chair" - 3,200/mo
Implicit demand (what clustering reveals)
"breathable desk chair" - synonym of mesh
"office chair for back pain" - same intent as lumbar
"adjustable task chair" - attribute overlap
"ergonomic WFH chair" - abbreviation (WFH = work from home)

Practical workflow: from raw keywords to topic clusters

A step-by-step process for implementing neuro-symbolic keyword clustering on an e-commerce site.

1

Export your product taxonomy

Start with your existing category tree. This is the symbolic backbone. If your site has 200 categories, those 200 nodes become the initial topic structure. Products that already live in a category tell you what keywords should cluster there.

2

Build your synonym and abbreviation dictionary

Collect regional variants (duvet/comforter), abbreviations (LED/light-emitting diode), and brand shorthands (Dyson V15 = Dyson V15 Detect). This dictionary does not need to be exhaustive on day one - it grows as you encounter misclassified keywords.

3

Define label groups

Create attribute labels that apply across your taxonomy: material (wood, metal, fabric), audience (kids, adults, pets), price tier (budget, premium), use case (indoor, outdoor, travel). These labels let you split or merge clusters at any granularity your site architecture demands.

4

Run keywords through the pipeline

Feed your keyword list through synonym normalization, then grammar normalization, then taxonomy mapping, then label assignment. The output is a set of topic clusters, each tied to a taxonomy node and annotated with labels. The final LLM verification step reviews borderline assignments.

5

Map clusters to pages

Each topic cluster becomes a candidate page. Compare against your existing pages: if a cluster matches an existing category, you strengthen that page with better keyword coverage. If no page exists, you have found a gap. The Topic Sieve automates this comparison and flags opportunities ranked by combined search volume.

6

Iterate on the dictionary

Review misclassified keywords monthly. Each correction feeds back into the synonym dictionary, grammar rules, or taxonomy. Over time the symbolic layers get more accurate and the LLM handles fewer edge cases. This feedback loop is what makes the system better with use, unlike a one-shot ML clustering run.

How the approaches compare

A side-by-side look at what you get from each clustering methodology.

DimensionEmbedding-only clusteringNeuro-symbolic clustering
Synonym handlingApproximate - depends on training dataDeterministic via explicit dictionary
Catalog-awareNo - clusters ignore site structureYes - taxonomy mirrors your categories
Abbreviation supportInconsistent across modelsExplicit rules with verified mappings
DebuggabilityOpaque - you cannot inspect why two keywords groupedTransparent - each layer logs its decision
Feedback loopRetrain the model (slow, expensive)Edit a synonym or rule (instant, no retraining)
Edge-case accuracyDegrades on niche or new vocabularyLLM referee resolves borderline cases

See neuro-symbolic clustering in action

Topic Sieve uses the methodology described in this guide to cluster keywords, identify gaps, and recommend new category pages for your e-commerce site. Request a demo to see it working on your own catalog data.