By Guy Gelbart
In 2018, we started asking a different question about keyword research.
The standard approach treated keywords as a list to rank for. You pulled terms from a tool, sorted by volume, and built content around them. But we kept running into the same problem: two keywords could look completely different on the surface and still trigger the exact same search results.
That observation changed how we thought about the entire process.
If Google serves the same pages for different queries, Google has already determined those queries share the same user intent. The search engine has spent billions figuring out what users want. The SERP isnât just a list of competitors â itâs a mirror reflecting user needs back at you.
We called it âsearch listening.â And for six years, we did it manually.
The Manual Process (And Why It Broke)
The methodology was straightforward in theory: gather all possible keywords, analyze the SERPs, and tag queries by shared intent. Group the ones that trigger the same results. Build content around those clusters.
In practice, it created two persistent problems.
Problem one: missing relevant queries. When youâre manually reviewing thousands of keywords, you miss connections. A query phrased differently, an industry synonym you didnât recognize, a long-tail variation that shares intent with your core term â these slip through. And every missed connection is a missed opportunity to consolidate content authority.
Problem two: misattributing intent. The opposite error. You tag two queries as related when theyâre not. You build one page targeting both. Google disagrees with your assessment. Now you have content cannibalization â two pages competing for the same intent, neither ranking well.
Both problems got worse at scale. A cybersecurity client came to us with thousands of technical terms. Industry jargon we didnât know. Acronyms with multiple meanings. The manual process would have taken weeks, and the error rate would have been high.
So we built something.
What We Actually Built
Two days in Google AI Studio. Thatâs how long it took to turn six years of methodology into a working tool.
The core logic mirrors what weâd been doing manually, but without the gaps:
Step one: Ingest the keyword universe. Every relevant term for the clientâs space â pulled from Ahrefs, Search Console, whatever sources apply.
Step two: Map each keyword to its SERP. For every keyword, we capture the top 10 ranking URLs. This is the raw signal â Googleâs interpretation of what users want when they type that query.
Step three: Calculate similarity. For any two keywords, we measure how much their SERPs overlap. The metric is Jaccard similarity: shared URLs divided by total unique URLs. If keyword A and keyword B return seven of the same pages in their top 10, they almost certainly share intent.
Step four: Cluster using Complete Linkage. This is the part that prevents bad groupings. Standard clustering can âchainâ keywords together â A connects to B, B connects to C, so A ends up grouped with C even if theyâre unrelated. Complete Linkage only merges clusters when every keyword in one group is sufficiently similar to every keyword in the other. Tighter clusters. Fewer errors.
Optional: AI Hybrid mode. For sparse niches where SERP overlap is low, we added a semantic layer. The tool embeds both the keywords and their SERP content (titles, snippets) using Googleâs text-embedding model. If two keywords have minimal SERP overlap but very high semantic similarity, they can still cluster â but only if the AI confidence is above 85%. Safety gates prevent hallucinated groupings.
The output: structured intent clusters, each with a representative keyword, supporting terms, and a clear signal for content strategy.
What Changed
The cybersecurity client had terminology weâd never encountered. Industry-specific acronyms. Technical phrases with precise meanings. Manually learning that landscape would have taken weeks of research before we could even begin clustering.
The tool handled it in hours.
It found synonym relationships we wouldnât have caught. It grouped queries by intent even when the phrasing was completely different. It surfaced clusters we didnât know existed—micro-intents within the broader topic space that deserved their own content.
Weâve since deployed the same approach across travel, fintech, web3, and insurance. B2B and B2C. The pattern holds: SERP overlap is a reliable intent signal regardless of industry.
The time savings are obvious. What used to take weeks now takes hours. But the accuracy improvement matters more. Weâre not guessing at intent anymore. Weâre reading Googleâs own interpretation and building strategy around it.
The Deeper Point
This isnât really a story about a tool. Tools are easy to build now â thatâs the whole premise of vibe coding.
The story is about methodology. Six years of developing an approach, testing it across clients, refining the logic. The tool just operationalized what we already knew worked.
Thatâs the pattern worth paying attention to. Domain expertise doesnât become obsolete when AI tools get powerful. It becomes more valuable. The people whoâve spent years understanding a problem are exactly the people who can build the right solution when the building becomes easy.
We didnât automate keyword research. We automated our keyword research â the specific methodology weâd developed through thousands of hours of client work.
The insight that made it possible wasnât technical. It was conceptual: the SERP is a signal layer. Googleâs results arenât random â theyâre a reflection of how the search engine interprets user needs. Read them correctly, and youâre not guessing at intent. Youâre observing it.
That insight took years to develop. The automation took two days.
What Weâd Add Next
The current version solves the clustering problem. But thereâs more signal in the SERP than just URL overlap.
Page types tell you what format users expect. If Google serves mostly comparison pages for a query, users want comparison content â not a product page, not a how-to guide. Weâre building that layer in.
SERP features signal intent nuances. Featured snippets, video carousels, People Also Ask boxes—each one indicates something about what users want and how they want it delivered.
Position stability over time shows which intents are settled versus contested. If the same pages have ranked for years, the intent is clear and the competition is established. If rankings churn constantly, thereâs an opportunity to define the space.
All of this is readable. All of it is a signal. The methodology keeps expanding because the SERP keeps telling us more than most people bother to listen to.
If youâre dealing with large-scale keyword research or struggling to map intent accurately, letâs talk. I share what weâre learning from real client work.
