Tutorial: SEO Forensic Audit with Ahrefs

ClickUp's blog dropped from 1.19 million monthly organic visitors to 28,790 in 15 months — and most SEO commentary misdiagnosed the cause. This intermediate tutorial walks through a 10-step forensic framework using Ahrefs to isolate section-level penalties, rule out backlinks, and identify the exact content patterns Google targeted. The methodology applies to any penalized site, any niche.


0

SEO Forensic Autopsy: How to Reverse-Engineer a 97.6% Traffic Collapse

ClickUp’s blog dropped from 1.19 million monthly organic visitors to 28,790 in 15 months — a collapse that most SEO commentary misdiagnosed by a factor of two. This walkthrough shows you how to run a full algorithmic post-mortem on any penalized site: isolating section-level penalties, eliminating false causes like backlinks, and surfacing the exact content patterns Google targeted. The methodology applies to any site, any niche, any penalty type.

  1. Pull the target site’s complete organic search history from Ahrefs, going back at least two years. A single-moment snapshot will misrepresent the scale — ClickUp’s crash was widely reported as a 50% decline; the full historical pull shows 97.6% and still falling. Export month-by-month data so you can plot exact inflection points against external events.

  2. Crawl the site’s XML sitemaps at multiple historical timestamps — the traffic peak and today, at minimum. Comparing which URLs existed at each point reveals what was added, removed, or silently redirected during the decline period.

  3. Map every confirmed Google algorithm update against the traffic timeline using Search Engine Land’s update log overlaid on your Ahrefs export. This step almost always contradicts the intuitive explanation: ClickUp’s blog survived four consecutive 2024 updates and was still growing when it peaked at 1.19M in January 2025 — the March 2025 Core Update broke it permanently, and every update since has compounded the damage.

The inflection point: ClickUp's blog survived four Google updates and kept growing — until March 2025 Core Update hit and every subsequent update compounded the damage
The inflection point: ClickUp’s blog survived four Google updates and kept growing — until March 2025 Core Update hit and every subsequent update compounded the damage
  1. Pull archived versions of the site’s highest-traffic pages from the Wayback Machine, capturing at least one snapshot per major update cycle. Document structural changes version by version: author attribution, content length, promotional density, and list depth. ClickUp’s ChatGPT alternatives article ran six versions over two years — the fifth version recovered traffic from 3,400 to 112,000 monthly visits; every subsequent version increased promotional density until March 2025 ended it permanently.
Page-level autopsy: the ChatGPT alternatives article survived two crashes, recovered to an all-time high of 112K in Feb 2025, then March 2025 Core Update permanently destroyed it
Page-level autopsy: the ChatGPT alternatives article survived two crashes, recovered to an all-time high of 112K in Feb 2025, then March 2025 Core Update permanently destroyed it
  1. Isolate blog subdirectory traffic from root domain traffic in Ahrefs to determine whether the penalty is section-level or sitewide. Calculate the blog’s share of total domain traffic at peak and today. ClickUp’s blog fell from 43.7% of domain traffic to 2.5%, while non-blog pages declined only 27% — confirming Google’s classifier targeted the subdirectory specifically, not the brand.
Blog vs. domain comparison: the full clickup.com domain fell 57.8% — but the blog fell 97.6% and dropped from 43.7% of domain traffic to just 2.5%
Blog vs. domain comparison: the full clickup.com domain fell 57.8% — but the blog fell 97.6% and dropped from 43.7% of domain traffic to just 2.5%
  1. Analyze referring domain growth across the same crash window to rule out backlinks as root cause. ClickUp’s referring domains grew 28% between January 2025 and January 2026 — the exact period traffic fell from 1.19M to 127,000. When domain rating rises while traffic craters, link acquisition is solving a problem that doesn’t exist.
Backlinks ruled out: ClickUp's domain rating rose and referring domains grew 28% during the same period traffic cratered — proving link profile was not the cause
Backlinks ruled out: ClickUp’s domain rating rose and referring domains grew 28% during the same period traffic cratered — proving link profile was not the cause
  1. Audit the six to ten biggest traffic losers for promotional density, template repetition, and asymmetric competitor framing. Count how often the host brand appears as a top recommendation regardless of keyword fit, measure word allocation to the brand versus any competitor entry, and tally promotional CTAs per page. Across 7,000-plus ClickUp posts, the brand’s listicle section ran four to seven times longer than every competitor entry — including a 2,500-word ClickUp section on the free project management software page where competitors received 300 words and a screenshot.
The smoking gun: ClickUp's own core product pages — 'task management software,' 'free project management software' — went to effectively zero traffic, not just off-topic content
The smoking gun: ClickUp’s own core product pages — ‘task management software,’ ‘free project management software’ — went to effectively zero traffic, not just off-topic content
  1. Fetch the JSON-LD structured data from penalized pages and compare it to top-ranking competitors targeting the same keywords. ClickUp’s schema contained a literal [year] placeholder in three fields — article headline, webpage name, and breadcrumb list — that Google’s structured data parser read as the string “year” rather than the rendered value.

Warning: this step may differ from current official documentation — see the verified version below.

  1. Pull current SERPs for the site’s former top commercial keywords to confirm traffic transferred to competitors rather than disappearing from search. Searching “task management software” today — ClickUp’s own product category — returns no ClickUp result in the top ten.
Proof of ranking collapse: search 'task management software' today — ClickUp's own product category — and they don't appear in the top 10
Proof of ranking collapse: search ‘task management software’ today — ClickUp’s own product category — and they don’t appear in the top 10
  1. Compare top-ranking competitor pages for each lost keyword against the penalized pages, documenting execution differences at the content structure level. Zapier’s ChatGPT alternatives page was hit by the same spam backlink network as ClickUp’s version and still ranks number two. The divergence is not links — it’s how each page treats the reader’s question relative to the brand’s promotional agenda.

How does this compare to the official docs?

Google’s public documentation on Helpful Content classifiers, scaled content abuse policies, and JSON-LD structured data requirements adds important precision to several of these steps — and reframes what recovery actually requires.

Here’s What the Official Docs Show

The video’s forensic framework is structurally sound — documentation fills a few navigational gaps and flags one platform change that meaningfully affects how you interpret results today. Where screenshots cover the tools used, the methodology holds; where they don’t, the video remains your primary guide.

Step 1 — Pull organic search history from Ahrefs

Navigate to Site Explorer > [target domain] > Organic search — the video doesn’t spell out this path, and the Ahrefs homepage shown in current documentation won’t get you there directly. Worth knowing: Ahrefs has expanded its positioning to “AI Marketing Platform” since most forensic SEO workflows like this were established, adding Brand Radar for tracking brand visibility across AI overviews, ChatGPT, and Perplexity. Site Explorer’s organic traffic history function is unchanged, but the platform is considerably broader than the tutorial implies.

Ahrefs homepage as of May 2026 — organic search history requires Site Explorer > [domain] > Organic search, not shown here.” style=”max-width: 100%; height: auto; border-radius: 8px; box-shadow: 0 4px 20px rgba(0,0,0,0.15);” /><figcaption style=📄 Ahrefs homepage as of May 2026 — organic search history requires Site Explorer > [domain] > Organic search, not shown here.

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 2 — Crawl XML sitemaps at historical timestamps

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 3 — Map algorithm updates against the traffic timeline

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 4 — Pull Wayback Machine snapshots

The video’s approach here matches the current docs exactly. One addition worth using: the Wayback Machine Availability API, listed under Tools on web.archive.org, enables programmatic multi-date snapshot queries — considerably more efficient than manual URL lookups when you’re tracking page state across five or more historical timestamps.

Wayback Machine homepage (web.archive.org) confirming the URL-based lookup interface for step 4, with the Availability API listed under Tools for bulk programmatic queries.
📄 Wayback Machine homepage (web.archive.org) confirming the URL-based lookup interface for step 4, with the Availability API listed under Tools for bulk programmatic queries.

Step 5 — Isolate blog subdirectory traffic from root domain

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 6 — Analyze referring domain growth

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 7 — Audit promotional density

No official documentation was found for this step — proceed using the video’s approach and verify independently.

Step 8 — Fetch and compare JSON-LD structured data

The video’s approach here matches the current docs exactly. Two additions from the current tool documentation: the Rich Results Test accepts pasted markup via the CODE tab — useful for testing extracted JSON-LD from archived pages when a live URL is no longer accessible, which is directly relevant to step 4’s archival pulls. The tool also defaults to mobile rendering (“Google Inspection Tool smartphone”). Step 8 doesn’t specify device context, and mobile vs. desktop can surface different structured data states for the same page — worth running both when the comparison is inconclusive.

Google Rich Results Test (search.google.com/test/rich-results) confirming URL and CODE input modes for step 8's structured data comparison — note the default mobile rendering context.
📄 Google Rich Results Test (search.google.com/test/rich-results) confirming URL and CODE input modes for step 8’s structured data comparison — note the default mobile rendering context.

Step 9 — Pull current SERPs for former top commercial keywords

The video’s approach here matches the current docs exactly. One live platform change to account for: Google Search now includes an AI Mode button in the search bar. For commercial keywords like “task management software,” AI Mode may generate summaries that surface brand visibility outside traditional ranked results entirely. The tutorial doesn’t address this, and it’s worth checking whether your target keyword triggers AI Mode before drawing conclusions about where the lost traffic redistributed.

Google Search homepage as of May 2026, showing the AI Mode button — a SERP format change with direct implications for commercial keyword analysis in steps 9 and 10.
📄 Google Search homepage as of May 2026, showing the AI Mode button — a SERP format change with direct implications for commercial keyword analysis in steps 9 and 10.

Step 10 — Compare competitor pages by content structure

No official documentation was found for this step — proceed using the video’s approach and verify independently.

  1. Ahrefs — AI Marketing Platform Powered by Big Data — Primary tool for organic traffic history and competitive analysis; access historical data via Site Explorer > [domain] > Organic search, not the homepage.
  2. Wayback Machine — Historical page snapshot tool for step 4; the Availability API (listed under Tools) supports programmatic multi-date queries across multiple URLs.
  3. Rich Results Test — Google Search Console — Official structured data auditing tool for step 8; accepts both live URLs and pasted code, and defaults to mobile rendering.
  4. Google Search — SERP verification interface for step 9; now includes AI Mode, which may alter how visibility is distributed for commercial keywords.

Like it? Share with your friends!

0

What's Your Reaction?

hate hate
0
hate
confused confused
0
confused
fail fail
0
fail
fun fun
0
fun
geeky geeky
0
geeky
love love
0
love
lol lol
0
lol
omg omg
0
omg
win win
0
win

0 Comments

Your email address will not be published. Required fields are marked *