Get AI summaries of any video or article — Sign up free
Abigail Sutherland - Organizing a Confluence hoard, or, does this page spark joy? thumbnail

Abigail Sutherland - Organizing a Confluence hoard, or, does this page spark joy?

Write the Docs·
5 min read

Based on Write the Docs's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Secure buy-in early by treating Confluence cleanup as a stakeholder-driven project, not a purely technical refactor.

Briefing

A sprawling Confluence setup—33 spaces and more than 20,000 pages—was turning into a daily irritant because people couldn’t find what they needed and teams kept duplicating or letting content rot. Abby Sutherland, TomTom’s Nav information architect, tackled the problem by redesigning the information architecture and then using usage data to decide what deserved to stay, move, or be archived—turning “does this page spark joy?” into a practical, measurable cleanup workflow.

The effort started with a commitment to tidying up, but the real leverage came from getting user buy-in. Once Sutherland took ownership of the Confluence mess, feedback poured in: people surfaced missing context, outdated pages, and pain points that could be used to shape the redesign. From there came a “future site” mindset—either building a greenfield space or restructuring an existing one—paired with a key diagnostic: look for “desire paths.” If users are creating workarounds (like walking off an official route), it signals an unmet need. In Confluence terms, that means fixing the underlying reasons people can’t reach content—often by improving search scope, simplifying browse paths, and surfacing deep content at the top.

Sutherland’s strategy centers on user-first design, but with a hard-nosed view of constraints. Confluence search doesn’t behave like Google, and learned helplessness sets in quickly when people can’t find old material. Tool limitations also shape behavior: Confluence requires unique page names within a space, which makes standardization tricky and can create menu clutter when teams namespace pages. Even tagging and workflow-based review are difficult, so content quality depends heavily on whether teams feel responsible.

To manage those constraints, she structured the hierarchy using a “golden line” between centralized control and team autonomy. The top layer is tightly controlled through a small set of category areas and standardized team landing pages. Those landing pages include a search box and page tree that search only within the team’s space—intentionally narrowing scope to improve signal-to-noise in Confluence search. Below that layer, teams own their internal page structures and can innovate without breaking the overall navigation model.

The cleanup itself relied on a “joy detector.” Instead of trying to manually judge 20,000 pages, Sutherland used Confluence’s API and internal reporting to pull metrics such as page view behavior and last-view timing. After testing candidate metrics, she built a weighted “view count” formula (average daily views divided by days since last viewed, with safeguards against division by zero). She then generated heat maps—color-coded page trees—to highlight which pages were actively used versus likely deadwood. Teams reviewed those maps, migrated what mattered, and archived what didn’t.

By the end of the work so far, five spaces were deleted outright, 13 spaces had their good content extracted and the rest archived, and three spaces were reclassified after organizational changes. The remaining work focuses on the last spaces, plus building a maintenance checklist and schedule so the system doesn’t drift back into chaos. The new consolidated space already exceeds 10,000 pages, but it’s structured to mitigate large-space weaknesses, with average content age and last-view metrics roughly halved compared with the old setup—evidence that the architecture and the data-driven triage are improving discoverability and freshness.

Cornell Notes

Confluence cleanup succeeded by combining user-centered information architecture with data-driven triage. Abby Sutherland redesigned the hierarchy so teams could own their content while centralized navigation and standardized team landing pages improved browse and constrained search. Because Confluence search is weaker than Google, she narrowed search scope (team-only) to boost signal-to-noise and reduce frustration. To decide what to keep, move, or archive, she built a “joy detector” using Confluence API metrics and created color-coded “joy maps” (heat maps) based on a weighted view-count formula. Teams then used those maps to migrate content and archive stale pages, producing a more maintainable “single source of truth.”

What problem did the Confluence overhaul target, beyond “too many pages”?

The setup had 33 Confluence spaces, many outdated or effectively unused, and a large volume of pages (over 20,000). The practical failure mode wasn’t just clutter—it was discoverability and maintenance: people couldn’t find relevant information, search felt unreliable, and teams responded by duplicating or creating new pages instead of updating old ones. That behavior created multiple versions of reference content and accelerated staleness, making Confluence a constant irritation.

How did “desire paths” translate into information architecture decisions?

In urban design, desire paths appear when people need a route that the official layout doesn’t provide. In Confluence, similar patterns show up when users take workarounds because the structure or navigation doesn’t support their goals. The remedy isn’t fencing off the workaround; it’s meeting the underlying need—often by paving the path (supporting the process users already follow) or by adding options that let users accomplish tasks within existing navigation patterns.

Why did constrained search matter so much, and how was it implemented?

Confluence search produces poor results when the scope is huge, creating noise that buries relevant pages. Sutherland’s fix was to add team landing pages with a search box and page tree that search only within that team’s space (the landing page and its child pages). That narrower scope improves the signal-to-noise ratio, making search more usable and reducing learned helplessness.

What was the “golden line” between centralized control and team autonomy?

The hierarchy was designed with a controlled top layer and an autonomous bottom layer. Central control covered the category structure and standardized team landing pages (including required sections like organization/outcomes and navigation elements). Below that golden line, teams could create whatever internal page structures they wanted. This preserved consistency where it mattered for navigation while allowing local flexibility for team-specific organization.

How did the “joy detector” work without manually reviewing thousands of pages?

Sutherland used Confluence’s API and internal reporting to extract page metrics such as who created/updated pages, last viewed time, revision history, and view behavior. She tested candidate metrics against known good and bad examples, then built a weighted “view count” formula: average daily views divided by (days since last viewed + 1) to avoid division by zero. She intentionally removed “content age” from the final metric because user behavior was more predictive than age alone.

What did the heat maps (“joy maps”) enable teams to do?

Heat maps were generated by creating a page tree listing, scraping the page tree children, calculating the weighted view count for each page, and applying conditional formatting to color-code pages. Teams could then quickly see which sections looked like “deadwood” versus “spark joy,” and use their domain knowledge to confirm migrations and archive decisions—turning the data into actionable cleanup.

Review Questions

  1. How did Sutherland’s hierarchy design reduce Confluence search frustration, and what role did search scope play?
  2. Explain why fencing a “desire path” is a bad strategy in both physical spaces and Confluence information architecture.
  3. Describe the weighted view-count formula used in the joy detector and why it performed better than relying on content age alone.

Key Points

  1. 1

    Secure buy-in early by treating Confluence cleanup as a stakeholder-driven project, not a purely technical refactor.

  2. 2

    Diagnose navigation failures by looking for “desire paths”—workarounds that reveal unmet user needs.

  3. 3

    Improve Confluence search by narrowing scope (e.g., team-only search) to increase signal-to-noise and counter learned helplessness.

  4. 4

    Use a two-level hierarchy: centralized, standardized team landing pages for consistent navigation, with autonomy below a “golden line” so teams can manage their own structures.

  5. 5

    Make cleanup measurable with a “joy detector” built from Confluence API metrics, then visualize results using heat maps for team review.

  6. 6

    Handle Confluence constraints (like unique page names) by enforcing naming conventions that keep relevant identifiers visible in menus.

  7. 7

    Plan for maintenance—create checklists and schedules—because backlog-driven cleanup without ongoing governance will drift back into clutter.

Highlights

The most effective search upgrade wasn’t better ranking—it was smaller search scope: team landing pages that search only within a team’s space.
A “joy detector” replaced subjective page triage by using a weighted view-count metric derived from Confluence usage data.
The “golden line” model balanced centralized navigation structure with team-level ownership to prevent chaos while enabling local innovation.
Heat maps turned raw metrics into a practical migration tool, letting teams confirm what should be kept, moved, or archived.

Topics

Mentioned

  • Abigail Sutherland
  • Marie Condo
  • API