Get AI summaries of any video or article — Sign up free
We Taught AI to 100+ Companies - Here's What We Learned thumbnail

We Taught AI to 100+ Companies - Here's What We Learned

Tiago Forte·
6 min read

Based on Tiago Forte's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Second Brain Enterprise’s strongest results came from process development, documentation, and context-rich research/personalization rather than broad automation.

Briefing

AI business transformation is less about “automating everything” and more about teaching companies a shared business language—especially process thinking—so they can feed that context into AI and rapidly redesign how work actually runs. Across two cohorts of Second Brain Enterprise, the program brought together 100+ companies worldwide, spanning solo founders, very small teams, and mid-sized businesses, and found that documentation, process development, and context-rich research/personalization are where AI delivers the most immediate leverage.

A key surprise was how broadly the approach fit different company sizes and industries. Participants ranged from roughly a quarter solo entrepreneurs to another quarter under 10 employees, with the remaining half between 10 and 300. Even larger businesses that initially looked like they might be “too big” ended up becoming top referral partners. The common thread wasn’t industry expertise—it was the ability to translate business problems into a structured framework that AI can use. That translation mattered because many participants didn’t just lack AI capability; they lacked a consistent way to describe their business in operational terms.

The program’s design centered on process as the “invisible structure of business.” Instead of treating process documentation as a niche consulting specialty, it became the gateway to effective AI use. Participants documented existing workflows with AI, assessed their current capabilities across business functions, and then used the results to redesign processes—creating a loop: assess → document → redesign. Over five weeks, their “master prompt” grew more detailed as the cohort added context session by session, which improved the quality of answers over time. By the end, the human coaching value dropped sharply because the AI, now armed with the company’s own context, could cover much of what coaches previously provided.

Another major finding: the work skewed toward “operations transformation,” not raw automation. Most companies started with operations as their target, but the goal wasn’t to remove humans from every step. Instead, AI was used to reimagine processes—making them faster, more efficient, and sometimes eliminating human involvement only where it truly wasn’t needed. Participants also reported that after operations work, their next steps often shifted toward marketing and sales, suggesting the operational clarity unlocked new leverage in revenue functions.

The curriculum also invested heavily in foundational business frameworks—pre-AI “business common sense” that many owners never learned formally. That shared language served two purposes: it aligned humans with each other, and it aligned AI with the same definitions and metrics. Without that alignment, AI can generate plausible answers from too many competing frameworks; with it, teams can compare tactics using consistent unit economics and customer acquisition metrics.

Finally, the cohorts highlighted a human role that’s changing rather than disappearing. Domain expertise remains crucial because AI is strong at producing the “top 20–30%” output, while thought leadership and high-impact differentiation require humans to push toward the “top 1–5%.” The emerging bottleneck becomes time and experimentation—creating an “AI operator” function one or two layers below ownership, responsible for running assessments, adding context to master prompts, and iterating across multiple AI-driven workflows. The broader implication is organizational: AI adoption isn’t just a tool rollout, but a mindset and commitment shift that can spread through entire teams, not only a single AI-enabled operator.

Cornell Notes

Second Brain Enterprise’s two cohorts trained 100+ companies to use AI for business transformation by starting with process and a shared business language. Participants weren’t limited by industry or company size; solo founders through 10–300 employee firms found the framework useful, with documentation and process development emerging as the strongest AI use cases. The program built a “master prompt” that improved over time as teams added context, enabling AI to answer with increasing specificity—often reducing the need for direct coaching by the end. Instead of chasing automation for its own sake, companies pursued operations transformation: redesigning workflows for speed, efficiency, and impact, then often moving next into marketing and sales. The human role shifts toward domain expertise and an “AI operator” who runs assessments, maintains context, and iterates experiments.

Why did “process” become the central lever for AI adoption in the cohort?

Participants treated process documentation and process improvement as the “invisible structure” of business. The program framed process as a shared language—defining what a process overview, blueprint, and guide look like—so humans and AI could reference the same steps, roles, and inputs. This mattered because AI implementation fails when teams can’t provide usable workflow context. The cohort also organized process into layers: strategic planning that informs the customer life cycle, enabled by enablement processes, plus recurring small-business needs like financial performance management, hire-to-retire (workforce culture), and product/service design.

What was the loop companies followed to turn AI from a tool into a transformation engine?

The recurring workflow was assess → document → redesign. First, teams assessed their current state across business functions and exported that context into their master prompt. Next, they used AI to document existing processes. Then they redesigned what the process should look like using AI’s output, feeding the improved understanding back into the master prompt. Over five weeks, each session added more context, so the quality of AI responses improved progressively.

How did the cohort’s AI use differ from the common “automate everything” trend?

Most companies targeted operations from the start, but not by simply automating every step. The emphasis was operations transformation: viewing everything through a process lens while using AI to ask how work could be faster, more efficient, more effective, and sometimes radically different. Human involvement was removed only when it wasn’t necessary, rather than treating automation as the default end goal.

Why did foundational business frameworks matter even after AI became available?

Many owners lacked a consistent set of business definitions and metrics, which made it hard to communicate precisely with AI and with each other. The program taught pre-AI frameworks as table stakes—so the master prompt could use the same language and metrics the team used internally. That alignment enabled precision on issues like customer acquisition costs and lifetime value, and helped teams identify the real bottleneck (for example, discovering a working capital problem they previously misattributed to marketing).

What did the cohort suggest about the best human roles in an AI-enabled company?

Domain expertise became more important, especially for thought leadership. AI was described as strong at producing the “top 20–30%” of outputs, but humans are needed to reach the “top 1–5%” that drives differentiation. The program also pointed to an “AI operator” role below ownership: someone who experiments with AI, consumes AI content, runs assessments, adds context to master prompts, and iterates experiments—because AI still requires time and active input to be effective.

How did “fractional coach” prompts change day-to-day work?

The program created assessment-driven prompts that acted like fractional roles—such as a fractional CMO, fractional CRO (chief revenue officer), and fractional CFO. Once the AI had full context, these prompts became faster than waiting for human experts, and in some cases participants preferred the AI’s output to meetings. The prompts were also personified by role to shape the AI’s perspective (e.g., a CFO’s conservative, risk-aware stance), and they were used through persistent Claude projects/workspaces that supported ongoing conversations.

Review Questions

  1. What specific evidence from the cohort suggests that process documentation improved AI output quality over time?
  2. How did the program’s approach to operations transformation differ from “automation-first” workflows?
  3. Why did teaching foundational business frameworks improve both human alignment and AI usefulness?

Key Points

  1. 1

    Second Brain Enterprise’s strongest results came from process development, documentation, and context-rich research/personalization rather than broad automation.

  2. 2

    The cohorts’ company mix (solo founders, under-10 teams, and 10–300 employee firms) didn’t limit effectiveness; the framework translated across industries.

  3. 3

    AI worked best when teams built a master prompt using shared business frameworks and consistent metrics like customer acquisition cost and lifetime value.

  4. 4

    Participants followed an assess → document → redesign loop, with the master prompt gaining more context each session and producing better answers over time.

  5. 5

    Most companies pursued operations transformation—reimagining workflows for speed and impact—then often shifted next toward marketing and sales.

  6. 6

    The program’s “fractional coach” prompts (e.g., fractional CMO/CRO/CFO) became practical substitutes for some expert Q&A once context was loaded into persistent Claude projects.

  7. 7

    AI adoption created new bottlenecks: domain expertise and time for experimentation, pointing to an emerging “AI operator” role below ownership.

Highlights

The program treated process as the “gateway” to AI value: once teams could describe workflows precisely, AI outputs became more specific and actionable.
Instead of automating every step, companies used AI to transform operations—redesigning processes while keeping humans only where they added necessary judgment.
Master prompts improved through an iterative context loop; by the end, coaching value often dropped because AI could answer with the company’s own specifics.
Fractional-role prompts (CMO/CRO/CFO) worked best when personified and paired with persistent Claude projects that preserved context across conversations.
The biggest human shift wasn’t replacing expertise—it was elevating domain expertise and creating an AI operator function to run assessments and iterate experiments.

Topics

  • AI Business Transformation
  • Process Documentation
  • Master Prompt
  • Operations Transformation
  • Fractional Coaching
  • AI Operator Role