First Block: Interview with Gabe Pereyra, Co-Founder and President of Harvey
Based on Notion's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Harvey’s scaling required shifting from document drafting to enterprise-grade orchestration: security, data scoping, and collaboration workflows.
Briefing
Harvey’s growth story hinges on a shift from early, high-value pilots to building the “broader machine” enterprises need: not just AI outputs, but the infrastructure, security, and collaboration workflows that let lawyers safely use models on real client matters. Gabe Pereira, co-founder and president of Harvey, describes landing unusually large seven-figure deals early for an enterprise startup—then learning that winning contracts wasn’t the same as having the operational muscle to scale. The key lesson was moving from a product that drafts or assists with legal work to an enterprise system that can orchestrate tools, manage data scope, and support long-running tasks across a firm.
Harvey began with co-pilots for lawyers and evolved into a platform for large law firms and their clients to collaborate on complex matters like transactions and litigation. The company’s origin traces to Pereira’s work on Meta’s large language model team during the release of GPT-3.5 and GPT-4-era capabilities, alongside Winston’s background as a litigation associate at a top firm. Their early traction came from identifying “customer zero” and “customer one” inside the legal ecosystem: Winston’s network and a senior technology partner at a major firm who quickly rolled the system out firmwide to thousands of lawyers after a pilot.
A recurring theme is that Harvey didn’t start as a narrow automation tool. Early demos that could draft NDAs sounded simple to lawyers, but users needed a mental model closer to “talk with this system and figure out what to do,” especially once they hit limits like missing access to case law or internal firm data. As model reasoning improved—particularly with step-change unlocks from stronger reasoning models—Harvey shifted toward agents that can chain tool calls for tasks such as legal research, reviewing emails, and synthesizing findings for litigation. Accuracy expectations also differed from consumer legal software: large law firms operate with layered review processes, similar to software engineering, where outputs are checked by multiple associates before reaching production.
Instead of relying primarily on custom model training, Pereira emphasizes that most value comes from infrastructure: an “IDE for lawyers” that connects the relevant tools and data for a specific matter (acquisition, litigation, fund formation) and constrains the agent to the correct scope. Domain improvements—post-training, retrieval-augmented generation, and other techniques—still matter, but the company’s emphasis is on orchestration and secure workflow design. He also argues that the next frontier is collaborative training between enterprises and their law firms, since client data can’t be reused broadly; the likely future is “inverted” training where legal models learn from the combined work between a specific enterprise and its counsel.
On go-to-market, the legal industry’s buyer-user split is compounded by an additional stakeholder: the law firm’s clients. Harvey’s early sales leaned on founder-led vision work with CIOs, partners, and clients to align on how teams would collaborate. Pereira’s advice to founders is to use models immediately—because the obvious applications are only the first wave—and to build domain expertise by pairing technical founders with real industry mental models. If rebuilding from scratch, he would invest earlier in foundational security and retention architecture, which became necessary as enterprise deals scaled, but could now be designed with clearer abstractions and collaboration workflows in mind.
Cornell Notes
Harvey’s early success came from deploying LLM-powered legal assistance, but scaling required a bigger shift: building the enterprise “machine” around the models—security, data scoping, collaboration workflows, and tool orchestration. Gabe Pereira credits traction to strong reasoning models that enabled agents to chain research and synthesis tasks, while also noting that large law firms tolerate imperfect drafts because multi-layer associate review functions like software testing. Most performance gains, he says, come from infrastructure that gives agents the right context (case law, emails, motions) and the right tools for each matter. Looking ahead, he expects model training to happen between an enterprise and its law firm, since client data can’t be reused generically across firms. The result is a platform approach rather than a narrow automation product.
Why did Harvey’s product need to evolve beyond “drafting NDAs” into something broader?
What changed as models improved, and how did that affect Harvey’s approach to legal work?
How does Harvey reconcile the need for legal accuracy with imperfect model outputs?
Why does Pereira emphasize infrastructure over custom model training?
What does “customer training” mean in a legal context, and why can’t it be generic?
What go-to-market challenge is unique to enterprise legal, and how did Harvey address it?
Review Questions
- What specific workflow limitations from early demos pushed Harvey toward agents and matter-scoped infrastructure?
- How do multi-layer associate review processes change the accuracy requirements for AI-assisted legal drafting?
- Why does Pereira believe model training will likely occur between an enterprise and its law firm rather than through generic customer fine-tuning?
Key Points
- 1
Harvey’s scaling required shifting from document drafting to enterprise-grade orchestration: security, data scoping, and collaboration workflows.
- 2
Early traction came from strong reasoning models enabling agents to chain tool calls for tasks like research, email review, and litigation synthesis.
- 3
Large law firms can tolerate imperfect AI outputs because layered associate review acts like a quality-control pipeline.
- 4
Most performance gains come from infrastructure that provides the right matter context and tool access, not from constant custom retraining.
- 5
Legal “training” is constrained by client data ownership, making enterprise–law-firm co-training a more realistic path than generic fine-tuning.
- 6
Go-to-market in legal is complicated by the buyer-user split plus the law firm’s client as an additional stakeholder, requiring founder-led alignment work.