Dendi vs. OpenAI at The International 2017
Based on OpenAI's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
OpenAI’s Shadowfiend bot beat Dendi in a constrained 1v1 at The International 2017, showing self-play training can reach elite human performance in Dota.
Briefing
OpenAI’s AI Shadowfiend crushed Dendi in a one-on-one match at The International 2017, using a training approach built on self-play rather than hand-coded Dota strategy. The result mattered because it demonstrated that a system trained through repeated matches against itself could reach—within a constrained ruleset—performance levels that beat elite human competition in a complex, real-time game.
The event set up a rare 1v1 format: no bottle, no runes, no neutrals, no raindrops, no shrines, and no Soul Ring. Dendi entered as the first contender, facing an OpenAI-built bot also using Shadowfiend. Before the match, the OpenAI team framed the project as a “one-versus-one Shadowfiend” designed to learn Dota by playing lifetimes of 1v1 games against itself. Instead of programming a playbook or importing expert tactics, the system was trained from randomness, then iteratively improved through small adjustments that gradually produced stronger decision-making.
During the first game, the bot’s laning and combat tempo stood out. It repeatedly pressured the lane, punished openings, and maintained a pace that made Dendi’s typical responses feel insufficient. Commentary from the stage emphasized how the bot didn’t merely win through a single trick; it kept applying pressure in ways that looked like a “smarter version” of the human player it was modeled after—while also showing behaviors that felt difficult to anticipate. Dendi’s attempts to stabilize the lane and find kill opportunities repeatedly ran into the bot’s speed and consistency.
After the match, OpenAI’s Greg and Jakob addressed how the learning worked. The bot was not hard-coded with strategies or trained directly from human expert data. It started from complete randomness, then improved through self-play, eventually reaching pro-level play within the match constraints. Early training looked chaotic—many games ended quickly with Shadowfiend dying to its own mistakes—before the system discovered more effective patterns. The “breakthrough” wasn’t a single leap; it was a chain of incremental improvements that eventually produced aggressive play, better timing, and more reliable execution.
A key question from the stage focused on scale: how could a computer learn so much without human experience? The answer pointed to training time and repetition. The system could learn from scratch in about two weeks of real time, and after roughly an hour it could already beat built-in bots; reaching the level seen on stage required far more experience. The team also positioned the project as a step toward general learning systems—tools that can master complicated, messy tasks—rather than a one-off game bot.
Dendi, despite losing, treated the experience as both unsettling and instructive. The bot’s behavior felt human-adjacent in its relentless lane control and instant reactions, yet also “something else,” suggesting a new kind of competence. The night ended with a promise of the next iteration—vU5—along with an invitation for early challengers to test their skills against the machine.
Cornell Notes
OpenAI’s Shadowfiend bot beat Dendi in a constrained 1v1 at The International 2017. The system wasn’t hand-programmed with Dota tactics or trained on human expert replays; it learned by playing 1v1 matches against itself, starting from randomness and improving through many small iterations. Early training produced chaotic games, but over time the bot developed aggressive lane pressure, better timing, and consistent execution. The match mattered because it showed self-play training can reach pro-level performance in a complex real-time game, and it was framed as a step toward broader “general learning” systems for difficult real-world tasks.
What rules made the 1v1 match a controlled test of laning and combat rather than item-based variety?
How did OpenAI describe the bot’s learning method?
What did the team say about what early training looked like?
Why was the bot’s performance considered more than a one-time trick?
What timeline did OpenAI give for learning from scratch to different skill levels?
How did the project connect to goals beyond Dota?
Review Questions
- What specific rule restrictions in the 1v1 match reduced item and map variability, and how might that change what skills the bot must learn?
- Explain the difference between self-play learning and hard-coded strategy. Why does starting from randomness matter for the bot’s eventual performance?
- According to the stage discussion, what were the observable signs that the bot improved over time (early chaos vs later aggression and consistency)?
Key Points
- 1
OpenAI’s Shadowfiend bot beat Dendi in a constrained 1v1 at The International 2017, showing self-play training can reach elite human performance in Dota.
- 2
The match removed major swing factors—no bottle, runes, neutrals, raindrops, shrines, or Soul Ring—placing more weight on lane control and fight execution.
- 3
The bot was not hard-coded with Dota strategy and wasn’t trained directly from human expert gameplay; it learned by playing 1v1 against itself.
- 4
Training began from complete randomness and improved through many small iterations, with early games often ending quickly due to mistakes.
- 5
OpenAI described a learning timeline: about two weeks to learn from scratch, with early strength against built-in bots after roughly an hour.
- 6
The project was positioned as a step toward general learning systems for complex real-world tasks, not just a one-off game achievement.