Skip to main content

Bot Tournaments

Welcome to the Reinforce Tactics Bot Tournament page! This page tracks official tournament results between different bot types, showcasing the performance of rule-based bots, LLM-powered bots, and trained RL models.

What are Tournaments?

Tournaments in Reinforce Tactics pit different bot types against each other in a round-robin format. Each matchup consists of multiple games where bots alternate playing as Player 1 and Player 2 to account for first-move advantage.

Tournament Format

  • Round-Robin: Every bot plays against every other bot
  • Fair Play: Equal games with each bot as Player 1 and Player 2
  • Map: Typically played on the 6x6 beginner map for quick matches
  • Results: Win/Loss/Draw records with win rate statistics

Bot Types

SimpleBot

A rule-based AI that follows a simple strategy:

  • Purchases the most expensive units first (one per turn)
  • Prioritizes capturing buildings and towers
  • Attacks nearby enemy units
  • Always included in tournaments

MediumBot

An improved rule-based AI with more sophisticated strategies:

  • Maximizes unit purchases: Buys multiple units per turn instead of just one
  • Coordinated attacks: Plans focus-fire to kill enemies in one turn
  • Smart prioritization: Captures structures closest to its own HQ first
  • Interrupts captures: Attacks enemies that are capturing structures
  • Efficient combat: Evaluates damage-per-cost to make better attack decisions
  • Included in tournaments by default alongside SimpleBot

AdvancedBot

A highly sophisticated AI extending MediumBot with enhanced strategic capabilities:

  • Inherits from MediumBot: Builds upon MediumBot's proven coordinated attack and capture strategies
  • Enhanced unit composition: Optimized purchasing (35% Warriors, 30% Archers, 25% Mages, 10% Clerics)
  • Mountain positioning: Positions units on mountains before attacking for bonus damage
  • Smart ranged combat: Prioritizes ranged attacks with Archers and Mages to minimize counter-attack damage
  • Special ability usage: Effectively uses Mage Paralyze on high-value targets and Cleric Heal on damaged frontline units
  • Map awareness: Analyzes HQ locations and defensive mountain positions on first turn
  • Aggressive tactics: More aggressive attack prioritization and pursuit of enemies
  • Included in tournaments by default alongside SimpleBot and MediumBot

LLM Bots

AI-powered by Large Language Models:

  • OpenAIBot: Powered by GPT models
  • ClaudeBot: Powered by Anthropic's Claude
  • GeminiBot: Powered by Google's Gemini
  • Uses natural language reasoning to make strategic decisions

Model Bots

Trained using Reinforcement Learning:

  • Uses Stable-Baselines3 (PPO, A2C, or DQN)
  • Trained through self-play and opponent challenges
  • Learns optimal strategies through experience

Running Your Own Tournament

Want to run a tournament? It's easy! See the Tournament System guide for detailed instructions.

Quick start:

python3 scripts/tournament.py

Official Tournament Results

Below are the results from official tournaments run on the Reinforce Tactics platform. Results are organized by game version.


Version 0.1.0

Tournament #1 - December 19, 2025

Configuration:

  • Games per matchup: 2 per side (16 total games per bot pair across all maps)
  • Map pool mode: All maps played
  • Starting ELO: 1500

Maps Used:

Beginner Map

beginner.csv
6×6 • 2 Players

Funnel Point Map

funnel_point.csv
13×11 • 2 Players

Center Mountains Map

center_mountains.csv
18×18 • 2 Players

Corner Points Map

corner_points.csv
12×10 • 2 Players

Map Legend:

  • 🟩 Green - Plains (walkable terrain)
  • 🟦 Blue - Ocean/Water (impassable)
  • Light Gray - Mountains (impassable, provides attack bonus)
  • 🌲 Dark Green - Forest (provides defense bonus)
  • 🟫 Brown - Roads (faster movement)
  • 🔴 Red - Player 1 structures (HQ, Buildings)
  • 🔵 Blue Tint - Player 2 structures (HQ, Buildings)
  • White/Gray - Neutral Towers/Buildings

Participants:

  • SimpleBot (rule-based)
  • MediumBot (rule-based)
  • AdvancedBot (rule-based)
  • Claude Haiku 4.5 (claude-haiku-4-5-20251001)

Final Standings

RankBot NameWinsLossesDrawsTotal GamesWin RateELOELO Change
🥇 1AdvancedBot362104875.0%1693+193
🥈 2MediumBot1910194839.6%1575+75
🥉 3SimpleBot12027482.1%1405-95
4Claude Haiku 4.522620484.2%1327-173

Head-to-Head Results

MatchupWinsLossesDrawsNotes
SimpleBot vs MediumBot0610MediumBot dominates
SimpleBot vs AdvancedBot0124Complete sweep by AdvancedBot
SimpleBot vs Claude Haiku 4.51213Mostly draws, slight LLM edge
MediumBot vs AdvancedBot2104AdvancedBot clearly superior
MediumBot vs Claude Haiku 4.51105MediumBot perfect record
AdvancedBot vs Claude Haiku 4.51402Complete domination

Per-Map Performance

Beginner Map
beginner.csv (6×6)
BotWinsLossesDraws
AdvancedBot1200
MediumBot444
Claude Haiku 4.5273
SimpleBot183
Funnel Point Map
funnel_point.csv (13×11)
BotWinsLossesDraws
MediumBot804
AdvancedBot624
SimpleBot048
Claude Haiku 4.5084
Center Mountains Map
center_mountains.csv (18×18)
BotWinsLossesDraws
AdvancedBot606
MediumBot129
SimpleBot0210
Claude Haiku 4.5039
Corner Points Map
corner_points.csv (12×10)
BotWinsLossesDraws
AdvancedBot1200
MediumBot642
SimpleBot066
Claude Haiku 4.5084

Analysis

Key Findings:

  1. AdvancedBot dominance: The AdvancedBot performed exceptionally well, achieving a 75% win rate and gaining +193 ELO. Its sophisticated strategies (mountain positioning, ranged combat, special abilities) proved highly effective.

  2. MediumBot solid performance: With a 39.6% win rate and +75 ELO gain, MediumBot showed the value of coordinated attacks and maximized unit production over SimpleBot's basic approach.

  3. LLM bot struggles: Claude Haiku placed last with only a 4.2% win rate (-173 ELO). This suggests that current LLM reasoning may not be optimized for the tactical decision-making required in this game format.

  4. High draw rates: Many matches ended in draws, particularly on center_mountains.csv (defensive terrain) and between lower-performing bots. This indicates the game's strategic depth and defensive viability.

  5. Map influence: AdvancedBot dominated on beginner.csv and corner_points.csv (100% win rate), while MediumBot actually outperformed on funnel_point.csv.

Notable Observations:

  • AdvancedBot never lost on 3 of 4 maps
  • The hierarchy is clear: AdvancedBot > MediumBot > SimpleBot/Claude Haiku
  • Center mountains map produces the most draws (defensive terrain advantage)

Version 0.1.1

Tournament #1 - December 29, 2025

Configuration:

  • Games per matchup: 8 games per bot pair (1 per side per map)
  • Map pool mode: All maps played
  • Starting ELO: 1500
  • Bots: 8 participants (3 rule-based, 5 LLM-powered)

Maps Used:

Beginner Map

beginner.csv
6×6 • 2 Players

Funnel Point Map

funnel_point.csv
13×11 • 2 Players

Center Mountains Map

center_mountains.csv
18×18 • 2 Players

Corner Points Map

corner_points.csv
12×10 • 2 Players

Participants:

  • SimpleBot (rule-based)
  • MediumBot (rule-based)
  • AdvancedBot (rule-based)
  • Claude Haiku 4.5 (claude-haiku-4-5-20251001)
  • Claude Sonnet 4.5 (claude-sonnet-4-5-20250929)
  • ChatGPT 5 Mini (gpt-5-mini-2025-08-07)
  • Gemini 2.0 Flash (gemini-2.0-flash)
  • Gemini 3.0 Flash (gemini-3-flash-preview)

Final Standings

RankBot NameWinsLossesDrawsTotal GamesWin RateELOELO Change
🥇 1AdvancedBot341215660.7%1702+202
🥈 2Gemini 3.0 Flash334195658.9%1646+146
🥉 3MediumBot268225646.4%1618+118
4Claude Sonnet 4.51913245633.9%1508+8
5SimpleBot12035561.8%1410-90
6Claude Haiku 4.532231565.4%1389-111
7ChatGPT 5 Mini22232563.6%1386-114
8Gemini 2.0 Flash02828560.0%1341-159

Head-to-Head Results

MatchupWinsLossesDrawsNotes
AdvancedBot vs Gemini 3.0 Flash404AdvancedBot undefeated
AdvancedBot vs MediumBot512AdvancedBot dominant
AdvancedBot vs Claude Sonnet 4.5503Complete domination
Gemini 3.0 Flash vs MediumBot305Gemini edges out
Gemini 3.0 Flash vs Claude Sonnet 4.5404Gemini superior
MediumBot vs Claude Sonnet 4.5404MediumBot perfect
Claude Sonnet 4.5 vs Claude Haiku 4.5503Sonnet dominates Haiku
Claude Sonnet 4.5 vs ChatGPT 5 Mini404Sonnet edges GPT
Claude Sonnet 4.5 vs Gemini 2.0 Flash602Sonnet sweeps
Claude Haiku 4.5 vs ChatGPT 5 Mini008Complete stalemate
Gemini 2.0 Flash vs SimpleBot008All draws

Per-Map Performance

Beginner Map
beginner.csv (6×6)
BotWinsLossesDraws
AdvancedBot1400
Gemini 3.0 Flash1220
Claude Sonnet 4.5752
MediumBot644
ChatGPT 5 Mini284
Claude Haiku 4.5284
SimpleBot185
Gemini 2.0 Flash095
Funnel Point Map
funnel_point.csv (13×11)
BotWinsLossesDraws
Gemini 3.0 Flash1112
MediumBot914
Claude Sonnet 4.5851
AdvancedBot419
Claude Haiku 4.5167
SimpleBot068
ChatGPT 5 Mini068
Gemini 2.0 Flash077
Center Mountains Map
center_mountains.csv (18×18)
BotWinsLossesDraws
AdvancedBot4010
Claude Sonnet 4.53011
Gemini 3.0 Flash2012
MediumBot2111
SimpleBot0113
ChatGPT 5 Mini0212
Claude Haiku 4.50212
Gemini 2.0 Flash059
Corner Points Map
corner_points.csv (12×10)
BotWinsLossesDraws
AdvancedBot1202
MediumBot923
Gemini 3.0 Flash815
Claude Sonnet 4.51310
SimpleBot059
ChatGPT 5 Mini068
Claude Haiku 4.5068
Gemini 2.0 Flash077

LLM Token Usage

ModelInput TokensOutput TokensTotal TokensGames
Gemini 3.0 Flash24,174,2401,295,61825,469,85858
ChatGPT 5 Mini13,899,7184,508,52118,408,23957
Claude Sonnet 4.516,585,568695,19917,280,76756
Claude Haiku 4.516,116,414552,45516,668,86956
Gemini 2.0 Flash12,057,436402,55512,459,99156

Analysis

Key Findings:

  1. AdvancedBot maintains dominance: With a 60.7% win rate and +202 ELO, AdvancedBot proves its sophisticated strategies remain effective even against a larger, more diverse field of competitors.

  2. Gemini 3.0 Flash emerges as top LLM: Gemini 3.0 Flash achieved an impressive 58.9% win rate (+146 ELO), becoming the first LLM bot to finish in the top 3. It went undefeated against all other LLM bots.

  3. Claude Sonnet 4.5 shows promise: With a 33.9% win rate, Claude Sonnet significantly outperformed other Claude and OpenAI models, demonstrating improved tactical reasoning over Haiku.

  4. LLM tier separation: Clear performance tiers emerged among LLMs: Gemini 3.0 Flash (top tier), Claude Sonnet 4.5 (mid tier), and Claude Haiku/ChatGPT 5 Mini/Gemini 2.0 Flash (bottom tier).

  5. Map-specific patterns:

    • beginner.csv: Decisive outcomes with few draws - fast-paced combat favors aggressive bots
    • center_mountains.csv: Extremely high draw rate (75%+ for most bots) - defensive terrain creates stalemates
    • funnel_point.csv: Gemini 3.0 Flash and MediumBot excelled with 11 and 9 wins respectively

Notable Observations:

  • AdvancedBot's only loss came from MediumBot (1 game out of 56)
  • Gemini 2.0 Flash won 0 games across all matchups - the only bot with no wins
  • Claude Haiku 4.5 vs ChatGPT 5 Mini resulted in 8 draws - complete stalemate
  • Token efficiency: Claude models used ~300K tokens per game vs ChatGPT's ~323K per game

How to Contribute Results

Have you run a tournament? Share your results with the community!

  1. Run a tournament using scripts/tournament.py
  2. Save the results (CSV/JSON files)
  3. Submit a pull request with your results
  4. Include replay files for verification

Insights and Analysis

Strategy Patterns

As tournaments are completed, we'll analyze:

  • Opening strategies and unit compositions
  • Economic vs. military balance
  • Successful tactical patterns
  • Common mistakes and pitfalls

Bot Strengths and Weaknesses

Each bot type has different characteristics:

  • SimpleBot: Predictable but consistent; makes basic strategic errors
  • MediumBot: More strategic than SimpleBot; uses coordinated tactics and better unit purchasing
  • AdvancedBot: Extends MediumBot with enhanced unit composition, mountain positioning, ranged combat prioritization, and special ability usage
  • LLM Bots: Creative but sometimes unpredictable
  • Model Bots: Optimized but may overfit to training conditions

Map-Specific Performance

Different maps favor different strategies:

  • Small maps (6x6): Fast, aggressive play
  • Medium maps (10x10): Balanced economic/military
  • Large maps (24x24+): Long-term strategy and positioning

Tournament Archive

All tournament replays are saved and can be watched using the game's replay system:

# Load a replay file
python3 main.py --mode play --replay path/to/replay.json

Replay files include:

  • Complete action history
  • Game metadata (bots, map, duration)
  • Can be analyzed programmatically

Future Tournaments

Upcoming tournament ideas:

  • Map Variety: Tournaments on different map sizes and layouts
  • Specialized Competitions: Economy-focused, combat-focused, speed-run
  • Evolution: Tournament of evolved/improved model versions
  • Human vs. Bot: Special exhibitions with human players

Resources


Last updated: December 2025

Tournament results will be updated as official competitions are completed.