Skip to content
NewsSubnetsApex

Apex Launches RL Tron Competition on Bittensor

Apex launched RL Tron, a new reinforcement learning competition on Bittensor where AI agents compete in recurring head-to-head Tron tournaments using decentralized bracket-based evaluations.

Apex launches RL Tron on Bittensor Subnet 1, introducing decentralized reinforcement learning tournaments where AI agents compete in recurring head-to-head Tron battles on the TAO network.

Table of Contents

Apex (SN1) has officially launched RL Tron, a new reinforcement learning competition on Bittensor that pits AI agents against each other in recurring head-to-head Tron tournaments.

The competition marks Apex’s first major step into reinforcement learning-focused evaluations, expanding beyond traditional static benchmarking tasks into adversarial environments where agents must continuously adapt against live opponents.

Inspired by the Tron light-cycle game concept, RL Tron challenges participants to train AI-controlled bikes that move through a digital arena while leaving behind permanent trails. Agents must avoid crashing into walls, their own trails, or opposing trails while attempting to trap and outmaneuver their opponent.

According to Apex, tournaments run every two days in a recurring single-elimination bracket format. Participants submit TorchScript-based reinforcement learning models, which are then seeded into bracket-style duels where winners advance until a single surviving miner remains.

“The rules are simple, but the strategy is not,” Apex wrote in the announcement. “Every move changes the map. Every opponent introduces a new strategy to exploit or adapt to.”

Each match consists of multiple Tron games played on a 30x30 grid, with players spawning in opposite corners and competing across up to 500 game ticks per round. Models must process real-time game-state information and return movement decisions within strict timing constraints during every tick of gameplay.

The subnet described RL Tron as more than an arcade-inspired competition, positioning it instead as a live testbed for studying miner behavior and reinforcement learning dynamics in decentralized tournament environments.

“Participants are no longer only optimizing against a fixed dataset or task definition,” Apex wrote. “They are building agents that must compete directly against other agents, adapt to emergent strategies, and survive in a constantly shifting environment.”

The competition is fully open source and playable, allowing the community to inspect submitted strategies, study replay files, and iterate on model designs over time. Miner code is revealed two days after evaluation rounds conclude.

Apex said RL Tron is intended to push the subnet toward “open, decentralized, reinforcement-learning-driven competition,” while introducing a more visual and spectator-friendly format compared to conventional AI evaluation systems.

Participants can join the competition through Apex’s dedicated page.


Disclaimer: This article is for informational purposes only and does not constitute financial, investment, or trading advice. The information provided should not be interpreted as an endorsement of any digital asset, security, or investment strategy. Readers should conduct their own research and consult with a licensed financial professional before making any investment decisions. The publisher and its contributors are not responsible for any losses that may arise from reliance on the information presented.

Comments

Latest