An open-source orchestrator for scalable, disposable development environments - built for training reinforcement learning models and running AI agents.
Already used by JetBrains Research

Training LLMs with reinforcement learning (RL) on coding tasks requires running tens or hundreds of thousands of generation-reward cycles. Each cycle needs a clean, isolated environment with the right source code and tools.
A single machine can't run hundreds of parallel environments that reset between iterations in seconds. You need a distributed system built for that from the ground up.
Agents must not interfere with each other. Every environment needs its own filesystem, process tree, and network to fully reset between episodes.
Spinning up a new container for every cycle is too slow at scale. Environments need to restart or reset in under a second to keep training throughput high.
IdeGYM is a Kubernetes-based orchestration framework that manages the full lifecycle of development environments, from image build to teardown.
Spins up isolated environments on demand and tears them down when finished — no manual cleanup.
Loads projects from a Git URL, archive, or mounted volume. Builds custom Docker images via a plugin API.
Proxies requests from your training loop directly to running pods and forwards responses back, so you can compute rewards offline or replay episodes.
IdeGYM enables multi-turn agentic rollouts with NeMo-Gym.
IdeGYM acts as the containerized resources server, providing the isolated execution environments.
IdeGYM is WebSocket-compatible with OpenEnv.
You can run OpenEnv environments directly on IdeGYM's Kubernetes infrastructure without changing the agent interface.

IdeGYM pairs with verl as the environment layer.
verl handles RL training; IdeGYM provides the scalable sandbox fleet.