Benchmarking Framework#
Isaac Lab provides a comprehensive benchmarking framework for measuring the performance of simulations, training workflows, and system resources. The framework is designed to work without depending on Isaac Sim’s benchmark services, enabling standalone benchmarking with pluggable output backends.
Overview#
The benchmarking framework consists of several key components:
┌─────────────────────────────────┐
│ BaseIsaacLabBenchmark │
│ (benchmark_core.py) │
└───────────────┬─────────────────┘
│
┌───────────┼───────────┐
│ │ │
▼ ▼ ▼
┌───────┐ ┌─────────┐ ┌──────────┐
│Phases │ │Recorders│ │ Backends │
└───────┘ └─────────┘ └──────────┘
Key Components:
BaseIsaacLabBenchmark: Main class for orchestrating benchmark execution
Measurements: Data classes for recording metrics (timing, counts, statistics)
Metadata: Data classes for recording context (hardware, versions, parameters)
TestPhase: Container for organizing measurements into logical groups
Recorders: System information collectors (CPU, GPU, memory, versions)
Backends: Output formatters (JSON, Osmo, OmniPerf, Summary)
See also
For method-level micro-benchmarks that measure asset setter/writer and property performance using mock interfaces (without running full simulations), see Micro-Benchmarks for Performance Testing.
Quick Start#
Basic usage with BaseIsaacLabBenchmark:
from isaaclab.test.benchmark import (
BaseIsaacLabBenchmark,
SingleMeasurement,
StatisticalMeasurement,
StringMetadata,
)
# Initialize benchmark
benchmark = BaseIsaacLabBenchmark(
benchmark_name="MyBenchmark",
backend_type="json",
output_path="./results",
)
# Record measurements
benchmark.add_measurement(
phase_name="simulation",
measurement=SingleMeasurement(
name="fps",
value=1234.5,
unit="frames/sec"
),
)
benchmark.add_measurement(
phase_name="simulation",
measurement=StatisticalMeasurement(
name="step_time",
mean=0.82,
std=0.05,
n=1000,
unit="ms"
),
)
# Add metadata
benchmark.add_measurement(
phase_name="simulation",
metadata=StringMetadata(name="task", data="Isaac-Cartpole-v0"),
)
# Finalize and write output
benchmark._finalize_impl()
Running Benchmark Scripts#
Isaac Lab provides shell scripts for running benchmark suites:
Non-RL Benchmarks#
Measure environment stepping performance without training:
# Run all non-RL benchmarks
./scripts/benchmarks/run_non_rl_benchmarks.sh ./output_dir
# Run a single benchmark manually
./isaaclab.sh -p scripts/benchmarks/benchmark_non_rl.py \
--task Isaac-Cartpole-v0 \
--num_envs 4096 \
--num_frames 100 \
--headless \
--benchmark_backend json \
--output_path ./results
RL Training Benchmarks#
Measure training performance with RSL-RL:
# Run training benchmarks
./scripts/benchmarks/run_training_benchmarks.sh ./output_dir
# Run manually with RSL-RL
./isaaclab.sh -p scripts/benchmarks/benchmark_rsl_rl.py \
--task Isaac-Cartpole-v0 \
--num_envs 4096 \
--max_iterations 500 \
--headless \
--benchmark_backend json \
--output_path ./results
PhysX Micro-Benchmarks#
Measure asset method and property performance using mock interfaces:
# Run PhysX micro-benchmarks
./scripts/benchmarks/run_physx_benchmarks.sh ./output_dir
# Run articulation benchmarks manually
./isaaclab.sh -p source/isaaclab_physx/benchmark/assets/benchmark_articulation.py \
--num_iterations 1000 \
--num_instances 4096
For detailed documentation on micro-benchmarks, including available benchmark files, input modes, and how to add new benchmarks, see Micro-Benchmarks for Performance Testing.
Startup Profiling Benchmark#
Profile the startup sequence of an IsaacLab environment using cProfile. Each
startup stage is wrapped in its own profiling session and the top functions by
own-time are reported. This is useful for investigating startup regressions and
understanding where time is spent during initialization.
# Basic usage — reports top 30 functions per phase
./isaaclab.sh -p scripts/benchmarks/benchmark_startup.py \
--task Isaac-Ant-v0 \
--num_envs 4096 \
--headless \
--benchmark_backend summary
The script profiles five phases independently:
app_launch:
launch_simulation()context entry (Kit/USD/PhysX init)python_imports: importing gymnasium, torch, isaaclab_tasks, etc.
task_config:
resolve_task_config()(Hydra config resolution)env_creation:
gym.make()+env.reset()(scene creation, sim start)first_step: a single
env.step()call
Each phase records a wall-clock time plus per-function own-time and cumulative
time as SingleMeasurement entries. Only IsaacLab functions and first-level
calls into external libraries are included (deep internals of torch, USD, etc.
are filtered out).
Whitelist mode — For dashboard time-series comparisons across runs, use a
YAML whitelist config to report a fixed set of functions instead of top-N.
Patterns use fnmatch syntax (* and ? wildcards):
# Example whitelist config
app_launch:
- "isaaclab.utils.configclass:_custom_post_init"
- "isaaclab.sim.*:__init__"
env_creation:
- "isaaclab.cloner.*:usd_replicate"
- "isaaclab.cloner.*:filter_collisions"
- "isaaclab.scene.*:_init_scene"
first_step:
- "isaaclab.actuators.*:compute"
- "warp.*:launch"
./isaaclab.sh -p scripts/benchmarks/benchmark_startup.py \
--task Isaac-Ant-v0 \
--num_envs 4096 \
--headless \
--benchmark_backend omniperf \
--whitelist_config scripts/benchmarks/startup_whitelist.yaml
Phases listed in the YAML use the whitelist; phases not listed fall back to
--top_n (default: 5 in whitelist mode, 30 otherwise). Patterns that match
no profiled function emit 0.0 placeholders so the output always contains
the same keys.
A default whitelist is provided at scripts/benchmarks/startup_whitelist.yaml.
Argument |
Default |
Description |
|---|---|---|
|
required |
Environment task name |
|
from config |
Number of parallel environments |
|
30 (5 with whitelist) |
Max functions per non-whitelisted phase |
|
None |
Path to YAML whitelist file |
|
|
Output backend ( |
|
|
Directory for output files |
Command Line Arguments#
Common Arguments#
Argument |
Default |
Description |
|---|---|---|
|
|
Output backend: |
|
|
Directory for output files |
|
|
Run without rendering |
Non-RL Benchmark Arguments#
Argument |
Default |
Description |
|---|---|---|
|
required |
Environment task name (e.g., |
|
|
Number of parallel environments |
|
|
Number of simulation frames to run |
|
|
Enable camera rendering (for RGB/depth tasks) |
RL Training Arguments#
Argument |
Default |
Description |
|---|---|---|
|
required |
Environment task name |
|
|
Number of parallel environments |
|
|
Number of training iterations |
Measurement Types#
The framework provides several measurement types for different data:
SingleMeasurement#
For single numeric values:
from isaaclab.test.benchmark import SingleMeasurement
measurement = SingleMeasurement(
name="total_frames",
value=100000,
unit="frames"
)
StatisticalMeasurement#
For statistical summaries:
from isaaclab.test.benchmark import StatisticalMeasurement
measurement = StatisticalMeasurement(
name="step_time",
mean=0.82,
std=0.05,
n=1000,
unit="ms"
)
BooleanMeasurement#
For pass/fail status:
from isaaclab.test.benchmark import BooleanMeasurement
measurement = BooleanMeasurement(
name="converged",
bvalue=True
)
DictMeasurement#
For structured data:
from isaaclab.test.benchmark import DictMeasurement
measurement = DictMeasurement(
name="config",
value={"learning_rate": 0.001, "batch_size": 64}
)
ListMeasurement#
For sequences of values:
from isaaclab.test.benchmark import ListMeasurement
measurement = ListMeasurement(
name="rewards_per_episode",
value=[100.5, 102.3, 98.7, 105.1]
)
Test Phases#
TestPhase organizes measurements and metadata
into logical groups. Common phases include:
benchmark_info: Workflow name, timestamp, configurationhardware_info: CPU, GPU, memory informationversion_info: Software versions (Isaac Sim, PyTorch, etc.)simulation: Environment stepping metricstraining: RL training metricsruntime: Execution time and resource usage
Example:
# Measurements are automatically grouped by phase
benchmark.add_measurement("simulation", measurement=fps_measurement)
benchmark.add_measurement("simulation", metadata=task_metadata)
benchmark.add_measurement("training", measurement=reward_measurement)
Output Backends#
JSON Backend#
Full output with all phases, measurements, and metadata:
./isaaclab.sh -p ... --benchmark_backend json --output_path ./results
Output structure:
[
{
"phase_name": "simulation",
"measurements": [
{
"name": "MyBenchmark simulation fps",
"value": 1234.5,
"unit": "frames/sec",
"type": "single"
}
],
"metadata": [
{"name": "MyBenchmark simulation task", "data": "Isaac-Cartpole-v0", "type": "string"}
]
}
]
Osmo Backend#
Simplified key-value format for CI/CD integration:
./isaaclab.sh -p ... --benchmark_backend osmo --output_path ./results
Output structure:
{
"workflow_name": "MyBenchmark",
"phase": "simulation",
"fps": 1234.5,
"task": "Isaac-Cartpole-v0"
}
OmniPerf Backend#
Format for database upload and performance tracking:
./isaaclab.sh -p ... --benchmark_backend omniperf --output_path ./results
Output structure:
{
"simulation": {
"workflow_name": "MyBenchmark",
"fps": 1234.5,
"step_time_mean": 0.82,
"step_time_std": 0.05
}
}
Summary Backend#
Human-readable console report plus JSON file. Prints a formatted summary to the
terminal while also writing the same data as JSON. Standard phases (runtime,
startup, train, frametime, system info) are rendered with specialized formatting;
any additional phases (e.g., from the startup profiling benchmark) are rendered
automatically with their SingleMeasurement and StatisticalMeasurement
entries. Use when you want a quick readout without opening the JSON:
./isaaclab.sh -p ... --benchmark_backend summary --output_path ./results
When summary is selected, frametime recorders are enabled automatically when
running with Isaac Sim (Kit).
BenchmarkMonitor#
BenchmarkMonitor enables continuous system
monitoring during blocking operations like RL training loops:
from isaaclab.test.benchmark import BaseIsaacLabBenchmark, BenchmarkMonitor
benchmark = BaseIsaacLabBenchmark(
benchmark_name="TrainingBenchmark",
backend_type="json",
output_path="./results",
)
# Monitor system resources during blocking training call
with BenchmarkMonitor(benchmark, interval=1.0):
runner.learn(num_learning_iterations=1000) # Blocking call
benchmark._finalize_impl()
The monitor runs in a background thread and periodically calls
update_manual_recorders() to capture CPU, GPU, and memory usage samples.
System Recorders#
The framework includes built-in recorders for system information:
CPUInfoRecorder#
Captures CPU model, core count, and usage statistics.
GPUInfoRecorder#
Captures GPU model, memory, and utilization via nvidia-smi.
MemoryInfoRecorder#
Captures system and GPU memory usage over time.
VersionInfoRecorder#
Captures software versions:
Isaac Sim version
Isaac Lab version
PyTorch version
CUDA version
Python version
Creating Custom Benchmarks#
Step 1: Initialize Benchmark#
import argparse
from isaaclab.test.benchmark import BaseIsaacLabBenchmark
parser = argparse.ArgumentParser()
parser.add_argument("--benchmark_backend", default="json")
parser.add_argument("--output_path", default="./")
args = parser.parse_args()
benchmark = BaseIsaacLabBenchmark(
benchmark_name="CustomBenchmark",
backend_type=args.benchmark_backend,
output_path=args.output_path,
)
Step 2: Run Your Workload#
import time
start_time = time.time()
# Your workload here
for i in range(num_iterations):
env.step(actions)
elapsed = time.time() - start_time
Step 3: Record Measurements#
from isaaclab.test.benchmark import SingleMeasurement, StringMetadata
benchmark.add_measurement(
phase_name="runtime",
measurement=SingleMeasurement(
name="total_time",
value=elapsed,
unit="seconds"
),
)
benchmark.add_measurement(
phase_name="runtime",
metadata=StringMetadata(name="num_iterations", data=str(num_iterations)),
)
Step 4: Finalize#
benchmark._finalize_impl()
Integration with CI/CD#
The shell scripts in scripts/benchmarks/ are designed for CI/CD integration:
# GitHub Actions / GitLab CI example
- name: Run Benchmarks
run: |
./scripts/benchmarks/run_non_rl_benchmarks.sh ./benchmark_results
- name: Upload Results
uses: actions/upload-artifact@v3
with:
name: benchmark-results
path: ./benchmark_results/
For Osmo integration, use the osmo backend:
./scripts/benchmarks/run_non_rl_benchmarks.sh ./results
# Results are in Osmo-compatible JSON format
Troubleshooting#
Import Errors#
Ensure you’re running through the Isaac Lab launcher:
./isaaclab.sh -p your_benchmark.py
Not:
python your_benchmark.py # Missing environment setup
Missing GPU Metrics#
Verify nvidia-smi is available and CUDA is configured:
nvidia-smi # Should show GPU info
Empty Output Files#
Ensure _finalize_impl() is called before the script exits:
try:
# Your benchmark code
pass
finally:
benchmark._finalize_impl()
Backend Not Recognized#
Valid backend types are: json, osmo, omniperf, summary
# Correct
--benchmark_backend json
# Incorrect
--benchmark_backend JSON # Case sensitive