OpAMP-native · v0.42

The hosted OpAMP control plane
for OpenTelemetry Collectors

Monitor collector health, inspect effective configs, sync from GitHub, and roll out changes safely — while your telemetry keeps flowing to the tools you already use.

No telemetry lock-in UI or Git, per configuration Free monitor-only tier
collectors · production · us-east
42 / 42
healthy 38 degraded 3 offline 1 ⌘K to filter
CollectorVersionConfigHeartbeatStatus
otel-edge-01.iad 0.105.0 prod-ingress@v14 2s ago healthy
otel-edge-02.iad 0.105.0 prod-ingress@v14 2s ago healthy
otel-edge-03.iad 0.105.0 prod-ingress@v14 3s ago healthy
otel-gateway-01 0.104.2 prod-gateway@v9 5s ago queue 84%
otel-gateway-02 0.104.2 prod-gateway@v9 14m ago offline
k8s-daemon-iad-a 0.105.0 k8s-daemon@v22 1s ago healthy
rollout · prod-ingress → v14
prod-ingress@v14
via commit a91f3b2
rolling
APPLIED
33
PENDING
8
FAILED
0
github · otel-configs/main
m
tighten batch processor timeout
a91f3b2 · maya · 6m ago
+3 −1 configs/prod-ingress.yaml
The problem

Collectors are easy to deploy.
Fleets are hard to operate.

Once collectors spread across clusters, regions, and teams, scattered YAML and ad hoc scripts stop scaling. You don't need a new place to send telemetry — you need a place to operate the layer that sends it.

01

Which collectors are healthy?

You added 30 nodes last sprint. Three of them haven't reported a heartbeat in an hour — and nobody noticed.

02

Which config is actually running?

The config in the repo is v14. Half the fleet is on v12. The intended state and effective state quietly drift apart.

03

What changed, and who changed it?

Three engineers pushed YAML last week. There is no timeline of who applied what, where, and when.

04

Which rollout failed?

You shipped a tail-sampling tweak. Some collectors picked it up, some rejected it, and the script exited 0 anyway.

05

Are collectors quietly dropping data?

Exporter queues are full. Spans are being dropped. Your backend dashboards look fine. The collector knows; nobody is listening.

Monitor-only adoption

Start with visibility.
No remote changes required.

Connect collectors in monitor-only mode to see health, heartbeat, version, capabilities, and effective configuration. Upgrade to managed mode only when you and your security team are ready.

monitor-only · fleet health · read-only
read-only
healthy 92 degraded 3 offline 1
workspace · managed mode
OPTIONAL · WHEN YOU'RE READY

Enable managed mode

Allow O11yFleet to push managed configurations to selected collectors. Existing monitor-only collectors are unaffected.

Enabled for 12 collectors in env=staging
production fleet remains monitor-only
Managed configurations

Version, roll out, and roll back
without touching a node.

Store and version collector configs in one place. Compare intended state with the effective config each collector reports, and revert instantly when something goes wrong.

prod-ingress · history
  1. v14
    tighten batch processor timeout
    a91f3b2 · maya · 6m ago
    applying
  2. v13
    add k8sattributes processor
    7c2ed11 · jordan · 2d ago
    applied · 42
  3. v12
    try otlp/http exporter
    3a01f9e · sam · 4d ago
    rolled back
  4. v11
    raise memory_limiter to 1500MiB
    b9d44a2 · maya · 1w ago
    applied · 42
prod-ingress.yaml · v13 → v14
12 processors:
13 batch:
14 timeout: 1s
15 send_batch_size: 8192
14+ timeout: 200ms
15+ send_batch_size: 4096
16+ send_batch_max_size: 4096
17 memory_limiter:
18 check_interval: 1s
19 limit_mib: 1500
intended → effective drift detected on 1 collector Resolve
UI or Git, per configuration

Pick the workflow that fits
each configuration.

Production gateways want pull requests and CODEOWNERS. Sandboxes and on-call tweaks want a UI editor. Choose per configuration — edit in the app, or back it with a GitHub repo and a watched path. Both land in the same rollout pipeline.

SOURCE A · UI
prod-ingress · edited inline
maya · 2m ago
+3 −1 · saved as v14
SOURCE B · GIT
a91f3b2
k8s-daemon · via PR
otel-configs/main · 6m ago
VALIDATE
schema · pass
processors compat
dry run · pass
ROLL OUT
edge-iad
12/12
edge-sfo
7/9
gateway
3/6
k8s-daemon
5/15
Flow & collector health

See collector problems before
telemetry disappears.

Pipeline throughput, queue pressure, exporter errors, and drops — the things the collector knows but your backend dashboards don't. Your application telemetry still flows to the tools you already use; we only surface what the collector says about itself.

otel-gateway-01 · pipeline · last 5m
queue 84%
RECEIVERS
otlp/grpc
▮▮▮▮▮▮▯ 142k spans/min
prometheus
▮▮▮▮▯▯▯ 38k pts/min
PROCESSORS
batch
avg 4.1k · p99 6.8k
memory_limiter
refused 132 · throttled
EXPORTER QUEUE
queue_size 8400 / 10000
send_failed11
EXPORTERS
otlp/datadog
→ your backend
otlp/grafana
→ your backend
ACCEPTED
180k /min
REFUSED
132
DROPPED
11
QUEUE
84%
Telemetry neutrality

Your telemetry stays yours.

We manage collectors. We don't replace your observability backend. Control the pipeline without moving the destination.

YOUR APPS
services
nodes
k8s pods
OPENTELEMETRY COLLECTORS
otel collector fleet
managed by O11yFleet
via OpAMP
YOUR BACKENDS
Datadog Grafana Honeycomb New Relic Splunk S3 ClickHouse

O11yFleet sits beside the data path — not in it. Your telemetry never traverses our infrastructure.

Pricing

Free visibility. Paid control.

Start free with monitor-only mode. Upgrade when you need managed configurations, rollout workflows, and team governance.

Free

monitor-only
$0

See every collector's health, version, and effective config. Forever free.

Start free

Business

most teams
From $199/mo

Managed configs, rollouts, GitHub sync, flow dashboards, RBAC, webhooks.

Start trial

Enterprise

custom
Custom

SSO, SCIM, advanced RBAC, audit export, extended retention, SLA.

Talk to sales
Compare all plans →

Connect your first collector in under a minute.

Free, monitor-only, no credit card. Add managed configs and team workflows when you're ready.