System health monitoring panel for the MCAP viewer
The Diagnostics panel displays real-time system health metrics inside the MCAP viewer. Monitor CPU, memory, disk, network, GPU, and battery data alongside your sensor recordings to correlate device performance issues with recording artifacts.
The Diagnostics Panel is in preview. Features described on this page may change.
The panel subscribes to standard diagnostics topics and renders each metric as a live gauge with sparkline history. All metrics synchronize with the shared viewer timeline.
Metric
Source Topic
Description
CPU Usage
/diagnostics/cpu
Per-core and aggregate CPU utilization
Memory
/diagnostics/memory
RSS, heap, available memory
Disk I/O
/diagnostics/disk
Read/write throughput, queue depth
Network
/diagnostics/network
Bandwidth, packet loss, latency
GPU
/diagnostics/gpu
GPU utilization, VRAM, temperature
Battery
/diagnostics/battery
Charge level, discharge rate, health
Each metric is rendered as a gauge showing the current value and a sparkline showing recent history. Hover over the sparkline to inspect values at a specific timestamp.
Displays per-core utilization bars and an aggregate percentage. The panel parses standard diagnostic_msgs/DiagnosticArray messages where the hardware ID matches cpu. Individual core values are extracted from key-value pairs in the values array.
The panel automatically monitors the health of every topic in the recording by tracking message rates, inter-message latency, and dropped messages.
Indicator
Meaning
Green
Healthy — messages arriving at expected rate
Yellow
Degraded — message rate below 80% of expected
Red
Unhealthy — no messages received for >5s
Gray
Inactive — topic has no recent data
The expected rate for each topic is inferred from the first 100 messages in the recording. You can also set explicit expected rates in the panel configuration.
The table view lists every topic with its current health status, message rate, average latency, and drop count. Click a column header to sort. Click a row to jump to the most recent message from that topic in the timeline.
By default, the panel infers expected rates from the first 100 messages on each topic. For topics with variable rates (such as event-driven triggers), this inference may not be accurate. Override the expected rate for specific topics in the panel settings:
The Diagnostics panel can be placed in any slot of the multi-window layout. A common arrangement is to dock it along the bottom of the viewer alongside the Plot and Log panels, providing a full system health overview beneath the primary 3D or image views.
When the panel is docked in a narrow strip, it automatically switches to compact mode. Compact mode shows only the gauge values without sparklines, fitting more metrics into a smaller area. Expand the panel to restore the full view with sparklines and the topic health table.
Connect diagnostics thresholds to fleet-level alert channels. When a metric crosses its configured threshold during a recording, Avala creates a timeline event and optionally triggers a notification.
Export diagnostics metrics from a recording as CSV or JSON for offline analysis. Open the panel settings menu and select Export Metrics. Choose the time range, metrics to include, and output format.
Format
Description
CSV
One row per sample, columns for each metric field
JSON
Array of timestamped metric objects
Exported files include all raw metric values at the original sample rate, without any downsampling applied by the panel’s display refresh rate.
Monitor robot CPU usage during complex manipulation tasks to identify compute bottlenecks that cause control loop delays.
Network latency debugging
Track network latency and packet loss to identify data transmission issues between the robot and remote operator station.
GPU thermal analysis
Correlate GPU temperature spikes with rendering quality drops or inference slowdowns in on-device perception pipelines.
Battery discharge profiling
Profile battery discharge rates across different operational modes to optimize mission duration and plan recharging schedules.
Place the Diagnostics panel next to the Plot panel to overlay system health metrics with sensor signals. This makes it straightforward to correlate a CPU spike with a dropped LiDAR frame or a network latency spike with a delayed camera image.