Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.cartesia.ai/llms.txt

Use this file to discover all available pages before exploring further.

Cartesia’s inference cluster includes support for Prometheus, an open source metrics and monitoring solution. All metrics are scraped every 5 seconds via PodMonitor on port 8080 /metrics.

Prometheus Metrics

Metric NameDescriptionNormal Range
inferno_worker_load# of concurrent chunks the worker is processing now< Capacity
inferno_worker_capacity# of concurrent chunks a worker can processhardware dependent
inferno_worker_ttfaTime to First Audio< 200 ms
inferno_worker_rtfReal time factor< 1
api_queue_sizeRequest queue size per offeringLow
api_unserviceable_requests_sizeUnserviceable requests count0