Back to Plano

Monitoring

docs/source/guides/observability/monitoring.rst

0.4.224.7 KB
Original Source

.. _monitoring:

Monitoring

OpenTelemetry <https://opentelemetry.io/>_ is an open-source observability framework providing APIs and instrumentation for generating, collecting, processing, and exporting telemetry data, such as traces, metrics, and logs. Its flexible design supports a wide range of backends and seamlessly integrates with modern application tools.

Plano acts a source for several monitoring metrics related to agents and LLMs natively integrated via OpenTelemetry <https://opentelemetry.io/>_ to help you understand three critical aspects of your application: latency, token usage, and error rates by an upstream LLM provider. Latency measures the speed at which your application is responding to users, which includes metrics like time to first token (TFT), time per output token (TOT) metrics, and the total latency as perceived by users. Below are some screenshots how Plano integrates natively with tools like Grafana <https://grafana.com/grafana/dashboards/>_ via Promethus <https://prometheus.io/>_

Metrics Dashboard (via Grafana)

.. image:: /_static/img/llm-request-metrics.png
   :width: 100%
   :align: center

.. image:: /_static/img/input-token-metrics.png
   :width: 100%
   :align: center

.. image:: /_static/img/output-token-metrics.png
   :width: 100%
   :align: center

Configure Monitoring
~~~~~~~~~~~~~~~~~~~~
Plano publishes stats endpoint at http://localhost:19901/stats. As noted above, Plano is a source for metrics. To view and manipulate dashbaords, you will
need to configiure `Promethus <https://prometheus.io/>`_ (as a metrics store) and `Grafana <https://grafana.com/grafana/dashboards/>`_ for dashboards. Below
are some sample configuration files for both, respectively.

.. code-block:: yaml
    :caption: Sample prometheus.yaml config file

    global:
    scrape_interval: 15s
    scrape_timeout: 10s
    evaluation_interval: 15s
    alerting:
    alertmanagers:
        - static_configs:
            - targets: []
        scheme: http
        timeout: 10s
        api_version: v2
    scrape_configs:
    - job_name: plano
        honor_timestamps: true
        scrape_interval: 15s
        scrape_timeout: 10s
        metrics_path: /stats
        scheme: http
        static_configs:
        - targets:
            - localhost:19901
        params:
        format: ["prometheus"]


.. code-block:: yaml
    :caption: Sample grafana datasource.yaml config file

    apiVersion: 1
    datasources:
    - name: Prometheus
        type: prometheus
        url: http://prometheus:9090
        isDefault: true
        access: proxy
        editable: true

Brightstaff metrics
~~~~~~~~~~~~~~~~~~~

In addition to Envoy's stats on ``:9901``, the brightstaff dataplane
process exposes its own Prometheus endpoint on ``0.0.0.0:9092`` (override
with ``METRICS_BIND_ADDRESS``). It publishes:

* HTTP RED — ``brightstaff_http_requests_total``,
  ``brightstaff_http_request_duration_seconds``,
  ``brightstaff_http_in_flight_requests`` (labels: ``handler``, ``method``,
  ``status_class``).
* LLM upstream — ``brightstaff_llm_upstream_requests_total``,
  ``brightstaff_llm_upstream_duration_seconds``,
  ``brightstaff_llm_time_to_first_token_seconds``,
  ``brightstaff_llm_tokens_total`` (labels: ``provider``, ``model``,
  ``error_class``, ``kind``).
* Routing — ``brightstaff_router_decisions_total``,
  ``brightstaff_router_decision_duration_seconds``,
  ``brightstaff_routing_service_requests_total``,
  ``brightstaff_session_cache_events_total``.
* Process & build — ``process_resident_memory_bytes``,
  ``process_cpu_seconds_total``, ``brightstaff_build_info``.

A self-contained Prometheus + Grafana stack is shipped under
``config/grafana/``. With Plano already running on the host, bring it up
with one command:

.. code-block:: bash

    cd config/grafana
    docker compose up -d
    open http://localhost:3000   # admin / admin (anonymous viewer also enabled)

Grafana auto-loads the Prometheus datasource and the brightstaff
dashboard (look under the *Plano* folder). Prometheus scrapes the host's
``:9092`` and ``:9901`` via ``host.docker.internal``.

Files:

* ``config/grafana/docker-compose.yaml`` — one-command Prom + Grafana
  stack with provisioning.
* ``config/grafana/prometheus_scrape.yaml`` — complete Prometheus config
  with ``envoy`` and ``brightstaff`` scrape jobs (mounted by the
  compose).
* ``config/grafana/brightstaff_dashboard.json`` — 19-panel dashboard
  across HTTP RED, LLM upstream, Routing service, and Process & Envoy
  link rows. Auto-provisioned by the compose; can also be imported by
  hand via *Dashboards → New → Import*.
* ``config/grafana/provisioning/`` — Grafana provisioning files for the
  datasource and dashboard provider.