Skip to content

Commit 62adfc8

Browse files
authored
[MLOB-3525] add setup instructions for llm obs litellm integration (DataDog#20911)
* add setup instructions for llm obs litellm integration * add apm section * remove detailed llm obs and apm sections in favor of linking to llm obs public documentation
1 parent aa0fc72 commit 62adfc8

File tree

1 file changed

+19
-9
lines changed

1 file changed

+19
-9
lines changed

litellm/README.md

Lines changed: 19 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,28 +1,37 @@
1-
# Agent Check: LiteLLM
1+
# LiteLLM
22

33
## Overview
44

5-
[LiteLLM][1] is a lightweight, open-source proxy and analytics layer for large language model (LLM) APIs. It enables unified access, observability, and cost control across multiple LLM providers.
5+
Monitor, troubleshoot, and evaluate your LLM-powered applications built using [LiteLLM][1]: a lightweight, open-source proxy and analytics layer for large language model (LLM) APIs. It enables unified access, observability, and cost control across multiple LLM providers.
66

7-
This integration provides real-time monitoring, alerting, and analytics for all LLM API usage through LiteLLM, helping customers optimize performance, manage costs, and ensure reliability across their AI-powered applications.
7+
Use LLM Observability to investigate the root cause of issues, monitor operational performance, and evaluate the quality, privacy, and safety of your LLM applications.
8+
9+
See the [LLM Observability tracing view video](https://imgix.datadoghq.com/video/products/llm-observability/expedite-troubleshooting.mp4?fm=webm&fit=max) for an example of how you can investigate a trace.
10+
11+
Get cost estimation, prompt and completion sampling, error tracking, performance metrics, and more out of [LiteLLM][1] Python library requests using Datadog metrics and APM.
812

913
Key metrics such as request/response counts, latency, error rates, token usage, and spend per provider or deployment are monitored. This data enables customers to track usage patterns, detect anomalies, control costs, and troubleshoot issues quickly, ensuring efficient and reliable LLM operations through LiteLLM's health check and Prometheus endpoints.
1014

1115
## Setup
1216

17+
### LLM Observability: Get end-to-end visibility into your LLM application using LiteLLM
18+
See the [LiteLLM integration docs][12] for details on how to get started with LLM Observability for LiteLLM.
19+
20+
21+
### Agent Check: LiteLLM
1322
Follow the instructions below to install and configure this check for an Agent running on a host. For containerized environments, see the [Autodiscovery Integration Templates][3] for guidance on applying these instructions.
1423

15-
### Installation
24+
#### Installation
1625

1726
Starting from Agent 7.68.0, the LiteLLM check is included in the [Datadog Agent][2] package. No additional installation is needed on your server.
1827

19-
### Configuration
28+
#### Configuration
2029

2130
This integration collects metrics through the Prometheus endpoint exposed by the LiteLLM Proxy. This feature is only available for enterprise users of LiteLLM. By default, the metrics are exposed on the `/metrics` endpoint. If connecting locally, the default port is 4000. For more information, see the [LiteLLM Prometheus documentation][10].
2231

2332
Note: The listed metrics can only be collected if they are available. Some metrics are generated only when certain actions are performed. For example, the `litellm.auth.failed_requests.count` metric might only be exposed after an authentication failed request has occurred.
2433

25-
#### Host-based
34+
##### Host-based
2635

2736
1. Edit the `litellm.d/conf.yaml` file in the `conf.d/` folder at the root of your Agent's configuration directory to start collecting your LiteLLM performance data. See the [sample litellm.d/conf.yaml][4] for all available configuration options. Example config:
2837

@@ -38,7 +47,7 @@ instances:
3847

3948
2. [Restart the Agent][5].
4049

41-
#### Kubernetes-based
50+
##### Kubernetes-based
4251

4352
For LiteLLM Proxy running on Kubernetes, configuration can be easily done via pod annotations. See the example below:
4453

@@ -69,11 +78,11 @@ spec:
6978

7079
For more information and alternative ways to configure the check in Kubernetes-based environments, see the [Kubernetes Integration Setup documentation][3].
7180

72-
#### Logs
81+
##### Logs
7382

7483
LiteLLM can send logs to Datadog through its callback system. You can configure various logging settings in LiteLLM to customize log formatting and delivery to Datadog for ingestion. For detailed configuration options and setup instructions, refer to the [LiteLLM Logging Documentation][11].
7584

76-
### Validation
85+
#### Validation
7786

7887
Run the Agent's status subcommand ([see documentation][6]) and look for `litellm` under the Checks section.
7988

@@ -109,3 +118,4 @@ Need help? Contact [Datadog support][9].
109118
[9]: https://docs.datadoghq.com/help/
110119
[10]: https://docs.litellm.ai/docs/proxy/prometheus
111120
[11]: https://docs.litellm.ai/docs/proxy/logging
121+
[12]: https://docs.datadoghq.com/llm_observability/instrumentation/auto_instrumentation?tab=python#litellm

0 commit comments

Comments
 (0)