This page is not yet available in Spanish. We are working on its translation. If you have any questions or feedback about our current translation project, feel free to reach out to us!
LLM Observability is not available in the selected site () at this time.
Overview
This guide uses the LLM Observability SDK for Python. If your application is written in another language, you can create traces by calling the API instead.
Setup
Jupyter notebooks
To better understand LLM Observability terms and concepts, you can explore the examples in the LLM Observability Jupyter Notebooks repository. These notebooks provide a hands-on experience, and allow you to apply these concepts in real time.
Command line
To generate an LLM Observability trace, you can run a Python script.
The following example script uses OpenAI, but you can modify it to use a different provider. To run the script as written, you need:
An OpenAI API key stored in your environment as OPENAI_API_KEY. To create one, see Account Setup and Set up your API key in the official OpenAI documentation.
The OpenAI Python library installed. See Setting up Python in the official OpenAI documentation for instructions.
Install the SDK by adding the ddtrace and openai packages:
pip install ddtrace
pip install openai
Create a Python script and save it as quickstart.py. This Python script makes a single OpenAI call.
quickstart.py
importosfromopenaiimportOpenAIoai_client=OpenAI(api_key=os.environ.get("OPENAI_API_KEY"))completion=oai_client.chat.completions.create(model="gpt-3.5-turbo",messages=[{"role":"system","content":"You are a helpful customer assistant for a furniture store."},{"role":"user","content":"I'd like to buy a chair for my living room."},],)
Run the Python script with the following shell command. This sends a trace of the OpenAI call to Datadog.
For more information about required environment variables, see the SDK documentation.
Note: DD_LLMOBS_AGENTLESS_ENABLED is only required if you do not have the Datadog Agent running. If the Agent is running in your production environment, make sure this environment variable is unset.
If your application consists of more elaborate prompting or complex chains or workflows involving LLMs, you can trace it using the Setup documentation and the SDK documentation.