- 필수 기능
- 시작하기
- Glossary
- 표준 속성
- Guides
- Agent
- 통합
- 개방형텔레메트리
- 개발자
- API
- Datadog Mobile App
- CoScreen
- Cloudcraft
- 앱 내
- 서비스 관리
- 인프라스트럭처
- 애플리케이션 성능
- APM
- Continuous Profiler
- 스팬 시각화
- 데이터 스트림 모니터링
- 데이터 작업 모니터링
- 디지털 경험
- 소프트웨어 제공
- 보안
- AI Observability
- 로그 관리
- 관리
Azure OpenAI enables development of copilots and generative AI applications using OpenAI’s library of models. Use the Datadog integration to track the performance and usage of the Azure OpenAI API and deployments.
If you haven’t already, set up the Microsoft Azure integration first. There are no other installation steps.
azure.cognitiveservices_accounts.active_tokens (gauge) | Total tokens minus cached tokens over a period of time. Applies to PTU and PTU-managed deployments. Use this metric to understand your TPS or TPM-based utilization for PTUs and compare to your benchmarks for target TPS or TPM for your scenarios. |
azure.cognitiveservices_accounts.azure_open_ai_requests (count) | Number of calls made to the Azure OpenAI API over a period of time. Applies to PTU, PTU-Managed, and Pay-as-you-go deployments. |
azure.cognitiveservices_accounts.blocked_volume (count) | Number of calls made to the Azure OpenAI API and rejected by a content filter applied over a period of time. You can add a filter or apply splitting by the following dimensions: ModelDeploymentName, ModelName, and TextType. |
azure.cognitiveservices_accounts.generated_completion_tokens (count) | Number of Generated Completion Tokens from an OpenAI model. |
azure.cognitiveservices_accounts.processed_fine_tuned_training_hours (count) | Number of training hours processed on an OpenAI fine-tuned model. |
azure.cognitiveservices_accounts.harmful_volume_detected (count) | Number of calls made to Azure OpenAI API and detected as harmful (both block model and annotate mode) by content filter applied over a period of time. |
azure.cognitiveservices_accounts.processed_prompt_tokens (count) | Number of prompt tokens processed on an OpenAI model. |
azure.cognitiveservices_accounts.processed_inference_tokens (count) | Number of inference tokens processed on an OpenAI model. |
azure.cognitiveservices_accounts.prompt_token_cache_match_rate (gauge) | Percentage of the prompt tokens that hit the cache. Shown as percent |
azure.cognitiveservices_accounts.provisioned_managed_utilization (gauge) | Utilization % for a provisoned-managed deployment, calculated as (PTUs consumed / PTUs deployed) x 100. When utilization is greater than or equal to 100%, calls are throttled and error code 429 is returned. Shown as percent |
azure.cognitiveservices_accounts.provisioned_managed_utilization_v2 (gauge) | Utilization % for a provisoned-managed deployment, calculated as (PTUs consumed / PTUs deployed) x 100. When utilization is greater than or equal to 100%, calls are throttled and error code 429 is returned. Shown as percent |
azure.cognitiveservices_accounts.time_to_response (gauge) | Recommended latency (responsiveness) measure for streaming requests. Applies to PTU and PTU-managed deployments. Calculated as time taken for the first response to appear after a user sends a prompt, as measured by the API gateway. Shown as millisecond |
azure.cognitiveservices_accounts.total_volume_sent_for_safety_check (count) | Number of calls made to the Azure OpenAI API and detected by a content filter applied over a period of time. |
The Azure OpenAI integration does not include any service checks.
The Azure OpenAI integration does not include any events.
Need help? Contact Datadog support.