Cette page n'est pas encore disponible en français, sa traduction est en cours. Si vous avez des questions ou des retours sur notre projet de traduction actuel, n'hésitez pas à nous contacter.
Join the Beta!
SaaS Cost Integrations are in public beta.
Overview
SaaS Cost Integrations allow you to send cost data directly from your providers by configuring the accounts associated with your cloud cost data in Datadog.
If your provider is not supported, use Custom Costs to upload any cost data source to Datadog and understand the total cost of your services.
Setup
To use SaaS Cost Integrations, you must configure Cloud Cost Management for AWS, Azure, or Google Cloud.
See the respective documentation for your cloud provider:
Enter the workspace name, url, and access token corresponding to your Databricks account.
Under the Select products to set up integration section, click the toggle for each account to enable Databricks Cloud Cost Management.
Enter a System Tables SQL Warehouse ID corresponding to your Databricks instance’s warehouse to query for system table billing data.
Click Save Databricks Workspace.
Your Databricks cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Create or acquire an API key with the billing admin role in Confluent Cloud.
Enter your Confluent Cloud account name, API key, API secret, and optionally, specify tags.
Under the Resources section, click the toggle for Collect cost data to view in Cloud Cost Management.
Click Save.
Your Confluent Cloud cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
If you wish to collect cluster-level tags or business metadata tags for your costs, you can add a Schema Registry API key and secret. Please look into Schema Management on Confluent Cloud for more information.
Create an API token in MongoDB with Organizational Billing Viewer permissions, and add Organizational Read Only permissions for cluster resource tags.
Enter your MongoDB account name, public key, private key, organizational ID, and optionally, specify tags.
Click Save.
Your MongoDB cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Enter your Snowflake account URL, for example: https://xyz12345.us-east-1.snowflakecomputing.com.
Under the Connect your Snowflake account section, click the toggle to enable Snowflake in Cloud Cost Management.
Enter your Snowflake user name in the User Name field.
Create a Datadog-specific role and user to monitor Snowflake.
Run the following in Snowflake to create a custom role:
-- Create a new role intended to monitor Snowflake usage.
create role DATADOG;-- Grant privileges on the SNOWFLAKE database to the new role.
grant imported privileges on database SNOWFLAKE to role DATADOG;-- Grant usage to your default warehouse to the role DATADOG.
grant usage on warehouse <WAREHOUSE> to role DATADOG;-- If you have cost usage collection enabled, ensure that your credentials have permission to view the ORGANIZATION_USAGE schema.
grant role orgadmin to role DATADOG
-- Create a user.
create user DATADOG_USER
LOGIN_NAME= DATADOG_USER
password= <PASSWORD>
default_warehouse= <WAREHOUSE>
default_role= DATADOG
-- Grant the monitor role to the user.
grant role DATADOG to user <USER>
Configure the key-value pair authentication:
Generate a private key by following the official Snowflake documentation and upload the private key file by clicking Upload Key.
Your Snowflake cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Go to the API Key section in your Elastic Cloud organization’s settings.
Click Create New Key.
Choose a Name and Expiration Date for your API key.
Enter your Elastic Cloud Organization ID and Billing API Key in the account table.
Your Elastic Cloud cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Enter your OpenAI account name, input your API key, and optionally, specify tags.
Under the Resources section, click the toggle for each account to enable OpenAI Billing Usage Data Collection.
Click Save.
Your OpenAI cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Create an API token with at least the "global:read" scope and "Billing" role on the Personal API tokens page in Fastly.
Your Fastly cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
Under the Resources section, click the toggle for each account to enable Twilio in Cloud Cost Management.
Enter an Account SID for your Twilio account.
Click Save.
Your Twilio cost data for the past 15 months can be accessed in Cloud Cost Management after 24 hours. To access the available data collected by each SaaS Cost Integration, see the Data Collected section.
The following table contains a non-exhaustive list of out-of-the-box tags associated with each SaaS Cost integration.
Tag Name
Tag Description
record_id
Unique ID for this record.
account_id
ID of the account this report was generated for.
workspace_id
ID of the Workspace this usage was associated with.
cloud
Cloud this usage is relevant for. Possible values are AWS, AZURE, and GCP.
billing_origin_product
Product or feature originating the billing event (for example, JOBS, CLUSTERS).
usage_type
Type of usage being billed (for example, COMPUTE_TIME).
job_run_id
Identifier for the specific job run (if applicable).
node_type
Type of node used in this billing record (for example, m5d.large).
destination_region
Region where the workload is directed (if applicable).
central_clean_room_id
ID of the central clean room associated with the workload (if applicable).
notebook_path
Path to the notebook in Databricks (if applicable).
job_name
Name of the job in Databricks (if applicable).
notebook_id
ID of the notebook used in this billing record (if applicable).
dlt_update_id
Delta Live Table update ID associated with this usage (if applicable).
job_id
Unique identifier for the job in Databricks.
dlt_maintenance_id
Maintenance ID for Delta Live Tables (if applicable).
run_name
Name of the current job or workflow run (if applicable).
instance_pool_id
ID of the instance pool used (if applicable).
cluster_id
ID of the cluster associated with this usage.
endpoint_id
ID of the endpoint for SQL-based or serving-related usage (if applicable).
warehouse_id
ID of the SQL warehouse (if applicable).
source_region
Originating region for this billing record (if applicable).
dlt_pipeline_id
ID of the Delta Live Tables pipeline (if applicable).
endpoint_name
Name of the SQL or serving endpoint (if applicable).
is_photon
Indicates whether Photon processing was used (true or false).
dlt_tier
Tier of Delta Live Tables service (if applicable).
jobs_tier
Tier of the job, such as CLASSIC or PREMIUM.
networking
Type of networking used for this job, if specified.
serving_type
Type of serving model used, if applicable (for example, Model Serving).
sql_tier
SQL tier associated with the usage (if applicable).
is_serverless
Indicates if the usage pertains to a serverless compute resource (true or false).
custom_tags
Custom tags applied to the usage, usually as key-value pairs for additional metadata or categorization.
usage_metadata
Metadata related to the usage, which might include details like usage type, service category, or other relevant information.
Tag Name
Tag Description
resource_id
The unique identifier of the Confluent resource.
resource_name
The name of the Confluent resource.
environment_id
The unique identifier for the environment.
network_access_type
Network access type for the cluster. Possible values are INTERNET, TRANSIT_GATEWAY, PRIVATE_LINK, and PEERED_VPC.
product
Product name. Possible values include KAFKA, CONNECT, KSQL, AUDIT_LOG, STREAM_GOVERNANCE, CLUSTER_LINK, CUSTOM_CONNECT, FLINK, SUPPORT_CLOUD_BASIC, SUPPORT_CLOUD_DEVELOPER, SUPPORT_CLOUD_BUSINESS, and SUPPORT_CLOUD_PREMIER.
Tag Name
Tag Description
organization_name
Name of the organization.
contract_number
Snowflake contract number for the organization.
account_name
Name of the account where the usage was consumed.
account_locator
Locator for the account where the usage was consumed.
region
Name of the region where the account is located.
service_level
Service level (edition) of the Snowflake account (Standard, Enterprise, or Business Critical).
user_name
Name of the user or service account associated with the query.
warehouse_id
Identifier for the warehouse generating the cost.
warehouse_name
Name of the warehouse associated with this usage.
warehouse_size
Size of the warehouse (for example, Large, Medium).
cost_type
Type of cost associated with the usage. Possible values include: - CLOUD_SERVICES: General costs related to Snowflake’s underlying cloud services, excluding warehouse usage. - IDLE_OR_LESS_100MS: Costs from warehouse idle time or queries that completed in under 100 milliseconds. Unattributed to specific queries. Falls under the warehouse_metering service type. - QUERY_ATTRIBUTION: Costs attributed to specific queries, grouped by the parameterized query hash. For these costs, the parameterized query associated with this cost can be found under charge description. Falls under the warehouse_metering service type.
query_hash
Unique hash representing a parameterized version of the query for attribution purposes. Only found for query attribution costs.
query_hash_version
Version of the Snowflake query hash algorithm used to generate query_hash. Only found for query attribution costs.
database_name
Name of the database in which the query was executed (if applicable). Only found for query attribution costs.
balance_source
Source of the funds used to pay for the daily usage. The source can be one of the following: - capacity: Usage paid with credits remaining on an organization’s capacity commitment. - rollover: Usage paid with rollover credits. When an organization renews a capacity commitment, unused credits are added to the balance of the new contract as rollover credits. - free usage: Usage covered by the free credits provided to the organization. - overage: Usage that was paid at on-demand pricing, which occurs when an organization has exhausted its capacity, rollover, and free credits. - rebate: Usage covered by the credits awarded to the organization when it shared data with another organization.
service_type
Type of usage. Possible service types include: - automatic_clustering: Refer to Automatic Clustering. - data_transfer: Refer to Understanding data transfer cost. - logging: Refer to Logging and Tracing Overview. - materialized_view: Refer to Working with Materialized Views. - replication: Refer to Introduction to replication and failover across multiple accounts. - query_acceleration: Refer to Using the Query Acceleration Service. - search_optimization: Refer to Search Optimization Service. - serverless_task: Refer to Introduction to tasks. - snowpipe: Refer to Snowpipe. - snowpipe_streaming: Refer to Snowpipe Streaming. - storage: Refer to Understanding storage cost. - warehouse_metering: Refer to Virtual warehouse credit usage. Does not indicate usage of serverless or cloud services compute.
rating_type
Indicates how the usage in the record is rated, or priced. Possible values include: - compute - data_transfer - storage - Other
billing_type
Indicates what is being charged or credited. Possible billing types include: - consumption: Usage associated with compute credits, storage costs, and data transfer costs. - rebate: Usage covered by the credits awarded to the organization when it shared data with another organization. - priority support: Charges for priority support services. This charge is associated with a stipulation in a contract, not with an account. - vps_deployment_fee: Charges for a Virtual Private Snowflake deployment. - support_credit: Snowflake Support credited the account to reverse charges attributed to an issue in Snowflake.
Tag Name
Tag Description
name
The unique identifier of the Elastic Cloud resource.
price_per_hour
The cost of the Elastic Cloud resource per hour.
kind
The type of resource.
Tag Name
Tag Description
invoice_id
The unique identifier of the invoice.
status
State of the payment.
mongo_org_id
MongoDB organization ID.
cluster_name
The name of the cluster that incurred the charge.
group_id
ID of the project with which the line item is associated.
replica_set_name
Name of the replica set with which the line item is associated.
resource_tags
Arbitrary tags on clusters set by users, usually as key-value pairs.
Tag Name
Tag Description
organization_id
The unique identifier of the organization.
project_id
The unique identifier of the project.
project_name
The name of the project.
organization_name
The name of the organization.
Tag Name
Tag Description
credit_coupon_code
Code of any coupon or credit applied to this cost entry (if available).
product_name
Name of the specific product being billed (for example, “North America Bandwidth”).
product_group
Group or category of the product, such as “Full Site Delivery”.
product_line
Line of products to which this item belongs, for example, “Network Services”.
usage_type
Type of usage being billed (for example, “Bandwidth”).
region
Region where the service usage occurred (for example, “North America”).
service_name
Name of the service associated with this cost entry, often matching the product_name.
usage_type_cd
Code or label representing the type of usage, such as “North America Bandwidth”.
plan_name
Name of the plan under which this service falls, often matching “product_line”.
Tag Name
Tag Description
account_sid
Alphanumeric string identifying the Twilio account.
category
The category of usage. For more information, see Usage Categories.
count_unit
The units in which count is measured, such as calls for calls or messages for SMS.
usage_unit
The units in which usage is measured, such as minutes for calls or messages for SMS.
Further reading
Documentation, liens et articles supplémentaires utiles: