Profile Types

In the Profiles tab, you can see all profile types available for a given language. Depending on the language and version, the information collected about your profile differs.

Once profiling is enabled, the following profile types are collected for supported Java versions:

CPU
The time each method spent running on the CPU. It includes your code that runs in the JVM (for example, Java, Kotlin), but not JVM operations or native code called from within the JVM.
Allocations
The number of heap allocations made by each method, including allocations which were subsequently freed.
Requires: Java 11
Allocated Memory
The amount of heap memory allocated by each method, including allocations which were subsequently freed.
Requires: Java 11
Heap Live Objects (in Preview, 1.17.0+)
The number of objects allocated by each method in heap memory that have not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Requires: Java 11
Heap Live Size (in Preview, 1.39.0+)
The amount of heap memory allocated by each method that has not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Requires: Java 11.0.23+, 17.0.11+, 21.0.3+ or 22+
Wall Time in Native Code
The elapsed time spent in native code. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the method is running. This profile does not include time spent running JVM bytecode, which is typically most of your application code.
Class Load
The number of classes loaded by each method.
Thrown Exceptions
The number of errors and exceptions thrown by each method, as well as their type.
File I/O
The time each method spent reading from and writing to files.
Lock
The time each method spent waiting for a lock.
Socket I/O
The time each method spent reading from and writing to socket I/O.

Once profiling is enabled, the following profile types are collected, depending on your Python version as noted:

Wall Time
The elapsed time used by each function. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the function is running.
Requires: Python 2.7+
Lock Wait Time
The time each function spent waiting for a lock.
Requires: Python 2.7+
Locked Time
The time each function spent holding a lock.
Requires: Python 2.7+
Lock Acquires
The number of times each function acquired a lock.
Requires: Python 2.7+
Lock Releases
The number of times each function released a lock.
Requires: Python 2.7+
CPU
The time each function spent running on the CPU, including Python and native code.
Requires: Python 2.7+, POSIX platform
Heap Live Size
The amount of heap memory allocated by each function that has not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Requires: Python 3.5+
Allocated Memory
The amount of heap memory allocated by each function, including allocations which were subsequently freed.
Requires: Python 3.5+
Allocations
The number of heap allocations made by each function, including allocations which were subsequently freed.
Requires: Python 3.5+
Thrown Exceptions
The number of caught or uncaught exceptions raised by each function, as well as their type.
Requires: Python 3.7+, POSIX platform

Once profiling is enabled, the following profile types are collected for supported Go versions:

CPU Time
The time each function spent running on the CPU. Off-CPU time such as waiting for Networking, Channels, Mutexes, and Sleep are not captured in this profile. See Mutex and Block profiles.
Allocations
The number of objects allocated by each function in heap memory during the profiling period (default: 60s), including allocations which were subsequently freed. Go calls this alloc_objects. Stack allocations are not tracked. This is useful for investigating garbage collection load. See also the note about how this measure changes in version 1.33.0 in Delta profiles.
Allocated Memory
The amount of heap memory allocated by each function during the profiling period (default: 60s), including allocations which were subsequently freed. Go calls this alloc_space. Stack allocations are not tracked. This is useful for investigating garbage collection load. See also the note about how this measure changes in version 1.33.0 in Delta profiles.
Heap Live Objects
The number of objects allocated by each function in heap memory that have not yet been garbage collected. Go calls this inuse_objects. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Heap Live Size
The amount of heap memory allocated by each function that has not yet been garbage collected. Go calls this inuse_space. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Mutex
The time functions have been waiting on mutexes during the profiling period (default: 60s). The stack traces in this profile point the Unlock() operation that allowed another goroutine blocked on the mutex to proceed. Short mutex contentions using spinlocks are not captured by this profile, but can be seen in the CPU profile. See also the note about how this measure changes in version 1.33.0 in Delta profiles.
Block
The time functions have been waiting on mutexes and channel operations during the profiling period (default: 60s). Sleep, GC, Network, and Syscall operations are not captured by this profile. Blocking operations are only captured after they become unblocked, so this profile cannot be used to debug applications that appear to be stuck. For mutex contentions, the stack traces in this profile point to blocked Lock() operations. This tells you where your program is getting blocked, while the mutex profile tells you what part of your program is causing the contention. See Datadog’s Block Profiling in Go research for more in-depth information. See also the note about how this measure changes in version 1.33.0 in Delta profiles. Note: The block profiler can cause noticeable overhead for production workloads. If enabling it in production, prefer high rates (such as 100000000, which is 100 milliseconds) and look for signs of increased latency or CPU utilization.
Goroutines
A snapshot of the number of goroutines currently executing the same functions (both on-CPU and waiting off-CPU). An increasing number of goroutines between snapshots can indicate that the program is leaking goroutines. In most healthy applications this profile is dominated by worker pools and the number of goroutines they use. Applications that are extremely latency-sensitive and use a large number of goroutines (> 10.000) should be aware that enabling this profile requires stop-the-world pauses. The pauses occur only once every profiling period (default 60s) and normally last for around 1µsec per goroutine. Typical applications with a p99 latency SLO of around 100ms can generally ignore this warning. See Datadog’s Goroutine Profiling in Go research for more in-depth information.

Delta profiles

Note: In Go profiler versions before 1.33.0, Allocations, Allocated Memory, Mutex, and Block metrics are shown as measures accumulated since the process was started, as opposed to during the profiling period. The change to delta profiles in version 1.33.0 lets you see how these measures are changing instead of accumulating. Delta profiling is on by default. Profiler version 1.35.0 allows you to disable delta profiles using the WithDeltaProfiles option.

As of profiler version 1.37.0, accumulated profiles are no longer uploaded when delta profiling is enabled to reduce upload bandwidth usage. Contact Support to discuss your use case if you rely on the full accumulated profiles.

Once profiling is enabled, the following profile types are collected for supported Ruby versions:

CPU
The time each function spent running on the CPU, including Ruby and native code.
Wall Time
The elapsed time used by each function. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the function is running.
Allocations (v2.3.0+)
The number of objects allocated by each method during the profiling period (default: 60s), including allocations which were subsequently freed. This is useful for investigating garbage collection load.
Requires: Manual enablement
Heap Live Objects (alpha, v2.3.0+)
The number of objects allocated by each method in heap memory that have not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Requires: Ruby 2.7+ and manual enablement
Heap Live Size (alpha, v2.3.0+)
The amount of heap memory allocated by each method that has not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.
Requires: Ruby 2.7+ and manual enablement

Once profiling is enabled, the following profile types are collected for supported Node.js versions:

CPU
The time each function spent running on the CPU, including JavaScript and native code.
Wall Time
The elapsed time used by each function. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the function is running.
Heap Live Size
The amount of heap memory allocated by each function that has not yet been garbage collected. This is useful for investigating the overall memory usage of your service and identifying potential memory leaks.

Once profiling is enabled, the following profile types are collected for supported .NET versions:

Wall Time
The elapsed time spent in managed methods. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the method is running.
CPU (v2.15+)
The time each method spent running on the CPU.
Thrown Exceptions (v2.31+)
The number of caught or uncaught exceptions raised by each method, as well as their type and message.
Allocations (in Preview, v2.18+)
The number and size of allocated objects by each method, as well as their type.
Requires: .NET Framework (with Datadog Agent 7.51+ and v3.2+) / .NET 6+
Lock (v2.49+)
The number of times threads are waiting for a lock and for how long.
Requires: Preview .NET Framework (requires Datadog Agent 7.51+) / .NET 5+
Live Heap (in Preview, v2.22+)
A subset of the allocated objects (with their class name) that are still in memory.
Requires: .NET 7+

Once profiling is enabled, the following profile types are collected for supported PHP versions:

Wall Time
The elapsed time used by each function. Elapsed time includes time when code is running on CPU, waiting for I/O, and anything else that happens while the function is running.
CPU
Shows the time each function spent running on the CPU.
Allocations (v0.88+)
The number of allocations by each function during the profiling period (default: 67s), including allocations which were subsequently freed. Stack allocations are not tracked.
Note: Not available when JIT is active on PHP 8.0.0-8.1.20 and 8.2.0-8.2.7
Allocated memory (v0.88+)
The amount of heap memory allocated by each function during the profiling period (default: 67s), including allocations which were subsequently freed. Stack allocations are not tracked.
Note: Not available when JIT is active on PHP 8.0.0-8.1.20 and 8.2.0-8.2.7
Thrown Exceptions (v0.92+)
The number of caught or uncaught exceptions raised by each method, as well as their type.

Once profiling is enabled, the following profile types are collected for supported languages and versions:

CPU
The time each function spent running on the CPU.
Allocations
The number of allocations by each function during the profiling period (default: 59s), including allocations which were subsequently freed. Stack allocations are not tracked.
Allocated memory
The amount of heap memory allocated by each function during the profiling period (default: 59s), including allocations which were subsequently freed. Stack allocations are not tracked.

Further Reading

PREVIEWING: mervebolat/span-id-preprocessing