Viewing a Raw Profile
A raw profile is a UI-generated profile that is a subset of the data that you can download and provides a summary of metrics collected for each executed query that can be used to monitor and analyze query performance.
To navigate to a raw profile:
- Click in the side navigation bar to open the Jobs page.
- On the Jobs page, click a job that you would like to see the raw profile for.
- At the top of the next page, click the Raw Profile tab to open a raw profile of the job in a separate dialog. The associated raw profile dialog shows a variety of information for review.
Views
Within the Raw Profile dialog, you can analyze the Job Metrics based on the following views:
View | Description |
---|---|
Query | Shows the selected query statement and job metrics. See if your SQL query is what you were expecting and the query is run against the source data. |
Visualized Plan | Shows a visualized diagram and job metrics. This view is useful in understanding the flow of the query and for analyzing out of memory issues and incorrect results. The detailed visualized pan diagram is always read from the bottom up. |
Planning | Shows planning metrics, query output schema, non default options, and job metrics. This view shows how query planning is executed, because it provides statistics about the actual cost of the query operations in terms of memory, input/output, and CPU processing. You can use this view to identify which operations consumed the majority of the resources during a query and to address the cost-intensive operations. In particular, the following information is useful:
|
Acceleration | Shows reflection outcome, canonicalized user query alternatives, reflection details, and job metrics.
|
Error | (If applicable) Shows information about an error. The Failure Node is always the coordinator node and the server name inside the error message is the actual affected node. |
Job Metrics
Each view displays the following metrics:
Job Summary
The job summary information includes:
- State
- Coordinator
- Threads
- Command Pool Wait
- Total Query Time
- # Joins in user query
- # Joins in final plan
- Considered Reflections
- Matched Reflections
- Chosen Reflections
Time in UTC
The Time in UTC section lists the job's start and end time, in UTC format.
State Durations
The State Durations section lists the length of time (in milliseconds) for each of the job states:
- Pending
- Metadata Retrieval
- Planning
- Engine Start
- Queued
- Execution Planning
- Starting
- Running
For descriptions of the job states, see Job States and Statuses.
Context
If you are querying an Iceberg catalog object, the Context section lists the Iceberg catalog and branch that is referenced in the query. Otherwise, the Context section is not populated. Read Iceberg Catalogs in Dremio for more information.
Threads
The Threads section provides an overview table and a major fragment block for each major fragment. Each row in the Overview table provides the number of minor fragments that Dremio parallelized from each major fragment, as well as aggregate time and memory metrics for the minor fragments.
Major fragment blocks correspond to a row in the Overview table. You can expand the blocks to see metrics for all of the minor fragments that were parallelized from each major fragment, including the host on which each minor fragment ran. Each row in the major fragment table presents the fragment state, time metrics, memory metrics, and aggregate input metrics of each minor fragment.
In particular, the following metrics are useful:
- Setup – Time opening and closing of files.
- Waiting – Time waiting on the CPU.
- Blocked on Downstream – Represents completed work whereas the next phase is not ready to accept work.
- Blocked on Upstream – Represents the phase before it is ready to give work though the cloud phase is not ready.
- Phase Metrics – Displays memory used per node (Phases can run in parallel).
Resource Allocation
The Resource Allocation section shows the following details for managed resources and workloads:
- Engine Name
- Queue Name
- Queue Id
- Query Cost
- Query Type
Nodes
The Nodes section includes host name, resource waiting time, and peak memory.
Operators
The Operators section shows aggregate metrics for each operator within a major fragment that performed relational operations during query execution.
Operator Overview Table
The following table lists descriptions for each column in the Operators Overview table:
Column Name | Description |
---|---|
SqlOperatorImpl ID | The coordinates of an operator that performed an operation during a particular phase of the query. For example, 02-xx-03 where 02 is the major fragment ID, xx corresponds to a minor fragment ID, and 03 is the Operator ID. |
Type | The operator type. Operators can be of type project, filter, hash join, single sender, or unordered receiver. |
Min Setup Time, Avg Setup Time, Max Setup Time | In general, the time spent opening and closing files. Specifically, the minimum, average, and maximum amount of time spent by the operator to set up before performing the operation. |
Min Process Time, Avg Process Time, Max Process Time | The shortest amount of time the operator spent processing a record, the average time the operator spent in processing each record, and the maximum time that the operator spent in processing a record. |
Wait (min, avg, max) | In general, the time spent waiting on Disk I/O. These fields represent the minimum, average, and maximum times spent by operators waiting on disk I/O. |
Avg Peak Memory | Represents the average of the peak direct memory allocated across minor fragments. Relates to the memory needed by operators to perform their operations, such as hash join or sort. |
Max Peak Memory | Represents the maximum of the peak direct memory allocated across minor fragments. Relates to the memory needed by operators to perform their operations, such as hash join or sort. |
Operator Block
The Operator Block shows time and memory metrics for each operator type within a major fragment. Examples of operator types include:
- SCREEN
- PROJECT
- WRITER_COMMITTER
- ARROW_WRITER
The following table describes each column in the Operator Block:
Column Name | Description |
---|---|
Thread | The coordinate ID of the minor fragment on which the operator ran. For example, 04-03-01 where 04 is the major fragment ID, 03 is the minor fragment ID, and 01 is the Operator ID. |
Setup Time | The amount of time spent by the operator to set up before performing its operation. This includes run-time code generation and opening a file. |
Process Time | The amount of time spent by the operator to perform its operation. |
Wait Time | The cumulative amount of time spent by an operator waiting for external resources. such as waiting to send records, waiting to receive records, waiting to write to disk, and waiting to read from disk. |
Max Batches | The maximum number of record batches consumed from a single input stream. |
Max Records | The maximum number of records consumed from a single input stream. |
Peak Memory | Represents the peak direct memory allocated. Relates to the memory needed by the operators to perform their operations, such as hash join and sort. |
Host Name | The hostname of the Executor the minor fragment is running on. |
Record Processing Rate | The rate at which records in the minor fragment are being processed. Combined with the Host Name, the Record Processing Rate can help find hot spots in the cluster, either from skewed data or a noisy query running on the same cluster. |
Operator State | The status of the minor fragment. |
Last Schedule Time | The last time at which work related to the minor fragment was scheduled to be executed. |
Operator blocks also contain three drop-down menus: Operator Metrics, Operator Details, and Host Metrics. Operator Metrics and Operator Details are unique to the type of operator and provide more detail about the operation of the minor fragments. Operator Metrics and Operator Details are intended to be consumed by Dremio engineers. Depending on the operator, both can be blank. Host Metrics provides high-level information about the host used when executing the operator.