Skip to main content

Viewing a Raw Profile

A raw profile is a UI-generated profile that is a subset of the data that you can download and provides a summary of metrics collected for each executed query that can be used to monitor and analyze query performance.

To navigate to a raw profile:

  1. Click This is the icon that represents the Jobs page. in the side navigation bar to open the Jobs page.
  2. On the Jobs page, click a job that you would like to see the raw profile for.
  3. At the top of the next page, click the Raw Profile tab to open a raw profile of the job in a separate dialog. The associated raw profile dialog shows a variety of information for review.

Views

Within the Raw Profile dialog, you can analyze the Job Metrics based on the following views:

ViewDescription
QueryShows the selected query statement and job metrics. See if your SQL query is what you were expecting and the query is run against the source data.
Visualized PlanShows a visualized diagram and job metrics. This view is useful in understanding the flow of the query and for analyzing out of memory issues and incorrect results. The detailed visualized pan diagram is always read from the bottom up.
PlanningShows planning metrics, query output schema, non default options, and job metrics. This view shows how query planning is executed, because it provides statistics about the actual cost of the query operations in terms of memory, input/output, and CPU processing. You can use this view to identify which operations consumed the majority of the resources during a query and to address the cost-intensive operations. In particular, the following information is useful:
  • Non Default Options – See if non-default parameters are being used.
  • Metadata Cache Hits and Misses with times
  • Final Physical Transformation – Look for pushdown queries for RDBMS, MongoDB, or Elasticsearch, filter pushdowns or partition pruning for parquet, and view usage of stripes for ORC.
  • Compare the estimated row count versus the actual scan, join, or aggregate result.
  • Row Count – See if row count (versus rows) is used. Row count can cause an expensive broadcast.
  • Build – See if build (versus probe) is used. Build loads data into memory.
AccelerationShows reflection outcome, canonicalized user query alternatives, reflection details, and job metrics.
  • Multiple substitutions – See if the substitutions are excessive.
  • System activity – See if sys.project.reflections, sys.project.materializations, and sys.project.refreshes are excessive.
  • Comparisons – Compare cumulative cost (found in Best Cost Replacement Plan) against Logical Planning, which is in the Planning view.
This view is useful for determining whether exceptions or matches are occurring. The following considerations determines the acceleration process:
  • Considered, Matched, Chosen – The query is accelerated.
  • Considered, Matched, Not Chosen – The query is not accelerated because either a costing issue or an exception during substitution occurred.
  • Considered, Not Matched, Not Chosen – The query is not accelerated because the reflection does not have the data to accelerate.
Error(If applicable) Shows information about an error. The Failure Node is always the coordinator node and the server name inside the error message is the actual affected node.

Job Metrics

Each view displays the following metrics:

Job Summary

The job summary information includes:

  • State
  • Coordinator
  • Threads
  • Command Pool Wait
  • Total Query Time
  • # Joins in user query
  • # Joins in final plan
  • Considered Reflections
  • Matched Reflections
  • Chosen Reflections

Time in UTC

The Time in UTC section lists the job's start and end time, in UTC format.

State Durations

The State Durations section lists the length of time (in milliseconds) for each of the job states:

  • Pending
  • Metadata Retrieval
  • Planning
  • Engine Start
  • Queued
  • Execution Planning
  • Starting
  • Running

For descriptions of the job states, see Job States and Statuses.

Context

If you are querying an Iceberg catalog object, the Context section lists the Iceberg catalog and branch that is referenced in the query. Otherwise, the Context section is not populated. Read Iceberg Catalogs in Dremio for more information.

Threads

The Threads section provides an overview table and a major fragment block for each major fragment. Each row in the Overview table provides the number of minor fragments that Dremio parallelized from each major fragment, as well as aggregate time and memory metrics for the minor fragments.

Major fragment blocks correspond to a row in the Overview table. You can expand the blocks to see metrics for all of the minor fragments that were parallelized from each major fragment, including the host on which each minor fragment ran. Each row in the major fragment table presents the fragment state, time metrics, memory metrics, and aggregate input metrics of each minor fragment.

In particular, the following metrics are useful:

  • Setup – Time opening and closing of files.
  • Waiting – Time waiting on the CPU.
  • Blocked on Downstream – Represents completed work whereas the next phase is not ready to accept work.
  • Blocked on Upstream – Represents the phase before it is ready to give work though the cloud phase is not ready.
  • Phase Metrics – Displays memory used per node (Phases can run in parallel).

Resource Allocation

The Resource Allocation section shows the following details for managed resources and workloads:

  • Engine Name
  • Queue Name
  • Queue Id
  • Query Cost
  • Query Type

Nodes

The Nodes section includes host name, resource waiting time, and peak memory.

Operators

The Operators section shows aggregate metrics for each operator within a major fragment that performed relational operations during query execution.

Operator Overview Table

The following table lists descriptions for each column in the Operators Overview table:

Column NameDescription
SqlOperatorImpl IDThe coordinates of an operator that performed an operation during a particular phase of the query. For example, 02-xx-03 where 02 is the major fragment ID, xx corresponds to a minor fragment ID, and 03 is the Operator ID.
TypeThe operator type. Operators can be of type project, filter, hash join, single sender, or unordered receiver.
Min Setup Time, Avg Setup Time, Max Setup TimeIn general, the time spent opening and closing files. Specifically, the minimum, average, and maximum amount of time spent by the operator to set up before performing the operation.
Min Process Time, Avg Process Time, Max Process TimeThe shortest amount of time the operator spent processing a record, the average time the operator spent in processing each record, and the maximum time that the operator spent in processing a record.
Wait (min, avg, max)In general, the time spent waiting on Disk I/O. These fields represent the minimum, average, and maximum times spent by operators waiting on disk I/O.
Avg Peak MemoryRepresents the average of the peak direct memory allocated across minor fragments. Relates to the memory needed by operators to perform their operations, such as hash join or sort.
Max Peak MemoryRepresents the maximum of the peak direct memory allocated across minor fragments. Relates to the memory needed by operators to perform their operations, such as hash join or sort.

Operator Block

The Operator Block shows time and memory metrics for each operator type within a major fragment. Examples of operator types include:

  • SCREEN
  • PROJECT
  • WRITER_COMMITTER
  • ARROW_WRITER

The following table describes each column in the Operator Block:

Column NameDescription
ThreadThe coordinate ID of the minor fragment on which the operator ran. For example, 04-03-01 where 04 is the major fragment ID, 03 is the minor fragment ID, and 01 is the Operator ID.
Setup TimeThe amount of time spent by the operator to set up before performing its operation. This includes run-time code generation and opening a file.
Process TimeThe amount of time spent by the operator to perform its operation.
Wait TimeThe cumulative amount of time spent by an operator waiting for external resources. such as waiting to send records, waiting to receive records, waiting to write to disk, and waiting to read from disk.
Max BatchesThe maximum number of record batches consumed from a single input stream.
Max RecordsThe maximum number of records consumed from a single input stream.
Peak MemoryRepresents the peak direct memory allocated. Relates to the memory needed by the operators to perform their operations, such as hash join and sort.
Host NameThe hostname of the Executor the minor fragment is running on.
Record Processing RateThe rate at which records in the minor fragment are being processed. Combined with the Host Name, the Record Processing Rate can help find hot spots in the cluster, either from skewed data or a noisy query running on the same cluster.
Operator StateThe status of the minor fragment.
Last Schedule TimeThe last time at which work related to the minor fragment was scheduled to be executed.

Operator blocks also contain three drop-down menus: Operator Metrics, Operator Details, and Host Metrics. Operator Metrics and Operator Details are unique to the type of operator and provide more detail about the operation of the minor fragments. Operator Metrics and Operator Details are intended to be consumed by Dremio engineers. Depending on the operator, both can be blank. Host Metrics provides high-level information about the host used when executing the operator.