On this page

    Viewing Jobs and Job Details

    All jobs run in Dremio Sonar are listed on a separate page, showing the job ID, type, status, and other attributes. To navigate to the Jobs page, click the Jobs This is the icon that represents the Jobs page. (list) icon in the side navigation bar.

    Search Filters and Columns

    By default, the Jobs page lists the jobs run within the last 30 days and the jobs are filtered by UI, External Tools job types. To change these defaults for your account, you can filter on values and manage columns directly on the Jobs page, as shown in this image:

    This is a screenshot showing the main components of the Jobs page.

    a. Search Jobs by typing the username or job ID.

    b. Start Time allows you to pick the date and time at which the job began.

    c. Status represents one or more job states. For descriptions, see Job States and Statuses.

    d. Type includes Accelerator, Downloads, External Tools, Internal, and UI. For descriptions, see Job Attributes.

    e. User can be searched by typing the username or checking the box next to the username in the dropdown.

    f. Manage Columns by checking the boxes next to additional columns that you want to see in the Jobs list. The grayed out checkboxes show the columns that are required by default. You can also rearrange the column order by clicking directly on a column to drag and drop.

    Job Attributes

    Each job has the following attributes, which can appear as columns in the list of jobs:

    AttributeDescription
    AcceleratedA purple lightning bolt in a row indicates that the job ran a query that was accelerated by one or more reflections.
    DatasetThe queried dataset, if one was queried. Hover over the dataset to see a metadata card appear with details about the dataset. For more information, see Viewing the Metadata of a Dataset.
    DurationThe length of time (in seconds) that a job required from start to completion.
    EngineThe engine used to run the query.
    Job IDA universally unique identifier.
    Planner Cost EstimateA cost estimate calculated by Dremio based on an evaluation of the resources that to be used in the execution of a query. The number is not in units, and is intended to give a an idea of the cost of executing a query relative to the costs of executing other queries. Values are derived by adding weighted estimates of required I/O, memory, and CPU load. In reported values, K = thousand, M = million, B = billion, and T = trillion. For example, a value of 12,543,765,321 is reported as 12.5B.
    Planning TimeThe length of time (in seconds) in which the query optimizer planned the execution of the query.
    Query TypeRepresents one of the five query types:
    • UI - queries issued from the SQL Runner in the Sonar UI.
    • External Tools - queries from client applications, such as Microsoft Power BI, Superset, Tableau, other third-party client applications, and custom applications.
    • Accelerator - queries related to creating, maintaining, and removing reflections.
    • Internal - queries that Sonar submits for internal operations.
    • Downloads - queries used to download datasets.
    Rows ReturnedNumber of output records.
    Rows ScannedNumber of input records.
    SQLThe SQL query that was submitted for the job.
    Start TimeThe date and time which the job began.
    StatusAn icon that represents one or more job states. This column is automatically shown at the start of each row. For descriptions, see Job States and Statuses.
    UserUsername of the user who ran the query and initiated the job.

    Job States and Statuses

    Each job passes through a sequence of states until it is complete, though the sequence can be interrupted if a query is canceled or if there is an error during a state. In this diagram, the states that a job passes through are in white, and the possible end states are in dark gray.

    This table lists the statuses that the UI lets you filter on and shows how they map to the states:

    IconStatusStateDescription
    SetupPendingRepresents a state where the query is waiting to be scheduled on the query pool.
    Metadata RetrievalRepresents a state where metadata schema is retrieved and the SQL command is parsed.
    PlanningRepresents a state where the following are done:
    • Physical and logical planning
    • Reflection matching
    • Partition metadata retrieval
    • Mapping the query to a queue based workload management rules
    • Pick the engine associated with the query to run the query.
    Engine StartEngine StartRepresents a state where the engine starts if it has stopped. If the engine is stopped, it takes time to restart for the executors to be active. If the engine is already started, then this state does not have a duration.
    QueuedQueuedRepresents a state where a job is queued. Each queue has a limit of concurrent queries. If the queries in progress exceed the concurrency limit, the query should wait in the queue until the jobs in progress complete.
    RunningExecution PlanningRepresents a state where executor nodes are selected from the chosen engine to run the query, and work is distributed to each executor.
    RunningRepresents a state where executor nodes execute and complete the fragments assigned to them. Typically, most queries spend more time in this state.
    StartingRepresents a state where the query is starting up.
    CanceledCanceledRepresents a terminal state that indicates that the query is canceled by the user or an intervention in the system.
    CompletedCompletedRepresents a terminal state that indicates that the query is successfully completed.
    FailedFailedRepresents a terminal state that indicates that the query has failed due to an error.

    Viewing Job Details

    You can view the details of a specific job by viewing the Job Overview, SQL, Query Profile, and Raw Profile pages.

    To navigate to the job details:

    1. Click the Jobs This is the icon that represents the Jobs page. (list) icon in the side navigation bar.
    2. On the Jobs page, click a job that you would like to see the job overview for.
    3. The Job Overview page then replaces the list of jobs.

    Job Overview

    By opening the Job Overview page, you can view the following:

    Summary

    Each job is summarized by the following information: Status, Total Memory, CPU Used, Query Type, Start Time, Duration, Wait on Client, User, Engine, Input, and Output.

    Total Execution Time

    The length of time for the total execution and the job state durations in the order they occur. Only the duration of the Engine Start state is in minutes and seconds. If the engine is stopped, it takes time to restart for the executors to be active. If the engine is already started, then Engine Start duration does not have a value.

    Submitted SQL

    The SQL query for the selected job.

    Queried Datasets

    The datasets queried for the selected job. These can be views or tables.

    Scans

    Scan details include the source type, scan thread count, IO wait time (in milliseconds), and the number of rows scanned.

    Results

    To see the job results, click the Open Results link in the top-right corner of the Job Overview page. As long as the engine that ran the job is up, the Open Results link is visible in the UI. It disappears when the engine that ran the job shuts down and is only visible for the jobs that are run through the UI.

    Download Profile

    To download the query profile, click the Download Profile button in the bottom-left corner of the Job Overview page. The profile will help you see more granular details about the job.

    The profile downloads as a ZIP file. When you extract the ZIP file, you will see the following JSON files:

    • profile_attempt_0.json – This file helps with troubleshooting out of memory and wrong results issues. Note that the start and end time of query is provided in EPOCH format. See the Epoch Converter utility for converting query time.
    • header.json – This file provides the full list of Dremio coordinators and executors, data sets, and sources. This information is useful when you are using REST calls.

    SQL

    Next to the Job Overview page is a tab for the SQL page, which shows the Submitted SQL and Dataset Graph.

    Submitted SQL

    You can view the SQL statement that was used for the selected job. Although the SQL statement is in read-only mode on the SQL Details page, the statement can be copied from the page and pasted into the SQL editor.

    Dataset Graph

    A dataset graph only appears if there is a queried dataset for the selected job. The dataset graph is a visual representation of the datasets used in the SQL statement.

    Query Profile

    The Profile page shows the nodes used to execute the job. On the left panel, you can sort by the runtime, total memory, and number of records for each node. The middle panel represents a visualized workflow for each phase, and the right panel provides the details for runtime, processed batches and records, peak memory, bytes sent, and aggregate statistics.

    Raw Profile

    Click Raw Profile to open a raw profile of the job in a separate dialog, which includes a job summary, state durations, threads, resource allocation, operators, visualized plan, acceleration, and other details.