Table: databricks_ml_experiment - Query Databricks ML Experiments using SQL
Databricks ML Experiments is a feature within Databricks that allows you to track, manage, and visualize machine learning experiments. It provides a centralized way to log parameters, metrics, and artifacts for each run in an experiment. Databricks ML Experiments helps you stay informed about the performance of your machine learning models and take appropriate actions based on the insights derived from the data.
Table Usage Guide
The databricks_ml_experiment
table provides insights into ML experiments within Databricks. As a data scientist, explore experiment-specific details through this table, including experiment ID, name, artifact location, and lifecycle stage. Utilize it to uncover information about experiments, such as their current lifecycle stage, the location of their artifacts, and other associated metadata.
Examples
Basic info
Gain insights into the creation and last update times of machine learning experiments within a specific account on Databricks. This is useful for tracking the progress and activity of various experiments over time.
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experiment;
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experiment;
List experiments created in the last 7 days
Discover the newly created experiments within the past week. This allows for a timely review and monitoring of the latest experiments, ensuring up-to-date insights and understanding.
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere creation_time >= now() - interval '7' day;
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere creation_time >= datetime('now', '-7 days');
List experiments that have not been modified in the last 90 days
Discover the segments that have not been updated or modified in the last 90 days in your Databricks machine learning experiments. This can help in identifying dormant or inactive experiments for potential clean up or review.
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere last_update_time <= now() - interval '90' day;
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere last_update_time <= datetime('now', '-90 day');
List all active experiments
Explore which experiments are currently active in your Databricks machine learning workflow. This allows you to keep track of ongoing studies and manage resources effectively.
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere lifecycle_stage = 'active';
select experiment_id, name, creation_time, last_update_time, account_idfrom databricks_ml_experimentwhere lifecycle_stage = 'active';
Find the account with the most experiments
Discover which account has conducted the most experiments, helping you identify the most active accounts and understand usage patterns. This can be beneficial in assessing resource allocation and planning future capacity needs.
select account_id, count(*) as experiment_countfrom databricks_ml_experimentgroup by account_idorder by experiment_count desclimit 1;
select account_id, count(*) as experiment_countfrom databricks_ml_experimentgroup by account_idorder by experiment_count desclimit 1;
Schema for databricks_ml_experiment
Name | Type | Operators | Description |
---|---|---|---|
_ctx | jsonb | Steampipe context in JSON form, e.g. connection_name. | |
account_id | text | The Databricks Account ID in which the resource is located. | |
artifact_location | text | Location where experiment artifacts are stored. | |
creation_time | timestamp with time zone | Time when the experiment was created. | |
experiment_id | text | = | Unique identifier for the experiment. |
last_update_time | timestamp with time zone | Time when the experiment was last updated. | |
lifecycle_stage | text | = | Current life cycle stage of the experiment. |
name | text | Human readable name that identifies the experiment. | |
tags | jsonb | Additional metadata key-value pairs. | |
title | text | The title of the resource. |
Export
This table is available as a standalone Exporter CLI. Steampipe exporters are stand-alone binaries that allow you to extract data using Steampipe plugins without a database.
You can download the tarball for your platform from the Releases page, but it is simplest to install them with the steampipe_export_installer.sh
script:
/bin/sh -c "$(curl -fsSL https://steampipe.io/install/export.sh)" -- databricks
You can pass the configuration to the command with the --config
argument:
steampipe_export_databricks --config '<your_config>' databricks_ml_experiment