Table: databricks_catalog_function - Query Databricks Catalog Functions using SQL
Databricks Catalog Functions are a part of Databricks' Unified Data Service, which provides a unified, collaborative workspace for data teams to build data pipelines, explore data, and perform machine learning tasks. Catalog Functions allow users to create, manage, and invoke functions that can be used in SQL expressions. These functions are stored in databases, which can be shared across multiple workspaces.
Table Usage Guide
The databricks_catalog_function
table provides insights into Catalog Functions within Databricks Unified Data Service. As a data engineer or data scientist, explore function-specific details through this table, including function names, associated databases, descriptions, and class names. Utilize it to uncover information about functions, such as their usage in SQL expressions, their storage in databases, and their shareability across multiple workspaces.
Examples
Basic info
Explore which functions have been created in a specific catalog and schema in Databricks, along with their details such as the creator and creation date. This can be useful for auditing, tracking changes, or understanding the structure of your data.
select function_id, name, comment, created_at, created_by, metastore_id, account_idfrom databricks_catalog_functionwhere catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, comment, created_at, created_by, metastore_id, account_idfrom databricks_catalog_functionwhere catalog_name = 'catalog' and schema_name = 'schema';
List functions modified in the last 7 days
Explore which functions have been updated in the past week. This can be useful in tracking recent changes and maintaining an understanding of ongoing modifications to your system.
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere updated_at >= now() - interval '7 days' and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere updated_at >= datetime('now', '-7 days') and catalog_name = 'catalog' and schema_name = 'schema';
List all scalar functions
Gain insights into the scalar functions present in your Databricks catalog to understand their creation, modification, and the users involved. This aids in managing and auditing the functions in your catalog.
select function_id, name, comment, created_at, created_by, updated_by, data_typefrom databricks_catalog_functionwhere data_type is not null and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, comment, created_at, created_by, updated_by, data_typefrom databricks_catalog_functionwhere data_type is not null and catalog_name = 'catalog' and schema_name = 'schema';
List all deterministic functions
Explore which functions within your Databricks catalog are deterministic, allowing you to understand the functions that will always produce the same results given the same input values. This is useful in maintaining consistency and predictability in your data processing and analysis tasks.
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere is_deterministic and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere is_deterministic and catalog_name = 'catalog' and schema_name = 'schema';
List all SQL functions
Explore the range of SQL functions within a specific catalog and schema. This is useful for understanding what functions are available and who created or updated them, providing a clearer view of your database's functionality and history.
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere routine_body = 'SQL' and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, comment, created_at, created_by, updated_by, account_idfrom databricks_catalog_functionwhere routine_body = 'SQL' and catalog_name = 'catalog' and schema_name = 'schema';
List all external functions
Discover the segments that utilize external functions within your Databricks catalog. This can be useful in understanding the dependencies and interactions within your data schema, aiding in efficient data management and optimization.
select function_id, name, routine_body, routine_definition, routine_dependencies ->> 'function' as routine_dependency_function, routine_dependencies ->> 'table' as routine_dependency_tablefrom databricks_catalog_functionwhere routine_body = 'EXTERNAL' and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, routine_body, routine_definition, json_extract(routine_dependencies, '$.function') as routine_dependency_function, json_extract(routine_dependencies, '$.table') as routine_dependency_tablefrom databricks_catalog_functionwhere routine_body = 'EXTERNAL' and catalog_name = 'catalog' and schema_name = 'schema';
List all functions that reads sql data
Explore which functions within a databricks catalog have access to read SQL data. This can be especially useful to identify potential data access points and maintain data security.
select function_id, name, sql_data_access, sql_pathfrom databricks_catalog_functionwhere sql_data_access = 'READS_SQL_DATA' and catalog_name = 'catalog' and schema_name = 'schema';
select function_id, name, sql_data_access, sql_pathfrom databricks_catalog_functionwhere sql_data_access = 'READS_SQL_DATA' and catalog_name = 'catalog' and schema_name = 'schema';
Get effective permissions for each function
Assess the elements within your Databricks catalog to gain insights into the effective permissions allocated to each function. This is useful to ensure correct access rights and maintain security within your system.
select name, p ->> 'principal' as principal_name, p ->> 'privileges' as permissionsfrom databricks_catalog_function, jsonb_array_elements(function_effective_permissions) pwhere catalog_name = 'catalog' and schema_name = 'schema';
select name, json_extract(p.value, '$.principal') as principal_name, json_extract(p.value, '$.privileges') as permissionsfrom databricks_catalog_function, json_each(function_effective_permissions) as pwhere catalog_name = 'catalog' and schema_name = 'schema';
Get details for a specific function
Explore the specifics of a particular function in your Databricks catalog, including its creator and creation date. This is useful for auditing purposes or understanding the history and purpose of a function in your data pipeline.
select function_id, name, comment, created_at, created_by, metastore_id, account_idfrom databricks_catalog_functionwhere full_name = '__catalog_name__.__schema_name__.__table_name__';
select function_id, name, comment, created_at, created_by, metastore_id, account_idfrom databricks_catalog_functionwhere full_name = '__catalog_name__.__schema_name__.__table_name__';
Schema for databricks_catalog_function
Name | Type | Operators | Description |
---|---|---|---|
_ctx | jsonb | Steampipe context in JSON form, e.g. connection_name. | |
account_id | text | The Databricks Account ID in which the resource is located. | |
catalog_name | text | = | Name of parent catalog. |
comment | text | User-provided free-form text description. | |
created_at | timestamp with time zone | Time at which this function was created. | |
created_by | text | The user who created this function. | |
data_type | text | Scalar function return data type. | |
external_language | text | External function language. | |
external_name | text | External function name. | |
full_data_type | text | Pretty printed function data type. | |
full_name | text | = | Full name of function, in form of __catalog_name__.__schema_name__.__function__name__. |
function_effective_permissions | jsonb | Function effective permissions. | |
function_id | text | Id of Function, relative to parent schema. | |
function_permissions | jsonb | Function permissions. | |
input_params | jsonb | The array of __FunctionParameterInfo__ definitions of the function's parameters. | |
is_deterministic | boolean | Whether the function is deterministic. | |
is_null_call | boolean | Whether the function is a null call. | |
metastore_id | text | Unique identifier of parent metastore. | |
name | text | Name of function, relative to parent schema. | |
owner | text | Owner of the function. | |
parameter_style | text | Parameter style of the function. | |
properties | jsonb | A map of key-value properties attached to the securable. | |
return_params | jsonb | Table function return parameters. | |
routine_body | text | Routine body of the function. | |
routine_definition | text | Routine definition of the function. | |
routine_dependencies | jsonb | Routine dependencies of the function. | |
schema_name | text | = | Name of parent schema relative to its parent catalog. |
security_type | text | Security type of the function. | |
specific_name | text | Specific name of the function. | |
sql_data_access | text | SQL data access of the function. | |
sql_path | text | List of schemes whose objects can be referenced without qualification. | |
title | text | The title of the resource. | |
updated_at | timestamp with time zone | Time at which this function was last updated. | |
updated_by | text | The user who last updated this function. |
Export
This table is available as a standalone Exporter CLI. Steampipe exporters are stand-alone binaries that allow you to extract data using Steampipe plugins without a database.
You can download the tarball for your platform from the Releases page, but it is simplest to install them with the steampipe_export_installer.sh
script:
/bin/sh -c "$(curl -fsSL https://steampipe.io/install/export.sh)" -- databricks
You can pass the configuration to the command with the --config
argument:
steampipe_export_databricks --config '<your_config>' databricks_catalog_function