layout | page_title | sidebar_current | description |
---|---|---|---|
databricks |
Provider: Databricks |
docs-databricks-index |
Terraform provider for the Databricks Lakehouse platform |
Use the Databricks Terraform provider to interact with almost all of Databricks resources. If you're new to Databricks, please follow guide to create a workspace on Azure or AWS and then this workspace management tutorial. If you're migrating from version 0.3.x, please follow this guide. Changelog is available on GitHub.
Compute resources
- Deploy databricks_cluster on selected databricks_node_type
- Schedule automated databricks_job
- Control cost and data access with databricks_cluster_policy
- Speedup job & cluster startup with databricks_instance_pool
- Customize clusters with databricks_global_init_script
- Manage few databricks_notebook, and even list them
- Manage databricks_repo
Storage
- Manage JAR, Wheel & Egg libraries through databricks_dbfs_file
- List entries on DBFS with databricks_dbfs_file_paths data source
- Get contents of small files with databricks_dbfs_file data source
- Mount storage with databricks_mount resource
Security
- Organize databricks_user into databricks_group through databricks_group_member, also reading metadata
- Create databricks_service_principal with databricks_obo_token to enable even more restricted access control.
- Manage data access with databricks_instance_profile, which can be assigned through databricks_group_instance_profile and databricks_user_instance_profile
- Control which networks can access workspace with databricks_ip_access_list
- Generically manage databricks_permissions
- Manage data object access control lists with databricks_sql_permissions
- Keep sensitive elements like passwords in databricks_secret, grouped into databricks_secret_scope and controlled by databricks_secret_acl
- Create workspaces in your VPC with DBFS using cross-account IAM roles, having your notebooks encrypted with CMK.
- Use predefined AWS IAM Policy Templates: databricks_aws_assume_role_policy, databricks_aws_crossaccount_policy, databricks_aws_bucket_policy
- Configure billing and audit databricks_mws_log_delivery
Databricks SQL
- Create databricks_sql_endpoint controlled by databricks_permissions.
- Manage queries and their visualizations.
- Manage dashboards and their widgets.
- Provide global configuration for all SQL Endpoints
MLFlow
- Create MLFlow models.
- Create MLFlow experiments.
provider "databricks" {
}
data "databricks_current_user" "me" {}
data "databricks_spark_version" "latest" {}
data "databricks_node_type" "smallest" {
local_disk = true
}
resource "databricks_notebook" "this" {
path = "${data.databricks_current_user.me.home}/Terraform"
language = "PYTHON"
content_base64 = base64encode(<<-EOT
# created from ${abspath(path.module)}
display(spark.range(10))
EOT
)
}
resource "databricks_job" "this" {
name = "Terraform Demo (${data.databricks_current_user.me.alphanumeric})"
new_cluster {
num_workers = 1
spark_version = data.databricks_spark_version.latest.id
node_type_id = data.databricks_node_type.smallest.id
}
notebook_task {
notebook_path = databricks_notebook.this.path
}
}
output "notebook_url" {
value = databricks_notebook.this.url
}
output "job_url" {
value = databricks_job.this.url
}
!> Warning Please be aware that hard coding any credentials in plain text is not something that is recommended. We strongly recommend using a Terraform backend that supports encryption. Please use environment variables, ~/.databrickscfg
file, encrypted .tfvars
files or secret store of your choice (Hashicorp Vault, AWS Secrets Manager, AWS Param Store, Azure Key Vault)
There are currently three supported methods to authenticate into the Databricks platform to create resources:
- PAT Tokens
- Username and password pair
- Azure Active Directory Tokens via Azure CLI, Service Principals, or Managed Service Identities
No configuration options given to your provider will look up configured credentials in ~/.databrickscfg
file. It is created by the databricks configure --token
command. Check this page
for more details. The provider uses config file credentials only when host
/token
or azure_auth
options are not specified.
It is the recommended way to use Databricks Terraform provider, in case you're already using the same approach with
AWS Shared Credentials File
or Azure CLI authentication.
provider "databricks" {
}
You can specify non-standard location of configuration file through config_file
parameter or DATABRICKS_CONFIG_FILE
environment variable:
provider "databricks" {
config_file = "/opt/databricks/cli-config"
}
You can specify a CLI connection profile through profile
parameter or DATABRICKS_CONFIG_PROFILE
environment variable:
provider "databricks" {
profile = "ML_WORKSPACE"
}
You can use host
and token
parameters to supply credentials to the workspace. When environment variables are preferred, then you can specify DATABRICKS_HOST
and DATABRICKS_TOKEN
instead. Environment variables are the second most recommended way of configuring this provider.
provider "databricks" {
host = "https://abc-cdef-ghi.cloud.databricks.com"
token = "dapitokenhere"
}
!> Warning This approach is currently recommended only for provisioning account-level resources, e.g. AWS workspaces and should be avoided for regular use.
You can use the username
+ password
attributes to authenticate provider for E2 workspace setup. Respective DATABRICKS_USERNAME
and DATABRICKS_PASSWORD
environment variables are applicable as well.
provider "databricks" {
host = "https://accounts.cloud.databricks.com"
username = var.user
password = var.password
}
-> Note If you experience technical difficulties with rolling out resources in this example, please make sure that environment variables don't conflict with other provider block attributes. When in doubt, please run TF_LOG=DEBUG terraform apply
to enable debug mode through the TF_LOG
environment variable. Look specifically for Explicit and implicit attributes
lines, that should indicate authentication attributes used.
The provider block supports the following arguments:
host
- (optional) This is the host of the Databricks workspace. It is a URL that you use to login to your workspace. Alternatively, you can provide this value as an environment variableDATABRICKS_HOST
.token
- (optional) This is the API token to authenticate into the workspace. Alternatively, you can provide this value as an environment variableDATABRICKS_TOKEN
.username
- (optional) This is the username of the user that can log into the workspace. Alternatively, you can provide this value as an environment variableDATABRICKS_USERNAME
. Recommended only for creating workspaces in AWS.password
- (optional) This is the user's password that can log into the workspace. Alternatively, you can provide this value as an environment variableDATABRICKS_PASSWORD
. Recommended only for creating workspaces in AWS.config_file
- (optional) Location of the Databricks CLI credentials file created bydatabricks configure --token
command (~/.databrickscfg by default). Check Databricks CLI documentation for more details. The provider uses configuration file credentials when you don't specify host/token/username/password/azure attributes. Alternatively, you can provide this value as an environment variableDATABRICKS_CONFIG_FILE
. This field defaults to~/.databrickscfg
.profile
- (optional) Connection profile specified within ~/.databrickscfg. Please check connection profiles section for more details. This field defaults toDEFAULT
.account_id
- (optional) Account Id that could be found in the bottom left corner of Accounts Console. Alternatively, you can provide this value as an environment variableDATABRICKS_ACCOUNT_ID
. Only has effect whenhost = "https://accounts.cloud.databricks.com/"
and currently used to provision account admins via databricks_user. In the future releases of the provider this property will also be used specify account fordatabricks_mws_*
resources as well.auth_type
- (optional) enforce specific auth type to be used in very rare cases, where a single Terraform state manages Databricks workspaces on more than one cloud andMore than one authorization method configured
error is a false positive. Valid values arepat
,basic
,azure-client-secret
,azure-msi
,azure-cli
, anddatabricks-cli
.
The provider works with Azure CLI authentication to facilitate local development workflows, though for automated scenarios a service principal auth is necessary (and specification of azure_use_msi
, azure_client_id
, azure_client_secret
and azure_tenant_id
parameters).
Since v0.3.8, it's possible to leverage Azure Managed Service Identity authentication, which is using the same environment variables as azurerm
provider. Both SystemAssigned
and UserAssigned
identities work, as long as they have Contributor
role on subscription level and created the workspace resource, or directly added to workspace through databricks_service_principal.
provider "databricks" {
host = data.azurerm_databricks_workspace.this.workspace_url
azure_workspace_resource_id = azurerm_databricks_workspace.this.id
# ARM_USE_MSI environment variable is recommended
azure_use_msi = true
}
It's possible to use Azure CLI authentication, where the provider would rely on access token cached by az login
command so that local development scenarios are possible. Technically, the provider will call az account get-access-token
each time before an access token is about to expire.
provider "azurerm" {
features {}
}
resource "azurerm_databricks_workspace" "this" {
location = "centralus"
name = "my-workspace-name"
resource_group_name = var.resource_group
sku = "premium"
}
provider "databricks" {
host = azurerm_databricks_workspace.this.workspace_url
}
resource "databricks_user" "my-user" {
user_name = "[email protected]"
display_name = "Test User"
}
provider "azurerm" {
client_id = var.client_id
client_secret = var.client_secret
tenant_id = var.tenant_id
subscription_id = var.subscription_id
}
resource "azurerm_databricks_workspace" "this" {
location = "centralus"
name = "my-workspace-name"
resource_group_name = var.resource_group
sku = "premium"
}
provider "databricks" {
host = azurerm_databricks_workspace.this.workspace_url
azure_workspace_resource_id = azurerm_databricks_workspace.this.id
azure_client_id = var.client_id
azure_client_secret = var.client_secret
azure_tenant_id = var.tenant_id
}
resource "databricks_user" "my-user" {
user_name = "[email protected]"
}
azure_workspace_resource_id
- (optional)id
attribute of azurerm_databricks_workspace resource. Combination of subscription id, resource group name, and workspace name. Required withauzre_use_msi
orazure_client_secret
.azure_client_secret
- (optional) This is the Azure Enterprise Application (Service principal) client secret. This service principal requires contributor access to your Azure Databricks deployment. Alternatively, you can provide this value as an environment variableARM_CLIENT_SECRET
.azure_client_id
- (optional) This is the Azure Enterprise Application (Service principal) client id. This service principal requires contributor access to your Azure Databricks deployment. Alternatively, you can provide this value as an environment variableARM_CLIENT_ID
.azure_tenant_id
- (optional) This is the Azure Active Directory Tenant id in which the Enterprise Application (Service Principal) resides. Alternatively, you can provide this value as an environment variableARM_TENANT_ID
.azure_environment
- (optional) This is the Azure Environment which defaults to thepublic
cloud. Other options aregerman
,china
andusgovernment
. Alternatively, you can provide this value as an environment variableARM_ENVIRONMENT
.azure_use_msi
- (optional) Use Azure Managed Service Identity authentication. Alternatively, you can provide this value as an environment variableARM_USE_MSI
.
There are ARM_*
environment variables provide a way to share authentication configuration using the databricks
provider alongside the azurerm
provider.
This section covers configuration parameters not related to authentication. They could be used when debugging problems, or do an additional tuning of provider's behaviour:
rate_limit
- defines maximum number of requests per second made to Databricks REST API by Terraform. Default is 15.debug_truncate_bytes
- Applicable only whenTF_LOG=DEBUG
is set. Truncate JSON fields in HTTP requests and responses above this limit. Default is 96.debug_headers
- Applicable only whenTF_LOG=DEBUG
is set. Debug HTTP headers of requests made by the provider. Default is false. We recommend to turn this flag on only under exceptional circumstances, when troubleshooting authentication issues. Turning this flag on will log firstdebug_truncate_bytes
of any HTTP header value in cleartext.skip_verify
- skips SSL certificate verification for HTTP calls. Use at your own risk. Default is false (don't skip verification).
The following configuration attributes can be passed via environment variables:
Argument | Environment variable |
---|---|
host |
DATABRICKS_HOST |
token |
DATABRICKS_TOKEN |
username |
DATABRICKS_USERNAME |
password |
DATABRICKS_PASSWORD |
account_id |
DATABRICKS_ACCOUNT_ID |
config_file |
DATABRICKS_CONFIG_FILE |
profile |
DATABRICKS_CONFIG_PROFILE |
azure_client_secret |
ARM_CLIENT_SECRET |
azure_client_id |
ARM_CLIENT_ID |
azure_tenant_id |
ARM_TENANT_ID |
azure_use_msi |
ARM_USE_MSI |
azure_environment |
ARM_ENVIRONMENT |
debug_truncate_bytes |
DATABRICKS_DEBUG_TRUNCATE_BYTES |
debug_headers |
DATABRICKS_DEBUG_HEADERS |
rate_limit |
DATABRICKS_RATE_LIMIT |
For example, with the following zero-argument configuration:
provider "databricks" {}
- Provider will check all the supported environment variables and set values of relevant arguments.
- In case any conflicting arguments are present, the plan will end with an error.
- Will check for the presence of
host
+token
pair, continue trying otherwise. - Will check for
host
+username
+password
presence, continue trying otherwise. - Will check for Azure workspace ID,
azure_client_secret
+azure_client_id
+azure_tenant_id
presence, continue trying otherwise. - Will check for availability of Azure MSI, if enabled via
azure_use_msi
, continue trying otherwise. - Will check for Azure workspace ID presence, and if
AZ CLI
returns an access token, continue trying otherwise. - Will check for the
~/.databrickscfg
file in the home directory, will fail otherwise. - Will check for
profile
presence and try picking from that file will fail otherwise. - Will check for
host
andtoken
orusername
+password
combination, will fail if nothing of these exist.
In Terraform 0.13 and later, data resources have the same dependency resolution behavior as defined for managed resources. Most data resources make an API call to a workspace. If a workspace doesn't exist yet, authentication is not configured for provider
error is raised. To work around this issue and guarantee a proper lazy authentication with data resources, you should add depends_on = [azurerm_databricks_workspace.this]
or depends_on = [databricks_mws_workspaces.this]
to the body. This issue doesn't occur if workspace is created in one module and resources within the workspace are created in another. We do not recommend using Terraform 0.12 and earlier, if your usage involves data resources.
The most common reason for technical difficulties might be related to missing alias
attribute in provider "databricks" {}
blocks or provider
attribute in resource "databricks_..." {}
blocks, when using multiple provider configurations. Please make sure to read alias
: Multiple Provider Configurations documentation article.
Error while installing hashicorp/databricks: provider registry
registry.terraform.io does not have a provider named
registry.terraform.io/hashicorp/databricks
If you notice below error, it might be due to the fact that required_providers block is not defined in every module, that uses Databricks Terraform Provider. Create versions.tf
file with the following contents:
# versions.tf
terraform {
required_providers {
databricks = {
source = "databrickslabs/databricks"
version = "0.5.4"
}
}
}
... and copy the file in every module in your codebase. Our recommendation is to skip version
field for versions.tf
file on module level, and keep it only on environment level.
├── environments
│ ├── sandbox
│ │ ├── README.md
│ │ ├── main.tf
│ │ └── versions.tf
│ └── production
│ ├── README.md
│ ├── main.tf
│ └── versions.tf
└── modules
├── first-module
│ ├── ...
│ └── versions.tf
└── second-module
├── ...
└── versions.tf
Important: Projects in the databrickslabs
GitHub account, including the Databricks Terraform Provider, are not formally supported by Databricks. They are maintained by Databricks Field teams and provided as-is. There is no service level agreement (SLA). Databricks makes no guarantees of any kind. If you discover an issue with the provider, please file a GitHub Issue on the repo, and it will be reviewed by project maintainers as time permits.