subcategory |
---|
Storage |
This is a resource that lets you manage relatively small files on Databricks File System (DBFS). The best use cases are libraries for databricks_cluster or databricks_job. You can also use databricks_dbfs_file and databricks_dbfs_file_paths data sources.
In order to manage file on Databricks File System with Terraform, you must specify source
attribute containing full path to the file on local filesystem.
resource "databricks_dbfs_file" "this" {
source = "${path.module}/main.tf"
path = "/tmp/main.tf"
}
Alternatively, you can create DBFS files with custom content, using filesystem functions.
resource "databricks_dbfs_file" "this" {
content_base64 = base64encode(<<-EOT
Hello, world!
Module is ${abspath(path.module)}
EOT
)
path = "/tmp/this.txt"
}
Install databricks_library on all databricks_clusters:
data "databricks_clusters" "all" {
}
resource "databricks_dbfs_file" "app" {
source = "${path.module}/baz.whl"
path = "/FileStore/baz.whl"
}
resource "databricks_library" "app" {
for_each = data.databricks_clusters.all.ids
cluster_id = each.key
whl = databricks_dbfs_file.app.dbfs_path
}
-> Note DBFS files would only be changed, if Terraform stage did change. This means that any manual changes to managed file won't be overwritten by Terraform, if there's no local change.
The following arguments are supported:
source
- The full absolute path to the file. Conflicts withcontent_base64
.content_base64
- Encoded file contents. Conflicts withsource
. Use ofcontent_base64
is discouraged, as it's increasing memory footprint of Terraform state and should only be used in exceptional circumstances, like creating a data pipeline configuration file.path
- (Required) The path of the file in which you wish to save.
In addition to all arguments above, the following attributes are exported:
id
- Same aspath
.file_size
- The file size of the file that is being tracked by this resource in bytes.dbfs_path
- Path, but withdbfs:
prefix
The resource dbfs file can be imported using the path of the file
$ terraform import databricks_dbfs_file.this <path>
The following resources are often used in the same context:
- End to end workspace management guide.
- databricks_dbfs_file data to get file content from Databricks File System (DBFS).
- databricks_dbfs_file_paths data to get list of file names from get file content from Databricks File System (DBFS).
- databricks_global_init_script to manage global init scripts, which are run on all databricks_cluster and databricks_job.
- databricks_library to install a library on databricks_cluster.
- databricks_mount to mount your cloud storage on
dbfs:/mnt/name
. - databricks_workspace_conf to manage workspace configuration for expert usage.