Announcement

Free to view yesterday and today
Customer Service: cat_manager

Databricks Terraform Provider: Automate Databricks Infrastructure with IaC

Manage your Databricks infrastructure and resources using HashiCorp Terraform. This provider enables declarative configuration of workspaces, clusters, jobs, permissions, and more.

Go
Added on 2025年5月10日
View on GitHub
Databricks Terraform Provider: Automate Databricks Infrastructure with IaC preview
513
Stars
428
Forks
Go
Language

Project Introduction

Summary

The Databricks Terraform Provider allows users to manage Databricks infrastructure and resources using Terraform's Infrastructure as Code approach. It integrates with the Databricks API to provision and manage resources like workspaces, clusters, jobs, notebooks, permissions, and more.

Problem Solved

Manually configuring Databricks workspaces, clusters, jobs, and permissions across multiple environments is time-consuming, inconsistent, and error-prone. This provider allows you to define your Databricks infrastructure as code, enabling reproducible deployments and version control.

Core Features

Workspace Management

Declaratively manage Databricks workspaces across cloud providers (AWS, Azure, GCP).

Cluster Automation

Automate the creation and configuration of Databricks clusters, including instance pools and libraries.

Jobs and MLOps Resource Management

Define and manage Databricks Jobs and MLOps resources as code.

Identity and Access Management

Control user, group, and service principal access and permissions within Databricks.

Tech Stack

Terraform
Go
Databricks API

Use Cases

The Databricks Terraform Provider can be used for a wide range of infrastructure automation tasks within Databricks.

Provisioning Databricks Workspaces

Details

Provision a new Databricks workspace in your preferred cloud provider and configure its initial settings, including VPC/VNet peering and storage.

User Value

Rapidly deploy new Databricks environments for projects, teams, or testing.

Managing Databricks Clusters and Compute Resources

Details

Define standard cluster configurations (size, autoscaling, libraries) and deploy them consistently for different workloads.

User Value

Ensure consistent, cost-effective, and performant compute resources for all users and jobs.

Automating Data Pipelines and ML Workflows

Details

Define and schedule data processing jobs or machine learning model training pipelines directly in your Terraform configuration.

User Value

Version control and automate the deployment of your operational workflows.

Implementing Access Control and Security Policies

Details

Manage user access, group memberships, and permissions on clusters, notebooks, and other workspace objects.

User Value

Enforce security best practices and compliance requirements programmatically.

Recommended Projects

You might be interested in these projects

eclipse-zenohzenoh

Zenoh is a unified data abstraction layer designed for efficiency and performance across data in motion, data in use, data at rest, and computations. It offers a unique blend of publish/subscribe, geo-distributed storage, querying, and computation capabilities.

Rust
1903196
View Details

agalwoodMotrix

Motrix is a full-featured download manager that supports downloading HTTP, FTP, BT, and Magnet links. It provides a clean and easy-to-use interface, offering high-speed and reliable downloads for various file types.

JavaScript
482544674
View Details

nothingsstb

A collection of high-quality, single-file public domain libraries for C/C++, covering image processing, font handling, audio, and more. Designed for easy integration and minimal dependencies, perfect for games, tools, and small projects.

C
292547846
View Details