Databricks framework to validate Data Quality of pySpark DataFrames and Tables
-
Updated
Jan 28, 2026 - Python
Databricks framework to validate Data Quality of pySpark DataFrames and Tables
Automated migrations to Unity Catalog
End-to-end Data Lakehouse project built on Databricks, following the Medallion Architecture (Bronze, Silver, Gold). Covers real-world data engineering and analytics workflows using Spark, PySpark, SQL, Delta Lake, and Unity Catalog. Designed for learning, portfolio building, and job interviews.
A cloud-native data pipeline and visualization project analyzing Formula 1 racing data using Azure, Databricks, Delta Lake, Tableau, and Python for insightful EDA and interactive dashboards.
Open, Multi-modal Catalog for Data & AI, written in Rust
Notebooks, terraform, tools to enable setting up Unity Catalog
A production-ready PySpark project template with medallion architecture, Python packaging, unit tests, integration tests, CI/CD automation, Databricks Asset Bundles, and DQX data quality framework.
Production-grade Databricks infrastructure templates for Azure. Deploy in 20 minutes with VNet injection, Unity Catalog, managed identity. Perfect for learning and prototyping. Free and open source.
Unity Catalog Explorer is a TypeScript + Next.js based Web UI for the Unity Catalog OSS.
How to Configure Azure Databricks Unity Catalog using Terraform
Production-ready support ticket classification using Unity Catalog AI Functions, Vector Search, and RAG. Features 6-phase workflow, knowledge base integration, and Streamlit dashboard.
Real Estate ELT pipeline using Databricks Asset Bundles on GCP. Ingests, transforms, and analyzes property data via Delta Live Tables. Follows medallion architecture (Bronze/Silver/Gold), modular Python design, CI/CD automation with GitHub Actions, and full Unit and Integration tests coverage.
End-to-end backend and data hub architecture on Azure, integrating Databricks and a suite of Azure services for seamless data processing, analytics, and deployment.
spark, databricks, kafka, batch and stream-processing
End-to-end Databricks Lakehouse pipeline using Auto Loader, Delta Lake, Unity Catalog, Bronze–Silver–Gold, and business marts (Daily Sales, Top Categories, Customer LTV).
This is a Web API to connect to your Lakehouse with Unity
A curated list of awesome Databricks resources, including tutorials, notebooks, libraries, and tools.
Add a description, image, and links to the unity-catalog topic page so that developers can more easily learn about it.
To associate your repository with the unity-catalog topic, visit your repo's landing page and select "manage topics."