You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A production-ready PySpark project template with medallion architecture, Python packaging, unit tests, integration tests, CI/CD automation, Databricks Asset Bundles, and DQX data quality framework.
Production-grade Databricks infrastructure templates for Azure. Deploy in 20 minutes with VNet injection, Unity Catalog, managed identity. Perfect for learning and prototyping. Free and open source.
Production-ready support ticket classification using Unity Catalog AI Functions, Vector Search, and RAG. Features 6-phase workflow, knowledge base integration, and Streamlit dashboard.
databricks-dab-lab is an end-to-end lab that shows how to deploy Databricks Asset Bundles (DABs) with GitHub Actions, using Terraform to provision an Azure Databricks workspace + cluster, then deploying and running three jobs in sequence (data setup → ETL → ML training).
End-to-end Azure Data Engineering project using ADF for incremental ingestion, Databricks (DLT) for Medallion Architecture, and Delta Lake for CDC (SCD Type 1). Managed via Databricks Asset Bundles (DABs) for professional CI/CD. Focuses on real-time streaming, scalability, and Star Schema modeling.
Real Estate ELT pipeline using Databricks Asset Bundles on GCP. Ingests, transforms, and analyzes property data via Delta Live Tables. Follows medallion architecture (Bronze/Silver/Gold), modular Python design, CI/CD automation with GitHub Actions, and full Unit and Integration tests coverage.
Databricks-native intelligent data transformation engine — coherence-scored Bronze/Silver/Gold with entity resolution and temporal reconciliation in a single deployable product.
Production-ready Databricks Asset Bundle for cross-region ML model serving using Delta Sharing. Deploy models and feature tables across workspaces with zero-copy data access and automated online feature store sync.
Enterprise Medallion Lakehouse processing IoT telemetry for heavy machinery. Built with PySpark, Azure Databricks, and Power BI. Features FinOps IaC, CI/CD via Databricks Asset Bundles, and a DirectQuery predictive maintenance semantic model.