Databricks SME

AI overview

Lead implementations of enterprise Databricks Lakehouse architectures while ensuring scalability, security, and compliance, collaborating with cross-functional teams to meet business needs.

Key Responsibilities:


Architecture & Platform Design

  • Design enterprise Databricks Lakehouse architectures aligned with the Databricks Well-Architected Framework
  • Define reference architectures for batch, streaming, analytics, and ML workloads
  • Select and standardize cluster, compute, and workspace architectures
  • Design multi-workspace strategies (dev/test/prod, shared vs. isolated)
  • Ensure architectures meet scalability, availability, and performance requirements


Well-Architected Framework Alignment

Apply Databricks best practices across all pillars, including:

  • Security & Governance (Unity Catalog, IAM, data access controls)
  • Reliability & Resilience (job retries, checkpointing, failure isolation)
  • Performance Efficiency (cluster sizing, autoscaling, caching)
  • Cost Optimization (compute policies, workload separation, monitoring)
  • Operational Excellence (monitoring, automation, CI/CD, runbooks)


Implementation & Engineering

  • Lead Databricks workspace, cluster, and Unity Catalog implementations
  • Implement Delta Lake, Delta Live Tables (DLT), and Structured Streaming
  • Build and optimize ETL/ELT pipelines using Spark and SQL
  • Integrate Databricks with cloud services (S3/ADLS/GCS, IAM, Key Vault, networking)
  • Establish CI/CD pipelines for notebooks, jobs, and infrastructure


Security, Governance & Compliance

  • Implement Unity Catalog for centralized governance
  • Define data classification, lineage, and audit strategies
  • Enforce least-privilege access and secure networking patterns
  • Support compliance requirements (HIPAA, SOC 2, PCI, GDPR as applicable)


Operations & Optimization

  • Monitor platform health, performance, and cost
  • Troubleshoot production issues across jobs, clusters, and data pipelines
  • Perform workload tuning and cost-performance optimization
  • Define SLOs, alerts, and operational metrics


Collaboration & Leadership

  • Partner with Data Engineering, Analytics, ML, Platform, and Security teams
  • Translate business requirements into technical architectures
  • Provide architectural guidance and technical mentorship
  • Communicate risks, tradeoffs, and recommendations to leadership


Required Qualifications:


Experience

  • 7+ years in data engineering, analytics, or platform architecture
  • 3–5+ years hands-on Databricks experience in production environments
  • Proven experience applying the Databricks Well-Architected Framework
  • Experience designing cloud-native lakehouse architectures
  • Experience supporting mission-critical data platforms


Technical Skills

  • Databricks Lakehouse Platform
  • Apache Spark (PySpark / Scala / Spark SQL)
  • Delta Lake, Delta Live Tables, Structured Streaming
  • Unity Catalog (governance, lineage, access controls)
  • Cloud platforms: AWS, Azure, or GCP
  • Infrastructure as Code (Terraform strongly preferred)
  • CI/CD tools (GitHub Actions, Azure DevOps, GitLab, etc.)
  • Data formats and protocols (Parquet, JSON, Avro)


Certifications Required:

  • Databricks Certified Data Engineer Professional
  • Databricks Certified Professional Architect (or equivalent advanced certification) 


Preferred / Additional Certifications

  • AWS Certified Solutions Architect (Associate or Professional)
  • Azure Solutions Architect Expert
  • Google Professional Data Engineer
  • Databricks Machine Learning Professional
  • Snowflake or other cloud data platform certifications


Soft Skills

  • Strong architectural decision-making and documentation skills
  • Excellent communication with technical and non-technical stakeholders
  • Ability to lead design reviews and architecture governance forums
  • Strong troubleshooting and performance-tuning mindset


Nice-to-Have Experience

  • MLflow and MLOps architectures
  • Real-time analytics and streaming pipelines
  • Multi-region or cross-account data architectures
  • Consulting or MSP delivery experience
Report this job
Apply for this job