Senior Data Engineer - Airflow and PySpark (Hybrid)

Auburn Hills , United States
full-time Hybrid

AI overview

Lead major data initiatives by developing scalable data pipelines and CI/CD processes using Airflow and PySpark for an industry-leading automotive client.

This role is based in Auburn Hills, Michigan, and follows a hybrid work model, requiring in-office presence three days per week.

Please note: Due to the in-office requirement, we will only be considering candidates that are local to the Metro Detroit area at this time.

Detroit Labs was founded in 2011 with a vision for building digital products, services, and the teams that power them. We create digital solutions that transform the way our clients do business. We build genuine relationships based on respect, trust, and results. We foster a diverse and inclusive culture that values people - providing them with the tools, resources, and support they need to thrive professionally, exceed client expectations, and be themselves at work. We have a variety of client teams we work with ranging from startups to Fortune 500 companies so there are always new and exciting projects going on.

Detroit Labs is looking for an experienced Data Engineer with expertise in Airflow and Pyspark to join an exciting project with an industry leading automotive client. As a Senior Data Engineer & Technical Lead, you will play a pivotal role in delivering major data engineering initiatives within the Data & Advanced Analytics space. This position requires hands-on expertise in building, deploying, and maintaining robust data pipelines using Python, PySpark, and Airflow, as well as designing and implementing CI/CD processes for data engineering projects.

Our Application

At Detroit Labs, a member of our team will read over every application (including yours), and will review your resume in addition to your responses to the application questions.

To help us get to know you better, we encourage you to answer these questions genuinely and honestly. We value each applicant and want to learn about the real you. Be yourself in your responses, and our team will look forward to understanding what you can bring to this role!

Requirements

7+ years of Data Engineering experience building production-grade data pipelines using Python and PySpark

    • Experience designing, deploying, and managing Airflow DAGs in enterprise environments
    • Experience maintaining CI/CD pipelines for data engineering workflows, including automated testing and deployment
    • Experience with cloud workflows and containerization, using Docker and cloud platforms (GCP preferred) for data engineering workloads
    • Knowledge and ability to follow twelve-factor design principles
    • Experience and ability to write object-oriented Python code, manage dependencies, and follow industry best practices
    • Proficiency with Git for source code management and collaboration (commits, branching, merging, GitHub/GitLab workflows).
    • Experience working with command lines in Unix/Linux-like environments
    • Solid understanding of SQL for data ingestion and analysis
    • Engineering mindset. Able to write code with an eye for maintainability and testability
    • Collaborative mindset. Comfortable with code reviews, paired programming, and using remote collaboration tools effectively
    • Detroit Labs is not currently able to hire candidates who will reside outside of the United States during their term of employment

Responsibilities

      • Design, develop, and optimize scalable data pipelines using Python and PySpark for batch and streaming workloads
      • Build, schedule, and monitor complex workflows using Airflow, ensuring reliability and maintainability
      • Architect and implement CI/CD pipelines for data engineering projects using GitHub, Docker, and cloud-native solutions (GCP preferred)
      • Apply test-driven development (TDD) practices and automate unit/integration tests for data pipelines
      • Implement secure coding best practices and design patterns throughout the development lifecycle
      • Work closely with Data Architects, QA teams, and business stakeholders to translate requirements into technical solutions
      • Create and maintain technical documentation, including process/data flow diagrams and system design artifacts
      • Lead and mentor junior engineers, providing guidance on coding, testing, and deployment best practices
      • Analyze and resolve technical issues across the data stack, including pipeline failures and performance bottlenecks
      • Cross-train team members outside the project team (e.g., operations support) for full knowledge coverage

Benefits

  • The salary range for this role is from $160,000 - $180,000 commensurate with experience
  • Full medical, dental, vision benefits
  • 401K contribution options
  • Quarterly outings and events
  • Paid holidays and vacation time
  • Parental leave program
  • Monthly budgets for “team fun” bonding events
  • Free lunch for various company meetings and Lunch & Learns
  • Access to our mentorship program and employee resource groups (ERGs)
  • Volunteer opportunities
  • All-company remote-friendly activities
  • Plenty of Detroit Labs swag

Perks & Benefits Extracted with AI

  • Health Insurance: Full medical, dental, vision benefits
  • Detroit Labs swag: Plenty of Detroit Labs swag
  • Paid Parental Leave: Parental leave program
  • Paid Time Off: Paid holidays and vacation time
  • Remote-Friendly: All-company remote-friendly activities

We are Detroit Labs, and we live to develop iPhone, iPad and Android apps. #mobiledevelopment

View all jobs
Salary
$160,000 – $180,000 per year
Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Senior Data Engineer Q&A's
Report this job
Apply for this job