DevOps Engineer (DataOps)

AI overview

As a DataOps Engineer, you will design and optimize resilient data pipelines empowering analytics and machine learning across the organization.

We’re seeking a DataOps Engineer who thrives at the intersection of data engineering, DevOps, and workflow orchestration. You’ll be instrumental in designing, automating, and optimizing data pipelines that power analytics, machine learning, and operational intelligence across the organization.

If you’re passionate about building resilient data systems, streamlining deployments, and enabling data teams to move faster with confidence—this role is for you.

Responsibilities

  • Administer a large-scale MongoDB cluster using a combination of bash, python, and Linux OS skills.
  • Work with data engineers to design and maintain scalable, automated data pipelines using tools like Apache Airflow, dbt, and Terraform.
  • Create abstractions around data workflows for self-service creation of data products in development teams.
  • Implement CI/CD workflows for data infrastructure and analytics code
  • Monitor and optimize data workflows for performance, reliability, and cost-efficiency
  • Integrate cloud-native services (e.g., S3, Redshift, BigQuery, Databricks) into unified data workflows
  • Develop disaster recovery strategies and backup automation for critical data assets
  • Champion DataOps best practices across teams, including version control, testing, and observability.
  • Participate in the team’s emergency on-call rotation, to ensure 24/7 uptime of our systems.
  • 3+ years of experience in data engineering, DevOps, or cloud infrastructure roles
  • Proficiency in automating administrative workflows using Bash and Python–with an emphasis on writing clean and maintainable code.
  • Intermediate knowledge of Linux system administration.
  • Basic proficiency writing queries in both relational (SQL) and NoSQL paradigms.
  • Experience administering Big Data querying engines like Hadoop, Apache Spark, or Google BigQuery.
  • Experience with data orchestration tools (Airflow, Prefect, Dagster)
  • Familiarity with cloud platforms (AWS, Azure, or GCP) and infrastructure-as-code (Terraform, CloudFormation)
  • Strong understanding of data lake and warehouse architectures
  • Experience working with containers.

Headquartered in Mountain View, California, with over 220 team members across the United States and Europe, DNAnexus is experiencing rapid growth and market adoption. With the support of leading investors including Google Ventures and Blackstone, and trusted by hundreds of the world's biomedical leaders,  the company is at the innovative forefront with our precision health data cloud to drive scientific breakthroughs.   If you are interested in joining our team, please apply today!

DNAnexus is a leading provider of secure, scalable, and intuitive biomedical data analysis software and bioinformatics applications for the life sciences and healthcare communities. The company actively manages and supports more than 80 petabytes of complex genomic, multi-omic, and clinical datasets on behalf of a growing network of collaborations with large-scale biobanks, as well as leading pharmaceutical, clinical diagnostic, academic research, and government organizations. Over 40,000 scientists across 48 countries are now using the highly collaborative, cloud-based, end-to-end platform to gain data-driven insights that can advance scientific discovery, accelerate precision medicine, and improve patient care.

View all jobs
Get hired quicker

Be the first to apply. Receive an email whenever similar jobs are posted.

Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Operations Engineer Q&A's
Report this job
Apply for this job