Data Engineer

AI overview

Contribute meaningful technical impact by designing, implementing, and maintaining scalable data integration solutions within a cloud-based environment.

About Us

Abacus Insights is transforming how data works for health plans. Our mission is simple: make healthcare data usable, so the people responsible for care and cost decisions can act faster, with confidence.  
We help health plans break down data silos to create a single, trusted data foundation. That foundation powers better decisions —so plans can improve outcomes, reduce waste, and deliver better experiences for members and providers alike.  

Backed by $100M from top investors, we’re tackling big challenges in an industry that’s ready for change.  Our platform enables GenAI use cases by delivering clean, connected, and reliable healthcare data that can support automation, prioritization, and decision workflows—and it’s why we are leading the way.

Our innovation begins with people. We are bold, curious, and collaborative—because the best ideas come from working together. Ready to make an impact? Join us and let's build the future together.

About the role 

We are seeking a skilled Data Engineer to join our growing Tech Ops division. As the organization continues to expand, this role provides the opportunity to contribute meaningful technical impact across our data delivery ecosystem. In this position, you will collaborate with customers, data vendors, and internal engineering teams to design, implement, and maintain reliable data integration solutions within a modern, cloudbased environment. 

You will apply advanced knowledge of distributed computing, data architecture, and cloudnative engineering to build scalable and resilient data ingestion and transformation pipelines. As a technical contributor, you will support the adoption of Abacus’s core data management platform and help ensure highquality, compliant data operations across the data lifecycle. 

Your daytoday 

Design and develop data pipelines 

  • Build and maintain batch and realtime data pipelines using PySpark, SparkSQL, Databricks Workflows, and distributed processing frameworks. 
  • Create ingestion frameworks integrating with Databricks, Snowflake, AWS services (S3, Lambda, SQS), and vendor data APIs, ensuring data quality, lineage, and schema evolution. 

Support data architecture & modeling 

  • Develop data models, including star/snowflake schemas, and apply optimization techniques for analytical workloads in cloud data environments. 
  • Participate in design discussions for reliable and faulttolerant architectures across multiaccount AWS environments. 

Translate business needs into engineering solutions 

  • Convert customer and internal business requirements into clear technical specifications, data flows, and engineering components. 
  • Contribute to reusable frameworks and standardized engineering patterns. 

Implement security & compliance workflows 

  • Apply data security best practices including RBAC, encryption at rest/in transit, PHI handling, tokenization, auditing, and adherence to HIPAA and SOC 2 compliance requirements. 
  • Support automation of secure data operations and governance frameworks. 

Advance engineering best practices 

  • Implement CI/CD workflows for data pipelines, including code versioning, automated testing, orchestration, observability, and logging standards. 
  • Conduct performance profiling and assist with compute optimization, cost efficiency, and tuning techniques for Databricks and Snowflake environments. 

Documentation & Collaboration 

  • Produce clear technical documentation, runbooks, architecture diagrams, and operational standards. 
  • Assist senior engineers through collaborative code reviews, knowledgesharing sessions, and crossteam engineering initiatives. 

What you bring to the team 

  • Bachelor’s degree in Computer Science, Computer Engineering, or a closely related technical field. 
  • 3+ years of handson experience as a Data Engineer working with distributed data processing systems in modern cloud environments. 
  • Working knowledge of U.S. healthcare data domains — including claims, eligibility, and provider datasets — and ability to apply this knowledge to ingestion and transformation workflows. 
  • Strong communication skills to articulate technical concepts across both technical and nontechnical stakeholders. 
  • Proficiency in Python, SQL, and PySpark, including distributed data transformations and performanceoptimized queries. 
  • Demonstrated experience building and operating production ETL/ELT pipelines using Databricks, Airflow, or similar orchestration frameworks. 
  • Familiarity with cloudnative data platforms and services such as dbt, Kafka, Delta Lake, and eventdriven/streaming architectures. 
  • Experience with structured and semistructured data formats (Parquet, JSON, Avro, ORC), including schema evolution techniques. 
  • Working knowledge of AWS data ecosystem components—S3, SQS, Lambda, Glue, IAM—used to support scalable data engineering workloads. 
  • Experience with Terraform and modern CI/CD pipelines (e.g., GitLab) for automating deployment and versioning. 
  • Understanding of SQL optimization strategies including partitioning, pruning, ZOrdering, clustering, and caching for large analytical workloads. 
  • Experience using cloud data warehouse technologies such as Snowflake, BigQuery, or Redshift, including performance tuning and data modeling practices. 

What we would like to see but not required: 

  • Experience in healthcare or payer data environments at scale. 

What you’ll get in return 

  • Competitive Leave & Benefits
  • Comprehensive health coverage
  • Equity for every employee – share in our success
  • Growth-focused environment – your development matters here

Working Arrangements

  • Standard hours: 9 hours/day, 5 working days
  • Location: Onsite
  • Shift: 10 AM – 7 PM local time

Our Commitment as an Equal Opportunity Employer

As a mission-led technology company helping to drive better healthcare outcomes, Abacus Insights believes that the best innovation and value we can bring to our customers comes from diverse ideas, thoughts, experiences, and perspectives. Therefore, we dedicate resources to building diverse teams and providing equal employment opportunities to all applicants. Abacus prohibits discrimination and harassment regarding race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws.

At the heart of who we are is a commitment to continuously and intentionally building an inclusive culture—one that empowers every team member across the globe to do their best work and bring their authentic selves. We carry that same commitment into our hiring process, aiming to create an interview experience where you feel comfortable and confident showcasing your strengths. If there’s anything we can do to support that—big or small—please let us know.

Perks & Benefits Extracted with AI

  • Health Insurance: Comprehensive health coverage
Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Data Engineer Q&A's
Report this job
Apply for this job