Data Engineer

Noida , India
full-time

AI overview

Collaborate with cross-functional teams to design and maintain robust data pipelines and systems that drive critical business decisions for Fortune 500 clients.
About ShyftLabs ShyftLabs is a fast-growing data product company founded in early 2020, working primarily with Fortune 500 clients. We design and deliver cutting-edge digital and data-driven solutions that help businesses accelerate growth, improve decision-making, and create measurable value through innovation. Position Overview We’re looking for an experienced Data Engineer who’s passionate about building scalable, high-performance data solutions. In this role, you’ll collaborate with cross-functional teams-including Data Engineers, Analysts, and Product Managers-to design, implement, and maintain robust data pipelines and systems that power our clients’ most critical business decisions. Key Responsibilities
  • Design, develop, and maintain data pipelines and ETL/ELT processes using Python.
  • Build and optimize scalable, high-performance data applications.
  • Collaborate with cross-functional teams to define requirements and deliver reliable solutions.
  • Develop and manage real-time streaming pipelines using Pub/Sub or Apache Beam.
  • Participate in code reviews, architecture discussions, and continuous improvement initiatives.
  • Monitor, troubleshoot, and optimize production data systems for reliability and performance.
  • Key Qualifications
  • 5+ years of professional experience in software or data engineering using Python.
  • Strong understanding of software engineering best practices (testing, version control, CI/CD).
  • Proven experience building and optimizing ETL/ELT pipelines and data workflows.
  • Proficiency in SQL and database concepts.
  • Experience with data processing frameworks (e.g., Pandas).
  • Understanding of software design patterns and scalable architecture principles.
  • Experience with cloud platforms (GCP preferred).
  • Knowledge of CI/CD pipelines and Infrastructure as Code tools.
  • Familiarity with containerization (Docker, Kubernetes).
  • Bachelor’s degree in Computer Science, Engineering, or a related field (or equivalent experience).
  • Excellent problem-solving, analytical, and communication skills.
  • Preferred Qualifications
  • Experience with GCP services such as Cloud Run and Dataflow.
  • Experience with stream processing technologies (e.g., Pub/Sub).
  • Familiarity with workflow orchestration tools (e.g., Airflow).
  • Exposure to data visualization tools or libraries.
  • Knowledge of GitLab CI/CD and Terraform.
  • Experience with Snowflake, BigQuery, or Databricks.
  • GCP Data Engineer Certification is a plus.
  • We are proud to offer a competitive salary alongside a strong insurance package. We pride ourselves on the growth of our employees, offering extensive learning and development resources.
    Get hired quicker

    Be the first to apply. Receive an email whenever similar jobs are posted.

    Ace your job interview

    Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

    Data Engineer Q&A's
    Report this job
    Apply for this job