Data Engineer

AI overview

Build and maintain core datasets and develop REST APIs and data pipelines, ensuring code quality and system performance while collaborating with cross-functional teams.
At Kpler, we are dedicated to helping our clients navigate complex markets with ease. By simplifying global trade information and providing valuable insights, we empower organisations to make informed decisions in commodities, energy, and maritime sectors. Since our founding in 2014, we have focused on delivering top-tier intelligence through user-friendly platforms. Our team of over 700 experts from 35+ countries works tirelessly to transform intricate data into actionable strategies, ensuring our clients stay ahead in a dynamic market landscape. Join us to leverage cutting-edge innovation for impactful results and experience unparalleled support on your journey to success. Build and maintain Kpler's core datasets (vessels characteristics, companies, geospatial data). You will be responsible for creating and maintaining REST APIs, streaming pipelines (Kafka Stream), and Spark batch pipelines.  The role involves end-to-end ownership of development tasks, beginning with a thorough understanding of assigned tickets and requirements. The individual designs and builds functionality—including APIs and data processing components—ensuring code is deployed to development environments and reviewed through peer and product testing.  They are responsible for writing and executing unit, integration, and functional tests aligned with defined test scenarios, while ensuring full compliance through detailed validation. After release, the role includes monitoring system performance, alerts, and SLOs to ensure optimal functionality and reliability. Responsibilities
  • Deliver well-documented, maintainable code following Test-Driven Development (TDD) principles, ensuring comprehensive unit, integration, and end-to-end testing. 
  • Design, operate, and document versioned RESTful APIs using FastAPI and JVM-based frameworks, ensuring scalability, reliability, and backward compatibility.
  • Implement and enforce data schema evolution and versioning strategies to support reliable data exchange across systems.
  • Develop and maintain batch and streaming data pipelines using technologies such as Kafka and Spark, handling backpressure, orchestration, retries, and data quality controls.
  • Instrument services with metrics, logs, and traces; contribute to CI/CD pipelines, automated testing, and participate in incident response to ensure system resilience and SLO adherence.
  • Partner closely with Product and cross-functional teams to translate requirements into high-quality technical solutions that deliver business outcomes.
  • Adhere to clean code and architectural standards through code reviews, testing, and Agile development practices, ensuring maintainable and compliant solutions. 
  • Skills and Experience
  • 3–5 years of experience in data-focused software engineering roles.
  • Good programming skills in Scala (or JVM) experience with Python preferred.
  • Proven experience designing and operating RESTful APIs, versioned interfaces.
  • Good understanding of data modeling, schema evolution, and serialization technologies such as Avro or Protobuf.
  • Experience building and maintaining batch or streaming data systems, with knowledge of streaming patterns and reliability concerns.
  • Familiarity CI/CD pipelines, and modern monitoring and alerting practices.
  • Proficiency with Git-based workflows, code reviews, and Agile development methodologies.
  • Good sense of ownership, with pragmatic problem-solving skills, constructive critique and the ability to deliver end-to-end solutions.
  • Excellent communication skills and fluency in English, with the ability to collaborate across product and engineering teams.
  • Nice to have
  • Experience with Apache Airflow for workflow orchestration.
  • Exposure to cloud platforms (preferably AWS) and infrastructure as code using Terraform.
  • Experience with Docker and Kubernetes in production environments.
  • Hands-on knowledge of Kafka and event-driven or microservices architectures.
  • Familiarity with JVM build and tooling ecosystems such as Gradle or Maven.
  • We are a dynamic company dedicated to nurturing connections and innovating solutions to tackle market challenges head-on. If you thrive on customer satisfaction and turning ideas into reality, then you’ve found your ideal destination. Are you ready to embark on this exciting journey with us?

    We make things happen
    We act decisively and with purpose, going the extra mile.

    We build
together
    We foster relationships and develop creative solutions to address market challenges.

    We are here to help
    We are accessible and supportive to colleagues and clients with a friendly approach.


    Our People Pledge

    Don’t meet every single requirement? Research shows that women and people of color are less likely than others to apply if they feel like they don’t match 100% of the job requirements. Don’t let the confidence gap stand in your way, we’d love to hear from you! We understand that experience comes in many different forms and are dedicated to adding new perspectives to the team.

    Kpler is committed to providing a fair, inclusive and diverse work-environment. We believe that different perspectives lead to better ideas, and better ideas allow us to better understand the needs and interests of our diverse, global community. We welcome people of different backgrounds, experiences, abilities and perspectives and are an equal opportunity employer.



    By applying, I confirm that I have read and accept the Staff Privacy Notice

    Founded in 2014, Kpler is an intelligence company providing transparency solutions in energy markets

    View all jobs
    Ace your job interview

    Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

    Data Engineer Q&A's
    Report this job
    Apply for this job