WATI.io is hiring a

Machine Learning / Data Engineer

Full-Time
Remote

WATI is an early-stage, venture-backed SaaS platform that is defining how companies communicate with their customers. Through our customer engagement software, built on top of WhatsApp’s Business API, businesses can easily engage with their customers in-real time - at scale!

We are growing fast, and we are now looking for a Machine Learning Data Engineer to work closely with Development, QA, DevOps, Support and Product. You will provide the necessary pipelines, infrastructure, and automation for Data Scientist to train and evaluate models and allow users to analyze and act on vast quantities of data effortlessly. The team has one of the most critical roles to play in ensuring our products are best-in-class in the industry.



What you’ll do:

  • Work with a team to design and build services that power industry-leading data products and transform data science prototypes
  • Build automation tools for scientists to deploy models
  • Develop, maintain and improve production machine learning applications according to requirements
  • Run machine learning tests and experiments
  • Work with peers on code reviews / PRs
  • Work across several teams responsible for different aspects of contribution to machine learning data at WATI


What will make you stand out:

  • Good knowledge of CI / CD and associated best practices
  • Familiarity with Docker and or Kubernetes based development and orchestration
  • Created automated / scalable infrastructure and pipelines for teams in the past
  • Contributed to the open-source community (GitHub, Stack Overflow, blogging)
  • Prior experience with the Big Data ecosystem like Spark
  • Prior experience in the chatbot, NLP or text mining fields
  • Good understanding of SOLID Principles


Requirements

  • Substantial experience in message queue, stream processing in these languages data streaming architectures: Google Dataflow/Apache Beam, Apache Samza, etc
  • Good knowledge and experience using and optimizing these GCP services: Google BigQuery, Google Dataflow, Google Cloud Composer, Google Cloud Databases, Google Cloud data store
  • Understanding of traditional NLP algorithms, deep learning algorithms and state of the art pre-trained NLP Models
  • Significant programming experience in “either” Python, Java, shell script “or” other like languages
  • Strong database manipulation skills in SQL and MongoDB with cost optimization concern and data tool / library including Pandas, matplotlib, Pytorch, Hugging face
  • Have worked with large volumes of data in the past
  • Bachelor’s degree in Computer Science, Computer Engineering, or equivalent work experience
  • Excellent communication skills in English, both written and verbal

This job is no longer available

Enter your email address below to get notified whenever we find a similar job post.

Unsubscribe at any time.