Binance is hiring a

Data Warehouse Engineer

Binance is the leading global blockchain ecosystem and cryptocurrency infrastructure provider whose suite of financial products includes the world’s largest digital-asset exchange.
Our mission is to accelerate cryptocurrency adoption and increase the freedom of money.

If you’re looking for a fast-paced, mission-driven organization where opportunities to learn and excel are endless, then Binance is the place for you.

Requirements

  • According to the company's data warehouse specifications and business understanding, build a universal and flexible data warehouse system that can quickly support the needs and reduce repetitive development work efforts.
  • Data model design, development, testing, deployment, online data job monitoring, and the ability to quickly solve complex problems, especially the optimization of complex calculation logic and performance tuning, etc.
  • Participate in Data governance, including the construction of the company’s metadata management system and data quality monitoring system. 
  • Design and implement a data platform integrated with data lake warehouse to support real-time data processing and analysis requirements.
  • Build knowledge graph, and provide in-depth business insight.
  • Participate in technical team building and learning growth, and contribute to the team’s overall knowledge accumulation and skill improvement.

Responsibilities

  • 5+ years experiences of data lake and data warehouse design and development experience.
  • Deeply understanding of data warehouse modeling and data governance. Solid knowledge of data warehouse development methodology, including dimensional modeling, information factory etc.
  • Proficient in Java / Scala / Python (at least one language) and Hive & Spark SQL programming languages.
  • Familiar with OLAP technology (such as: kylin, impala, presto, druid, etc.).
  • Proficient in Big Data batch pipeline development.
  • Familiar with Big Data components including but not limited to Hadoop, Hive, Spark, Delta lake, Hudi, Presto, Hbase, Kafka, Zookeeper, Airflow, Elastic search, Redis, etc.
  • Experiences with AWS Big Data services are a plus.
  • Have a strong team collaboration attitude and develop partnerships with other teams and businesses.
  • Rich experience in real-time data processing, familiar with stream processing frameworks such as Apache Kafka, Apache Flink, in-depth knowledge of Lakehouse technology, practical project experience, proficiency in StarRocks, including its data model design, query optimization and performance tuning.
  • Experience in knowledge graph construction and application, and knowledge of graph databases such as Nebula, etc.

Apply for this job

Please mention you found this job on AI Jobs. It helps us get more startups to hire on our site. Thanks and good luck!

Get hired quicker

Be the first to apply. Receive an email whenever similar jobs are posted.

Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Engineer Q&A's
Report this job
Apply for this job