As a Data Engineer on our Professional Services team, you will be responsible for working with customers, understanding their needs and requirements, and be able to discuss with them the “art of the possible”. You will also design and implement solutions for data warehouses, data lakes, ETL jobs, and data pipelines using AWS Services, as well as designing and implementing AI/ML solutions using AWS or IBM services such as Bedrock and WatsonX.
You will be responsible for:
Consulting with customers to:
o Understand their data management strategy
o Provide insights into optimizing their data strategy
o Architect and Design data management processes
o Excite customers on how AI/ML services can enhance their business
· Implementing and Deploying machine learning models and enable advanced analytics
· Documenting data architectures and data flows
· Driving innovation internally and for customers
· Contributing to R&D projects which may turn into new services offerings
How you will be successful:
· Living and breathing the “cloud-first” approach
· Thinking analytically to solve complex business problems
· Obsessively delivering amazing customer experiences
· Continually tracking new developments in one or more cloud platforms
· Building trusting relationships with all team members
· Comfortable with pushing boundaries and technical limits
· Keeping up to date on industry trends
· Always be learning
Qualifications for the Sr- level role:
· 5+ years consulting/professional service experience, with customer facing responsibilities
· 4+ years professional AWS and/or WatsonX experience
· At least (1) AWS Professional Level Certification
· Proficient in one or more of these languages: Python, R, Java, Scala
· Experience in:
o Designing and building data pipelines,
o Creating machine learning models or using LLMs
o SQL and NoSQL databases like PostgreSQL, MySQL, MongoDB, Cassandra to store and query large datasets
o Data modeling - ability to design conceptual, logical and physical data models
o Dimensional modeling, star schemas, data warehouses
o ETL (Extract, Transform, Load) tools like Glue, Informatica, Talend, Pentaho etc. to integrate and move data between systems
o Big data frameworks like Hadoop, Spark, or Kafka. for distributed data processing and building data lakes
o Machine learning frameworks like Tensorflow, PyTorch, Keras, Scikit-Learn for building ML models
o Business Intelligence experience with Power BI and/or QuickSight is a great addition
· Knowledge of data architecture patterns like lambda, kappa architecture.
· Ability to design scalable and flexible data pipelines are also a great addition
· Experience working within standard agile methodologies