Shape LLM system behavior by designing prompting strategies and evaluation pipelines, while collaborating with cross-functional teams on research challenges in AI.
About Mistral
At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.
We democratize AI through high-performance, optimized, open-source and cutting-edge models, products and solutions. Our comprehensive AI platform is designed to meet enterprise as well as personal needs. Our offerings include Le Chat, La Plateforme, Mistral Code and Mistral Compute - a suite that brings frontier intelligence to end-users.
We are a dynamic, collaborative team passionate about AI and its potential to transform society. Our diverse workforce thrives in competitive environments and is committed to driving innovation. Our teams are distributed between France, USA, UK, Germany and Singapore. We are creative, low-ego and team-spirited.
Join us to be part of a pioneering company shaping the future of AI. Together, we can make a meaningful impact. See more about our culture on https://mistral.ai/careers.
Mistral AI participates in the E-Verify program
About the role
As a Model Behavior Architect, you’re at the forefront of shaping LLM system behaviour to align to Mistral’s values.
We’re looking for people who have experience with Literature and Translation, who are experts in model evaluation, prompt engineering, and policy writing. Your role would include helping our Science team create AI systems that respond with good generations across diverse scenarios.
Join us if you are passionate about tackling cutting-edge, open-ended research challenges and transforming your findings into best-in-class model responses.
What you will do
Interact with models to identify where model behavior can be improved
Gather internal and external feedback on model behavior to scope areas for improvement
Design and implement subtle prompting strategies, data generation, and evaluation pipelines that steer better model responses
Identify and fix edge case behaviors through rigorous testing of your data generation pipelines
Develop evaluations of language model behaviors
Work collaboratively with AI Scientists on related teams like Alignment, RAG, Multimodal, Audio, and Code Generation
About you
You have a deep understanding of the humanities; whether it’s literature, language, anthropology, philosophy, political science, film, etc.
You have prior knowledge in training and optimising model behaviour and building evaluations
You are keen to dive into prompt generations and unit tests
You thrive in dynamic and technically complex environments
You have a track record of delivering innovative, out-of-the-box solutions to address real-world constraints
Please mention you found this job on AI Jobs. It helps us get more startups to hire on our site. Thanks and good luck!
Get hired quicker
Be the first to apply. Receive an email whenever similar jobs are posted.
Ace your job interview
Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.