Research Scientist

AI overview

Lead impactful research enhancing the safety and reliability of frontier AI systems, designing experiments, and collaborating with top researchers in the field.
The Center for AI Safety (CAIS) is a leading research and advocacy organization focused on mitigating societal-scale risks from AI. We address AI’s toughest challenges through technical research, field-building initiatives, and policy engagement, along with our sister organization, Center for AI Safety Action Fund. As a Research Scientist here, you will lead and execute high-impact research that advances the safety and reliability of frontier AI systems. You'll design and run experiments on large language models, build the tooling needed to train and evaluate models at scale, and turn results into publishable research. You'll collaborate closely with CAIS researchers and external academic and commercial partners, using our compute cluster to run large-scale training and evaluation. The work spans areas like AI honesty, robustness, transparency, and trojan/backdoor behaviors, aimed at reducing real-world risks from advanced AI systems. Key Responsibilities Include:
  • Help set and lead research agenda.
  • Own end-to-end research experiments.
  • Train and fine-tune large transformer models across domains.
  • Build and maintain datasets and benchmarks.
  • Run distributed training and evaluation at scale.
  • Write and ship research, collaborating with co-authors, and supporting submissions of papers to top conferences.
  • Collaborate with researchers and external partners while contributing to shared research direction and responding quickly in research cycles.
  • Mentor and guide others on the team.
  • You might be a good fit if you:
  • Ph.D. in computer science, machine learning, or a related field, with 5+ years of related research experience.
  • Familiar with relevant frameworks and libraries (e.g., pytorch and huggingface).
  • Have experience launching and training distributed ML jobs.
  • Communicate clearly and promptly with teammates.
  • Have co-authored an NLP or RL paper in a top conference.
  • Know someone who could be a great fit for this role? Submit their details through our Referral Form. If we end up hiring your referral, you’ll receive a $1,500 bonus once they’ve been with CAIS for 90 days.

    The Center for AI Safety is an Equal Opportunity Employer. We consider all qualified applicants without regard to race, color, religion, sex, sexual orientation, gender identity or expression, national origin, ancestry, age, disability, medical condition, marital status, military or veteran status, or any other protected status in accordance with applicable federal, state, and local laws. In alignment with the San Francisco Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records for employment.​

    If you require a reasonable accommodation during the application or interview process, please contact [email protected].​

    We value diversity and encourage individuals from all backgrounds to apply.

    Perks & Benefits Extracted with AI

    • Free Meals & Snacks: Lunch and dinner at the office
    • Health Insurance: Health insurance for you and your dependents
    • Paid Time Off: Unlimited PTO

    The Center for AI Safety (CAIS) is dedicated to tackling the complex risks posed by artificial intelligence at a societal level. We combine rigorous research with proactive policy initiatives to confront AI's most pressing challenges, making us a crucial player in the conversation around safe AI development and implementation.

    View all jobs
    Salary
    $170,000 – $220,000 per year
    Ace your job interview

    Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

    Research Scientist Q&A's
    Report this job
    Apply for this job