Staff/Senior Applied Scientist, AI/GenAI & ML Systems

Wilmington , United States

AI overview

Lead the design and deployment of production-grade GenAI and ML systems, optimizing for agentic multi-agent architectures in a collaborative cloud environment.

Company overview:

TraceLink’s software solutions and Opus Platform help the pharmaceutical industry digitize their supply chain and enable greater compliance, visibility, and decision making. It reduces disruption to the supply of medicines to patients who need them, anywhere in the world.

 

Founded in 2009 with the simple mission of protecting patients, today Tracelink has 8 offices, over 800 employees and more than 1300 customers in over 60 countries around the world. Our expanding product suite continues to protect patients and now also enhances multi-enterprise collaboration through innovative new applications such as MINT.

 

Tracelink is recognized as an industry leader by Gartner and IDC, and for having a great company culture by Comparably.

Staff / Senior Applied Scientist, GenAI & ML Systems

Location: Boston, MA (US)

About the Role

We are hiring a Staff / Senior Applied Scientist to lead the design and deployment of production-grade GenAI and ML systems with a strong emphasis on being hands-on. You will personally build, iterate, and ship systems focused on LLM/SLM optimization for agentic, multi-agent architectures in cloud environments.

This role is ideal for someone with deep expertise in one or more areas of LLM/SLM optimization for agent-based systems, and hands-on experience in designing, implementing, and operating large-scale multi-agent systems in the cloud.

Key Responsibilities

  • Hands-on ownership of building and shipping multi-agent systems (planner/executor, tool-using agents, supervisor patterns, routing, role-based agents) from prototype to production.
  • Write production-quality code for agent orchestration, tool integration, memory/state design, and context management.
  • Lead context engineering strategies for multi-agent coordination: prompt design, state persistence, agent handoffs, grounding, constraints, and safety controls.
  • Hands-on fine-tune and deploy SLM models for production usage: dataset creation, training workflows, evaluation, and inference serving.
  • Build Advanced RAG pipelines end-to-end, including semantic search, embeddings, hybrid retrieval, and cross-encoder reranking.
  • Implement evaluation frameworks for multi-agent systems covering quality, latency, cost, robustness, and failure mode detection.
  • Collaborate with platform and product engineering to ensure solutions are cloud-native, secure, observable, and scalable (monitoring, logging, CI/CD).
  • Optimize for cost and latency via model routing, caching, compression strategies, and inference efficiency improvements.
  • Mentor peers through code reviews, architecture sessions, and hands-on technical leadership.

Required Knowledge & Experience

  • Context engineering for complex multi-agent systems
    (prompt orchestration, tool calling, memory/state design, routing, constraint handling)
  • Fine-tuning of SLMs and delivering them to production
    (training strategies, validation, deployment, monitoring, rollback readiness)
  • Experience with Advanced RAG, semantic search, embeddings, and cross-encoders
    (retrieval tuning, chunking strategies, query rewriting/planning, reranking)
  • Ability to translate ambiguous requirements into concrete architectures, metrics, and deliverables
  • Hands-on inference optimization experience: quantization, distillation, batching, caching, model routing, speculative decoding
  • Experience building retrieval systems at scale using vector DBs and search stacks
  • Comfort working across the full lifecycle: research → prototype → A/B test → production hardening

 

Preferred Qualifications

  • Familiarity with enterprise constraints: privacy, security, data governance, permissions, auditability
  • Experience designing and running GenAI observability: traces, prompt/versioning, tool call logging, feedback loops
  • Strong ability to implement production-quality systems in Python (and/or adjacent backend languages)
  • Proven experience deploying GenAI/ML systems in cloud environments (AWS/Azure/GCP)
  • Experience with scalable inference and service operations: containers, APIs, observability, reliability practices
  • MS/PhD in CS/ML/NLP/Stats (or equivalent applied experience building production systems)



TraceLink is committed to providing competitive compensation and benefits to all employees. This is the estimated base salary range for this role and should serve only as a guide. Final compensation offered may vary based on a variety of factors including but not limited to experience level, fit for the role, skills, domain knowledge, internal equity, budget, and location.

US Pay Range
$151,999.74$189,263.73 USD

Please see the Tracelink Privacy Policy for more information on how Tracelink processes your personal information during the recruitment process and, if applicable based on your location, how you can exercise your privacy rights. If you have questions about this privacy notice or need to contact us in connection with your personal data, including any requests to exercise your legal rights referred to at the end of this notice, please contact [email protected].  

 

Join our passionate and dynamic team and be part of a company that is reshaping the supply chain landscape. Explore our current job openings and discover how you can contribute to our mission of ensuring a safer, more connected future.

View all jobs
Salary
$151,999 – $189,263 per year
Ace your job interview

Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

Scientist Q&A's
Report this job
Apply for this job