Senior Engineering Manager - Accelerated Compute Memory Systems

AI overview

Lead the development of cloud-native ingestion, retrieval, and inference layers to support mission-critical AI deployments while navigating startup dynamics.
About Pryon:  We’re a team of AI, technology, and language experts whose DNA lives in Alexa, Siri, Watson, and virtually every human language technology product on the market. Now we’re building an industry-leading knowledge management and Retrieval-Augmented Generation (RAG) platform. Our proprietary, cutting-edge natural language processing capabilities transform unstructured data into meaningful experiences that increase productivity with unmatched accuracy and speed. Pryon is building one of the industry's most ambitious cloud-native AI infrastructure platforms: a petabyte-scale ingestion and inference system powering mission-critical government and enterprise deployments. We need an Engineering Manager who excels at designing distributed systems for large-scale AI memory workloads in modern cloud and on-premise environments. You'll lead the team building our ingestion, retrieval, and inference layers, ensuring scalability, reliability, and compliance while navigating the ambiguity inherent in a fast-growing startup.  You will be a founding member of our Super Compute Memory (SCM) team, reporting to the VP of Engineering. This team's charter is to build the high-performance computing infrastructure that enables Pryon's AI memory layer to scale to petabytes of knowledge while maintaining real-time retrieval performance.  This is a high-visibility role with significant ownership. You'll work closely with the Research, AI/ML Engineering, and Platform teams.  In This Role You Will:
  • Build and lead a team delivering cloud-native ingestion, retrieval, and inference layers that will power mission-critical deployments for commercial and federal entities with millions of public users. 
  • Architect and deliver scalable, fault-tolerant distributed systems capable of handling billions of documents and burst loads of 30K+ concurrent users on managed cloud infrastructure and on-premises deployments. 
  • Guide implementation of multimodal ingestion pipelines (PDF, HTML, DOCX, JSON, XML, PPTX, TIFF) optimized for cloud-scale AI memory workloads. 
  • Oversee design and optimization of LLM-driven data ingestion and retrieval workflows using modern orchestration frameworks. 
  • Own optimization and tuning of high-throughput, low-latency production environments via async orchestration and resource management. 
  • Establish performance benchmarking, compliance frameworks, and automated testing strategies for petabyte-scale systems. 
  • Balance technical leadership with people leadership—guiding architecture decisions at the application and service layer while scaling and mentoring a high-performing team. 
  • Collaborate cross-functionally with Product, Executive Leadership, and Customer Success in a dynamic startup environment. 
  • What You'll Need to Be Successful:
  • 10+ years in software engineering, 5+ years in management roles delivering large-scale AI/ML systems and cloud infrastructure. 
  • Expert-level proficiency in Python, with strong experience in at least one systems language (Go, Rust, C++, or Java). 
  • 5+ years building production-grade distributed systems on cloud platforms (AWS, GCP, or Azure). 
  • Hands-on experience with modern ML orchestration frameworks (Ray, Kubeflow, Airflow, or similar open-source tools). 
  • Production experience with vector databases (Pinecone, Weaviate, Qdrant, Milvus, or similar). 
  • Deep understanding of message queuing and streaming systems (Kafka, Pulsar, RabbitMQ, Kinesis). 
  • Proven track record designing and operating scalable, fault-tolerant distributed architectures in cloud environments. 
  • Direct experience building multimodal ingestion pipelines for knowledge management platforms. 
  • Experience optimizing LLM inference and retrieval workloads at the application/framework level (PyTorch, TensorFlow, vLLM, or similar). 
  • Previous success managing engineering teams delivering production-scale AI infrastructure in startup or high-growth environments. 
  • Technical Depth
  • Deep understanding of cloud-native distributed systems architecture: compute orchestration (Kubernetes/EKS/GKE), storage systems, networking, observability, security, disaster recovery, and cost optimization. 
  • Strong knowledge of AI memory and knowledge management system design patterns, embedding models, retrieval strategies, and LLM integration patterns. 
  • Experience with infrastructure-as-code (Terraform, CloudFormation, Pulumi) and modern DevOps practices. 
  • Familiarity with distributed tracing, metrics, and logging systems (Datadog, Prometheus, Grafana, CloudWatch). 
  • Experience with parallel programming models (e.g., MPI, OpenMP, CUDA) 
  • Custom hardware accelerator design or bare-metal cluster management 
  • On-premises datacenter operations or HPC cluster management tools (e.g., Slurm) 
  • Leadership & Culture
  • Demonstrated ability to mentor engineers and build high-performing teams in ambiguous, fast-paced environments. 
  • Strong communication skills with ability to translate technical decisions for executive and product stakeholders. 
  • Comfort with startup dynamics: rapid iteration, evolving requirements, and wearing multiple hats. 
  • Proven track record of candor and transparency when discussing technical tradeoffs and knowledge boundaries. 
  • Experience balancing technical excellence with pragmatic delivery in resource-constrained environments. 
  • Benefits for Full Time Employees: 
    - Remote first organization
    - 100% Company paid Health/Dental/Vision benefits for you and your dependents
    - Life Insurance, Short-term and Long-term Disability
    - 401k
    - Unlimited PTO

    We are interested in every qualified candidate who is authorized to work in the United States. However, we are not able to sponsor or take over sponsorship of employment visas at this time.

    Pryon will not consider race, religion, sex, sexual preference, or national origin in ways that violate the Nation's civil rights laws.

    Perks & Benefits Extracted with AI

    • Health Insurance: 100% Company paid Health/Dental/Vision benefits for you and your dependents
    • Other Benefit: 401k
    • Paid Time Off: Unlimited PTO
    • Remote-Friendly: Remote first organization
    Salary
    $220,000 – $250,000 per year
    Ace your job interview

    Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.

    Senior Engineering Manager Q&A's
    Report this job
    Apply for this job