
Senior Applied Research Engineer
Job Description
Zep is the memory and context layer for AI agents. As a Senior Applied Research Engineer, you'll explore novel approaches to memory, context, and context generation, then own those ideas all the way to production.
This is a research role with a hard applied bent. We're not hiring ML researchers chasing publications. We're hiring engineers who can run rigorous experiments, train and evaluate models, and ship the result as production code our customers depend on.
What you'll do
- Explore novel approaches to memory, context, and context generation. Define the problem, run the experiments, ship the result.
- Own research to production end-to-end: dataset creation and curation, experiment design, evaluation, training and finetuning, and production deployment.
- Train, finetune, and evaluate models on Zep's domain. Build the eval harnesses that catch regressions before they ship.
- Work with our model serving stack to operate inference at low latency and reasonable cost on AWS.
What we're looking for
- 6+ years of production engineering with a strong backend systems background. You've shipped services with real throughput and latency requirements.
- Master's in Computer Science or equivalent.
- Strong research skills: methodology, dataset creation and curation, experiment design, and evaluation. You can frame an open problem and design experiments that actually answer the question.
- Hands-on experience with model finetuning. Working familiarity with transformer architectures, training and finetuning workflows, and evaluation. PyTorch and OpenAI Triton for experimentation.
- Working experience with model serving technologies: vLLM, SGLang, or Triton Inference Server. You've operated inference in production.
- Python, plus high proficiency in one of Rust, C++, or Go. You can work in critical-path code and on performance. Python-only is not enough.
- Hands-on AWS experience in production: deployments, monitoring, scaling, cost and reliability tradeoffs.
Nice to have
- Published or open-source work in retrieval, memory systems, or LLM evaluation.
Tech stack: Python, Rust/C++/Go, PyTorch, vLLM/SGLang, AWS.
This role is probably NOT a fit if:
- You're an ML researcher or model trainer who hasn't shipped research to production.
- Your background is primarily Python application work without lower-level systems experience.
- You haven't operated production backend systems with real latency or throughput requirements.
Interview Process
We respect your time and keep our interview process tight and focussed.
Screening Call (w/ Daniel, our Founder) → Team Calls (2-3 hours back-to-back, may include a presentation) → Decision Call (Daniel, again)
Optimize Your Resume for This Job
Get a match score and see exactly which keywords you're missing
Job Details
- Category
- Aerospace Engineering
- Employment Type
- Full Time
- Location
- San Francisco, United States / Remote (US) (Remote Available)
- Posted
- May 7, 2026, 07:40 PM
- Listed
- May 7, 2026, 07:40 PM
- Compensation
- $180,000 - $250,000 per year
About Zep AI
Part of the growing frontier tech ecosystem pushing the edges of what's possible.
More Roles at Zep AI
Similar Aerospace Engineering Roles



Found this role interesting?
Career Guides
Inside guide to Airbus Defence & Space careers: Ariane 6, Eurostar, Orion ESM programs, salary ranges across Europe, Graduate Programme, and hiring process.
Inside guide to Thales Alenia Space careers: Galileo, MetOp, SpaceRider programs, salary ranges, locations across France and Italy, hiring process, and work culture.
Practical guide to writing resumes for aerospace and space jobs: ATS optimization, keywords by role, translating experience from other industries, clearance listing, and cover letters.