
Forward Deployed Engineer - LLM Systems
Job Description
The most important scientific discoveries of our time won't happen in a traditional lab. We're an AI and physical sciences company building state-of-the-art models to accelerate breakthroughs across materials, energy, and beyond. Backed by world-class investors and growing rapidly, we operate at the pace the frontier requires. Our team brings deep expertise, genuine ownership, and an insatiable drive to push the boundaries of what's scientifically possible.
About the Role
You will be a key builder behind the world's first on-prem LLM system for atoms, deploying inference and reinforcement learning systems directly into semiconductor fabs where the science happens. The role splits roughly 80% LLM system development and deployment, and 20% semiconductor customer interaction — translating fab requirements into engineering specs and ensuring our systems meet the realities of production.
You will move fluidly between improving LLM systems, managing Kubernetes clusters, and interacting with semiconductor experts — owning deployments end-to-end and serving as the technical face of our system to vendor partners. You will also work closely with LLM systems and modeling experts from OpenAI, Anthorpic, xAI, Google, and other frontier labs.
What You'll Do
Deploy and operate inference and reinforcement learning systems on-site at semiconductor partner facilities, from bring-up through ongoing operation
Build and maintain the on-prem LLM platform powering our atomic-scale science workflows, including orchestration, scheduling, observability, and reliability
Develop and extend open-source LLM frameworks (SGLang, vLLM, Megatron, Slime) to meet the performance and integration needs of on-prem deployments
Build custom Kubernetes operators and Slurm integrations to run ML workloads in heterogeneous on-prem environments
Own metrics, dashboards, and alerting in Prometheus, Grafana, and PagerDuty
Partner with semiconductor process engineers to translate fab requirements into engineering specs, with a focus on New Product Introduction (NPI) flows
You will thrive in this role if you have experience in:
Deploying inference and/or RL systems in production, including new-cluster bring-up and integration with existing infrastructure
Kubernetes and/or Slurm — for example, building a custom Kubernetes operator for ML systems, or running large-scale workloads on Slurm
Prometheus, Grafana, and PagerDuty, with a strong grasp of how to set up dashboards and reason about system performance
Hands-on framework-level work with SGLang, vLLM, Megatron, Slime, or other open-source inference and RL engines
Systems engineering fundamentals: Linux, networking, distributed systems, GPU computing, and performance debugging
Direct semiconductor process experience across wafer processing modules (deposition, etch, litho, packaging, metrology), FEOL/MEOL/BEOL integration, NPI spec definition, device performance/yield/reliability, and DOE/SPC/APC, with the ability to engage directly with customers
Especially Strong Candidates May Also Have
Deployed ML or LLM systems in air-gapped, on-prem, or otherwise constrained environments
Contributed upstream to open-source inference or training frameworks
Shipped real systems in both software systems and semiconductor process engineering
Worked on RL infrastructure at scale, including rollout systems, training/inference co-location, and reward modeling pipelines
Prior forward-deployed or field engineering experience at a research lab, AI company, or deep tech startup
Mechanics:
Minimum education: bachelor's degree or an equivalent combination of education and training or experience
Location: Our lab is located in Menlo Park and we prefer folks to be located in Menlo Park or San Francisco but can be flexible based on role
Compensation: The annual compensation range for this role - $350,000-$400,000
Visa sponsorship: Yes, we sponsor visas and will do everything we can to assist in this process with our legal support.
We're building a team of the world's best — the scientists, engineers, and problem-solvers who don't just follow the frontier, they define it. If you're driven to bring AI to life in the physical world and make discoveries that have never been made before, you belong here.
Optimize Your Resume for This Job
Get a match score and see exactly which keywords you're missing
Job Details
- Department
- Manufacturing
- Category
- Software
- Employment Type
- Full Time
- Location
- Menlo Park
- Posted
- Apr 22, 2026, 09:27 PM
- Listed
- Apr 22, 2026, 09:40 PM
- Last updated
- Apr 27, 2026, 06:44 PM
- Compensation
- $350,000 - $400,000 per year
About Periodic Labs
Part of the growing frontier tech ecosystem pushing the edges of what's possible.
More Roles at Periodic Labs





Similar Software Roles



Found this role interesting?