Founding LLM Inference Engineer
Full-time | On-site | San Francisco, CA
Compensation: $200K – $300K + 0.10%–1.00% Equity
About the Role
We’re looking for a Founding LLM Inference Engineer to architect and optimize large-scale inference systems powering cutting-edge AI applications. You’ll be building the backbone of an AI platform used by top enterprises, with a focus on performance, scalability, and reliability.
This is a hands-on, high-impact role where you’ll collaborate closely with research and product teams, moving fast to bring breakthrough model capabilities into production. If you’re excited about low-latency systems, high-throughput pipelines, and deploying bleeding-edge LLMs, this role is for you.
Tech stack: Python, CUDA, LLMs, API integrations, TGI, vLLM, TensorRT-LLM
What You’ll Do
What We’re Looking For
Benefits
👉 Ready to take the next step?
Apply now or email Jenn at Recruiter@CareDynamicsFL.com to learn more.