Design and build the AI inference infrastructure for Cloudflare, optimizing systems for high availability and performance while mentoring junior engineers.
Available Locations: Austin, TX or London, UK (Hybrid) About the role
You'll design and build the core infrastructure that powers AI inference across Cloudflare's global network - real-time voice, frontier open LLMs, and customer-deployed models running on a heterogeneous fleet of GPUs and next-generation accelerators in hundreds of cities worldwide. Working alongside AI/ML engineers, hardware partners, and Cloudflare product teams, you'll solve hard problems in distributed systems and high-performance computing: sub-second model cold starts, multi-accelerator workload scheduling, efficient KV cache management, and a model deployment platform serving both Cloudflare and customers bringing their own models. We're building an AI inference platform embedded in the fabric of the internet - something that doesn't exist yet - and this role puts you at the center of it. We're looking for high-agency systems engineers who are energized by foundational infrastructure problems and want to define how AI runs at the edge of the network.
Role Responsibilities
Must-Have Skills
Nice-to-Have Skills
You'll design and build the core infrastructure that powers AI inference across Cloudflare's global network - real-time voice, frontier open LLMs, and customer-deployed models running on a heterogeneous fleet of GPUs and next-generation accelerators in hundreds of cities worldwide. Working alongside AI/ML engineers, hardware partners, and Cloudflare product teams, you'll solve hard problems in distributed systems and high-performance computing: sub-second model cold starts, multi-accelerator workload scheduling, efficient KV cache management, and a model deployment platform serving both Cloudflare and customers bringing their own models. We're building an AI inference platform embedded in the fabric of the internet - something that doesn't exist yet - and this role puts you at the center of it. We're looking for high-agency systems engineers who are energized by foundational infrastructure problems and want to define how AI runs at the edge of the network.
Role Responsibilities
- Develop and maintain core components of the serverless inference platform to ensure high availability and scalability for Cloudflare users.
- Optimize the model scheduling system to significantly increase efficiency and resource utilization across our inference infrastructure.
- Implement improvements to the inference request routing logic to enhance overall performance and reduce latency for end-users.
- Drive significant, measurable improvements in the platform's reliability and resilience by identifying and mitigating systemic risks.
- Expand and refine the observability stack, including metrics, logging, and tracing, and fine-tune alerts to proactively identify and resolve production issues.
- Lead complex, cross-functional technical projects from initial concept and design through final deployment and operationalization.
- Act as a mentor to junior engineers and actively contribute to cultivating a strong, collaborative engineering culture within the team.
Must-Have Skills
- Experience in systems engineering, with a focus on distributed, high-performance systems.
- Expert proficiency in Rust programming, particularly in an asynchronous environment.
- Deep understanding and hands-on experience with relevant networking and application protocols (e.g., TCP, HTTP, WebSocket).
- Experience with scaling and performance optimization techniques, including load balancing and caching in a distributed environment.
Nice-to-Have Skills
- Demonstrable experience with container orchestration platforms, specifically Kubernetes and/or Nomad.
- Familiarity with the challenges and architectures involved in large-scale inference serving (e.g., LLM and diffusion models).
Top Skills
HTTP
Kubernetes
Nomad
Rust
Tcp
Websocket
Cloudflare San Francisco, California, USA Office
101 Townsend St, San Francisco, CA, United States, 94107
Similar Jobs at Cloudflare
Cloud • Information Technology • Security • Software • Cybersecurity
Lead technical projects for the Workers AI team focused on deploying AI inference, building innovative features, and enhancing developer experience.
Top Skills:
PythonPyTorchTensorFlow
Cloud • Information Technology • Security • Software • Cybersecurity
The Talent Analytics Intern will conduct research on hiring processes and employee retention while creating frameworks to enhance future analytics. Responsibilities include data analysis and collaboration with stakeholders for actionable insights.
Top Skills:
BigQueryPythonRSQL
Cloud • Information Technology • Security • Software • Cybersecurity
The Customer Advocacy Intern will support customer marketing programs, engage with customers, assist teams, and contribute to content creation and analysis.
Top Skills:
AICustomer Marketing Tools
What you need to know about the San Francisco Tech Scene
San Francisco and the surrounding Bay Area attracts more startup funding than any other region in the world. Home to Stanford University and UC Berkeley, leading VC firms and several of the world’s most valuable companies, the Bay Area is the place to go for anyone looking to make it big in the tech industry. That said, San Francisco has a lot to offer beyond technology thanks to a thriving art and music scene, excellent food and a short drive to several of the country’s most beautiful recreational areas.
Key Facts About San Francisco Tech
- Number of Tech Workers: 365,500; 13.9% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Google, Apple, Salesforce, Meta
- Key Industries: Artificial intelligence, cloud computing, fintech, consumer technology, software
- Funding Landscape: $50.5 billion in venture capital funding in 2024 (Pitchbook)
- Notable Investors: Sequoia Capital, Andreessen Horowitz, Bessemer Venture Partners, Greylock Partners, Khosla Ventures, Kleiner Perkins
- Research Centers and Universities: Stanford University; University of California, Berkeley; University of San Francisco; Santa Clara University; Ames Research Center; Center for AI Safety; California Institute for Regenerative Medicine

