Persimmons, Inc.
Technical Lead, Runtime Software/Hardware (Spatial AI Accelerator)
Who we are:
Persimmons is building the infrastructure that will power the next decade of AI. Founded in 2023 by veteran technologists from the worlds of semiconductors, AI systems, and software innovation, We’re on a mission to enable smarter devices, more sustainable data centers, and entirely new applications the world hasn’t imagined yet.
Why join us:
We’re growing fast and looking for bold thinkers, builders, and curious problem-solvers who want to push the limits of AI hardware and software. If you're ready to join a world-class team and play a critical role in making a global impact - we want to talk to you.
Summary of Role:
Persimmons.ai seeks a multidisciplinary Technical Lead for runtime software/hardware and compiler integration, focused on our next-generation custom spatial AI accelerator. You will architect and guide the runtime system bridging compiler, host, driver, device firmware, and control hardware: enabling high-performance, robust, and scalable execution of modern AI workloads.
This is a hands-on and technical leadership role spanning system design, cross-stack engineering, technical mentorship, and collaboration with compiler, ML framework, and hardware teams.
What you’ll do:
- Architect, design, and implement the runtime stack for Persimmons' custom spatial accelerator, covering host drivers, device runtime, and hardware/firmware control loops.
- Lead technical direction and decisions for runtime–hardware interface, device work and command queue infrastructure, and memory management.
- Coordinate with compiler/backend, ML systems, and hardware architects to ensure seamless end-to-end ML model execution.
- Define and co-design hardware support features essential to runtime: queueing structures, synchronization primitives, interrupt/event signaling, dispatching and orchestrating ML workloads on spatial execution fabric.
- Drive performance analysis, development tools for tracing, bottleneck identification, and runtime-level optimizations for latency, throughput, and hardware utilization.
- Build and mentor a cross-disciplinary engineering team focused on runtime and system validation—establishing best practices, technical standards, and robust software-hardware collaboration.
- Champion efficient tooling, simulation/emulation environments, and test infrastructure for system validation and robust runtime dev/debug.
Requirements
*We do not expect candidates to meet all of the requirements listed below; strong candidates will demonstrate expertise in several key areas.*
- Deep experience architecting runtime software, device firmware, hardware interfaces, or control systems for AI accelerators and/or high-performance SoCs.
- Hands-on expertise developing drivers, resource managers, command/queue control, and dispatching and synchronization primitives (queues, barriers, event notifications) for custom hardware.
- Strong understanding of C/C++ multi-threaded programming and concurrent system design, including experience developing and debugging software that leverages threads, synchronization primitives, and parallel runtime constructs to maximize hardware utilization and performance in latency- and throughput-sensitive environments.
- Solid understanding of hardware–software co-design principles: memory hierarchies, DMA engines, interconnects, job scheduling, on-device synchronization.
- Experience integrating kernel libraries into device runtime stacks—connecting optimized compute kernels (such as SIMD operations and common AI operator libraries) to runtime software through seamless invocation and well-defined APIs, efficient scheduling and memory/resource management.
- Experience with modern large language model (LLM) inference servers and serving stacks (e.g., vLLM, TensorRT-LLM, Triton Inference Server, Hugging Face Text Generation Inference, Ray Serve), including their architecture, runtime scheduling, memory management, batching, streaming, and distributed deployment. Understanding of how runtime design, kernel integration, and hardware acceleration impact performance, scalability, and latency in LLM serving workloads.
- Experience with system-level performance tuning, debugging complex hardware–software interactions, and building scalable test/validation infrastructure.
- High level of understanding and 5+ years of experience with in C/C++; familiarity with hardware description languages (Verilog/VHDL/SystemVerilog), or firmware development is a strong plus.
- Drive for innovation—keeping up with new architectures, techniques, and runtime models in ML or spatial computing.
Benefits
- Competitive salary and benefits package.
- Flexible PTO
- 401k
Please note: Our organization does not accept unsolicited candidate submissions from external recruiters or agencies. Any such submissions, regardless of form (including but not limited to email, direct messaging, or social media), shall be deemed voluntary and shall not create any express or implied obligation on the part of the organization to pay any fees, commissions, or other compensation. Direct contact of employees, officers, or board members regarding employment opportunities is strictly prohibited and will not receive a response.
Top Skills
Persimmons, Inc. San Jose, California, USA Office
San Jose, California, United States, 95054
Similar Jobs
What you need to know about the San Francisco Tech Scene
Key Facts About San Francisco Tech
- Number of Tech Workers: 365,500; 13.9% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Google, Apple, Salesforce, Meta
- Key Industries: Artificial intelligence, cloud computing, fintech, consumer technology, software
- Funding Landscape: $50.5 billion in venture capital funding in 2024 (Pitchbook)
- Notable Investors: Sequoia Capital, Andreessen Horowitz, Bessemer Venture Partners, Greylock Partners, Khosla Ventures, Kleiner Perkins
- Research Centers and Universities: Stanford University; University of California, Berkeley; University of San Francisco; Santa Clara University; Ames Research Center; Center for AI Safety; California Institute for Regenerative Medicine


.png)