Eli Lilly and Company Logo

Eli Lilly and Company

Engineer - MLOps & Scientific Platforms - Data Foundry

Posted 3 Days Ago
In-Office
San Francisco, CA, USA
66K-165K Annually
Mid level
In-Office
San Francisco, CA, USA
66K-165K Annually
Mid level
Engineer MLOps & Scientific Platforms role at Lilly involves building ML deployment pipelines, API layers, and ensuring scientific tools are reliable and scalable.
The summary above was generated by AI

At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our employees around the world work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the world.

Locations: San Diego, CA; San Francisco, CA; Boston, MA; Louisville, CO; Indianapolis, IN
 

Lilly Small Molecule Discovery is purpose-built to create molecules that make life better for people. Discovery Technology and Platforms (DTP) accelerates molecule discovery by building optimized foundational platforms, streamlining lab operations through advanced technologies and data connectivity, and investing in novel capabilities.

Data Foundry is a multidisciplinary team within DTP that enables AI-native drug discovery through four integrated pillars: Architecture4Insight (data infrastructure and scientific software), Methods4Insight (analytical and computational methods), Automation & Scale4Insight (lab automation and agentic workflows), and Preparedness4Insight (data governance and readiness). These pillars empower every Lilly scientist to make optimal decisions by providing seamless access to data, insights, and AI-driven capabilities—serving both human scientists and autonomous AI agents.

Position Summary

We are seeking an Engineer - MLOps & Scientific Platforms - Data Foundry to operationalize Data Foundry’s scientific tools and analytical methods into actionable-prototypes. You will build the ML deployment pipelines, model serving infrastructure, API layers, and observability guardrails that make our scientific discovery methods and tools reliable, scalable, and consumable, both by discovery scientists and by the Frontier AI group’s autonomous agents.

This role sits at the interface between Methods4Insight (which develops analytical methods) and Architecture4Insight(which provides the agile data infrastructure). Your job is to ensure every scientific tool Data Foundry produces are analytics-ready, well-monitored, and exposed through APIs with the response-time guarantees and error handling that both human users and AI agents require.

Responsibilities

MLOps & Model Lifecycle Management
  • Build and maintain end-to-end ML deployment pipelines: experiment tracking, model versioning (MLflow, Weights & Biases), containerized model serving, and automated retraining triggers.

  • Develop model registry infrastructure and feature engineering pipelines that enable computational scientists to access models.

  • Implement monitoring and alerting for data pipelines, APIs, ML models, and agentic systems (LLMOps) to ensure system reliability and performance at scale.

  • Build dashboards and metrics tracking for pipeline execution, API latency, token usage, model prediction quality, and system health

  • Establish structured logging and tracing infrastructure for debugging and performance optimization across scientific data systems

Scientific Tool Agile Deployment
  • Deploy predictive and analytical methods from Methods4Insight (e.g. cheminformatics, structural biology, bioinformatics, reaction informatics) with versioning, structured error handling, and response-time guarantees that enable insight generation in agile manner. Productionize when and where needed in partnerships with Tech@Lilly.

  • Build serving infrastructure supporting both synchronous (interactive scientist queries) and asynchronous (batch and agent-invoked) workloads in partnership with Tech@Lilly and Frontier AI.

  • Define and implement API contracts, documentation standards, and testing frameworks that ensure scientific tools are analysis ready, robust and consumable by external teams including Frontier AI.

Platform Engineering & Integration
  • Build and operate cloud-native model serving infrastructure (AWS, Azure, or GCP) using containers, Kubernetes, and infrastructure-as-code.

  • Develop CI/CD pipelines for ML models: automated validation, A/B testing, canary deployments, and rollback procedures.

  • Integrate model serving with Data Foundry’s data pipelines, ensuring models have access to properly formatted, versioned training and inference data.

Frontier AI Interface & Collaboration
  • Partner with the Frontier AI team and Tech@Lilly to ensure Data Foundry’s scientific tools are exposed via well-defined interfaces (REST APIs, MCP-compatible endpoints) that agents can invoke programmatically.

  • Collaborate on API performance requirements: latency targets, throughput guarantees, and graceful degradation under load.

  • Work with Methods4Insight scientists to ensure deployed models include appropriate uncertainty quantification and confidence metrics.

Basic Requirements

  • B.S. or M.S. in Computer Science, Data Science, Machine Learning, Bioinformatics, Computational Biology, or related field.

  • 3+ years of experience in MLOps, ML engineering, or scientific platform development

  • Qualified applicants must be authorized to work in the United States on a full-time basis. Lilly will not provide support for or sponsor work authorization or visas for this role, including but not limited to F-1 CPT, F-1 OPT, F-1 STEM OPT, J-1, H-1B, TN, O-1, E-3, H-1B1, or L-1.

Preferred Qualifications

  • Pharmaceutical or biotech research industry experience.

  • Strong Python skills; experience with ML frameworks (PyTorch, TensorFlow, scikit-learn) and ML lifecycle tools (MLflow, W&B, Kubeflow, or similar).

  • Proven track record building and deploying production model serving infrastructure — containerized endpoints, RESTful/gRPC APIs, and operational monitoring

  • Working knowledge of cloud platforms (AWS, Azure, or GCP), Kubernetes, and CI/CD automation.

  • Strong communication skills with ability to collaborate across computational scientists, software engineers, and partner teams.

  • Experience operationalizing scientific or computational models (cheminformatics, bioinformatics, structural biology, QSAR, molecular simulations, PK/PD, systems biology, or ODE-based models).

  • Hands-on experience with model monitoring, drift detection, and automated retraining systems.

  • Familiarity with API gateway patterns, event-driven architectures, and service mesh technologies.

  • Experience with feature stores, data versioning (DVC), or experiment tracking at scale.

  • Exposure to AI agent frameworks (MCP, LangChain) or building APIs that AI systems invoke programmatically.

  • Experience with C, C++, CUDA, or GPU-accelerated computing for optimizing model training/inference performance; familiarity with containerizing HPC workloads (Singularity/Apptainer).

Lilly is dedicated to helping individuals with disabilities to actively engage in the workforce, ensuring equal opportunities when vying for positions. If you require accommodation to submit a resume for a position at Lilly, please complete the accommodation request form (https://careers.lilly.com/us/en/workplace-accommodation) for further assistance. Please note this is for individuals to request an accommodation as part of the application process and any other correspondence will not receive a response.

Lilly is proud to be an EEO Employer and does not discriminate on the basis of age, race, color, religion, gender identity, sex, gender expression, sexual orientation, genetic information, ancestry, national origin, protected veteran status, disability, or any other legally protected status.


Our employee resource groups (ERGs) offer strong support networks for their members and are open to all employees. Our current groups include: Africa, Middle East, Central Asia Network, Black Employees at Lilly, Chinese Culture Network, Japanese International Leadership Network (JILN), Lilly India Network, Organization of Latinx at Lilly (OLA), PRIDE (LGBTQ+ Allies), Veterans Leadership Network (VLN), Women’s Initiative for Leading at Lilly (WILL), enAble (for people with disabilities). Learn more about all of our groups.

Actual compensation will depend on a candidate’s education, experience, skills, and geographic location.  The anticipated wage for this position is

$66,000 - $165,000

Full-time equivalent employees also will be eligible for a company bonus (depending, in part, on company and individual performance). In addition, Lilly offers a comprehensive benefit program to eligible employees, including eligibility to participate in a company-sponsored 401(k); pension; vacation benefits; eligibility for medical, dental, vision and prescription drug benefits; flexible benefits (e.g., healthcare and/or dependent day care flexible spending accounts); life insurance and death benefits; certain time off and leave of absence benefits; and well-being benefits (e.g., employee assistance program, fitness benefits, and employee clubs and activities).Lilly reserves the right to amend, modify, or terminate its compensation and benefit programs in its sole discretion and Lilly’s compensation practices and guidelines will apply regarding the details of any promotion or transfer of Lilly employees.

#WeAreLilly

Top Skills

AWS
Azure
Ci/Cd
GCP
Grpc
Kubernetes
Mlflow
Python
PyTorch
Rest Apis
Scikit-Learn
TensorFlow
Weights & Biases

Similar Jobs

42 Minutes Ago
Remote or Hybrid
United States
216K-320K Annually
Expert/Leader
216K-320K Annually
Expert/Leader
Automotive • Big Data • Information Technology • Robotics • Software • Transportation • Manufacturing
As a Principal Software Engineer, you will architect and develop virtualization technology for vehicle testing, mentor teams, and enhance code quality across GM products.
Top Skills: C/C++GoJava
42 Minutes Ago
Remote or Hybrid
United States
124K-165K Annually
Mid level
124K-165K Annually
Mid level
Automotive • Big Data • Information Technology • Robotics • Software • Transportation • Manufacturing
The District Sales Manager manages Buick GMC dealers, focusing on sales growth, customer satisfaction, and effective training, requiring frequent travel.
42 Minutes Ago
Remote or Hybrid
United States
106K-141K Annually
Mid level
106K-141K Annually
Mid level
Automotive • Big Data • Information Technology • Robotics • Software • Transportation • Manufacturing
The HRSD Developer customizes the ServiceNow HRSD platform, develops HR services, gathers requirements, and collaborates with stakeholders to enhance the platform.
Top Skills: Agentic AiFlow DesignerJIRAMoveworksServicenowWorkday Hcm

What you need to know about the San Francisco Tech Scene

San Francisco and the surrounding Bay Area attracts more startup funding than any other region in the world. Home to Stanford University and UC Berkeley, leading VC firms and several of the world’s most valuable companies, the Bay Area is the place to go for anyone looking to make it big in the tech industry. That said, San Francisco has a lot to offer beyond technology thanks to a thriving art and music scene, excellent food and a short drive to several of the country’s most beautiful recreational areas.

Key Facts About San Francisco Tech

  • Number of Tech Workers: 365,500; 13.9% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Google, Apple, Salesforce, Meta
  • Key Industries: Artificial intelligence, cloud computing, fintech, consumer technology, software
  • Funding Landscape: $50.5 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Sequoia Capital, Andreessen Horowitz, Bessemer Venture Partners, Greylock Partners, Khosla Ventures, Kleiner Perkins
  • Research Centers and Universities: Stanford University; University of California, Berkeley; University of San Francisco; Santa Clara University; Ames Research Center; Center for AI Safety; California Institute for Regenerative Medicine

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account