anyscale
anyscale2d ago
New
$170,112 – $247,000 • Offers Equity/yr

Distributed LLM Inference Engineer

San Francisco, Palo Altofull-timemid
OtherEngineer
0 views0 saves0 applied

Quick Summary

Overview

About Anyscale At Anyscale, we're on a mission to democratize distributed computing and make it accessible to software developers of all skill levels. We’re commercializing Ray, a popular open-source project that's creating an ecosystem of libraries for scalable machine learning.

Technical Tools
openaipytorchtensorflowdeep-learningdistributed-systemsmachine-learning

At Anyscale, we're on a mission to democratize distributed computing and make it accessible to software developers of all skill levels. We’re commercializing Ray, a popular open-source project that's creating an ecosystem of libraries for scalable machine learning. Companies like OpenAIUberSpotifyInstacartCruise, and many more, have Ray in their tech stacks to accelerate the progress of AI applications out into the real world.

About the Role

~1 min read

As a Distributed LLM Inference Engineer, you will help systems and optimizations that push the boundaries of performance for inference at large scale. This is an incredibly critical role to Anyscale as it allows us to achieve a market leading position for  AI infrastructure.

  • Iterate very quickly with product teams to ship the end to end solutions for Batch and Online inference at high scale which will be used by open-source Ray users and customers of Anyscale

  • Work across the stack integrating Ray Data and LLM engine providing optimizations achieving low cost solutions for large scale ML inference 

  • Integrate with Open source software like vLLM, work closely with the community to adopt these techniques in Anyscale solutions, and also contribute improvements to open source

  • Follow the latest state-of-the-art in the open source and the research community, implementing and extending best practices

  • Familiarity with running ML inference at large scale with high throughput and low latency

  • Familiarity with deep learning and deep learning frameworks (e.g. PyTorch)

  • Solid understanding of distributed systems, ML inference challenges

Nice to Have

~1 min read
  • ML Systems knowledge

  • Experience using Ray 

  • Work closely with community on LLM engines like vLLM, TensorRT-LLM

  • Contributions to deep learning frameworks (PyTorch, TensorFlow)

  • Contributions to deep learning compilers (Triton, TVM, MLIR)

  • Prior experience working on GPUs / CUDA

What We Offer

~1 min read
Stock Options
Healthcare plans, with premiums covered by Anyscale at 99% for both employees and dependents
401k Retirement Plan
Education & Wellbeing Stipend
Paid Parental Leave
Fertility Benefits
Paid Time Off
Commute reimbursement
100% of in-office meals covered

Location & Eligibility

Where is the job
Location terms not specified
Who can apply
Same as job location

Listing Details

Posted
May 4, 2026
First seen
May 5, 2026
Last seen
May 6, 2026

Posting Health

Days active
0
Repost count
0
Trust Level
66%
Scored at
May 6, 2026

Signal breakdown

freshnesssource trustcontent trustemployer trust
Newsletter

Stay ahead of the market

Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.

A
B
C
D
Join 12,000+ marketers

No spam. Unsubscribe at any time.

anyscaleDistributed LLM Inference Engineer$170,112 – $247,000 • Offers Equity