There are still lots of open positions. Let's find the one that's right for you.
We are looking for a skilled developer to work on LLM inference, optimization, and deployment at scale. The ideal candidate will have experience in backend development and will be responsible for ensuring the reliability and scalability of backend services that leverage machine learning models. This role involves working with teams to optimize LLM inference performance and resource utilization in production systems, as well as developing high-performance, low-latency code to support machine learning applications and inference tasks.