Vertiv - Westerville, OH
posted 4 months ago
As an LLMOps Engineer - Cloud/Gen AI, you will play a crucial role in building and maintaining the infrastructure and pipelines for cutting-edge Large Language Models (LLMs). This position requires close collaboration with Generative AI Architects to ensure the efficiency, scalability, and reliability of Generative AI models in production. Your expertise in automating and streamlining the LLM lifecycle will be instrumental in achieving these goals. The role is based onsite at Vertiv's Westerville, OH - HQ location, where you will be expected to conceptualize, develop, and execute Machine Learning (ML)/LLM pipelines specifically tailored for Large Language Models. This includes tasks such as data acquisition, pre-processing, model training and tuning, deployment, and monitoring. In this position, you will utilize automation tools such as GitOps, CI/CD pipelines, and containerization technologies like Docker and Kubernetes to streamline ML/LLM tasks across the Large Language Model lifecycle. Establishing robust monitoring and alerting systems will be essential to track Large Language Model performance, data drift, and other key metrics, allowing you to proactively identify and resolve issues. You will also perform truth analysis to assess the accuracy and effectiveness of Large Language Model outputs, comparing them to known, accurate data. Collaboration is key in this role, as you will work closely with infrastructure and DevOps teams, as well as Generative AI Architects, to optimize model performance and resource utilization. Additionally, you will oversee and maintain cloud infrastructure (e.g., AWS, Azure) specifically for Large Language Model workloads, ensuring cost-efficiency and scalability. Staying current with the latest advancements in ML/LLM Ops will be crucial, as you will integrate these developments into generative AI platforms and processes. Effective communication with both technical and non-technical stakeholders will be necessary to provide updates on the performance and status of Large Language Models.