AMD - Bellevue, WA

posted 6 months ago

Full-time - Director
Bellevue, WA
Computer and Electronic Product Manufacturing

About the position

At AMD, we are committed to transforming lives through our technology, and we are looking for a Principal Research Scientist to join our team. This role is pivotal in exploring and advancing the state-of-the-art in large language models (LLMs) and large multimodal models (LMMs). As part of a world-class research team, you will engage in pre-training, fine-tuning, and aligning these models while staying abreast of the latest trends and advancements in the field. Your work will not only contribute to the development of cutting-edge products but also influence the future direction of AMD's AI platform. In this position, you will be responsible for designing and implementing innovative research ideas aimed at improving the quality and performance of LLMs and LMMs. You will work on accelerating the training and inference speeds of these models, ensuring that AMD remains at the forefront of technology. The ideal candidate will possess a deep understanding of hyper-parameter tuning, data preprocessing, and tokenization methods, as well as the latest training approaches for LLMs and LMMs. A strong background in transformer architectures is essential, as is a proven track record of publishing research in top-tier venues. We value collaboration, humility, and inclusivity, and we believe that diverse perspectives drive innovation. If you are passionate about pushing the limits of what is possible in AI and machine learning, and if you thrive in a dynamic and collaborative environment, we encourage you to apply for this exciting opportunity at AMD.

Responsibilities

  • Train and finetune LLMs/LMMs.
  • Improve on the state-of-the-art LLMs/LMMs.
  • Accelerate the training and inference speed of LLMs/LMMs.
  • Research novel ML techniques and model architectures.
  • Influence the direction of AMD AI platform.
  • Publish your work at top-tier venues.

Requirements

  • PhD degree or equivalent in machine learning, computer science, artificial intelligence, or a related field.
  • Expertise and hands-on experience in training LLMs/LMMs.
  • Familiarity with hyper-parameter tuning techniques, data preprocessing, and tokenization methods.
  • Knowledge of the latest transformer architectures.
  • Strong publication record in top-tier conferences and journals.
  • Strong communication and problem-solving skills.

Nice-to-haves

  • Experience in developing and debugging in Python.
  • Experience in ML frameworks such as PyTorch, JAX, or TensorFlow.
  • Experience with distributed training.
  • Expertise on LLM/LMM pretraining, finetuning, and/or RLHF.

Benefits

  • Base pay dependent on skills, qualifications, experience, and location.
  • Eligibility for annual bonuses or sales incentives.
  • Opportunity to own shares of AMD stock through the Employee Stock Purchase Plan with discounts.
  • Competitive benefits package.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service