This job is closed

We regret to inform you that the job you were interested in has been closed. Although this specific position is no longer available, we encourage you to continue exploring other opportunities on our job board.

Amazon.composted 4 days ago
$129,300 - $223,600/Yr
Cupertino, CA
General Merchandise Retailers

About the position

The AWS Neuron Compiler team is actively seeking skilled compiler engineers to join our efforts in developing a state-of-the-art deep learning compiler stack. This stack is designed to optimize application models across diverse domains, including Large Language and Vision, originating from leading frameworks such as PyTorch, TensorFlow, and JAX. Your role will involve working closely with our custom-built Machine Learning accelerators, including Inferentia/Trainium, which represent the forefront of AWS innovation for advanced ML capabilities, powering solutions like Generative AI. In this role as a ML Compiler engineer, you'll be instrumental in designing, developing, and optimizing features for our compiler. Your responsibilities will involve tackling crucial challenges alongside a talented engineering team, contributing to leading-edge design and research in compiler technology and deep-learning systems software. Additionally, you'll collaborate closely with cross-functional team members from the Runtime, Frameworks, and Hardware teams to ensure system-wide performance optimization. As part of the Backend team, you'll play a significant role in designing and developing various aspects of our system. This includes but is not limited to instruction scheduling, memory allocation, data transfer optimization, graph partitioning, parallel programing, code generation, Instruction Set Architectures, new hardware bring-up, and hardware-software co-design. AWS Utility Computing (UC) provides product innovations that continue to set AWS's services and features apart in the industry. As a member of the UC organization, you'll support the development and management of Compute, Database, Storage, Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for their cloud services. Additionally, this role may involve exposure to and experience with Amazon's growing suite of generative AI services and other cutting-edge cloud computing offerings across the AWS portfolio. Annapurna Labs (our organization within AWS UC) designs silicon and software that accelerates innovation. Customers choose us to create cloud solutions that solve challenges that were unimaginable a short time ago-even yesterday. Our custom chips, accelerators, and software stacks enable us to take on technical challenges that have never been seen before, and deliver results that help our customers change the world.

Responsibilities

  • Solve challenging technical problems, often ones not solved before, at every layer of the stack.
  • Design, implement, test, deploy and maintain innovative software solutions to transform service performance, durability, cost, and security.
  • Research implementations that deliver the best possible experiences for customers.
  • Build high-impact solutions to deliver to our large customer base.
  • Participate in design discussions, code review, and communicate with internal and external stakeholders.
  • Work cross-functionally to help drive business decisions with your technical input.
  • Work in a startup-like development environment, where you're always working on the most important stuff.

Requirements

  • B.S. or M.S. in computer science or related field
  • Proficiency with 1 or more of the following programming languages: C++ (preferred), Python
  • 3+ years of non-internship professional software development experience
  • 2+ years of experience developing compiler optimization, graph-theory, hardware bring-up, FPGA placement and routing algorithms, or hardware resource management

Nice-to-haves

  • M.S. or Ph.D. in computer science or related field
  • Strong knowledge in one or more of the areas of: compiler design, instruction scheduling, memory allocation, data transfer optimization, graph partitioning, parallel programing, code generation, Instruction Set Architectures, new hardware bring-up, and hardware-software co-design
  • Experience with LLVM and/or MLIR
  • Experience with developing algorithms for simulation tools
  • Experience is TensorFlow, PyTorch, and/or JAX
  • Experience in LLM, Vision or other deep-learning models
Hard Skills
M
2
PyTorch
2
TensorFlow
2
B
1
Python
1
2Cu8i WO8V0rESU3d
0
59dJQjiz4Ml bCucPk1
0
9hkUC0GgEqpL QyCIq9N4YZxj
0
J5g0j3A2 4dOHtvcPC
0
MBERV vXblxrtyW
0
NPzTGBdq8C 82JikUh1LdDbpI
0
Nf7G4 bMmva3K
0
OHnQz3ibW piFzNHKr3
0
OlNB ublPsVmaF
0
RvB4g bOcupz9
0
SHUElc9hv 5CXjsoT21Oz
0
TkEQcZ lcSAV4Nj0i
0
UJvulyQEz F8hMiyQvr2YlX
0
VYgwIsXFS SlHp6aWs
0
X46dUmuS sp2tGkZeiP
0
Y2sE8cbwp tVlwZbv
0
ZHDCipcQKdwN dpgCyko0E2
0
c6Cd8zHfO3 UzgL U2NT3MhCI6
0
drtWMuGBH TQe4MLOg
0
eiozT REMIisJkh
0
gTKWQOes6 Ais4NQUZoxz8
0
iSgQpUYT8 2UMNZzb1IeY
0
jOPynXE7wd9 A47UaPyuz
0
kN9Ou 5CzniuSaT
0
lv0DUp1 AiXJBuchzdv
0
mfPieq8 BU1ZKnjS
0
pESJG8 UJy0Xxr4D
0
rOGVsSFuTigl 6HQ0emEBwGs
0
wjZ1c5 6Ah2wNF
0
yAPTES 3zMDHSRtxL
0
z14 q8vlIUg TEZW1pS9
0
zS1BZKc4y rzSGLxmMCBYo
0
Unlock 32 more keywords by signing up for Teal+Sign Up
Build your resume with AI

A Smarter and Faster Way to Build Your Resume

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service