Snowflakeposted about 2 months ago
$157,000 - $230,000/Yr
Full-time - Mid Level
Bellevue, WA
Publishing Industries

About the position

Snowflake's vision is to enable every organization to be data-driven. We're at the forefront of innovation, helping customers realize the full potential of their data with our AI data cloud. We are now going far beyond the traditional data warehouse and helping customers unlock the power of the open data lakehouse architecture with significant investment in Open Source Analytics! Snowflake engineers are leading the way with innovations directly in OSS projects like Apache Iceberg, Apache Polaris (incubating), Apache Parquet and more! As a Software Engineer on the Open Source Analytics team, you'll play a key role in building and evolving our open and interoperable data lake ecosystem. You'll work on some of the most complex and exciting challenges in the enterprise data lake analytics area, all while collaborating closely with some of the best minds in the open source community! You will have a direct impact towards Snowflake's mission of providing a truly open data lake architecture, free from vendor lock-in.

Responsibilities

  • Pioneer new and innovative technical capabilities in the Open Source Analytics community.
  • Design and implement features and enhancements for Apache Iceberg and Apache Polaris focusing on scalability, performance and usability.
  • Collaborate with the Open source community by contributing code, participating in discussions and reviewing pull requests.
  • Architect and build systems that integrate open source technologies seamlessly with Snowflake.
  • Collaborate with Snowflake's open-source team and the Apache Iceberg community to contribute new features and enhance the Iceberg table format and REST specification.
  • Work on core data access control and governance features for Apache Polaris.
  • Contribute to our managed Polaris service, Snowflake Open Catalog, enabling customers to manage and expand their data lake.
  • Build tooling and services that automate data lake table maintenance.

Requirements

  • 4+ years of experience designing and building scalable, distributed systems.
  • Strong programming skills in Java, Scala, or C++ with an emphasis on performance and reliability.
  • Deep understanding of distributed transaction processing, concurrency control, and high-performance query engines.
  • Experience with open-source data lake formats (e.g., Apache Iceberg, Parquet, Delta).
  • Experience building cloud-native services and working with public cloud providers like AWS, Azure, or GCP.
  • A passion for open-source software and community engagement, particularly in the data ecosystem.
  • Familiarity with data governance, security, and access control models in distributed data systems.

Nice-to-haves

  • Contributing to open-source projects, especially in the data infrastructure space.
  • Designing or implementing REST APIs, particularly in the context of distributed systems.
  • Managing large-scale data lakes or data catalogs in production environments.
  • Working on highly-performant and scalable query engines such as Spark, Flink, or Trino.

Benefits

  • Medical, dental, vision, life, and disability insurance
  • 401(k) retirement plan
  • Flexible spending & health savings account
  • At least 12 paid holidays
  • Paid time off
  • Parental leave
  • Employee assistance program
  • Other company benefits
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service