Pubmatic - Redwood City, CA

posted about 1 month ago

Full-time - Mid Level
Redwood City, CA
Computing Infrastructure Providers, Data Processing, Web Hosting, and Related Services

About the position

PubMatic is seeking a Senior Machine Learning Engineer with big data experience to build the next generation machine learning platform. This role involves designing and implementing a scalable, fault-tolerant big data platform that processes large datasets and provides in-depth analytics. The ideal candidate will thrive in a creative and fast-paced environment, working collaboratively with teams to enhance data processing and analytics capabilities.

Responsibilities

  • Build, design and implement a highly scalable, fault-tolerant big data platform to process terabytes of data.
  • Develop Big Data pipelines using technologies such as Spark, Hadoop, Kafka, HBase, and Hive.
  • Create analytics applications using Java, Spring, Tomcat, Jenkins, REST APIs, JDBC, AWS, and Hibernate.
  • Automate high-volume data collection and processing for real-time analytics.
  • Collaborate with Machine Learning and monetization teams to democratize data for analysis.
  • Develop solutions to help the monetization team run experiments and analyze data accurately.
  • Monitor the impact of experimental changes on machine learning algorithms.
  • Manage Hadoop Map Reduce and Spark Jobs, addressing ongoing cluster issues.
  • Implement software engineering best practices throughout the software development life cycle.
  • Participate in Agile/Scrum processes including Sprint Planning and Backlog grooming.
  • Support customer issues via email or JIRA, providing updates and patches as needed.
  • Coordinate with the Technical Writing team on technical documentation.
  • Conduct code and design reviews for peer implementations.

Requirements

  • 3-5 years coding experience in Java.
  • Solid computer science fundamentals including data structures and algorithm design.
  • Expertise in software engineering best practices for the full software development life cycle.
  • Passion for maintaining high-quality code and enabling contributions from team members.
  • Expertise in big data technologies like Hadoop, Spark, Kafka, and HBase.
  • Experience in developing large scale big data pipelines and real-time systems.
  • Demonstrated ability to achieve stretch goals in a fast-paced environment.
  • Ability to learn new technologies quickly and independently.
  • Excellent verbal and written communication skills, especially in technical contexts.
  • Strong interpersonal skills and a desire to work collaboratively.

Nice-to-haves

  • Experience with data warehouses.
  • Familiarity with automated testing and operations.

Benefits

  • Paid leave programs
  • Paid holidays
  • Healthcare, dental, and vision insurance
  • Disability and life insurance
  • Commuter benefits
  • Physical and financial wellness programs
  • Unlimited DTO in the US
  • Reimbursement for mobile expenses
  • Fully stocked pantries
  • In-office catered lunches 4 days per week
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service