Aroha Technologies - Mountain View, CA

posted about 2 months ago

Full-time - Mid Level
Mountain View, CA
Professional, Scientific, and Technical Services

About the position

We are seeking an experienced Senior Engineer, Analytics with a focus on Streaming to join our team. As a Senior Engineer, you will play a key role in designing, building, and maintaining our real-time analytics infrastructure. You will work closely with our data scientists, product managers, and other engineers to develop and deploy scalable, efficient, and reliable data pipelines. This position requires a deep understanding of streaming data technologies and the ability to implement solutions that meet the needs of our analytics team and the broader organization. In this role, you will be responsible for designing, building, and maintaining large-scale streaming data pipelines using technologies such as Apache Beam, Apache Kafka, Apache Kinesis, Google Cloud Pub/Sub, and Google Cloud DataFlow. You will develop and implement streaming data processing jobs using programming languages like Java or Python, and your experience with both will be a significant advantage. Collaboration is key, as you will work with data scientists and product managers to develop and deploy real-time analytics applications that drive business insights. You will also collaborate with other engineers to integrate streaming data pipelines with our data warehouse and data lake, ensuring that data flows seamlessly across our systems. Your expertise in public cloud providers such as Google Cloud Platform, Amazon Web Services, or Microsoft Azure will be essential as you monitor and troubleshoot streaming pipelines to ensure high availability and performance. Additionally, you will implement DevOps principles and practices to ensure efficient and reliable deployment of batch processing systems, leveraging containerization technologies such as Docker and Kubernetes to enhance our infrastructure.

Responsibilities

  • Design, build, and maintain large-scale streaming data pipelines using technologies such as Apache Beam, Apache Kafka, Apache Kinesis, Google Cloud Pub/Sub, and Google Cloud DataFlow
  • Develop and implement streaming data processing jobs using programming languages like Java or Python
  • Work with data scientists and product managers to develop and deploy real-time analytics applications
  • Collaborate with other engineers to integrate streaming data pipelines with our data warehouse and data lake
  • Work with public cloud providers such as Google Cloud Platform, Amazon Web Services, or Microsoft Azure
  • Monitor and troubleshoot streaming pipelines to ensure high availability and performance
  • Implement DevOps principles and practices to ensure efficient and reliable deployment of batch processing systems
  • Utilize containerization technologies such as Docker and Kubernetes

Requirements

  • Bachelor's degree in Computer Science or a related field
  • 5+ years of experience in software engineering, with a focus on streaming data pipelines and analytics
  • Strong programming skills in Java or Python
  • Experience with a public cloud provider, with a focus on Google Cloud Platform
  • Strong experience with messaging/stream processing systems such as Apache Kafka, Apache Kinesis, Google Cloud Pub/Sub, and Google Cloud DataFlow
  • Experience with data warehousing and data lake technologies
  • Strong understanding of data modeling, data governance, and data security
  • Excellent problem-solving skills, with the ability to work independently and collaboratively
  • Excellent communication skills, with the ability to explain complex technical concepts to non-technical stakeholders

Nice-to-haves

  • Experience with machine learning and data science technologies
  • Certification in a public cloud provider or a relevant technology
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service