Unclassified - Reston, VA

posted 30 days ago

Full-time
Reston, VA

About the position

The Java/Big Data Developer position is a critical role within our team, focusing on the development and integration of data solutions using a variety of technologies. The ideal candidate will have a strong background in software development, particularly in Java, Python, and Scala, with a minimum of 7 years of experience in the field. This role requires a deep understanding of Big Data technologies, including Hadoop, Spark, Hive, and HBase, as well as experience in data integration projects. The successful candidate will work closely with Data Analysts and other team members to review business requirements and translate them into technical specifications. In this position, you will be responsible for designing and building data integration pipelines using the Cloudera Hadoop platform. You will transform data to create a consumable data layer for various applications, ensuring that the data is accessible and usable for business needs. Additionally, you will support the data pipeline by addressing bug fixes and implementing enhancements as necessary. Documentation is also a key aspect of this role, as you will be required to document technical designs and operational runbooks to ensure clarity and continuity within the team. The role demands a rigorous approach to code quality, automated testing, and adherence to engineering best practices. You will be expected to write reusable code components and have strong SQL query writing and data analysis skills. Good shell scripting experience is also essential to facilitate various tasks within the data integration process. This position offers an exciting opportunity to work with cutting-edge technologies in a collaborative environment, contributing to the overall success of our data initiatives.

Responsibilities

  • Work with Data Analysts and other team members to review business requirements and translate them into technical requirements.
  • Collaborate with application architects and data solution architects.
  • Design and build data integration pipelines using the Cloudera Hadoop platform.
  • Transform data to create a consumable data layer for various application uses.
  • Support data pipeline with bug fixes and additional enhancements.
  • Document technical design, operational runbook, etc.

Requirements

  • BA/BS in Computer Science, Information Systems, Information Technology or related field.
  • 7+ years of prior experience in software development, Data Engineering, and Business Intelligence or equivalent experience.
  • 7+ years of strong programming background with Java/Python/Scala.
  • At least 3+ years of experience working on Data Integration projects using Hadoop MapReduce, Spark, Hive, HBase, and other related Big Data technologies.
  • Some working experience building Kafka-based data ingestion/retrieval programs.
  • Experience tuning Hadoop/Spark/Hive parameters for optimal performance.
  • Strong SQL query writing and data analysis skills.
  • Good shell scripting experience.
  • Rigor in high code quality, automated testing, and other engineering best practices, ability to write reusable code components.

Nice-to-haves

  • Healthcare experience
  • Cloudera Developer certification
  • Cloud Development experience
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service