Brixton Group - Washington, DC

posted about 2 months ago

Full-time - Mid Level
Remote - Washington, DC
Performing Arts, Spectator Sports, and Related Industries

About the position

We seek a skilled Cloudera Data Engineer to join our data engineering team. The ideal candidate will have a strong background in data engineering, hands-on experience with the Cloudera platform, and expertise in building, maintaining, and optimizing large-scale data processing systems. As a Cloudera Data Engineer, you will be critical in designing and implementing data pipelines, ensuring data quality, and supporting data-driven decision-making processes. You will work closely with data scientists, analysts, and other stakeholders to understand their data needs and provide solutions that enhance data accessibility and usability. Your role will involve not only technical implementation but also collaboration with cross-functional teams to ensure that the data architecture aligns with business objectives and supports analytical initiatives. You will be responsible for troubleshooting and resolving data-related issues, optimizing existing data workflows, and implementing best practices for data governance and security. This position requires a proactive approach to problem-solving and a commitment to continuous improvement in data processes.

Responsibilities

  • Design and implement data pipelines using Cloudera technologies.
  • Ensure data quality and integrity throughout the data lifecycle.
  • Collaborate with data scientists and analysts to understand data requirements.
  • Optimize existing data processing systems for performance and scalability.
  • Troubleshoot and resolve data-related issues in a timely manner.
  • Implement best practices for data governance and security.
  • Support data-driven decision-making processes across the organization.

Requirements

  • Proven experience as a Data Engineer, specifically with Cloudera platforms (CDP, CDH, or HDP).
  • Strong proficiency in big data technologies such as Apache Spark, Hive, Impala, HDFS, Kafka, and HBase.
  • Experience with data modeling, ETL/ELT processes, and data pipeline orchestration tools (e.g., Apache NiFi, Airflow).
  • Proficient in programming languages such as Python, Java, or Scala.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service