Charter Global - Reston, VA

posted 3 months ago

Full-time - Senior
Remote - Reston, VA
Professional, Scientific, and Technical Services

About the position

The Big Data Administrator Lead position is a hands-on role focused on orchestrating, deploying, maintaining, and scaling cloud infrastructure specifically targeting big data and platform data management. This includes managing data warehouses and data lakes, as well as developing data access APIs. The role emphasizes high availability, reliability, automation, and performance in the management of big data systems. The successful candidate will lead the automation and improvement of components such as Cloudera Solr and HBase within the Cloudera CDP public cloud, integrating these with other AWS services. The Lead Data Engineer will be responsible for preparing and manipulating data using Hadoop or equivalent technologies. This position requires advanced expertise in administrating and engineering relational databases like MySQL and PostgreSQL, as well as Big Data systems including Cloudera Data Platform in both private and public cloud environments. The candidate should have a strong background in Apache Solr and automation tools such as Ansible, Terraform, and Bit Bucket, along with experience in cloud solutions, particularly data products on AWS. Candidates must have at least 10 years of experience in the administration of big data and Meta Data Hub, specifically with Cloudera, Solr, and HBase. Familiarity with additional technologies such as Ab Initio, EMR, S3, Dynamo DB, Mongo DB, PostgreSQL, RDS, and DB2 is considered a plus. Knowledge of DevOps practices, particularly CI/CD pipelines, is also advantageous. The role requires advanced knowledge of UNIX and SQL, as well as experience managing metadata hubs and troubleshooting environmental issues affecting these components. The Lead Data Engineer will represent the team in architectural and design discussions, providing credible feedback and input across all impacted areas, while also tracking and monitoring projects and tasks as the lead.

Responsibilities

  • Orchestrate, deploy, maintain, and scale cloud infrastructure targeting big data and platform data management.
  • Lead the automation and improvement of Cloudera components such as Solr and HBase.
  • Prepare and manipulate data using Hadoop or equivalent technologies.
  • Administer and engineer relational databases including MySQL and PostgreSQL.
  • Manage Big Data systems including Cloudera Data Platform in both private and public cloud environments.
  • Utilize automation tools such as Ansible, Terraform, and Bit Bucket for operational efficiency.
  • Integrate big data solutions with AWS services.
  • Troubleshoot environmental issues affecting metadata hubs and operational consoles.
  • Represent the team in architectural and design discussions, providing subject matter expertise.
  • Track and monitor projects and tasks while also engaging in hands-on work.

Requirements

  • Advanced experience in administrating and engineering relational databases (e.g., MySQL, PostgreSQL).
  • Expertise in Big Data systems, particularly Cloudera Data Platform (both private and public cloud).
  • Strong knowledge of Apache Solr and HBase as a subject matter expert.
  • Experience with automation tools such as Ansible, Terraform, and Bit Bucket.
  • Familiarity with cloud solutions, specifically data products on AWS.
  • At least 10 years of experience in big data administration and Meta Data Hub management.
  • Advanced knowledge of UNIX and SQL.
  • Experience managing metadata hubs and troubleshooting environmental issues.

Nice-to-haves

  • Experience with Ab Initio, EMR, S3, Dynamo DB, Mongo DB, PostgreSQL, RDS, and DB2.
  • Knowledge of DevOps practices, particularly CI/CD pipelines.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service