Cerecore - Nashville, TN

posted 2 months ago

Full-time - Mid Level
Nashville, TN
501-1,000 employees
Professional, Scientific, and Technical Services

About the position

CereCore® is seeking a Data Engineer to join our team in Nashville, TN. This position serves as a primary development resource responsible for writing code, testing, implementing, documenting functionality, and maintaining NextGen solutions for our GCP Cloud enterprise data initiatives. The Data Engineer will work closely with data teams in a matrixed environment as part of a broader project team. Given the fast-evolving nature of GCP/Hadoop technology, the candidate must stay well-informed of technological advancements and be proficient in applying new innovations effectively. The role requires analyzing business requirements, performing design tasks, constructing, testing, and implementing solutions with minimal supervision. The ideal candidate will have a proven track record of participating in successful projects within a fast-paced, mixed team environment. In this role, the Data Engineer will provide application development tailored to specific business environments. Responsibilities include building and supporting a GCP-based ecosystem designed for enterprise-wide analysis of structured, semi-structured, and unstructured data. The engineer will be responsible for bringing new data sources into GCP, transforming, and loading them into databases, as well as supporting regular requests to move data between clusters. A strong understanding of the relevant product area, codebase, and systems is essential, along with proficiency in data analysis, programming, and software engineering. The Data Engineer will collaborate closely with the Lead Architect and Product Owner to define, design, and build new features while improving existing products. High-quality code production with good test coverage is expected, utilizing modern abstractions and frameworks. The position also involves working independently to complete tasks on schedule, exercising strong judgment and problem-solving skills. The Data Engineer will closely collaborate with team members to execute development initiatives using Agile practices and principles. Participation in the deployment, change, configuration, management, administration, and maintenance of deployment processes and systems is also required. The candidate must effectively prioritize workload to meet deadlines and work objectives in an environment with rapidly changing business requirements and priorities. Collaboration with Data Scientists and business and IT leaders throughout the company is crucial to understand their needs and use cases, as well as working closely with management, architects, and other teams to develop and implement projects. Active participation in technical discussions and the adoption of new technologies to enhance development and operations is expected.

Responsibilities

  • Provide application development for specific business environments.
  • Build and support a GCP-based ecosystem for enterprise-wide analysis of structured, semi-structured, and unstructured data.
  • Bring new data sources into GCP, transform, and load to databases.
  • Support regular requests to move data from one cluster to another.
  • Develop a strong understanding of relevant product area, codebase, and/or systems.
  • Demonstrate proficiency in data analysis, programming, and software engineering.
  • Work closely with the Lead Architect and Product Owner to define, design, and build new features and improve existing products.
  • Produce high quality code with good test coverage, using modern abstractions and frameworks.
  • Work independently and complete tasks on-schedule by exercising strong judgment and problem-solving skills.
  • Collaborate with team members to successfully execute development initiatives using Agile practices and principles.
  • Participate in the deployment, change, configuration, management, administration, and maintenance of deployment processes and systems.
  • Effectively prioritize workload to meet deadlines and work objectives.
  • Work collaboratively with Data Scientists and business and IT leaders to understand their needs and use cases.
  • Work closely with management, architects, and other teams to develop and implement projects.
  • Actively participate in technical group discussions and adopt new technologies to improve development and operations.

Requirements

  • Good understanding of best practices and standards for GCP Data process design and implementation.
  • Two plus years of hands-on experience with GCP platform and experience with components such as Cloud Run, GKE, Cloud Functions, Spark Streaming, Kafka, Pub/Sub, Bigtable, Firestore, Cloud SQL, Cloud Spanner, JSON, Avro, Parquet, Python, Java, Terraform, BigQuery, Dataflow, Data Fusion, Cloud Composer, DataProc, CI/CD, Cloud Logging, Vertex AI, NLP, GitHub.
  • Ability to multitask and balance competing priorities.
  • Ability to define and utilize best practice techniques and impose order in a fast-changing environment.
  • Strong problem-solving skills.
  • Strong verbal, written, and interpersonal skills, with a desire to work in a highly matrixed, team-oriented environment.

Nice-to-haves

  • Experience in Healthcare Domain
  • Experience in Patient Data
  • Experience with Linux, UNIX
  • Experience in GCP Distributed, highly scalable processing environments.
  • Experience with RDBMS - MS SQL Server/Teradata/Oracle
  • Experience with NoSQL, Hbase, Cassandra, MongoDB, In-memory, Columnar, and other emerging technologies.
  • Experience with Source Control Systems - Git, Mercurial
  • Experience with Continuous Integration Systems - Jenkins or Bamboo
  • Certifications such as GCP Cloud Professional Data Engineer.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service