Zen & Art - Alpharetta, GA

posted about 2 months ago

Full-time - Entry Level
Alpharetta, GA

About the position

The Data System Engineer will play a crucial role in the data engineering landscape, focusing on data modeling, ETL processes, data warehousing, and data analytics & science. This position is hybrid, requiring three days a week onsite in Alpharetta, GA, and offers the potential for conversion to a permanent role. The engineer will be responsible for establishing, modifying, and maintaining data structures and associated components according to design specifications. They will work closely with senior data engineers and architects to create platform-level data models and database designs, ensuring that business data requirements are well understood and documented. The role involves developing conceptual and logical data models at various levels, including enterprise and business unit/domain levels. The engineer will also need to have a solid understanding of XML/JSON, schema development, and database concepts, as well as open-source and NoSQL technologies. Participation in the software delivery methodology of the assigned team, whether Agile, Scrum, Test-Driven Development, or Waterfall, is essential to support the development of data engineering pipelines. In addition to data modeling, the engineer will write code to develop, maintain, and optimize batch and event-driven processes for managing and analyzing large volumes of structured and unstructured data. They will also be responsible for metadata integration in data pipelines and automating build and deployment processes using Jenkins across all environments to enable faster, high-quality releases. This position requires a proactive approach to learning and adapting to new technologies and methodologies in the ever-evolving field of data engineering.

Responsibilities

  • Establish, modify, and maintain data structures and associated components according to design.
  • Understand and document business data requirements.
  • Develop conceptual and logical data models at the enterprise and business unit/domain level.
  • Collaborate with senior data engineers and architects to create platform-level data models and database designs.
  • Participate in reviews of own work and colleagues' work.
  • Write code to develop, maintain, and optimize batch and event-driven processes for managing and analyzing large volumes of data.
  • Integrate metadata in data pipelines.
  • Automate build and deployment processes using Jenkins across all environments.

Requirements

  • Up to 4 years of software development experience in a professional environment.
  • Understanding of Agile or other rapid application development methods.
  • Exposure to design and development across one or more database management systems such as DB2, SybaseIQ, or Snowflake.
  • Understanding of application and database design, development, and automated testing methods.
  • Knowledge of big data technology and NoSQL design and development with various data stores (document, column family, graph, etc.).
  • General knowledge of distributed systems, algorithms, and relational & non-relational databases.
  • Experience with Linux and Python scripting.
  • Experience with large scale data processing technology such as Spark.
  • Exposure to cloud technologies such as AWS and Azure, including deployment, management, and optimization of data analytics & science pipelines.

Nice-to-haves

  • Collibra
  • Terraform
  • Java
  • Golang
  • Ruby
  • Machine Learning Operation deployment
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service