This job is closed

We regret to inform you that the job you were interested in has been closed. Although this specific position is no longer available, we encourage you to continue exploring other opportunities on our job board.

Paycom Softwareposted 10 days ago
Oklahoma City, OK
Professional, Scientific, and Technical Services

About the position

This position will be located within the Development and IT space and work closely with computer scientists, IT and data scientists to build, deploy and optimize data pipelines and integrate data infrastructure for enabling analytics, reporting, and machine learning workloads at scale.

Responsibilities

  • Build, test, and validate robust production-grade data pipelines that can ingest, aggregate, and transform large datasets according to the specifications of the internal teams who will be consuming the data.
  • Configure connections to source data systems and validates schema definitions with the teams responsible for the source data.
  • Monitor data pipelines and troubleshoots issues as they arise.
  • Monitor data lake environment for performance and data integrity.
  • Collaborate with IT and database teams to maintain the overall data ecosystem.
  • Assist data science, business intelligence, and other teams in using the data provided by the data pipelines.
  • Serve as on-call for production issues related to data pipelines and other data infrastructure maintained by the data engineering team.

Requirements

  • BS degree in Computer Science or related field
  • 3+ years of data engineering work experience
  • Experience coding in Java or Scala and build tools such as Maven, Gradle, and SBT
  • Experience with SQL databases
  • Experience working with HDFS or S3 storage environments
  • Experience with Apache Spark or Databricks and reading and writing Parquet, Avro and JSON
  • Experience working in a Unix or Linux environment, including writing shell scripts
  • Experience with ETL and ELT processes in data pipelines
  • Experience with Docker and Kubernetes highly preferred
  • Experience with workflow orchestration tools like Apache Airflow, Control-M, or Arrow highly preferred
  • Experience with Apache Kafka or Confluent is preferred

Nice-to-haves

  • Experience coding in Python
  • Experience with NoSQL solutions is helpful

Job Keywords

Hard Skills
  • Apache Kafka
  • Docker
  • Gradle
  • JSON
  • Kubernetes
  • 4kRmo qCbnKgSxJoyLv5F
  • 54INz f53O1P2zbr
  • 7oEt81h wfrHOq
  • AslowvL 4zpC8
  • bedk82IzTQL
  • CXMAQ 4IQXEomAdk
  • DZgcrF
  • EAQ8J
  • f6cg8 BAPhy4SYZo
  • hvx4e nM5k4VZLqg
  • jiCR0vw WfAIF
  • K1gxTsLAC pQESaK5m
  • kc1VzJO4 mOkBuXQc3
  • LFVbQ YXurNGJI9Z7w
  • MFTj5 xrSH7mfCok
  • n6BS548
  • of1JVOw0GyA
  • Q4j6L WcwKaBv4Lr
  • rbD1
  • V24p1v
  • Xl1zpkim
  • yi1Kq
  • YIRJ86myD CbtY7IAM
  • ZUgSJ7
Build your resume with AI

A Smarter and Faster Way to Build Your Resume

Go to AI Resume Builder
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service