NOV - Houston, TX

posted about 2 months ago

Full-time
Houston, TX
Machinery Manufacturing

About the position

Are you passionate about helping people improve the impact of their business using data and analytics? At NOV, we strive to lead technology innovation that delivers significant value to our customers. We are hiring a Data Engineer to create and support data architecture that moves and translates data used to inform our most critical strategic and real-time decisions. This position will be part of NOV Rig Technologies, Digital Products and Automation located in Houston, Texas. In this role, you will assist with developing a data ecosystem that is robust, fast, and scalable. You will design, build, and launch efficient and reliable data pipelines to move and transform data, whether in large or small amounts. Your responsibilities will also include optimizing existing pipelines and maintaining all domain-related data pipelines. You will deploy inclusive data quality checks to ensure high quality of data and use DevOps methodologies to create automated, efficient CI/CD processes to reduce the time to promote, test, and deploy analytics models and analyses. Additionally, you will develop, test, deploy, and maintain efficient and reusable patterns of streaming and batch data ingestion pipeline architectures. Documenting and maintaining architecture and coding standards for supported platforms will be essential. You will participate in all phases of the software development lifecycle, including requirements gathering, technical planning, design, development, testing, sustaining support, and documentation. It is important to seek guidance when a direction is needed and to speak up about technology risks identified. Following agile practices, as well as quality management procedures as defined by precedents, standards, or policies, will be crucial. Collaboration with the Analytics team members and teams across NOV to deliver solutions and evolve products is also a key aspect of this role. Compliance with all NOV Company and HSE policies and procedures is expected.

Responsibilities

  • Assist with developing a data ecosystem that is robust, fast, and scalable.
  • Design, build, and launch efficient and reliable data pipelines to move and transform data.
  • Optimize existing pipelines and maintain all domain-related data pipelines.
  • Deploy inclusive data quality checks to ensure high quality of data.
  • Use DevOps methodologies to create automated, efficient CI/CD processes to reduce the time to promote, test, and deploy analytics models and analyses.
  • Develop, test, deploy, and maintain efficient and reusable patterns of streaming and batch data ingestion pipeline architectures.
  • Document and maintain architecture and coding standards for supported platforms.
  • Participate in all phases of the software development lifecycle, including requirements gathering, technical planning, design, development, testing, sustaining support, and documentation.
  • Seek guidance when a direction is needed and speak up about technology risks identified.
  • Follow agile practices, as well as quality management procedures as defined by precedents, standards, or policies.
  • Collaborate with the Analytics team members and teams across NOV to deliver solutions and evolve products.
  • Comply with all NOV Company and HSE policies and procedures.

Requirements

  • Bachelor's degree in computer science, Computer Engineering, relevant technical field, or equivalent practical experience.
  • 3 years of Python, PySpark or other modern programming language development experience.
  • Hands-on experience working with data pipelines using a variety of source and target locations (e.g., Databricks, SQL Server, Data Lake, file-based, SQL and No-SQL database).
  • 3 years of experience in custom ETL design, implementation, and maintenance.
  • Experience developing batch ETL pipelines; real-time pipelines are a plus.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service