Prodapt North America (Formerly Pacific Crest Technology) - Irving, TX

posted 4 months ago

Full-time - Senior
Irving, TX
Professional, Scientific, and Technical Services

About the position

The Google Cloud Platform Teradata Engineer will be responsible for developing and supporting database applications to drive automated data collection, storage, visualization, and transformation as per business needs. This role is crucial in ensuring that the data infrastructure is robust and scalable, enabling the organization to leverage data effectively for decision-making. The engineer will work closely with various stakeholders to understand their data needs and translate them into technical solutions that align with the company's vision and values. The position requires a strong background in Teradata and Google Cloud Platform, as well as experience in ETL processes and data modeling. In this role, the engineer will be tasked with writing Teradata and SQL/Oracle code according to established design patterns, ensuring that the data pipelines are efficient and reliable. The candidate will analyze, design, code, and test complex ETL processes for data warehouses and operational data stores, implementing data pipelines as per the design document. Additionally, the engineer will consolidate data across multiple sources and databases, making it easier for users to locate and access the information they need. The role also involves developing and deploying applications for large-scale data processing, computationally intensive statistical modeling, and advanced analytics. The engineer will write complex SQL queries and stored procedures, and must have experience with Google Cloud Platform services such as Big Query, Composer, Dataflow, and Pub-Sub. Familiarity with industry-standard technology stacks for data management and ingestion is essential, as is the ability to create custom templates for Google Cloud Platform data flow. The engineer will also be responsible for scheduling tasks using tools like Airflow and Cloud Composer, ensuring that data processing jobs run smoothly and efficiently.

Responsibilities

  • Write Teradata, SQL/Oracle code according to established design patterns.
  • Analyze, design, code, and test complex ETL processes for data warehouses and operational data stores.
  • Implement data pipelines as per the design document.
  • Consolidate data across multiple sources and databases to make it easier to locate and access.
  • Implement automated data collection and data storage systems.
  • Provide database support by coding utilities, respond to and resolve user problems.
  • Develop and deploy applications for large-scale data processing, computationally intensive statistical modeling, and advanced analytics.
  • Write complex SQL queries (T-SQL/PL-SQL) and stored procedures.
  • Utilize Google Cloud Platform services like Big Query, Composer, Dataflow, Pub-Sub, and Cloud Monitoring.
  • Create Google Cloud Platform data flow custom templates.
  • Build data pipelines supporting both batch and real-time streams for data collection, storage, processing, transformation, and aggregation.
  • Schedule tasks using Airflow, Cloud Composer, etc.

Requirements

  • Bachelor's degree in any field.
  • 5+ years of experience in Teradata utilities like BTEQ, FastLoad, MultiLoad, and TPT (Teradata Parallel Transporter).
  • ETL development experience with a strong SQL background, analyzing huge data sets, trends, and issues, and creating structured outputs.
  • Expert in Teradata utilities and Hive Queries to point external tables.
  • Experience in building high-performing data processing frameworks leveraging Google Cloud Platform and Teradata.
  • Experience in OOzi scheduling and Google Cloud Platform Airflow scheduling.
  • Experience in building data pipelines supporting both batch and real-time streams.
  • Experience in utilizing Google Cloud Platform Services like Big Query, Composer, Dataflow, Pub-Sub, and Cloud Monitoring.
  • Experience in performing ETL and data engineering work using Dataflow, Data Proc, BigQuery.
  • Experience in CI/CD automation pipeline facilitating automated deployment and testing.
  • Experience in bash shell scripts, UNIX utilities & UNIX Commands.

Nice-to-haves

  • Familiarity with the technology stack available in the industry for data management, data ingestion, capture, processing, and curation.
  • Experience in JIRA or any other Project Management Tools.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service