Americas International - Frisco, TX

posted 2 months ago

Full-time - Senior
Remote - Frisco, TX
10,001+ employees
Nursing and Residential Care Facilities

About the position

As a Data Architect specializing in Microsoft Fabric at NTT DATA, you will play a pivotal role in designing and implementing advanced data architectures that support complex analytics and data processing needs. Your primary responsibility will be to own the design and implementation of Microsoft Fabric architectures, ensuring they are tailored to meet sophisticated analytics requirements. You will be tasked with developing and optimizing MS Fabric Job Definitions, which are crucial for efficient data processing and transformation. In this role, you will design the OneLake platform to effectively handle and integrate diverse data sources into a cohesive data platform. Your expertise in SQL will be essential as you develop and maintain SQL queries and scripts that support data extraction, transformation, and loading processes. Additionally, you will leverage PySpark to build scalable data pipelines and perform intricate data transformations, ensuring that the data architecture is robust and efficient. Collaboration will be key in this position, as you will work closely with multi-functional teams to understand business requirements and translate them into effective technical solutions. You will also be responsible for ensuring data quality and integrity by implementing comprehensive data validation and cleansing processes. Monitoring and fixing data processing workflows will be part of your daily tasks to guarantee optimal performance and reliability. As a mentor, you will guide junior team members in best practices for data architecture and engineering, fostering a culture of learning and excellence. Your technical leadership will extend to the Healthcare and Health Plan domains, where your domain knowledge will enhance data solutions. Furthermore, you will contribute to the development of data governance policies and procedures, ensuring compliance with regulatory requirements. Effective communication with stakeholders will be essential as you present technical solutions and gather feedback to refine processes and architectures.

Responsibilities

  • Own the design and implementation of Microsoft Fabric architectures to support sophisticated analytics and data processing requirements
  • Be responsible for the development and optimization of MS Fabric Job Definitions to ensure efficient data processing and transformation
  • Design the OneLake to handle and integrate diverse data sources into a unified data platform
  • Develop and maintain SQL queries and scripts to support data extraction transformation and loading processes
  • Apply PySpark to build scalable data pipelines and perform complex data transformations
  • Collaborate with multi-functional teams to understand business requirements and translate them into technical solutions
  • Ensure data quality and integrity by implementing robust data validation and cleansing processes
  • Monitor and fix data processing workflows to ensure optimal performance and reliability
  • Mentor and guide junior team members in standard methodologies for data architecture and engineering
  • Provide technical leadership in the domain of Healthcare and Health Plan domains demonstrating domain knowledge to enhance data solutions
  • Contribute to the development of data governance policies and procedures to ensure compliance with regulatory requirements
  • Communicate effectively with stakeholders to present technical solutions and gather feedback

Requirements

  • 12+ years of experience in implementing large data and analytics platforms
  • 5+ years of experience architecting and building data platforms on Azure
  • Hands-on experience with cloud data platforms such as Azure Synapse Analytics, Azure Databricks, or Microsoft Fabric
  • Ability to architect, implement, and optimize data workflows and pipelines on these platforms
  • In-depth understanding of data engineering principles, data architecture, and data management practices
  • Proficiency in designing and implementing scalable data solutions
  • Exhibit strong to expert skills in PySpark for building scalable data pipelines
  • Demonstrate proficiency in developing and optimizing Spark job definitions
  • Have in-depth knowledge of OneLake for data integration and management
  • Show expertise in writing and optimizing SQL queries and scripts
  • Understanding of data modelling, data mappings, understand the data model changes and propose the best practices and guidelines
  • Experience in Healthcare or Health Plans is a plus
  • Strong problem-solving skills and attention to detail

Nice-to-haves

  • Experience in Healthcare or Health Plans is a plus
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service