Autodesk - San Francisco, CA

posted 2 months ago

Full-time - Mid Level
Remote - San Francisco, CA
Publishing Industries

About the position

The Lead Data Engineer will be responsible for designing and implementing strategies for enterprise databases and data warehouse systems. This role involves leading large-scale data engineering projects, managing teams, and collaborating with cross-functional teams to build scalable data pipelines. The position requires a deep understanding of data systems, software development, and data analytics to enhance performance and reliability.

Responsibilities

  • Design strategies for enterprise databases and data warehouse systems.
  • Propose and implement enhancements to improve system performance and reliability.
  • Analyze information and evaluate results to choose the best solutions and solve problems.
  • Modify existing software to correct errors and improve performance.
  • Develop or direct software system testing or validation procedures.
  • Lead large-scale data engineering projects and manage teams of Data Engineers and data analytics.
  • Collaborate with cross-functional teams to build scalable and high-performing data pipelines.
  • Maintain scalable ETL pipelines on distributed software systems and cloud platforms.
  • Collaborate on the design and maintenance of operations data mart systems.
  • Locate and acquire data from warehouse and transactional source systems.
  • Design and develop software to automate data reporting ETL functions.
  • Develop and maintain large-scale data analytics systems.
  • Manage data-related activities and design processes with internal developers and external consultants.
  • Interact with internal users regarding data system availability and potential issues.
  • Review reports and analysis to identify efficient data summarization techniques.
  • Detect underlying data quality issues and implement fixes.

Requirements

  • Bachelor's degree in Computer Science, Engineering, Computer Information Systems, or related field.
  • 5 years of progressive, post-baccalaureate experience in the job offered or in a software engineering/technical lead-related occupation.
  • Experience in building scalable Massive Parallel Processing (MPP) data pipelines in big data platforms.
  • Experience in designing and modeling STAR and Snowflake schemas for data warehousing.
  • Proficiency in SQL, including advanced functions and query optimization techniques.
  • Programming experience in Python, Java, or Scala, including libraries like PANDAS and NUMPY.
  • Experience in building products in a cloud-based environment, especially AWS.
  • Familiarity with HIVE and SPARK or similar distributed processing platforms.
  • Experience in scaling and optimizing schemas and performance tuning SQL and ETL pipelines.
  • Experience with enterprise ETL tools like Informatica or SSIS.
  • Familiarity with Agile development processes and tools like JIRA and GitHub.
Job Description Matching

Match and compare your resume to any job description

Start Matching
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service