Ernst & Young - Akron, OH

posted 4 days ago

Full-time - Senior
Akron, OH
Professional, Scientific, and Technical Services

About the position

The Data Architect position at EY focuses on leading the design, development, and management of data architecture within the Industrials and Energy sector. This senior role requires a strong understanding of technology and data, with a proven ability to deliver scalable, efficient, and secure data solutions that align with business goals. The architect will be responsible for creating and maintaining the data platform and infrastructure that supports analytics, ensuring high-volume data processing and database management.

Responsibilities

  • Design, develop, optimize, and maintain data architecture and pipelines adhering to ETL principles and business goals.
  • Develop and maintain scalable data pipelines, building new integrations using AWS native technologies.
  • Define data requirements, gather and mine large-scale structured and unstructured data, and validate data using various tools in a Big Data environment.
  • Support standardization, customization, and ad hoc data analysis, developing mechanisms to ingest, analyze, validate, normalize, and clean data.
  • Write unit/integration/performance test scripts and perform data analysis to troubleshoot and resolve data-related issues.
  • Implement processes and systems for data reconciliation and monitor data quality, ensuring accuracy for stakeholders and business processes.
  • Lead the evaluation, implementation, and deployment of emerging tools and processes for analytic data engineering.
  • Develop and deliver communication and education plans on analytic data engineering capabilities, standards, and processes.
  • Learn about machine learning, data science, computer vision, artificial intelligence, statistics, and applied mathematics.
  • Solve complex data problems to deliver insights that help achieve business objectives.
  • Implement statistical data quality procedures on new data sources through rigorous iterative data analytics.

Requirements

  • Experience in the development of Hadoop APIs and MapReduce for large scale data processing.
  • Hands-on programming experience in Apache Spark using SparkSQL and Spark Streaming or Apache Storm.
  • Hands-on experience with major components like Hive, Spark, and MapReduce.
  • Experience working with NoSQL in at least one of the data stores - HBase, Cassandra, MongoDB.
  • Experienced in Hadoop clustering and Auto scaling.
  • Good knowledge in Apache Kafka & Apache Flume.
  • Knowledge of Spark and Kafka integration with multiple Spark to consume messages from multiple Kafka partitions.
  • Advanced experience and understanding of data/Big Data, data integration, data modelling, AWS, and cloud technologies.
  • Strong business acumen with knowledge of the Industrial Products sector is preferred, but not required.
  • Ability to build processes that support data transformation, workload management, data structures, dependency, and metadata.
  • Ability to build and optimize queries (SQL), data sets, 'Big Data' pipelines, and architectures for structured and unstructured data.
  • Experience with or knowledge of Agile Software Development methodologies.
  • Demonstrated understanding and experience using Data Engineering Programming Languages (i.e., Python).
  • Experience with Distributed Data Technologies (e.g., Pyspark).
  • Experience with cloud platform deployment and tools (e.g., Kubernetes).
  • Experience with relational SQL databases.
  • Experience with DevOps and continuous integration.
  • Experience with AWS cloud services and technologies (i.e., Lambda, S3, DMS, Step Functions, Event Bridge, Cloud Watch, RDS).
  • Experience with Databricks/ETL, IICS/DMS, GitHub, Event Bridge, Tidal.

Nice-to-haves

  • Experience in leading and influencing teams, with a focus on mentorship and professional development.
  • A passion for innovation and the strategic application of emerging technologies to solve real-world challenges.
  • The ability to foster an inclusive environment that values diverse perspectives and empowers team members.

Benefits

  • Comprehensive compensation and benefits package based on performance.
  • Medical and dental coverage.
  • Pension and 401(k) plans.
  • Wide range of paid time off options including flexible vacation policy, designated paid holidays, and other leaves of absence.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service