Ernst & Young - Greenville, SC

posted about 2 months ago

Full-time - Senior
Greenville, SC
Professional, Scientific, and Technical Services

About the position

At EY, you'll have the chance to build a career as unique as you are, with the global scale, support, inclusive culture and technology to become the best version of you. And we're counting on your unique voice and perspective to help EY become even better. Join us and build an exceptional experience for yourself, and a better working world for all. The exceptional EY experience. It's yours to build. EY focuses on high-ethical standards and integrity among its employees and expects all candidates to demonstrate these qualities. EY is seeking a Data Architect with strong technology and data understanding having proven delivery capability. Lead the design, development, and management of the organization's data architecture, ensuring scalable, efficient, and secure data solutions that align with business goals and support enterprise-wide data initiatives. In this role, you will create, maintain, and support the data platform and infrastructure that enables the analytics front-end; this includes the testing, maintenance, construction, and development of architectures such as high-volume, large-scale data processing and databases with proper verification and validation processes.

Responsibilities

  • Design, develop, optimize, and maintain data architecture and pipelines that adheres to ETL principles and business goals
  • Develop and maintain scalable data pipelines, build out new integrations using AWS native technologies to support continuing increases in data source, volume, and complexity
  • Define data requirements, gather and mine large scale of structured and unstructured data, and validate data by running various data tools in the Big Data Environment
  • Support standardization, customization and ad hoc data analysis and develop the mechanisms to ingest, analyse, validate, normalize, and clean data
  • Write unit/integration/performance test scripts and perform data analysis required to troubleshoot data related issues and assist in the resolution of data issues
  • Implement processes and systems to drive data reconciliation and monitor data quality, ensuring production data is always accurate and available for key stakeholders, downstream systems, and business processes
  • Lead the evaluation, implementation and deployment of emerging tools and processes for analytic data engineering to improve productivity
  • Develop and deliver communication and education plans on analytic data engineering capabilities, standards, and processes
  • Learn about machine learning, data science, computer vision, artificial intelligence, statistics, and/or applied mathematics
  • Solve complex data problems to deliver insights that help achieve business objectives
  • Implement statistical data quality procedures on new data sources by applying rigorous iterative data analytics
  • Strong understanding & familiarity with all Hadoop Ecosystem components and Hadoop Administrative Fundamentals
  • Strong understanding of underlying Hadoop Architectural concepts and distributed computing paradigms

Requirements

  • Experience in the development of Hadoop APIs and MapReduce jobs for large scale data processing
  • Hands-on programming experience in Apache Spark using SparkSQL and Spark Streaming or Apache Storm
  • Hands on experience with major components like Hive, Spark, and MapReduce
  • Experience working with NoSQL in at least one of the data stores - HBase, Cassandra, MongoDB
  • Experienced in Hadoop clustering and Auto scaling
  • Good knowledge in apache Kafka & Apache Flume
  • Knowledge of Spark and Kafka integration with multiple Spark jobs to consume messages from multiple Kafka partitions
  • Advanced experience and understanding of data/Big Data, data integration, data modelling, AWS, and cloud technologies
  • Strong business acumen with knowledge of the Industrial Products sector is preferred, but not required
  • Ability to build processes that support data transformation, workload management, data structures, dependency, and metadata
  • Ability to build and optimize queries (SQL), data sets, 'Big Data' pipelines, and architectures for structured and unstructured data
  • Experience with or knowledge of Agile Software Development methodologies
  • Demonstrated understanding and experience using Data Engineering Programming Languages (i.e., Python)
  • Distributed Data Technologies (e.g., Pyspark)
  • Cloud platform deployment and tools (e.g., Kubernetes)
  • Relational SQL databases
  • DevOps and continuous integration
  • AWS cloud services and technologies (i.e., Lambda, S3, DMS, Step Functions, Event Bridge, Cloud Watch, RDS)
  • Databricks/ETL
  • IICS/DMS
  • GitHub
  • Event Bridge, Tidal

Nice-to-haves

  • Experience in leading and influencing teams, with a focus on mentorship and professional development
  • A passion for innovation and the strategic application of emerging technologies to solve real-world challenges
  • The ability to foster an inclusive environment that values diverse perspectives and empowers team members

Benefits

  • Comprehensive compensation and benefits package
  • Medical and dental coverage
  • Pension and 401(k) plans
  • Wide range of paid time off options
  • Flexible vacation policy allowing employees to decide how much vacation time they need
  • Time off for designated EY Paid Holidays, Winter/Summer breaks, Personal/Family Care, and other leaves of absence
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service