JPMorgan Chase - Plano, TX

posted 2 months ago

Full-time
Plano, TX
Credit Intermediation and Related Activities

About the position

Embrace this pivotal role as an essential member of a high performing team dedicated to reaching new heights in data engineering. Your contributions will be instrumental in shaping the future of one of the world's largest and most influential companies. As a Lead Data Engineer at JPMorgan Chase within the Consumer & Community Banking, Cards Technology Team, you are an integral part of an agile team that works to enhance, build, and deliver data collection, storage, access, and analytics in a secure, stable, and scalable way. Leverage your deep technical expertise and problem-solving capabilities to drive significant business impact and tackle a diverse array of challenges that span multiple data pipelines, data architectures, and other data consumers. In this role, you will provide recommendations and insight on data management, governance procedures, and intricacies applicable to the acquisition, maintenance, validation, quality, anomaly detection, and utilization of data. You will design and deliver trusted data collection, storage, access, and analytics data platform solutions in a secure, stable, and scalable way. Additionally, you will define database back-up, recovery, and archiving strategies, generate advanced data models for one or more teams using firmwide tooling, linear algebra, statistics, and geometrical algorithms, and approve data analysis tools and processes. You will also create functional and technical documentation supporting best practices and proactively identify hidden problems and patterns in data to drive improvements to coding hygiene and system architecture. Your responsibilities will include designing and developing data pipelines end to end using Spark SQL, Java, and AWS Services. You will utilize programming languages like Java, Python, NoSQL databases, SQL, Container Orchestration services including Kubernetes, and a variety of AWS tools and services. You will contribute to software engineering communities of practice and events that explore new and emerging technologies, evaluate and report on access control processes to determine the effectiveness of data asset security, and add to the team culture of diversity, equity, inclusion, and respect.

Responsibilities

  • Provides recommendations and insight on data management, governance procedures, and intricacies applicable to the acquisition, maintenance, validation, quality, anomaly detection and utilization of data.
  • Designs and delivers trusted data collection, storage, access, and analytics data platform solutions in a secure, stable, and scalable way.
  • Defines database back-up, recovery, and archiving strategy.
  • Generates advanced data models for one or more teams using firmwide tooling, linear algebra, statistics, and geometrical algorithms.
  • Approves data analysis tools and processes.
  • Creates functional and technical documentation supporting best practices.
  • Proactively identifies hidden problems and patterns in data and uses these insights to drive improvements to coding hygiene and system architecture.
  • Design & develop data pipelines end to end using Spark SQL, Java and AWS Services.
  • Utilizes programming languages like Java, Python, NoSQL databases, SQL, Container Orchestration services including Kubernetes, and a variety of AWS tools and services.
  • Contributes to software engineering communities of practice and events that explore new and emerging technologies.
  • Evaluates and reports on access control processes to determine effectiveness of data asset security.
  • Adds to team culture of diversity, equity, inclusion, and respect.

Requirements

  • Formal training or certification on software engineering concepts and 5 years of applied experience.
  • Experience in developing, debugging, and maintaining code in a large corporate environment with one or more modern programming languages and database querying language.
  • Hands-on practical experience in developing spark-based Frameworks for end-to-end ETL, ELT & reporting solutions using key components like Spark SQL & Spark Streaming.
  • Proficient in coding in one or more Coding languages - Java, Python.
  • Experience with Relational and No SQL databases, Cloud implementation experience with AWS including: AWS Data Services: Proficiency in Lake formation, Glue ETL (or) EMR, S3, Glue Catalog, Athena, Kinesis (or) MSK, Airflow (or) Lambda Step Functions Event Bridge.
  • Data De/Serialization: Expertise in at least 2 of the formats: Parquet, Iceberg, AVRO, JSON-LD.
  • AWS Data Security: Good Understanding of security concepts such as: Lake formation, IAM, Service roles, Encryption, KMS, Secrets Manager.
  • Proficiency in automation and continuous delivery methods.
  • Solid understanding of agile methodologies such as CI/CD, Applicant Resiliency, and Security.

Nice-to-haves

  • Snowflake knowledge or experience preferred.
  • In-depth knowledge of the financial services industry and their IT systems.
  • Worked with building Data lake, built Data platforms, built Data frameworks, Built/Design of Data as a Service AP.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service