LOG IN
SIGN UP
Tech Job Finder - Find Software, Technology Sales and Product Manager Jobs.
Sign In
OR continue with e-mail and password
E-mail address
Password
Don't have an account?
Reset password
Join Tech Job Finder
OR continue with e-mail and password
E-mail address
First name
Last name
Username
Password
Confirm Password
How did you hear about us?
By signing up, you agree to our Terms & Conditions and Privacy Policy.

Lead Software Engineer - Databricks, Spark, AWS

at J.P. Morgan

Back to all Data Engineering jobs
J.P. Morgan logo
Industry not specified

Lead Software Engineer - Databricks, Spark, AWS

at J.P. Morgan

Tech LeadNo visa sponsorshipData Engineering

Posted 17 hours ago

No clicks

Compensation
Not specified USD

Currency: $ (USD)

City
Not specified
Country
United States

Lead Software Engineer responsible for architecting and delivering high-throughput, low-latency data pipelines using Databricks and Apache Spark, with a focus on lakehouse patterns and scalable data platforms. Own Databricks cluster strategy, job orchestration with Databricks Workflows, and integration with AWS services (S3, Glue, IAM, CloudWatch, Kinesis/MSK) for secure data ingestion and processing. Develop reusable libraries in Python/Java, implement CI/CD for data projects, enforce data governance and security, and drive reliability, observability, and cost optimization across production pipelines. Lead multiple teams in a fast-paced CTO environment at JPMorgan Chase, shaping data architectures and ensuring SLAs and readiness for production.

Location: Plano, TX, United States

This is your chance to change the path of your career and guide multiple teams to success at one of the world's leading financial institutions.

As a Lead Software Engineer at JPMorgan Chase within Corporate Sector, Chief Technology Office, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives.

Job Responsibilities:

  • Lead architecture and delivery of high-throughput, low-latency data pipelines using Databricks and Apache Spark (Core, SQL, Structured Streaming).
  • Establish lakehouse patterns with Delta Lake (ACID transactions, schema evolution, time travel, Z-ordering, compaction) and ensure performance at scale.
  • Own Databricks cluster strategy and setup: runtime selection, autoscaling, driver/executor sizing, Spark configs, init scripts, cluster policies, pools, and instance profiles.
  • Orchestrate jobs with Databricks Workflows; integrate with AWS eventing and orchestration as needed.
  • Design secure data ingestion and transformation frameworks leveraging AWS services:
    • S3 for data lake storage and lifecycle management
    • Glue for catalog/metadata and ETL jobs
    • IAM and Secrets Manager for role-based access and credential management
    • CloudWatch for logging, metrics, and alerting
    • Lambda for serverless utilities
    • Kinesis and/or Kafka/MSK for streaming ingestion
  • Enforce data quality, lineage, and governance using Unity Catalog and/or Glue Catalog; embed expectations and validation into pipelines.
  • Drive Spark performance engineering: partitioning strategies, file sizing, AQE, broadcast joins, shuffle tuning, caching, spill/memory control, and job right-sizing to optimize cost.
  • Build reusable libraries, frameworks, and APIs in Python and/or Java; oversee unit, integration, and data validation testing.
  • Implement CI/CD for data projects (Git-based workflows), Terraform Infrastructure deployments environment promotion, and automated deployments; champion engineering standards and code reviews.
  • Partner with platform security and networking teams to enforce encryption, network controls, and least-privilege access; ensure compliance with organizational policies.
  • Lead incident response and root-cause analysis; establish SLAs, observability, and runbooks; drive continuous improvement in reliability and cost efficiency.

Required qualifications, capabilities, and skills:

  • Formal training or certification on software engineering concepts and 5+ years applied experience. 
  • 10+ years of professional software/data engineering experience, including substantial production work with Spark on Databricks or EMR.
  • Strong proficiency in Python and/or Java for data processing, platform tooling, and automation.
  • Hands-on Databricks expertise (Delta Lake, Unity Catalog, Workflows, Repos/notebooks, SQL Warehouses).
  • Solid AWS experience: S3, IAM, Glue, CloudWatch, Kinesis / MSK, DynamoDB
  • Proven track record architecting and operating ETL/ELT pipelines (batch and streaming), with schema design/evolution, SLAs, and reliability engineering.
  • Deep skills in Spark performance tuning and Databricks cluster setup/optimization.
  • Strong SQL and analytics data modeling (dimensional/star schema; lakehouse best practices).
  • Security-first mindset: roles/instance profiles, secret management, encryption-at-rest/in-transit, and network controls.
  • Demonstrated leadership in code quality, reviews, testing strategy, CI/CD, and technical mentorship; excellent communication with stakeholders.

Preferred qualifications, capabilities, and skills:

  • Experience with Delta Live Tables and advanced governance (catalogs, grants, auditing) in Databricks.
  • AWS networking knowledge (VPC, subnets, routing, security groups) and data egress controls.
  • Experience with Terraform for Infra deployments
  • Cost optimization experience: autoscaling strategies, spot vs on-demand, auto-termination, storage layouts and compaction.
  • Familiarity with Kafka/MSK or Kinesis Data Streams/Firehose for real-time ingestion.
  • CI/CD and automation tooling for data (Git workflows, artifact management) and testing frameworks (pytest, JUnit).
  • Observability for data systems (freshness/completeness metrics, lineage, SLAs, alerting).
  • Experience in financial services or other regulated industries.
We are looking for a Lead Software Engineer with a strong focus on Data engineering with AWS Cloud and Databricks Data pipelines.

Lead Software Engineer - Databricks, Spark, AWS

at J.P. Morgan

Back to all Data Engineering jobs
J.P. Morgan logo
Industry not specified

Lead Software Engineer - Databricks, Spark, AWS

at J.P. Morgan

Tech LeadNo visa sponsorshipData Engineering

Posted 17 hours ago

No clicks

Compensation
Not specified USD

Currency: $ (USD)

City
Not specified
Country
United States

Lead Software Engineer responsible for architecting and delivering high-throughput, low-latency data pipelines using Databricks and Apache Spark, with a focus on lakehouse patterns and scalable data platforms. Own Databricks cluster strategy, job orchestration with Databricks Workflows, and integration with AWS services (S3, Glue, IAM, CloudWatch, Kinesis/MSK) for secure data ingestion and processing. Develop reusable libraries in Python/Java, implement CI/CD for data projects, enforce data governance and security, and drive reliability, observability, and cost optimization across production pipelines. Lead multiple teams in a fast-paced CTO environment at JPMorgan Chase, shaping data architectures and ensuring SLAs and readiness for production.

Location: Plano, TX, United States

This is your chance to change the path of your career and guide multiple teams to success at one of the world's leading financial institutions.

As a Lead Software Engineer at JPMorgan Chase within Corporate Sector, Chief Technology Office, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives.

Job Responsibilities:

  • Lead architecture and delivery of high-throughput, low-latency data pipelines using Databricks and Apache Spark (Core, SQL, Structured Streaming).
  • Establish lakehouse patterns with Delta Lake (ACID transactions, schema evolution, time travel, Z-ordering, compaction) and ensure performance at scale.
  • Own Databricks cluster strategy and setup: runtime selection, autoscaling, driver/executor sizing, Spark configs, init scripts, cluster policies, pools, and instance profiles.
  • Orchestrate jobs with Databricks Workflows; integrate with AWS eventing and orchestration as needed.
  • Design secure data ingestion and transformation frameworks leveraging AWS services:
    • S3 for data lake storage and lifecycle management
    • Glue for catalog/metadata and ETL jobs
    • IAM and Secrets Manager for role-based access and credential management
    • CloudWatch for logging, metrics, and alerting
    • Lambda for serverless utilities
    • Kinesis and/or Kafka/MSK for streaming ingestion
  • Enforce data quality, lineage, and governance using Unity Catalog and/or Glue Catalog; embed expectations and validation into pipelines.
  • Drive Spark performance engineering: partitioning strategies, file sizing, AQE, broadcast joins, shuffle tuning, caching, spill/memory control, and job right-sizing to optimize cost.
  • Build reusable libraries, frameworks, and APIs in Python and/or Java; oversee unit, integration, and data validation testing.
  • Implement CI/CD for data projects (Git-based workflows), Terraform Infrastructure deployments environment promotion, and automated deployments; champion engineering standards and code reviews.
  • Partner with platform security and networking teams to enforce encryption, network controls, and least-privilege access; ensure compliance with organizational policies.
  • Lead incident response and root-cause analysis; establish SLAs, observability, and runbooks; drive continuous improvement in reliability and cost efficiency.

Required qualifications, capabilities, and skills:

  • Formal training or certification on software engineering concepts and 5+ years applied experience. 
  • 10+ years of professional software/data engineering experience, including substantial production work with Spark on Databricks or EMR.
  • Strong proficiency in Python and/or Java for data processing, platform tooling, and automation.
  • Hands-on Databricks expertise (Delta Lake, Unity Catalog, Workflows, Repos/notebooks, SQL Warehouses).
  • Solid AWS experience: S3, IAM, Glue, CloudWatch, Kinesis / MSK, DynamoDB
  • Proven track record architecting and operating ETL/ELT pipelines (batch and streaming), with schema design/evolution, SLAs, and reliability engineering.
  • Deep skills in Spark performance tuning and Databricks cluster setup/optimization.
  • Strong SQL and analytics data modeling (dimensional/star schema; lakehouse best practices).
  • Security-first mindset: roles/instance profiles, secret management, encryption-at-rest/in-transit, and network controls.
  • Demonstrated leadership in code quality, reviews, testing strategy, CI/CD, and technical mentorship; excellent communication with stakeholders.

Preferred qualifications, capabilities, and skills:

  • Experience with Delta Live Tables and advanced governance (catalogs, grants, auditing) in Databricks.
  • AWS networking knowledge (VPC, subnets, routing, security groups) and data egress controls.
  • Experience with Terraform for Infra deployments
  • Cost optimization experience: autoscaling strategies, spot vs on-demand, auto-termination, storage layouts and compaction.
  • Familiarity with Kafka/MSK or Kinesis Data Streams/Firehose for real-time ingestion.
  • CI/CD and automation tooling for data (Git workflows, artifact management) and testing frameworks (pytest, JUnit).
  • Observability for data systems (freshness/completeness metrics, lineage, SLAs, alerting).
  • Experience in financial services or other regulated industries.
We are looking for a Lead Software Engineer with a strong focus on Data engineering with AWS Cloud and Databricks Data pipelines.

SIMILAR OPPORTUNITIES

No similar jobs available at the moment.