refresh

热门公司

Trending

招聘

JobsJPMorgan Chase

Lead Software Engineer - Databricks, Spark, AWS

JPMorgan Chase

Lead Software Engineer - Databricks, Spark, AWS

JPMorgan Chase

Plano, TX, United States, US

·

On-site

·

Full-time

·

5d ago

This is your chance to change the path of your career and guide multiple teams to success at one of the world's leading financial institutions.

As a Lead Software Engineer at JPMorgan Chase within Corporate Sector, Chief Technology Office, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives.

Job Responsibilities:

  • Lead architecture and delivery of high-throughput, low-latency data pipelines using Databricks and Apache Spark (Core, SQL, Structured Streaming).
  • Establish lakehouse patterns with Delta Lake (ACID transactions, schema evolution, time travel, Z-ordering, compaction) and ensure performance at scale.
  • Own Databricks cluster strategy and setup: runtime selection, autoscaling, driver/executor sizing, Spark configs, init scripts, cluster policies, pools, and instance profiles.
  • Orchestrate jobs with Databricks Workflows; integrate with AWS eventing and orchestration as needed.
  • Design secure data ingestion and transformation frameworks leveraging AWS services:S3 for data lake storage and lifecycle management
  • Glue for catalog/metadata and ETL jobs
  • IAM and Secrets Manager for role-based access and credential management
  • CloudWatch for logging, metrics, and alerting
  • Lambda for serverless utilities
  • Kinesis and/or Kafka/MSK for streaming ingestion
  • Enforce data quality, lineage, and governance using Unity Catalog and/or Glue Catalog; embed expectations and validation into pipelines.
  • Drive Spark performance engineering: partitioning strategies, file sizing, AQE, broadcast joins, shuffle tuning, caching, spill/memory control, and job right-sizing to optimize cost.
  • Build reusable libraries, frameworks, and APIs in Python and/or Java; oversee unit, integration, and data validation testing.
  • Implement CI/CD for data projects (Git-based workflows), Terraform Infrastructure deployments environment promotion, and automated deployments; champion engineering standards and code reviews.
  • Partner with platform security and networking teams to enforce encryption, network controls, and least-privilege access; ensure compliance with organizational policies.
  • Lead incident response and root-cause analysis; establish SLAs, observability, and runbooks; drive continuous improvement in reliability and cost efficiency.

Required qualifications, capabilities, and skills:

  • Formal training or certification on software engineering concepts and 5+ years applied experience.
  • 10+ years of professional software/data engineering experience, including substantial production work with Spark on Databricks or EMR.
  • Strong proficiency in Python and/or Java for data processing, platform tooling, and automation.
  • Hands-on Databricks expertise (Delta Lake, Unity Catalog, Workflows, Repos/notebooks, SQL Warehouses).
  • Solid AWS experience: S3, IAM, Glue, CloudWatch, Kinesis / MSK, DynamoDB
  • Proven track record architecting and operating ETL/ELT pipelines (batch and streaming), with schema design/evolution, SLAs, and reliability engineering.
  • Deep skills in Spark performance tuning and Databricks cluster setup/optimization.
  • Strong SQL and analytics data modeling (dimensional/star schema; lakehouse best practices).
  • Security-first mindset: roles/instance profiles, secret management, encryption-at-rest/in-transit, and network controls.
  • Demonstrated leadership in code quality, reviews, testing strategy, CI/CD, and technical mentorship; excellent communication with stakeholders.

Preferred qualifications, capabilities, and skills:

  • Experience with Delta Live Tables and advanced governance (catalogs, grants, auditing) in Databricks.
  • AWS networking knowledge (VPC, subnets, routing, security groups) and data egress controls.
  • Experience with Terraform for Infra deployments
  • Cost optimization experience: autoscaling strategies, spot vs on-demand, auto-termination, storage layouts and compaction.
  • Familiarity with Kafka/MSK or Kinesis Data Streams/Firehose for real-time ingestion.
  • CI/CD and automation tooling for data (Git workflows, artifact management) and testing frameworks (pytest, JUnit).
  • Observability for data systems (freshness/completeness metrics, lineage, SLAs, alerting).
  • Experience in financial services or other regulated industries.

Total Views

0

Apply Clicks

0

Mock Applicants

0

Scraps

0

About JPMorgan Chase

JPMorgan Chase

JPMorgan Chase is a multinational investment bank and financial services company that provides banking, investment, and asset management services globally. It is one of the largest banks in the United States by assets and market capitalization.

300,000+

Employees

New York City

Headquarters

Reviews

4.2

10 reviews

Work Life Balance

4.2

Compensation

4.3

Culture

4.5

Career

4.4

Management

4.1

75%

Recommend to a Friend

Pros

Good pay and benefits

Work-life balance

Career advancement opportunities

Cons

Heavy workload at times

Career advancement takes time

Pay could be better in some roles

Salary Ranges

47 data points

Junior/L3

Mid/L4

Senior/L5

Junior/L3 · Analyst

21 reports

$126,500

total / year

Base

$110,000

Stock

-

Bonus

-

$95,450

$155,250

Interview Experience

4 interviews

Difficulty

2.8

/ 5

Duration

14-28 weeks

Interview Process

1

Application Review

2

HireVue Video Interview

3

Technical/Behavioral Assessment

4

Final Interview Round

5

Offer Decision

Common Questions

Behavioral/STAR

Technical Knowledge

Past Experience

Culture Fit

Case Study