refresh

トレンド企業

トレンド企業

採用

求人JPMorgan Chase

SR Principal Software Engineer - LLM Engineering

JPMorgan Chase

SR Principal Software Engineer - LLM Engineering

JPMorgan Chase

Palo Alto, CA, United States, US

·

On-site

·

Full-time

·

4w ago

必須スキル

AWS

GCP

Azure

We’re looking for a tech leader ready to take their career to new heights. Join the ranks of top talent at one of the world’s most influential companies.

As a Senior Principal Software Engineer at JPMorgan Chase within the Commercial & Investment Bank Trust & Safety Fraud Prevention team, you provide deep engineering expertise and work across agile teams to enhance, build, and deliver trusted market‑leading technology products in a secure, stable, and scalable way. Leverage your deep expertise to consistently challenge the status quo, innovate for business impact, lead the strategic development behind new and existing products and technology portfolios, and remain at the forefront of industry trends, best practices, and technological advances.

Job responsibilities

  • Advises and leads on the strategy, architecture, and development of Model serving solutions for different model architectures including LLMs & GNNs, across cloud and on‑premises environments, aligning initiatives to business outcomes.
  • Defines and implements MLOps and LLMOps strategies for end‑to‑end model lifecycle management, including training, versioning, deployment, monitoring, and governance.
  • Drives optimization of Model inferencing for high throughput and low latency using quantization, model parallelism, intelligent batching, and hardware acceleration for all model architectures
  • Creates durable, reusable software and platform frameworks to standardize ML Engineering services, enabling scale across teams and functions.
  • Establishes best practices for automation, CI/CD, and infrastructure‑as‑code using containerization and orchestration technologies.
  • Partners closely with data science, platform engineering, and SRE teams to productionize the models on AWS, ensuring observability, reliability, and cost efficiency.
  • Leads deployment and optimization using Model Inference servers such as Triton Inference Server and vLLM for high‑throughput, low‑latency serving at scale.
  • Oversees production operations for AI workloads, including monitoring, incident response, security, and compliance, with continuous improvement.
  • Translates highly complex technical concepts and emerging trends into actionable strategies for executive and product leadership.
  • Influences senior stakeholders and cross‑functional partners to prioritize and deliver AI/ML capabilities that drive measurable business impact.
  • Promotes the firm’s culture of diversity, opportunity, inclusion, and respect across teams and communities.

Required qualifications, capabilities, and skills

  • Formal training or certification on software engineering concepts and 10+ years of applied experience.
  • 8+ years of AI/ML engineering experience with significant expertise in LLMs, GNNs and other model architectures (e.g., GPT, Llama, Falcon, Mistral).
  • Demonstrated success architecting and deploying LLM & GNN solutions on AWS (e.g., Sage Maker, Bedrock, EKS) at enterprise scale; experience with Azure ML or GCP Vertex AI.
  • Experience building LLM, GNN serving platforms in large‑scale environments typical of major tech firms.
  • Hands‑on experience building LLM inference engines using Triton Inference Server and vLLM, including autoscaling, caching, and throughput optimization.
  • Advanced proficiency in Python and optimization techniques applied to deep learning frameworks (Py Torch, Tensor Flow, Hugging Face Transformers).
  • Deep understanding of LLMOps/MLOps (e.g., MLflow, Sage Maker Pipelines, Kubeflow) with a track record of implementing best practices at scale.
  • Expertise in inference optimization and distributed systems for large models focused on high‑throughput, low‑latency applications.
  • Practical experience delivering system design, application development, testing, and operational stability for enterprise AI platforms.
  • Proven collaboration with SRE to implement observability, incident response, and SLIs/SLOs for LLM services.
  • Excellent communication skills with the ability to influence both technical and non‑technical stakeholders and deliver value across functions at scale.

Preferred qualifications, capabilities, and skills

  • Master’s or PhD in Computer Science, Engineering, or a related field (or equivalent experience).
  • Practical cloud‑native experience, including containerization (Docker), orchestration (Kubernetes), and infrastructure‑as‑code (Terraform, CloudFormation).
  • Expertise in security, compliance, and governance for AI/ML deployments in regulated environments.
  • Experience in trust and safety or fraud prevention domains; familiarity with payments platforms is a plus.
  • Track record of contributions to open‑source LLM projects or peer‑reviewed research and/or experience presenting at industry conferences or leading technical communities.
  • Familiarity with hardware acceleration strategies across GPUs, TPUs, and specialized inference runtimes.
  • Experience in building java based applications

This position is subject to Section 19 of the Federal Deposit Insurance Act. As such, an employment offer for this position is contingent on JPMorgan Chase’s review of criminal conviction history, including pretrial diversions or program entries.

総閲覧数

0

応募クリック数

0

模擬応募者数

0

スクラップ

0

JPMorgan Chaseについて

JPMorgan Chase

JPMorgan Chase & Co. is an American multinational banking institution headquartered in New York City and incorporated in Delaware. It is the largest bank in the United States, and the world's largest bank by market capitalization as of 2025.

300,000+

従業員数

New York City

本社所在地

$500B

企業価値

レビュー

3.8

10件のレビュー

ワークライフバランス

3.2

報酬

4.1

企業文化

3.8

キャリア

3.0

経営陣

2.5

65%

友人に勧める

良い点

Good benefits and compensation

Supportive and collaborative environment

Flexible work arrangements

改善点

Long hours and heavy workload

Management issues and lack of direction

High stress during peak times

給与レンジ

41件のデータ

Junior/L3

Mid/L4

Senior/L5

Junior/L3 · Analytics Solutions Associate

1件のレポート

$139,000

年収総額

基本給

$107,000

ストック

-

ボーナス

-

$139,000

$139,000

面接体験

5件の面接

難易度

3.0

/ 5

期間

14-28週間

内定率

40%

体験

ポジティブ 20%

普通 80%

ネガティブ 0%

面接プロセス

1

Application Review

2

HireVue Video Interview

3

Recruiter Screen

4

Superday/Panel Interview

5

Final Interview

6

Offer

よくある質問

Behavioral/STAR

Technical Knowledge

Culture Fit

Past Experience

Case Study