採用
必須スキル
Python
Terraform
AWS SageMaker
Kubernetes
MLOps
SRE
Docker
At Navan, we aren't building a single, generic chatbot. We are building a Composable AI Microservice Architecture, a swarm of hundreds of hyper-specialized AI services, each meticulously "programmed" to solve small, focused tasks with high precision. This fleet powers Ava, our AI support engine, and a suite of cutting-edge generative tools for travel and expense management.
As a Senior AI Operations (AI Ops) Engineer, you are the architect of the platform that makes this scale possible. You will move beyond traditional MLOps to manage a "factory" of Language Models. Your challenge is one of orchestration and standardization, ensuring that every service in the swarm meets a rigorous bar for quality, reliability, and cost-efficiency.
What You’ll Do
-
Orchestrate the AI Fleet: Build and own the runtime environment for 100+ specialized AI services. Manage model routing, context versioning, and standardized memory/history stores.
-
High-Density Inference Optimization: Design and implement Sage Maker Multi-Model Endpoints (MME) and Inference Components to serve multiple tuned SLMs per GPU, maximizing hardware utilization while minimizing latency.
-
Deterministic Service Excellence: Treat reliability as a layered engineering problem. Build deterministic "shells" around probabilistic LM outputs, prioritizing data-layer validation and strict serialization.
-
Automated Evaluation & Observability: Implement "LLM-as-a-judge" patterns and automated benchmarking to detect semantic drift and hallucinations across the fleet before they impact the user.
-
Standardize the Workflow: Obsess over building reusable patterns and Terraform-based infrastructure that eliminate "snowflake" configurations, allowing us to deploy new specialized AI tasks in minutes.
-
Agency Strategy: Partner with AI Researchers to find the "Goldilocks zone" for agentic autonomy—balancing the flexibility of LLM tool-use with the precision required for production stability.
What We’re Looking For
-
Experience: 5+ years in SRE, Platform Engineering, or MLOps, with at least 2 years focused on deploying LLMs/SLMs in production environments.
-
Sage Maker Mastery: Deep hands-on expertise with AWS Sage Maker, specifically configuring Multi-Model Endpoints (MME), Inference Components, and GPU-backed instances (G5/P4).
-
SLM Expertise: Proven experience with Small Language Models (e.g., Mistral, Llama 3, Phi) and parameter-efficient fine-tuning (PEFT) deployment strategies like LoRA/QLoRA.
-
Technical Stack: * *Languages: Strong proficiency in Python and Terraform.
-
Orchestration: Experience with Docker, Kubernetes (EKS), or AWS ECS/Fargate.
-
Data: Familiarity with Snowflake and Vector Databases.
-
The "AI Ops" Mindset: You understand that AI at scale is a statistical challenge. You are comfortable debugging issues at the data/serialization layer rather than defaulting to prompt tweaks.
-
CI/CD & Automation: Experience building robust pipelines (Jenkins, GitHub Actions) for non-deterministic software, including automated "eval" stages.
-
Education: BS or MS in Computer Science, Engineering, Mathematics, or a related technical field.
The posted pay range represents the anticipated low and high end of the compensation for this position and is subject to change based on business need. To determine a successful candidate’s starting pay, we carefully consider a variety of factors, including primary work location, an evaluation of the candidate’s skills and experience, market demands, and internal parity.
For roles with on-target-earnings (OTE), the pay range includes both base salary and target incentive compensation. Target incentive compensation for some roles may include a ramping draw period. Compensation is higher for those who exceed targets. Candidates may receive more information from the recruiter.
Pay Range**$116,100—$258,000 USD**
総閲覧数
0
応募クリック数
0
模擬応募者数
0
スクラップ
0
類似の求人

Senior AI Systems Performance Engineer
SambaNova · Palo Alto, California, United States

Sr. Electrical Engineer, Power Electronics
Tesla · Palo Alto, California

Staff Android Framework Engineer - GNSS HAL & Energy HAL
Ford · Palo Alto, CA, United States, US

Principal Compiler Engineer - ML Systems
SambaNova · Palo Alto, California, United States

Principal Strategic Architect
Workato · Palo Alto, California
Navanについて

Navan
Series F+Navan is a corporate travel and expense management platform that combines travel booking, expense reporting, and payment solutions for businesses.
1,001-5,000
従業員数
Palo Alto
本社所在地
$9.2B
企業価値
レビュー
3.9
10件のレビュー
ワークライフバランス
3.5
報酬
2.8
企業文化
4.2
キャリア
3.0
経営陣
2.5
72%
友人に勧める
良い点
Flexible work hours
Great team and colleagues
Good culture and inclusive workplace
改善点
Poor compensation/salary
Heavy workload
Poor management and communication
給与レンジ
42件のデータ
Junior/L3
Mid/L4
Junior/L3 · Data Analyst
0件のレポート
$169,150
年収総額
基本給
-
ストック
-
ボーナス
-
$143,778
$194,522
面接体験
2件の面接
難易度
3.5
/ 5
期間
14-28週間
体験
ポジティブ 0%
普通 50%
ネガティブ 50%
面接プロセス
1
Application Review
2
Recruiter Screen
3
Online Assessment
4
Technical Phone Screen
5
Onsite/Virtual Interviews
6
Offer
よくある質問
Coding/Algorithm
Technical Knowledge
Behavioral/STAR
System Design
ニュース&話題
Navan Shareholder Alert: ClaimsFiler Reminds Investors With Losses In Excess Of $100,000 Of Lead Plaintiff Deadline In Class Action Lawsuit Against Navan, Inc. - NAVN - PR Newswire
PR Newswire
News
·
3d ago
Navan announces departure of chief accounting officer, CFO to assume duties - Investing.com
Investing.com
News
·
3d ago
Navan Announces Finance Leadership Transition and Role Consolidation - TipRanks
TipRanks
News
·
3d ago
ARC, Navan, Travelfusion and SAP Concur Join NDC FastTrack - Business Travel Executive
Business Travel Executive
News
·
4d ago