refresh

트렌딩 기업

트렌딩

채용

JobsApple

Senior Software Engineer, Model Inference

Apple

Senior Software Engineer, Model Inference

Apple

Reedley, CA

·

On-site

·

Full-time

·

1mo ago

Benefits & Perks

Competitive salary and equity package

Flexible work arrangements

Comprehensive health, dental, and vision insurance

Generous paid time off and holidays

Team events and activities

Professional development budget

Equity

Flexible Hours

Healthcare

Learning

Required Skills

JavaScript

React

PostgreSQL

About the Role

Join Apple Maps to help build the best map in the world. In this role on ML Platform, you will help bring advanced deep learning and large language models into high-volume, low-latency, highly available production serving, improving search quality and powering experiences across Maps. You will partner closely with research and product teams, take end-to-end ownership, and deliver measurable results at global scale.

As a Software Engineer on the Apple Maps team, you will lead the design and implementation of large-scale, high-performance inference services that support a wide range of models used across Maps, including deep learning and large language models. You will collaborate closely with research and product partners to bring models into production, with a strong focus on efficiency, reliability, and scalability. Your responsibilities span the full server stack, including onboarding new use cases, optimizing inference across heterogeneous accelerated compute hardware, deploying services on Kubernetes, building and integrating inference engines and control-plane components, and ensuring seamless integration with Maps infrastructure.

Responsibilities

  • Own the technical architecture of large-scale ML inference platforms, defining long-term design direction for serving deep learning and large language models across Apple Maps.
  • Lead system-level optimization efforts across the inference stack, balancing latency, throughput, accuracy, and cost through advanced techniques such as quantization, kernel fusion, speculative decoding, and efficient runtime scheduling.
  • Design and evolve control-plane services responsible for model lifecycle management, including deployment orchestration, versioning, traffic routing, rollout strategies, capacity planning, and failure handling in production environments.
  • Drive adoption of platform abstractions and standards that enable partner teams to onboard, deploy, and operate models reliably and efficiently at scale.
  • Partner closely with research, product, and infrastructure teams to translate model requirements into production-ready systems, providing technical guidance and feedback to influence upstream model design.
  • Optimize inference execution across heterogeneous compute environments, including GPUs and specialized accelerators, collaborating with runtime, compiler, and kernel teams to maximize hardware utilization.
  • Establish robust observability and performance diagnostics, defining metrics, dashboards, and profiling workflows to proactively identify bottlenecks and guide optimization decisions.
  • Provide technical leadership and mentorship, reviewing designs, setting engineering best practices, and raising the quality bar across teams contributing to the inference ecosystem.
  • Continuously evaluate emerging research and industry trends in LLM inference, distributed systems, and ML infrastructure, driving the transition of high-impact ideas into production systems.

Minimum Qualifications

  • Bachelor's degree in Computer Science, Engineering, or related field (or equivalent experience).
  • 5+ years in software engineering focused on ML inference, GPU acceleration, and large-scale systems.
  • Expertise in deploying and optimizing LLMs for high-performance, production-scale inference.
  • Proficiency in Python, Java or C++.
  • Experience with deep learning frameworks like Py Torch, Tensor Flow, and Hugging Face Transformers.
  • Experience with model serving tools (e.g., NVIDIA Triton, Tensor Flow Serving, VLLM, etc).
  • Experience with optimization techniques like Attention Fusion, Quantization, and Speculative Decoding.
  • Skilled in GPU optimization (e.g., CUDA, TensorRT-LLM, cuDNN) to accelerate inference tasks.
  • Skilled in cloud technologies like Kubernetes, Ingress, HAProxy for scalable deployment.

Preferred Qualifications

  • Master's or PhD in Computer Science, Machine Learning, or a related field.
  • Understanding of ML Ops practices, continuous integration, and deployment pipelines for machine learning models.
  • Familiarity with model distillation, low-rank approximations, and other model compression techniques for reducing memory footprint and improving inference speed.
  • Strong understanding of distributed systems, multi-GPU/multi-node parallelism, and system-level optimization for large-scale inference.

Equal Opportunity

Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant.

Total Views

0

Apply Clicks

0

Mock Applicants

0

Scraps

0

About Apple

Apple

Apple

Public

A technology company that designs, manufactures, and markets consumer electronics, personal computers, and software.

10,001+

Employees

Cupertino

Headquarters

$3.5T

Valuation

Reviews

4.0

10 reviews

Work Life Balance

4.0

Compensation

4.2

Culture

3.8

Career

3.5

Management

3.2

75%

Recommend to a Friend

Pros

Great coworkers and people

Excellent benefits and perks

Fast-paced and engaging work environment

Cons

High expectations and pressure

Management quality varies

Limited career progression opportunities

Salary Ranges

17,968 data points

L2

L3

L4

L5

L6

M3

M4

M5

M6

L2 · Industrial Designer L2

0 reports

$320,450

total / year

Base

$128,180

Stock

$160,225

Bonus

$32,045

$224,315

$416,585

Interview Experience

5 interviews

Difficulty

3.4

/ 5

Duration

28-42 weeks

Offer Rate

20%

Experience

Positive 20%

Neutral 40%

Negative 40%

Interview Process

1

Application Review

2

Recruiter Screen

3

Technical Phone Screen

4

Behavioral Interview

5

Onsite/Virtual Interviews

6

Team Matching

7

Offer

Common Questions

Coding/Algorithm

System Design

Behavioral/STAR

Technical Knowledge

Culture Fit