refresh

Trending Companies

Trending

Jobs

JobsDatabricks

Software Engineer - GenAI inference

Databricks

Software Engineer - GenAI inference

Databricks

San Francisco, California

·

On-site

·

Full-time

·

1mo ago

Compensation

$142,200 - $204,600

Benefits & Perks

Flexible PTO policy

Remote work flexibility

Health, dental, and vision coverage

Annual team offsites

Learning and development stipend

Required Skills

TensorFlow

Python

Airflow

P-1284

About This Role

As a software engineer for GenAI inference, you will help design, develop, and optimize the inference engine that powers Databricks’ Foundation Model API. You’ll work at the intersection of research and production, ensuring our large language model (LLM) serving systems are fast, scalable, and efficient. Your work will touch the full GenAI inference stack — from kernels and runtimes to orchestration and memory management.

What You Will Do

  • Contribute to the design and implementation of the inference engine, and collaborate on model-serving stack optimized for large-scale LLMs inference

  • Collaborate with researchers to bring new model architectures or features (sparsity, activation compression, mixture-of-experts) into the engine

  • Optimize for latency, throughput, memory efficiency, and hardware utilization across GPUs, and accelerators

  • Build and maintain instrumentation, profiling, and tracing tooling to uncover bottlenecks and guide optimizations

  • Develop and enhance scalable routing, batching, scheduling, memory management, and dynamic loading mechanisms for inference workloads

  • Support reliability, reproducibility, and fault tolerance in the inference pipelines, including A/B launches, rollback, and model versioning

  • Integrate with federated, distributed inference infrastructure – orchestrate across nodes, balance load, handle communication overhead

  • Collaborate cross-functionally: with platform engineers, cloud infrastructure, and security/compliance teams

  • Document and share learnings, contributing to internal best practices and open-source efforts when possible

What We Look For

  • BS/MS/PhD in Computer Science, or a related field

  • Strong software engineering background (3+ years or equivalent) in performance-critical systems

  • Solid understanding of ML inference internals: attention, MLPs, recurrent modules, quantization, sparse operations, etc.

  • Hands-on experience with CUDA, GPU programming, and key libraries (cuBLAS, cuDNN, NCCL, etc.)

  • Comfortable designing and operating distributed systems, including RPC frameworks, queuing, RPC batching, sharding, memory partitioning

  • Demonstrated ability to uncover and solve performance bottlenecks across layers (kernel, memory, networking, scheduler)

  • Experience building instrumentation, tracing, and profiling tools for ML models

  • Ability to work closely with ML researchers, translate novel model ideas into production systems

  • Ownership mindset and eagerness to dive deep into complex system challenges

  • Bonus: published research or open-source contributions in ML systems, inference optimization, or model serving

Pay Range Transparency

Databricks is committed to fair and equitable compensation practices. The pay range(s) for this role is listed below and represents the expected salary range for non-commissionable roles or on-target earnings for commissionable roles. Actual compensation packages are based on several factors that are unique to each candidate, including but not limited to job-related skills, depth of experience, relevant certifications and training, and specific work location. Based on the factors above, Databricks anticipates utilizing the full width of the range. The total compensation package for this position may also include eligibility for annual performance bonus, equity, and the benefits listed above. For more information regarding which range your location is in visit our page here.

Local Pay Range:

$142,200—$204,600 USD

About Databricks

Databricks is the data and AI company. More than 10,000 organizations worldwide — including Comcast, Condé Nast, Grammarly, and over 50% of the Fortune 500 — rely on the Databricks Data Intelligence Platform to unify and democratize data, analytics and AI. Databricks is headquartered in San Francisco, with offices around the globe and was founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow. To learn more, follow Databricks on Twitter, LinkedIn and Facebook.

Benefits:

At Databricks, we strive to provide comprehensive benefits and perks that meet the needs of all of our employees. For specific details on the benefits offered in your region, please visit https://www.mybenefitsnow.com/databricks.

Our Commitment to Diversity and Inclusion

At Databricks, we are committed to fostering a diverse and inclusive culture where everyone can excel. We take great care to ensure that our hiring practices are inclusive and meet equal employment opportunity standards. Individuals looking for employment at Databricks are considered without regard to age, color, disability, ethnicity, family or marital status, gender identity or expression, language, national origin, physical and mental ability, political affiliation, race, religion, sexual orientation, socio-economic status, veteran status, and other protected characteristics.

Compliance If access to export-controlled technology or source code is required for performance of job duties, it is within Employer's discretion whether to apply for a U.S. government license for such positions, and Employer may decline to proceed with an applicant on this basis alone.

Total Views

0

Apply Clicks

0

Mock Applicants

0

Scraps

0

About Databricks

Databricks

Databricks

Series I

Databricks, Inc. is an American software company based in San Francisco. It was founded in 2013 by the original creators of Apache Spark. It offers a cloud-based platform for data analytics and artificial intelligence.

6,000+

Employees

San Francisco

Headquarters

$43B

Valuation

Reviews

4.2

9 reviews

Work Life Balance

3.5

Compensation

4.7

Culture

4.3

Career

4.5

Management

4.0

86%

Recommend to a Friend

Pros

Working on industry-leading data and AI platform

Excellent compensation with high equity upside

Strong engineering culture with Apache Spark creators

Cons

High intensity work environment with demanding deadlines

Work-life balance can suffer during key releases

Growing pains as company scales rapidly

Salary Ranges

25 data points

Junior/L3

Mid/L4

Staff/L6

Junior/L3 · Data Scientist L3

0 reports

$244,818

total / year

Base

-

Stock

-

Bonus

-

$208,096

$281,540

Interview Experience

9 interviews

Difficulty

3.0

/ 5

Duration

21-35 weeks

Offer Rate

22%

Experience

Positive 22%

Neutral 67%

Negative 11%

Interview Process

1

Application Review

2

Recruiter/Phone Screen

3

Technical Interview/Coding Round

4

System Design Interview

5

Behavioral Interview

6

Final Round/Hiring Manager Interview

Common Questions

Coding/Algorithm

System Design

Behavioral/STAR

Technical Knowledge