채용

LLM Inference Frameworks and Optimization Engineer
San Francisco, Singapore, Amsterdam
·
On-site
·
Full-time
·
1mo ago
필수 스킬
Python
PyTorch
About the Role
At Together.ai, we are building state-of-the-art infrastructure to enable efficient and scalable inference for large language models (LLMs). Our mission is to optimize inference frameworks, algorithms, and infrastructure, pushing the boundaries of performance, scalability, and cost-efficiency.
We are seeking an Inference Frameworks and Optimization Engineer to design, develop, and optimize distributed inference engines that support multimodal and language models at scale. This role will focus on low-latency, high-throughput inference, GPU/accelerator optimizations, and software-hardware co-design, ensuring efficient large-scale deployment of LLMs and vision models.
This role offers a unique opportunity to shape the future of LLM inference infrastructure, ensuring scalable, high-performance AI deployment across a diverse range of applications. If you're passionate about pushing the boundaries of AI inference, we’d love to hear from you!
Responsibilities
Inference Framework Development and Optimization
-
Design and develop fault-tolerant, high-concurrency distributed inference engine for text, image, and multimodal generation models.
-
Implement and optimize distributed inference strategies, including Mixture of Experts (MoE) parallelism, tensor parallelism, pipeline parallelism for high-performance serving.
-
Apply CUDA graph optimizations, TensorRT/TRT-LLM graph optimizations, and Py Torch-based compilation (torch.compile), and speculative decoding to enhance efficiency and scalability.
Software-Hardware Co-Design and AI Infrastructure
-
Collaborate with hardware teams on performance bottleneck analysis, co-optimize inference performance for GPUs, TPUs, or custom accelerators.
-
Work closely with AI researchers and infrastructure engineers to develop efficient model execution plans and optimize E2E model serving pipelines.
Requirements
Must-Have:
Experience:
- 3+ years of experience in deep learning inference frameworks, distributed systems, or high-performance computing.
Technical Skills:
-
Familiar with at least one LLM inference frameworks (e.g., TensorRT-LLM, vLLM, SGLang, TGI(Text Generation Inference)).
-
Background knowledge and experience in at least one of the following: GPU programming (CUDA/Triton/TensorRT), compiler, model quantization, and GPU cluster scheduling.
-
Deep understanding of KV cache systems like Mooncake, Paged Attention, or custom in-house variants.
Programming:
- Proficient in Python and C++/CUDA for high-performance deep learning inference.
Optimization Techniques:
-
Deep understanding of Transformer architectures and LLM/VLM/Diffusion model optimization.
-
Knowledge of inference optimization, such as workload scheduling, CUDA graph, compiled, efficient kernels
Soft Skills:
-
Strong analytical problem-solving skills with a performance-driven mindset.
-
Excellent collaboration and communication skills across teams.
Nice-to-Have:
-
Experience in developing software systems for large-scale data center networks with RDMA/RoCE
-
Familiar with distributed filesystem(e.g., 3FS, HDFS, Ceph)
-
Familiar with open source distributed scheduling/orchestration frameworks, such as Kubernetes (K8S)
-
Contributions to open-source deep learning inference projects.
About Together AI
Together AI is a research-driven artificial intelligence company. We believe open and transparent AI systems will drive innovation and create the best outcomes for society, and together we are on a mission to significantly lower the cost of modern AI systems by co-designing software, hardware, algorithms, and models. We have contributed to leading open-source research, models, and datasets to advance the frontier of AI, and our team has been behind technological advancement such as Flash Attention, Hyena, Flex Gen, and Red Pajama. We invite you to join a passionate group of researchers in our journey in building the next generation AI infrastructure.
Compensation
We offer competitive compensation, startup equity, health insurance and other competitive benefits. The US base salary range for this full-time position is: $160,000 - $230,000 + equity + benefits. Our salary ranges are determined by location, level and role. Individual compensation will be determined by experience, skills, and job-related knowledge.
Equal Opportunity
Together AI is an Equal Opportunity Employer and is proud to offer equal employment opportunity to everyone regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, veteran status, and more.
Please see our privacy policy at https://www.together.ai/privacy
총 조회수
2
총 지원 클릭 수
0
모의 지원자 수
0
스크랩
0
비슷한 채용공고

Release Engineer, Consumer Products
OpenAI · San Francisco

Pipeline Engineer (Graphics/3D)
World Labs · San Francisco

Forward Deployed Engineer, GenAI
Scale AI · San Francisco, CA; New York, NY

Hardware / Software CoDesign Engineer
OpenAI · San Francisco

Test EA Sensitive Question Check
LinkedIn Learning · San Francisco, CA
Together AI 소개

Together AI
Series BData annotation company.
51-200
직원 수
San Francisco
본사 위치
$1.25B
기업 가치
리뷰
3.8
10개 리뷰
워라밸
3.5
보상
2.8
문화
4.2
커리어
3.0
경영진
3.2
65%
친구에게 추천
장점
Great team culture and collaboration
Flexible work arrangements and remote options
Good work-life balance
단점
Below industry standard compensation
High workload and overwhelming demands
Limited career advancement opportunities
연봉 정보
0개 데이터
Mid/L4
Senior
Mid/L4 · Product Designer
0개 리포트
$156,800
총 연봉
기본급
$156,800
주식
-
보너스
-
$133,280
$180,320
면접 경험
3개 면접
난이도
3.0
/ 5
소요 기간
14-28주
면접 과정
1
Application Review
2
Recruiter Screen
3
Technical Phone Screen
4
Coding Rounds
5
System Design Interview
6
Final Interview
자주 나오는 질문
Coding/Algorithm
System Design
Technical Knowledge
Behavioral/STAR
Infrastructure/SRE
뉴스 & 버즈
Amazon launches AI Store, showcasing range of AI-powered consumer devices - connectedtoindia.com
connectedtoindia.com
News
·
1w ago
Together AI - Forbes
Forbes
News
·
1w ago
Together AI lands massive new headquarters in San Francisco while 'in this hypergrowth phase' - The Business Journals
The Business Journals
News
·
1w ago
Annual TraceGains ‘Together’ Conference to Showcase AI, Connected Data in Food and Beverage Industry - Quality Assurance & Food Safety
Quality Assurance & Food Safety
News
·
1w ago