채용
WHAT YOU DO AT AMD CHANGES EVERYTHING
At AMD, our mission is to build great products that accelerate next-generation computing experiences—from AI and data centers, to PCs, gaming and embedded systems. Grounded in a culture of innovation and collaboration, we believe real progress comes from bold ideas, human ingenuity and a shared passion to create something extraordinary. When you join AMD, you’ll discover the real differentiator is our culture. We push the limits of innovation to solve the world’s most important challenges—striving for execution excellence, while being direct, humble, collaborative, and inclusive of diverse perspectives. Join us as we shape the future of AI and beyond. Together, we advance your career.THE ROLE
We are seeking a Principal GenAI Inference Optimization Engineer to join our Models and Applications team. This role focuses on improving performance, efficiency, and scalability of generative AI inference workloads on AMD GPU platforms. You will contribute to optimizing latency, throughput, and cost efficiency for real-world deployment of large-scale models, working across the software-hardware stack.
THE PERSON
The ideal candidate is a strong technical contributor with expertise in GenAI inference optimization, GPU performance, and large-scale serving systems. You have a solid understanding of GPU architecture, memory systems, and communication patterns, and can apply this knowledge to improve inference efficiency.
You are comfortable working across multiple layers—from kernels and runtimes to frameworks and serving systems—and can independently drive optimization efforts while collaborating with cross-functional teams.
KEY RESPONSIBILITIES
-
Optimize performance of GenAI inference workloads on AMD GPU platforms across single-node and distributed environments.
-
Improve latency, throughput, and cost efficiency for LLM and multimodal model serving in production.
-
Analyze and resolve bottlenecks across compute, memory, and communication (e.g., kernel efficiency, KV-cache usage, memory bandwidth, scheduling).
-
Contribute to cross-stack optimizations spanning kernels, runtimes, communication libraries, and inference/serving frameworks (e.g., vLLM, SGLang, Triton, or similar systems).
-
Implement and evaluate inference optimization techniques such as batching strategies, quantization, prefix caching, and speculative decoding.
-
Support development and optimization of scalable serving systems, including request scheduling and resource utilization.
-
Develop and use profiling, benchmarking, and performance analysis tools for inference workloads.
-
Collaborate with hardware, compiler, and framework teams to improve overall system performance.
-
Contribute to internal tools and, where applicable, open-source projects for inference optimization on AMD platforms.
-
Document best practices and contribute to performance guidelines for GenAI deployment.
PREFERRED EXPERIENCE
-
Strong understanding of GPU architecture and performance fundamentals (compute, memory hierarchy, interconnects such as PCIe/Infinity Fabric/RDMA).
-
Experience with GenAI inference optimization techniques (e.g., quantization, KV-cache optimization, batching).
-
Hands-on experience with inference/serving frameworks such as vLLM, SGLang, Triton, TensorRT-LLM, or similar.
-
Experience working on LLM or multimodal inference workloads.
-
Familiarity with distributed systems and serving architectures.
-
Experience with ML frameworks (Py Torch, JAX, or Tensor Flow), especially for inference.
-
Proficiency in Python and at least one systems language (C++/CUDA/HIP).
-
Experience with profiling, debugging, and performance tuning tools.
-
Ability to work collaboratively across teams and deliver impactful optimizations.
ACADEMIC CREDENTIALS
- B.S., M.S. or Ph.D. in Computer Science, Computer Engineering, or a related field preferred, or equivalent industry experience.
LOCATION
- San Jose, CA preferred. May consider other US locations where AMD has offices such as Bellevue, WA or Austin, TX.
Benefits offered are described: AMD benefits at a glance.
AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.
AMD may use Artificial Intelligence to help screen, assess or select applicants for this position. AMD’s “Responsible AI Policy” is available here.
This posting is for an existing vacancy.
THE ROLE
We are seeking a Principal GenAI Inference Optimization Engineer to join our Models and Applications team. This role focuses on improving performance, efficiency, and scalability of generative AI inference workloads on AMD GPU platforms. You will contribute to optimizing latency, throughput, and cost efficiency for real-world deployment of large-scale models, working across the software-hardware stack.
THE PERSON
The ideal candidate is a strong technical contributor with expertise in GenAI inference optimization, GPU performance, and large-scale serving systems. You have a solid understanding of GPU architecture, memory systems, and communication patterns, and can apply this knowledge to improve inference efficiency.
You are comfortable working across multiple layers—from kernels and runtimes to frameworks and serving systems—and can independently drive optimization efforts while collaborating with cross-functional teams.
KEY RESPONSIBILITIES
-
Optimize performance of GenAI inference workloads on AMD GPU platforms across single-node and distributed environments.
-
Improve latency, throughput, and cost efficiency for LLM and multimodal model serving in production.
-
Analyze and resolve bottlenecks across compute, memory, and communication (e.g., kernel efficiency, KV-cache usage, memory bandwidth, scheduling).
-
Contribute to cross-stack optimizations spanning kernels, runtimes, communication libraries, and inference/serving frameworks (e.g., vLLM, SGLang, Triton, or similar systems).
-
Implement and evaluate inference optimization techniques such as batching strategies, quantization, prefix caching, and speculative decoding.
-
Support development and optimization of scalable serving systems, including request scheduling and resource utilization.
-
Develop and use profiling, benchmarking, and performance analysis tools for inference workloads.
-
Collaborate with hardware, compiler, and framework teams to improve overall system performance.
-
Contribute to internal tools and, where applicable, open-source projects for inference optimization on AMD platforms.
-
Document best practices and contribute to performance guidelines for GenAI deployment.
PREFERRED EXPERIENCE
-
Strong understanding of GPU architecture and performance fundamentals (compute, memory hierarchy, interconnects such as PCIe/Infinity Fabric/RDMA).
-
Experience with GenAI inference optimization techniques (e.g., quantization, KV-cache optimization, batching).
-
Hands-on experience with inference/serving frameworks such as vLLM, SGLang, Triton, TensorRT-LLM, or similar.
-
Experience working on LLM or multimodal inference workloads.
-
Familiarity with distributed systems and serving architectures.
-
Experience with ML frameworks (Py Torch, JAX, or Tensor Flow), especially for inference.
-
Proficiency in Python and at least one systems language (C++/CUDA/HIP).
-
Experience with profiling, debugging, and performance tuning tools.
-
Ability to work collaboratively across teams and deliver impactful optimizations.
ACADEMIC CREDENTIALS
- B.S., M.S. or Ph.D. in Computer Science, Computer Engineering, or a related field preferred, or equivalent industry experience.
LOCATION
- San Jose, CA preferred. May consider other US locations where AMD has offices such as Bellevue, WA or Austin, TX.
Benefits offered are described: AMD benefits at a glance.
AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.
AMD may use Artificial Intelligence to help screen, assess or select applicants for this position. AMD’s “Responsible AI Policy” is available here.
This posting is for an existing vacancy.
총 조회수
0
총 지원 클릭 수
0
모의 지원자 수
0
스크랩
0
비슷한 채용공고

Staff Design Integration Lead
Micron · San Jose, CA

Staff Software Engineer
PayPal · San Jose, California, United States of America

Principal Design Engineer
Micron · San Jose, CA

Senior Software Engineer — Experience League Knowledge Platform & GenAI Services
Adobe · San Jose

Senior Technical Lead - Test/Interface/UC Board Bring up
HCL Technologies · San Jose, United States
AMD 소개

AMD
PublicAdvanced Micro Devices, Inc. (AMD) is an American multinational semiconductor company headquartered in Santa Clara, California.
10,001+
직원 수
Santa Clara
본사 위치
$240B
기업 가치
리뷰
3.7
10개 리뷰
워라밸
2.8
보상
3.2
문화
4.1
커리어
3.4
경영진
3.8
68%
친구에게 추천
장점
Great team culture and spirit
Innovative projects and cutting-edge technology
Supportive management and leadership
단점
High workload and overwhelming demands
Work-life balance challenges
High pressure and stressful deadlines
연봉 정보
6개 데이터
L2
L3
L4
L5
L6
L2 · Data Analyst L2
0개 리포트
$76,430
총 연봉
기본급
$30,572
주식
$38,215
보너스
$7,643
$53,501
$99,359
면접 경험
2개 면접
난이도
3.0
/ 5
소요 기간
14-28주
합격률
50%
면접 과정
1
Application Review
2
Recruiter Screen
3
Hiring Manager Interview
4
Technical Interview
5
Offer
자주 나오는 질문
Technical Knowledge
Behavioral/STAR
Past Experience
Problem Solving
뉴스 & 버즈
Anything Intel Can Do, Advanced Micro Devices Might Be Able to Do Better - Barron's
Barron's
News
·
1d ago
These stocks are today’s movers: Intel, AMD, Arm, SAP, Comfort Systems, and more - MSN
MSN
News
·
1d ago
Many Intel & AMD Laptop Improvements Merged For Linux 7.1 - Phoronix
Phoronix
News
·
1d ago
AMD’s stock is red-hot — and just hit a major milestone - MarketWatch
MarketWatch
News
·
1d ago