refresh

트렌딩 기업

트렌딩 기업

채용

채용NVIDIA

Deep Learning Algorithms Engineer - ACOT

NVIDIA

Deep Learning Algorithms Engineer - ACOT

NVIDIA

Vietnam

·

On-site

·

Full-time

·

2w ago

NVIDIA is seeking a motivated AI Acceleration & Optimization Engineer to join our Acceleration Computing, Optimization and Tools (ACOT) team. In this role, you will help improve the performance, scalability, and efficiency of modern AI models across NVIDIA GPU platforms. You will work with engineers across algorithms, systems, and hardware to support high-performance model deployment and development for real-world AI workloads.

As part of ACOT, you will collaborate with architecture, research, CUDA, compiler, and framework teams to help bring next-generation AI workloads from research to production with strong performance and reliability.

What you will be doing

  • Assist in optimizing AI models such as LLMs, VLMs, diffusion models, and multimodal models for inference and training on NVIDIA GPUs.
  • Profile workloads and help identify performance bottlenecks across GPU compute, memory, networking, and storage.
  • Support the development and integration of optimization techniques such as quantization, kernel fusion, parallelism, and memory efficiency improvements.
  • Use tools including CUDA, TensorRT, Nsight, and NVIDIA acceleration libraries to analyze and improve model performance.
  • Work with deep learning frameworks including Py Torch, JAX, and Tensor Flow, as well as open-source inference frameworks like vLLM and SGLang.
  • Contribute to performance benchmarking, testing, and internal tooling to improve optimization workflows.
  • Partner with senior engineers and multi-functional teams to evaluate workload behavior and support future performance improvements.

What we want to see

  • Bachelor’s or Master’s degree in Computer Science, Electrical Engineering, Computer Engineering, or related field (or equivalent experience).
  • 2–4 years of experience, or strong academic/project experience, in deep learning, performance engineering, systems, or high-performance computing.
  • Good understanding of deep learning fundamentals and modern AI model architectures, especially transformers.
  • Familiarity with GPU architecture and parallel computing concepts such as CUDA, kernels, memory hierarchy, and streams.
  • Exposure to profiling and performance analysis tools.
  • Programming skills in Python.
  • Experience with at least one major ML framework such as Py Torch, Tensor Flow, or JAX.

Ways to stand out from the crowd

  • Internship, research, or project experience optimizing AI/ML workloads on GPUs.
  • Hands-on experience with TensorRT, TensorRT-LLM, vLLM, SGLang, or similar inference/runtime frameworks.
  • Familiarity with quantization, sparsity, or mixed-precision techniques.
  • Experience with distributed training or inference concepts. Contributions to open-source ML systems, performance tools, or infrastructure projects.
  • Proficiency in C++, strong debugging skills and interest in low-level performance optimization.

총 조회수

0

총 지원 클릭 수

0

모의 지원자 수

0

스크랩

0

NVIDIA 소개

NVIDIA

NVIDIA

Public

A computing platform company operating at the intersection of graphics, HPC, and AI.

10,001+

직원 수

Santa Clara

본사 위치

$4.57T

기업 가치

리뷰

4.1

10개 리뷰

워라밸

3.5

보상

4.2

문화

4.3

커리어

4.5

경영진

4.0

75%

친구에게 추천

장점

Great culture and supportive environment

Smart colleagues and excellent people

Cutting-edge technology and learning opportunities

단점

Team-dependent experience and outcomes

Work-life balance issues with long hours

Politics and influence over competence

연봉 정보

73개 데이터

L3

L4

L5

L3 · Data Scientist IC2

0개 리포트

$177,542

총 연봉

기본급

-

주식

-

보너스

-

$150,910

$204,174

면접 경험

7개 면접

난이도

3.1

/ 5

경험

긍정 0%

보통 86%

부정 14%

면접 과정

1

Application Review

2

Recruiter Screen

3

Online Assessment

4

Technical Interview

5

System Design Interview

6

Team Review

자주 나오는 질문

Coding/Algorithm

System Design

Technical Knowledge

Behavioral/STAR