채용
ABOUT LIQUID AI:
Spun out of MIT CSAIL, we build general-purpose AI systems that run efficiently across deployment targets, from data center accelerators to on-device hardware, ensuring low latency, minimal memory usage, privacy, and reliability. We partner with enterprises across consumer electronics, automotive, life sciences, and financial services. We are scaling rapidly and need exceptional people to help us get there.
THE OPPORTUNITY:
This is a rare chance to sit at the intersection of frontier vision-language models and real-world deployment. You'll own applied post-training work for VLMs end-to-end for some of the world's largest enterprises, while still contributing directly to Liquid's core multimodal model development.
Unlike most roles that force a trade-off between customer impact and foundational work, this role gives you both: deep ownership over how vision-language models are adapted, evaluated, and shipped, and a direct line into the evolution of Liquid's multimodal post-training stack.
If you care about visual understanding, data quality, evaluation, and making VLMs actually work in production, this is a chance to shape how applied multimodal AI is done at a foundation model company.
WHAT WE'RE LOOKING FOR:
We need someone who:
-
Takes ownership: Owns VLM post-training projects end-to-end, from customer requirements through delivery and evaluation.
-
Thinks end-to-end: Can reason across visual data curation, training, alignment, and evaluation as a single system.
-
Is pragmatic: Optimizes for model quality and customer outcomes over publications or theory.
-
Communicates clearly: Can translate between customer needs and internal technical teams, and push back when needed.
THE WORK
-
Act as the technical owner for enterprise customer VLM post-training engagements.
-
Translate customer requirements into concrete multimodal post-training specifications and workflows.
-
Design and execute visual data generation, filtering, and quality assessment processes, including image-text pair curation, annotation pipelines, and synthetic data generation for visual tasks.
-
Run supervised fine-tuning, preference alignment, and reinforcement learning workflows for vision-language models.
-
Design task-specific evaluations for visual understanding, grounding, OCR, document parsing, and other multimodal capabilities. Interpret results and feed learnings back into core post-training pipelines.
DESIRED EXPERIENCE:
Must-have:
-
Hands-on experience with data generation and evaluation for VLM or multimodal post-training.
-
Experience training or fine-tuning vision-language models using SFT, preference alignment, and/or RL.
-
Strong intuition for visual data quality, annotation design, and multimodal evaluation.
-
Familiarity with vision encoders, image-text architectures, and how visual representations interact with language model backbones.
Nice-to-have:
-
Experience with visual grounding, document understanding, OCR, or video understanding tasks.
-
Experience contributing to shared or general-purpose multimodal post-training infrastructure.
-
Prior exposure to customer-facing or applied ML delivery environments.
-
Familiarity with alignment or RL techniques beyond basic supervised fine-tuning in the multimodal setting.
WHAT SUCCESS LOOKS LIKE (YEAR ONE)
-
Independently owns and delivers enterprise VLM post-training projects with minimal oversight.
-
Is trusted by customers as the technical owner, demonstrating strong judgment and delivery quality on multimodal workloads.
-
Has made durable contributions to Liquid's general-purpose multimodal post-training pipelines by feeding applied learnings back into baseline model development.
WHAT WE OFFER:
-
Real ML work: You will fine-tune vision-language models, generate multimodal data, and ship solutions, not configure API calls. Your work feeds directly back into our core model development.
-
Compensation: Competitive base salary with equity in a unicorn-stage company.
-
Health: We pay 100% of medical, dental, and vision premiums for employees and dependents.
-
Financial: 401(k) matching up to 4% of base pay.
-
Time Off: Unlimited PTO plus company-wide Refill Days throughout the year.
총 조회수
1
총 지원 클릭 수
0
모의 지원자 수
0
스크랩
0
비슷한 채용공고

Sr. Machine Learning Engineer, tvScientific
Pinterest · San Francisco, CA, US; Remote, US

Senior Machine Learning Engineer II, NLU & Agentic AI
Moveworks · San Francisco, CA

Senior Machine Learning Engineer (Research Scientist) - Data Foundation & AI
Plaid · San Francisco

Staff Machine Learning Engineer
Databricks · San Francisco, California

Staff Machine Learning Engineer, Content Quality Signals
Pinterest · San Francisco, CA, US; Remote, US
Liquid AI 소개

Liquid AI
Series ALiquid AI is an artificial intelligence company focused on developing liquid neural networks and dynamic AI systems. The company specializes in creating adaptive neural architectures inspired by biological systems.
51-200
직원 수
Cambridge
본사 위치
연봉 정보
4개 데이터
Staff/L6
Staff/L6 · GTM STAFF - STRATEGIC PARTNERSHIPS
1개 리포트
$455,000
총 연봉
기본급
$350,000
주식
-
보너스
-
$455,000
$455,000
뉴스 & 버즈
Vertiv Stock: The $15 Billion Backlog, Liquid Cooling Dominance, And The AI Trade (VRT) - Seeking Alpha
Seeking Alpha
News
·
6d ago
Taiwan cooling suppliers post record March revenue as AI demand lifts liquid cooling - digitimes
digitimes
News
·
6d ago
Best practices for deploying liquid-cooled servers in AI data centers - Data Center Dynamics
Data Center Dynamics
News
·
1w ago
Liquid AI Releases LFM2.5-VL-450M: a 450M-Parameter Vision-Language Model with Bounding Box Prediction, Multilingual Support, and Sub-250ms Edge Inference - MarkTechPost
MarkTechPost
News
·
1w ago