company

Member of Technical Staff - Multi-Modal Vision at Liquid AI | San Francisco

Liquid AISan Francisco
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Experience Level

Experience

Qualifications

Minimum Qualifications:Proven hands-on experience in training or evaluating Vision-Language Models (VLMs) with a strong emphasis on experimental rigor. Capability to convert research concepts into scalable implementations, with an iterative approach to hypothesis refinement. Strong proficiency in Python and at least one deep learning framework. M. S. or Ph. D. in Computer Science, Mathematics, or a related domain; or equivalent practical experience. Preferred Experience:Experience in building or optimizing multimodal training or data pipelines. Familiarity with distributed training frameworks (e.g., DeepSpeed, FSDP, Megatron-LM). Experience with multimodal post-training techniques (e.g., SFT, preference optimization, reinforcement learning methods). Expertise in dataset design and data quality assessment (including quality and diversity evaluation, long-tail mining). Prior contributions to open-source projects (code, data, models) on platforms like GitHub or Hugging Face. Published research in leading AI conferences (e.g., NeurIPS, ICML, CVPR, ECCV, ICLR, ACL). Experience in computer vision or visual representation learning.

About the job

About Liquid AI

Originating from MIT CSAIL, Liquid AI specializes in creating versatile AI systems that operate efficiently across various platforms, from data center accelerators to on-device hardware, focusing on low latency, minimal memory consumption, privacy, and dependability. Our collaborations extend across industries including consumer electronics, automotive, life sciences, and financial services. As we undergo rapid expansion, we are on the lookout for outstanding individuals to join our journey.

The Opportunity

The Vision-Language Models (VLM) team is dedicated to developing cutting-edge vision-language models that function seamlessly on devices, adhering to stringent latency and memory requirements without compromising quality. Having already launched four premier models, we are excited about what lies ahead.

This team is responsible for the complete VLM pipeline, encompassing research on novel architectures, training algorithms, data curation, evaluation, and deployment. You will be part of a dedicated, hands-on team that directly engages with models and works closely with our pretraining, post-training, and infrastructure teams. Your success will be gauged by the performance of the models we deliver.

About Liquid AI

Liquid AI is a pioneering organization emerging from MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL), dedicated to crafting advanced AI systems that meet the demands of modern technology across various sectors. Our mission is to provide privacy-focused, reliable, and efficient AI solutions that enhance user experiences across industries, including automotive, life sciences, and financial services.

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.