About the job
About NomadicML
NomadicML is at the forefront of revolutionizing machine intelligence through video data. In a world where Americans drive over 5 trillion miles annually, we are uncovering insights from the vast amounts of recorded footage. Our cutting-edge Vision-Language Models (VLMs) serve as a transformative tool, converting raw video into structured intelligence that fuels advancements in autonomy and robotics.
By collaborating with industry leaders in self-driving technology, robotics, and industrial automation, we extract valuable insights from petabytes of previously unusable data. Founded by Mustafa Bal and Varun Krishnan, who met while studying Computer Science at Harvard University, we are driven by innovation and impact.
Mustafa is a key contributor to ONNX Runtime and DeepSpeed, specializing in distributed systems and large-scale model training.
Varun is recognized as an INFORMS Wagner Prize Finalist for his pioneering research in large-scale driver navigation AI models and is among the top chess players in the U. S.
Our team comprises experts who have developed mission-critical AI systems at leading companies such as Snowflake, Lyft, Microsoft, Amazon, and IBM Research. We are proud of our numerous top-tier publications in VLMs and AI at prestigious conferences like CVPR, and we operate with the agility and focus of a startup.
About the Role
We are seeking a talented Backend / Infrastructure Engineer who excels in the realms of cloud systems, SDK design, and large-scale inference infrastructure. In this role, you will enhance and scale the backbone of NomadicML’s video intelligence platform, overseeing processes from secure cloud ingestion to distributed GPU inference pipelines that support our largest foundation models.
You will work closely with ML researchers to streamline the production of their models, automate deployment and scaling, and create clean APIs and SDKs that empower enterprises globally.
This position encompasses systems engineering, distributed compute orchestration, and developer experience. Your responsibilities will include managing cloud storage, inference scheduling, GPU clusters, and the NomadicML SDK.
What You’ll Build
GPU Inference Workflows: Design pipelines for executing extensive multi-GPU inference tasks on foundational video models, with a focus on optimizing throughput, cost, and reliability.
API Development: Develop and maintain robust APIs that facilitate seamless integration with various enterprise applications.

