About the job
Join Cartesia as a Model Architecture Researcher
At Cartesia, our vision is to revolutionize AI by creating interactive intelligence that is seamlessly integrated into your daily life. Unlike current models, our goal is to develop systems capable of processing extensive streams of audio, video, and text—1 billion text tokens, 10 billion audio tokens, and 1 trillion video tokens—directly on devices.
As pioneers in innovative model architectures, our founding team, which originated from the Stanford AI Lab, has developed State Space Models (SSMs)—a groundbreaking foundation for training efficient, large-scale models. Our diverse team merges deep expertise in model innovation with a design-focused engineering approach, allowing us to create and deploy state-of-the-art models and applications.
Backed by leading investors such as Index Ventures, Lightspeed Venture Partners, and many others, including industry veterans and advisors, we are poised to shape the future of AI.
Your Contribution
In this role, you will drive forward-thinking research in neural network architecture, focusing on alternative models like state space models, efficient transformers, and hybrid architectures.
Create innovative architectures that enhance model performance, inference speed, and adaptability in various environments, from cloud infrastructures to on-device implementations.
Develop advanced capabilities for models, including statefulness, long-range memory, and novel conditioning mechanisms to boost expressiveness and generalization.
Analyze architectural decisions and their effects on model characteristics such as scalability, robustness, latency, and energy consumption.
Create frameworks and tools to assess architectural advancements, benchmarking their performance in both research and production contexts.
Collaborate with interdisciplinary teams to translate architectural insights into scalable systems that deliver real-world impact.
Your Qualifications
Extensive experience in architecture design with a focus on advanced models such as state space models, transformers, and RNN/CNN variants.
In-depth understanding of the interplay between architectural designs and system constraints, particularly in cloud and on-device deployments.
Strong proficiency in the design and evaluation of neural network architectures.

