companyCartesia logo

Inference Engineer at Cartesia | San Francisco, CA

Cartesia*HQ - San Francisco, CA
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Experience Level

Experience

Qualifications

Prospective candidates should possess strong engineering skills and demonstrate a capacity for clean code practices, experience in building distributed systems, technical leadership abilities, and familiarity with machine learning inference pipelines.

About the job

Join Cartesia as an Inference Engineer

At Cartesia, our vision is to create the next evolution of AI: an interactive, omnipresent intelligence that operates seamlessly across all environments. Currently, even the most advanced models struggle to continuously analyze a year's worth of audio, video, and text data—comprising 1 billion text tokens, 10 billion audio tokens, and 1 trillion video tokens—much less perform these tasks on-device.

We are at the forefront of developing the model architectures that will make this a reality. Our founding team, who met as PhD candidates at the Stanford AI Lab, pioneered State Space Models (SSMs), a groundbreaking framework for training efficient, large-scale foundation models. Our talented team merges deep expertise in model innovation and systems engineering with a design-focused product engineering approach, enabling us to build and launch state-of-the-art models and user experiences.

Supported by leading investors such as Index Ventures and Lightspeed Venture Partners, along with contributions from Factory, Conviction, A Star, General Catalyst, SV Angel, Databricks, and others, we are fortunate to be guided by numerous exceptional advisors and over 90 angel investors from diverse industries, including some of the world’s foremost experts in AI.

About the Role

We are actively seeking an Inference Engineer to propel our mission of creating real-time multimodal intelligence.

Your Impact

  • Develop and implement a low-latency, scalable, and dependable model inference and serving stack for our innovative foundation models utilizing Transformers, SSMs, and hybrid models.

  • Collaborate closely with our research team and product engineers to efficiently deliver our product suite in a fast, cost-effective, and reliable manner.

  • Construct robust inference infrastructure and monitoring systems for our product offerings.

  • Enjoy substantial autonomy in shaping our products and directly influencing how cutting-edge AI is integrated across diverse devices and applications.

What You Bring

At Cartesia, we prioritize strong engineering skills due to the complexity and scale of the challenges we tackle.

  • Proficient engineering skills with a comfort level in navigating intricate codebases, and a commitment to producing clean, maintainable code.

  • Experience in developing large-scale distributed systems with strict performance, reliability, and observability requirements.

  • Proven technical leadership, capable of executing and delivering results from zero to one amidst uncertainty.

  • A background in or experience with inference pipelines, machine learning, and generative models.

About Cartesia

Cartesia is committed to pioneering the future of AI through innovative model architectures, guided by a team of experts with a strong foundation in academic research and practical application.

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.