AnthropicRemote-Friendly (Travel-Required) | San Francisco, CA | New York City, NYNew
Remote Full-time
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Entry Level
Qualifications
Strong background in machine learning and AI model evaluation techniques. Experience with programming languages such as Python or similar. Ability to work collaboratively in a remote team environment. Excellent analytical and problem-solving skills.
About the job
Anthropic is looking for a Research Engineer focused on model evaluations. This position involves research and development to assess and strengthen the performance of AI models. Teams are based in San Francisco and New York City, and the role supports remote work with required travel.
Key responsibilities
Design and implement evaluations for Anthropic's AI models
Collaborate with team members to enhance model performance
Contribute to research that pushes the boundaries of AI systems
Location
Remote-friendly (travel required)
San Francisco, CA
New York City, NY
About Anthropic
Anthropic is at the forefront of AI research, dedicated to developing safe and beneficial AI technologies. Our mission is to ensure that AI systems are aligned with human intentions and values. We foster a collaborative and innovative environment that encourages curiosity and creativity.
Zyphra is an innovative artificial intelligence company located in the heart of San Francisco, California.The Opportunity:Join our dynamic team as a Research Engineer - Audio & Speech Models, where you will play a pivotal role in advancing Zyphra’s Audio Team. You will be instrumental in developing cutting-edge open-source text-to-speech and audio models. Your contributions will span the full spectrum of the model training process, from data collection and processing to the design of innovative architectures and training approaches.Your Responsibilities:Conduct large-scale audio training operationsOptimize the performance of our training infrastructureCollect, process, and evaluate audio datasetsImplement architectural and methodological improvements through rigorous testingWhat We Seek:A strong research mindset with the ability to navigate projects from ideation to implementation and documentation.Proficiency in rapid prototyping and implementation, allowing for swift experimentation.Effective collaboration skills in a fast-paced research environment.A quick learner who is eager to embrace and implement new concepts.Excellent communication abilities, enabling you to contribute to both research and engineering tasks at scale.Preferred Qualifications:Expertise in training audio models, such as text-to-speech, ASR, speech-to-speech, or emotion recognition.Experience with training audio autoencoders.Solid understanding of signal processing, particularly in audio.Familiarity with diffusion models, consistency models, or GANs.Experience with large-scale (multi-node) GPU training environments.Strong understanding of experimental methodologies for conducting rigorous tests and ablations.Interest in large-scale, parallel data processing pipelines.Competence in PyTorch and Python programming.Experience contributing to large, established codebases with rapid adaptation.
Company OverviewEcho Neurotechnologies is a pioneering startup in the Brain-Computer Interface (BCI) sector, dedicated to revolutionizing the lives of individuals with disabilities through advanced hardware engineering and artificial intelligence solutions. Our vision is to develop innovative technologies that empower users, restoring autonomy and enhancing their quality of life.Team CultureWe pride ourselves on cultivating an inclusive and dynamic team of skilled professionals who are passionate about their work. Our startup environment encourages ownership of impactful decisions and fosters continuous learning and collaboration, where every contribution is essential to our collective success.Job SummaryWe are on the lookout for a talented Machine Learning Research Engineer specialized in speech modeling to join our innovative team. The successful candidate will leverage ML/AI methodologies to create and refine adaptable speech models aimed at brain-computer interface applications, ultimately making a difference in the lives of patients facing severe disabilities. Candidates should possess significant expertise in speech modeling, feature engineering, time-series analysis, and the development of custom ML models.Key ResponsibilitiesDesign and evaluate diverse model architectures and strategies to enhance the accuracy and resilience of models for interpreting speech from brain activity.Investigate and implement cutting-edge speech features and representations within neural-decoding frameworks, informed by speech science and functional neurophysiology.Create pipelines for generating personalized and naturalistic speech from both text and brain activity inputs.Develop algorithms to analyze both intact and compromised speech signals, identifying biomarkers linked to various diseases and disabilities.Collaborate within a tight-knit team to build models, define R&D workflows, and translate scientific discoveries into practical applications.Contribute to best practices ensuring reliability, observability, reproducibility, and scientific rigor across the R&D landscape.Maintain well-documented, versioned code, analysis pipelines, and results for maximum interpretability and reproducibility.
Full-time|Remote|Remote-Friendly (Travel-Required) | San Francisco, CA | New York City, NY
Anthropic is looking for a Research Engineer focused on model evaluations. This position involves research and development to assess and strengthen the performance of AI models. Teams are based in San Francisco and New York City, and the role supports remote work with required travel. Key responsibilities Design and implement evaluations for Anthropic's AI models Collaborate with team members to enhance model performance Contribute to research that pushes the boundaries of AI systems Location Remote-friendly (travel required) San Francisco, CA New York City, NY
About Liquid AILiquid AI, a pioneering company spun out of MIT CSAIL, is at the forefront of developing general-purpose AI systems that operate efficiently across various platforms, from data center accelerators to on-device hardware. Our commitment to low latency, minimal memory usage, privacy, and reliability allows us to partner with some of the most esteemed enterprises in consumer electronics, automotive, life sciences, and financial services. As we experience rapid growth, we are seeking exceptional talent to join our innovative journey.The OpportunityJoin our cutting-edge Audio team, where we are developing advanced speech-language models capable of handling Speech-to-Text (STT), Text-to-Speech (TTS), and speech-to-speech tasks within a unified architecture. This pivotal role supports applied audio model development, directly collaborating with the technical lead to deliver production systems that operate on-device under real-time constraints. You will take ownership of key workstreams encompassing data pipelines, evaluation systems, and customer deployments. If you are eager to tackle unique technical challenges within a small, elite team where your contributions are impactful, this is the role for you.What We're Looking ForWe are seeking an individual who:Builds first, theorizes later: You prioritize shipping working systems over theoretical models; production-grade code is your default.Owns outcomes end-to-end: You take full responsibility for everything from data pipelines to customer deployments and don't shy away from challenges.Thrives under constraints: On-device, low-latency, memory-constrained environments motivate you. You view constraints as opportunities for innovative design.Ramps quickly on new territory: You are comfortable closing knowledge gaps swiftly and actively seek feedback to drive results.The WorkDevelop and scale data pipelines for audio model training, including preprocessing, augmentation, and quality filtering at scale.Design, implement, and maintain evaluation systems that assess multimodal performance across both internal and public benchmarks.Fine-tune and adapt audio models to cater to customer-specific use cases, taking charge from requirement gathering through to deployment.Contribute production code to the core audio repository while collaborating closely with infrastructure and research teams.Facilitate experimentation under real hardware constraints, transitioning smoothly between customer-focused projects and core development initiatives.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to enhance humanity's potential through the advancement of collaborative general intelligence. We envision a future where everyone can access the knowledge and tools necessary to leverage AI for their individual needs and objectives.As a team of scientists, engineers, and innovators, we have developed some of the most widely utilized AI products, such as ChatGPT and Character.ai, along with open-weight models like Mistral, and popular open-source projects including PyTorch, OpenAI Gym, Fairseq, and Segment Anything.Role OverviewAt Thinking Machines, we adopt a multimodal-first approach, where multimodality is integral to our scientific goals and infrastructure. We are seeking skilled researchers to push the boundaries of audio capabilities. In this role, you will delve into how audio models can facilitate more natural and efficient communication and collaboration by preserving information and accurately capturing user intent.This position requires strong collaboration across pre-training, post-training, and product development with top-tier researchers, infrastructure engineers, and designers. Here, you will have the chance to influence the foundational capabilities of AI systems that will be utilized by millions globally.This role marries fundamental research with practical engineering, as we do not separate these functions internally. You will be expected to write high-performance code as well as engage with technical reports. It is an ideal position for someone who enjoys both extensive theoretical research and hands-on experimentation while laying the groundwork for how AI learns.Note: This is an evergreen role that remains open continuously to gauge interest in this research area. We receive numerous applications, and there may not always be an immediate match for your skills and experience. Nevertheless, we encourage you to apply, as we routinely review applications and reach out as new opportunities arise. You are welcome to reapply if you gain additional experience, but please avoid applying more frequently than every six months. Occasionally, we also post specific roles for distinct project or team needs, in which case you are welcome to apply directly alongside an evergreen submission.
Join Our Innovative Team at David AIDavid AI is pioneering the audio data research landscape. We adopt a rigorous R&D methodology for developing datasets that parallels the standards upheld by leading AI laboratories. Our vision is to seamlessly integrate AI into everyday experiences, with audio serving as the perfect conduit. The evolution of audio AI is rapidly unfolding, yet the availability of high-quality training data remains a critical challenge. This is where David AI steps in.Founded in 2024 by a talented group of former engineers and operators from Scale AI, we have quickly become a trusted partner to numerous FAANG companies and AI research labs. Recently, we secured $50 million in a Series B funding round with notable investors, including Meritech, NVIDIA, and Alt Capital.Our culture is built on sharp intellect, humility, ambition, and a close-knit community. We invite exceptional minds in research, engineering, product development, and operations to join us as we advance the field of audio AI.Research Team OverviewAt David AI, we are convinced that superior model capabilities stem from high-quality, differentiated data. Our research team is dedicated to conducting ambitious, long-term studies into audio technology while collaborating with both internal and external partners to implement cutting-edge research insights into practical applications.Your Role as a Founding Audio AI Research EngineerIn this position, you will establish the research framework that influences how premier AI labs develop their audio models. You will have access to a top-tier team of human AI trainers, robust computing resources, and the autonomy to shape your research agenda.Key ResponsibilitiesCreate and implement comprehensive evaluation frameworks for assessing audio AI capabilities in areas such as speech, emotion detection, conversational dynamics, and acoustic patterns.Investigate and prototype innovative methodologies for audio quality assessment, automated labeling, and optimizing data collection processes.Design focused data collection pipelines aimed at capturing novel, high-value audio capabilities.Develop automated systems for ongoing classifier enhancement and prompt engineering evaluation.Assess cutting-edge models and formulate actionable research strategies.Publish your findings in prestigious conferences.
Join Perplexity as a Research Engineering Manager, where you will spearhead a team of exceptional AI researchers and engineers dedicated to crafting the advanced models that power our innovative products. Our talented team has pioneered some of the most sophisticated models in agentic research, query understanding, and other critical domains that demand precision and depth. As we broaden our user base and expand our product offerings, our proprietary models are increasingly essential for delivering a premium experience to the world's most discerning users.You will explore our extensive datasets of conversational and agentic queries, applying state-of-the-art training methodologies to enhance AI model performance. Through proactive technical and organizational leadership, you will empower your team to create cutting-edge models for the applications that are most significant to our business and our users.
About SesameAt Sesame, we envision a world where technology seamlessly integrates into our daily lives, making interactions with computers feel intuitive and human. Our mission is to create groundbreaking voice agents that enhance everyday experiences. Our talented team comprises trailblazers from Oculus, Ubiquity6, and industry giants like Meta, Google, and Apple, bringing extensive knowledge in both hardware and software. Join us in crafting a future where technology truly comes to life.About the RoleAs an Audio Systems Engineer, you will play a pivotal role in developing innovative smart wearable devices with exceptional audio capabilities. We seek a skilled engineer to define, calibrate, assess, and deliver high-fidelity speech reproduction and spatial audio capture systems.Key Responsibilities:Lead the development and release of consumer electronics that set new standards for audio performance, either independently or with guidance from the Audio Lead.Translate Product Level Requirement Documents into precise audio system and transducer specifications that ensure product excellence.Design, evaluate, and enhance real-time audio capture and rendering systems.Investigate and adopt cutting-edge technologies to improve audio system integration.Stay updated with emerging technologies and establish benchmarks for optimal user experience.Contribute to system architecture discussions, including algorithms and tools, and oversee their implementation.Assess third-party audio technologies and remain informed about industry advancements.Understand the trade-offs related to various DSP and ML approaches for embedded solutions.Collaborate with cross-functional teams to drive innovation in audio solutions.Support factory testing and product builds throughout the development lifecycle.Essential Qualifications:Minimum of 4 years of experience in audio hardware development.Proficient in designing, calibrating, and testing microphone arrays and real-time audio capture systems.Demonstrated success in developing comprehensive audio solutions that encompass acoustics, transducers, and DSP technologies.
Zyphra is a cutting-edge artificial intelligence firm headquartered in the vibrant city of San Francisco, California.Position Overview:As a Research Scientist specializing in Model Architectures, you will play a pivotal role in Zyphra’s AI Architecture Research Team. Your responsibilities will include the design and thorough evaluation of innovative model architectures and training methodologies aimed at enhancing essential modeling capabilities (e.g., loss per flop or loss per parameter) and tackling core limitations inherent in current models. You will collaborate closely with our pre-training team to ensure that your findings are seamlessly integrated into our next-generation models.Qualifications:A strong research acumen and intuition.Proven ability to navigate research projects from initial conception to execution and final write-up.Exceptional implementation and prototyping skills, with the capability to swiftly transform ideas into experimental outcomes.A collaborative spirit and the ability to thrive in a fast-paced research environment.A deep curiosity and enthusiasm for understanding intelligence.Requirements:Experience with long-term memory, RAG/retrieval systems, dynamic/adaptive computation, and alternative credit assignment strategies.Knowledge of reinforcement learning, control theory, and signal processing techniques.A passion for exploring and critically evaluating unconventional ideas, with the ability to maintain a unique perspective.Familiarity with modern training pipelines and the hardware necessities for designing efficient architectures compatible with GPU hardware.Strong understanding of experimental methodologies for conducting rigorous ablations and hypothesis testing.High proficiency in PyTorch and Python programming.Ability to quickly assimilate into large pre-existing codebases and contribute effectively.Prior publication of machine learning research in reputable venues.Postgraduate degree in a scientific discipline (e.g., Computer Science, Electrical Engineering, Mathematics, Physics).Why Join Zyphra?We emphasize a structured research methodology that systematically addresses ambitious challenges in AI.
Join our innovative team at Zyphra as a Research Engineer specializing in Brain-Computer Interface (BCI) Models. In this pivotal role, you will contribute to groundbreaking research and development initiatives in the field of neuroscience and artificial intelligence. Your expertise will help shape the future of communication between humans and machines, enhancing the quality of life for countless individuals.As a Research Engineer, you will be responsible for designing, implementing, and testing advanced BCI models, collaborating closely with a diverse team of scientists and engineers. Your work will play a crucial role in advancing our understanding of neural dynamics and their applications in technology.
bland is looking for a Machine Learning Researcher with a focus on audio. This position is based in San Francisco and centers on advancing how machines process and understand sound. The team works on pushing the boundaries of audio technology for a range of platforms. Responsibilities Research and develop new machine learning techniques for audio applications Contribute to projects that improve audio processing and analysis Collaborate with colleagues to bring research ideas into real-world audio products Location This role requires working onsite in San Francisco.
Join Cartesia as a Model Architecture ResearcherAt Cartesia, our vision is to revolutionize AI by creating interactive intelligence that is seamlessly integrated into your daily life. Unlike current models, our goal is to develop systems capable of processing extensive streams of audio, video, and text—1 billion text tokens, 10 billion audio tokens, and 1 trillion video tokens—directly on devices.As pioneers in innovative model architectures, our founding team, which originated from the Stanford AI Lab, has developed State Space Models (SSMs)—a groundbreaking foundation for training efficient, large-scale models. Our diverse team merges deep expertise in model innovation with a design-focused engineering approach, allowing us to create and deploy state-of-the-art models and applications.Backed by leading investors such as Index Ventures, Lightspeed Venture Partners, and many others, including industry veterans and advisors, we are poised to shape the future of AI.Your ContributionIn this role, you will drive forward-thinking research in neural network architecture, focusing on alternative models like state space models, efficient transformers, and hybrid architectures.Create innovative architectures that enhance model performance, inference speed, and adaptability in various environments, from cloud infrastructures to on-device implementations.Develop advanced capabilities for models, including statefulness, long-range memory, and novel conditioning mechanisms to boost expressiveness and generalization.Analyze architectural decisions and their effects on model characteristics such as scalability, robustness, latency, and energy consumption.Create frameworks and tools to assess architectural advancements, benchmarking their performance in both research and production contexts.Collaborate with interdisciplinary teams to translate architectural insights into scalable systems that deliver real-world impact.Your QualificationsExtensive experience in architecture design with a focus on advanced models such as state space models, transformers, and RNN/CNN variants.In-depth understanding of the interplay between architectural designs and system constraints, particularly in cloud and on-device deployments.Strong proficiency in the design and evaluation of neural network architectures.
Zyphra is an innovative leader in artificial intelligence, located in the heart of San Francisco, California.Role Overview:As a Research Engineer specializing in Language Model Pre-Training, you will play a pivotal role in defining our language model strategy through comprehensive pretraining development. Your close collaboration with our pretraining team will ensure that your insights contribute to the advancement of our next-generation models.Key Responsibilities:Conduct large-scale training runs and implement model parallelization techniques.Optimize the performance of our pretraining stack.Oversee dataset collection, processing, and evaluation.Research architecture and methodologies, including optimizer ablations.Qualifications:Demonstrated engineering prowess in developing reliable and robust systems.A quick learner with a passion for implementing innovative ideas.Exceptional communication and collaboration skills, capable of working effectively on both research and engineering implementations at scale.Preferred Skills:Profound expertise in addressing machine learning challenges and training models.Experience training on large-scale (multi-node) GPU clusters.In-depth understanding of model training pipelines, including model/data parallelism and distributed optimizers.Strong methodology for conducting rigorous ablations and hypothesis testing.Familiarity with large-scale, high-performance data processing pipelines.High proficiency in PyTorch and Python programming.Ability to navigate and understand extensive pre-existing codebases swiftly.Published research in machine learning in reputable venues is an advantage.Postgraduate degree in a relevant scientific field (Computer Science, Electrical Engineering, Mathematics, Physics).Why Join Zyphra?We value a research methodology that emphasizes thoughtful, methodical progress towards ambitious objectives. Both deep research and engineering excellence are given equal importance.Join us in an environment that fosters innovation, collaboration, and professional growth.
Full-time|On-site|San Francisco (London/Europe - OK)
Tavus – Multimodal AI Model OptimizationResearch EngineerAt Tavus, we are pioneering the human aspect of AI technology. Our objective is to make human-AI interactions as seamless and natural as in-person conversations, allowing for a human touch in areas that were once considered unscalable.We accomplish this through groundbreaking research in multimodal AI, focusing on human-to-human communication modeling (encompassing language, audio, and video) and the development of audio-visual avatar behaviors. Our innovative models drive applications ranging from text-to-video AI avatars to real-time conversational video experiences across sectors such as healthcare, recruitment, sales, and education.By empowering AI to perceive, listen, and engage with an authentic human-like presence, we are laying the groundwork for the next generation of AI workers, assistants, and companions.As a Series B company, we are supported by renowned investors, including Sequoia, Y Combinator, and Scale VC. Join us as we shape the future of human-AI interaction.The RoleWe are seeking an accomplished Research Scientist/Engineer with expertise in model optimization to be a vital part of our core AI team.The ideal candidate thrives in dynamic startup environments, is adept at setting priorities independently, and is open to making calculated decisions. We are moving swiftly and need individuals who can help navigate our path forward.Your MissionTransform state-of-the-art research models into fast, efficient, and production-ready systems through techniques such as sparsification, distillation, and quantization.Oversee the optimization lifecycle for critical models: establish metrics, conduct experiments, and evaluate trade-offs among latency, cost, and quality.Collaborate closely with researchers and engineers to convert innovative concepts into deployable solutions.RequirementsExtensive experience in deep learning with PyTorch.Practical experience in model optimization and compression, including knowledge distillation, pruning/sparsification, quantization, and mixed precision.Familiarity with efficient architectures such as low-rank adapters.Strong grasp of inference performance and GPU/accelerator fundamentals.Proficient in Python coding and adherence to best practices in research engineering.Experience with large models and datasets in cloud environments.Capability to read ML literature, reproduce results, and modify ideas accordingly.
Join David AI as a Product ManagerDavid AI is pioneering the audio data research landscape, harnessing an R&D methodology to develop high-quality datasets that support AI models. Our vision is to seamlessly integrate AI into everyday experiences, with audio serving as the perfect entry point. As the demand for advanced audio AI solutions grows, the necessity for superior training data becomes paramount, and David AI is at the forefront of this evolution.Founded in 2024 by a talented team of former engineers from Scale AI, we’ve quickly established ourselves as a trusted partner for leading FAANG companies and AI labs. Our recent $50M Series B funding round, led by top-tier investors including Meritech and NVIDIA, is a testament to our trajectory and potential.We pride ourselves on our team’s intelligence, humility, and ambition. If you are passionate about research, engineering, product development, or operations, we invite you to contribute to our mission of advancing audio AI.The Product TeamOur Product team operates at the intersection of research, engineering, and operations, crafting audio data products that empower cutting-edge machine learning models. We translate groundbreaking research into scalable and thoughtful data products that prioritize quality, coverage, and clarity for model creators.Your RoleAs a Product Manager on our Product team, you will take charge of the strategy and execution of key components within our audio data portfolio. Collaborating closely with the Research, ML, Engineering, and Operations teams, you will convert model requirements into actionable data roadmaps, quality standards, and deliverables ready for clients.Key ResponsibilitiesSpecialize in areas such as data roadmap expansion, quality metrics, and evaluation based on your interests and background.Lead product strategy and development for audio data offerings, guiding them from research to full-scale production.Convert research insights into specific requirements for data collection, quality assurance, and documentation.Work in partnership with Research, ML, Engineering, and Operations teams to operationalize new datasets and capabilities.Establish metrics and success criteria to prioritize initiatives and evaluate impact across quality, coverage, cost, and speed.
Remote|Remote|Remote-Friendly (Travel Required) | San Francisco, CA
Join Anthropic as a Senior Research Scientist on our Reward Models team, where you will spearhead groundbreaking research aimed at enhancing our understanding of human preferences at scale. Your innovative contributions will directly influence how our AI models, including Claude, align with human values and optimize for user needs. You will delve into the forefront of reward modeling for large language models, designing novel architectures and training methodologies for Reinforcement Learning from Human Feedback (RLHF). Your research will explore advanced evaluation techniques, including rubric-based grading, and tackle challenges such as reward hacking. Collaboration is key, as you'll work alongside teams in Finetuning, Alignment Science, and our broader research organization to ensure your findings result in tangible advancements in AI capabilities and safety. This role offers you an opportunity to address critical AI alignment challenges, leveraging cutting-edge models and substantial computational resources to further the science of safe and capable AI systems.
Full-time|On-site|San Francisco, California, United States
Join Us at Amplifier Health!We are pioneering healthcare innovations with the world's first Large Acoustic Model (LAM), a groundbreaking foundation model that utilizes human voice to identify health conditions. This is where science fiction meets reality, and we have secured substantial funding from leading investors to establish a transformative new category in healthcare.We are in search of a passionate AI researcher who is ready to break free from the traditional "publish or perish" mindset and focus on creating impactful intelligence that truly works in real-world applications.The Reality of Our WorkWe are entering an exhilarating phase of rapid growth. Our commitment to pushing the boundaries of technology is matched only by our dedication to saving lives at scale.Our team collaborates in person in San Francisco, believing that the most challenging problems are best tackled together at a whiteboard rather than through virtual meetings.We operate at a fast pace, quickly transitioning from hypothesis to code, training, and validation with immediate feedback.We enjoy our work and thrive as a close-knit team on an exciting journey, driven by our passion for what we do.Your MissionAs part of our elite AI Research team, you will elevate the state-of-the-art in acoustic modeling. Your role will involve designing innovative architectures to extract clinical-grade biomarkers from raw audio data, not just fine-tuning existing models.The Challenges Ahead:Novel Architectures: You will explore how Transformer architectures can be adapted to process complex acoustic signals and long-range dependencies.Biomarker Discovery: You will conduct experiments to identify specific acoustic features (such as jitter, shimmer, and respiratory rate) that correlate with health conditions, often uncovering new signals that have yet to be recognized by medical science.Data Efficiency: You will contribute to building a foundation model, utilizing self-supervised learning techniques to harness vast amounts of unlabeled audio data.
About TavusTavus is at the forefront of innovation in human computing. Our mission is to develop AI Humans: an advanced interface that bridges the gap between individuals and machines, eliminating the friction found in current technologies. Our state-of-the-art human simulation models empower machines to see, hear, respond, and even exhibit realistic appearances—facilitating genuine, face-to-face interactions. AI Humans integrate the emotional insight of humans with the scalability and dependability of machines, making them reliable agents accessible 24/7, in any language, on our terms.Imagine having access to an affordable therapist, a personal trainer that fits your schedule, or a team of medical assistants dedicated to providing personalized care for every patient. With Tavus, individuals, enterprises, and developers have the tools to create AI Humans that connect, comprehend, and act with empathy on a large scale.We are a Series A company supported by esteemed investors such as Sequoia Capital, Y Combinator, and Scale Venture Partners.Join us in shaping a future where machines and humans genuinely understand one another.The PositionWe are seeking an AI Researcher to join our core AI team and advance the frontiers of multimodal conversational intelligence. If you excel in dynamic environments, enjoy transforming abstract concepts into functional code, and derive motivation from pushing the boundaries of possibility, this role is designed for you.Your Responsibilities Engage in research focusing on Foundational Multimodal Models specifically in the realm of Conversational Avatars (such as Neural Avatars and Talking-Heads).Develop models for video, audio, and language sequences utilizing Autoregressive and Predictive Architectures (e.g., V-JEPA) and/or Diffusion methodologies, with a focus on temporal and sequential data rather than static images.Collaborate closely with the Applied ML team to implement your research into production systems.Remain at the forefront of multimodal learning and assist us in defining what “cutting edge” will mean in the future.Ideal Candidate ProfilePhD (or nearing completion) in a relevant field, or equivalent practical research experience.Experience in multimodal machine learning, particularly focused on conversational interfaces.
Full-time|$250K/yr - $325K/yr|On-site|San Francisco
About World Labs: At World Labs, we create foundational world models capable of perceiving, generating, reasoning, and interacting with the 3D environment. Our mission is to unlock the full potential of AI through spatial intelligence, transforming perception into action, reasoning into insight, and imagination into creation. We believe that spatial intelligence will revolutionize storytelling, creativity, design, simulation, and immersive experiences across both virtual and physical realms. Our world-class team is driven by curiosity and passion, boasting diverse backgrounds in technology, from AI research and systems engineering to product design. This synergy fosters a tight feedback loop between our cutting-edge research and user-empowering products. Role Overview We are seeking an innovative Research Scientist specializing in generative modeling, especially diffusion models, to join our modeling team. This position is ideal for individuals with extensive expertise in applying diffusion models to images, videos, or 3D assets and scenes. While not mandatory, experience in any of the following areas will be considered a significant advantage: Large-scale model trainingResearch in 3D computer vision In this role, you will work closely with researchers, engineers, and product teams to translate advanced 3D modeling and machine learning techniques into practical applications, ensuring our technology stays at the forefront of visual innovation. This position entails substantial hands-on research and engineering work, taking projects from conception to production deployment. Key Responsibilities Design, implement, and train large-scale diffusion models for generating 3D worlds. Develop and experiment with large-scale diffusion models to introduce novel control signals, align with target aesthetic preferences, or optimize for efficient inference. Collaborate closely with research and product teams to comprehend and translate product requirements into actionable technical roadmaps. Contribute actively to all phases of model development, including data curation, experimentation, evaluation, and deployment. Continuously investigate and integrate the latest research in diffusion and generative AI. Serve as a key technical resource within the team, mentoring peers and promoting best practices in generative modeling and machine learning engineering.
About UsTavus is an innovative research lab at the forefront of human computing technology. Our mission is to create AI Humans—advanced interfaces that bridge the gap between individuals and machines, eliminating the friction found in current systems. Our real-time human simulation models empower machines to see, hear, respond, and appear realistic, facilitating genuine, face-to-face conversations. With AI Humans, we blend the emotional intelligence inherent in humans with the extensive reach and reliability of machines, enabling them to serve as capable and trusted agents available 24/7, capable of communicating in any language.Envision a therapist accessible to everyone, a personal trainer that tailors sessions to your schedule, or a fleet of medical assistants dedicated to providing personalized attention to every patient. Tavus enables individuals, enterprises, and developers to create AI Humans that connect, empathize, and act with understanding on a large scale.Backed by prestigious investors like Sequoia Capital, Y Combinator, and Scale Venture Partners, we are a Series A company ready to shape the future of human-machine interaction.Join us in transforming a future where humans and machines genuinely comprehend one another.The RoleWe are seeking a passionate AI Researcher to join our core AI team and advance the science of audio-visual avatar generation. If you thrive in dynamic startup environments, enjoy experimenting with generative models, and are excited to see your research translated into production, you will find a welcoming home here.Your Mission Conduct research and develop cutting-edge audio-visual generation models for conversational agents (e.g., Neural Avatars, Talking Heads).Focus on models that intricately align with conversation flows, ensuring seamless integration of verbal and non-verbal cues.Experiment with diffusion models (DDPMs, LDMs, etc.), long-video generation, and audio synthesis.Collaborate closely with the Applied ML team to transition your research into practical applications.Stay updated on the latest breakthroughs in multimodal generation and contribute to the evolution of this field.You Will Excel If You Have:A PhD (or nearing completion) in a relevant discipline, or equivalent hands-on research experience.Proficiency in applying image/video generation techniques and a solid understanding of machine learning principles.
Oct 16, 2025
Sign in to browse more jobs
Create account — see all 5,512 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.