Senior Applied AI Engineer - Multimodal Perception & Reasoning
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Senior
Qualifications
About voltai
VOLT is revolutionizing the AI landscape by developing next-generation perception systems that prioritize safety, security, and real-time risk detection. Our innovative approach aims to bridge the gap between digital intelligence and physical reality.
Similar jobs
Search for Research Engineer Multimodal Data
6,116 results
Eventual Computing
Eventual Computing builds tools that help AI teams work with large, complex datasets. Based in San Francisco, the company supports projects in robotics, autonomous vehicles, and advanced video generation. Its open-source engine, Daft, is already in use at organizations with demanding data needs. The team focuses on making data curation and model training more efficient, so the right datasets are always within reach. The office is located in the Mission district, where collaboration with leading AI labs and infrastructure companies is part of daily work. Role overview The Research Engineer - Multimodal Data will join the Visual Understanding team. This position centers on building solutions to make vast amounts of video and sensor data accessible and easy to query. The work directly supports researchers who need to find and use specific datasets quickly. What you will do Develop and refine systems that process petabytes of multimodal data, including video and sensor streams. Apply vision-language models to improve how data is discovered and retrieved. Define and influence the roadmap for visual understanding features. Train models to streamline large-scale data annotation and improve efficiency for research teams.
Join worldlabs as a Research Engineer focused on scaling multimodal data. In this dynamic role, you will leverage cutting-edge technologies and methodologies to enhance data processing capabilities. You will be responsible for developing innovative solutions that integrate various data types and drive impactful research outcomes.
Zyphra is a cutting-edge artificial intelligence firm located in the heart of San Francisco, California, dedicated to advancing technology across various modalities.About the Position:We are seeking a Data Engineer - Multimodal Systems to play a pivotal role in the enhancement and expansion of Zyphra's datasets and data pipelines. This position offers a unique opportunity to collaborate with diverse teams and contribute to innovative data solutions. You will engage in the collection of extensive datasets and the development and optimization of high-performance parallel data pipelines.Your Responsibilities Will Include:Executing large-scale data collection across multiple modalities, including text, audio, and image.Designing and implementing highly efficient, parallelized data processing pipelines that integrate various modalities.Conducting rigorous experimental ablations to evaluate the effectiveness of new data enhancements.Candidate Requirements:Proven ability in implementation and prototyping.Capability to transform ideas into experimental frameworks swiftly.Strong collaborative skills, thriving in a dynamic research environment.Eagerness to learn and apply new concepts effectively.Exceptional communication and teamwork skills, capable of contributing to both research and large-scale engineering projects.Preferred Qualifications:Experience in the collection, management, and processing of large datasets.Familiarity with parallel programming frameworks in Python, such as Dask.In-depth understanding of state-of-the-art dataset curation practices.A detail-oriented mindset with a passion for data integrity and verification.Strong foundation in experimental methodologies for conducting thorough ablation studies and hypothesis testing.Knowledge and interest in large-scale, highly parallel data processing systems.Proficiency in PyTorch and Python.Experience with large, complex codebases and the ability to quickly become productive within them.Published research in respected machine learning venues.Postgraduate degree in a relevant field is a plus.
Tavus – Multimodal AI Model OptimizationResearch EngineerAt Tavus, we are pioneering the human aspect of AI technology. Our objective is to make human-AI interactions as seamless and natural as in-person conversations, allowing for a human touch in areas that were once considered unscalable.We accomplish this through groundbreaking research in multimodal AI, focusing on human-to-human communication modeling (encompassing language, audio, and video) and the development of audio-visual avatar behaviors. Our innovative models drive applications ranging from text-to-video AI avatars to real-time conversational video experiences across sectors such as healthcare, recruitment, sales, and education.By empowering AI to perceive, listen, and engage with an authentic human-like presence, we are laying the groundwork for the next generation of AI workers, assistants, and companions.As a Series B company, we are supported by renowned investors, including Sequoia, Y Combinator, and Scale VC. Join us as we shape the future of human-AI interaction.The RoleWe are seeking an accomplished Research Scientist/Engineer with expertise in model optimization to be a vital part of our core AI team.The ideal candidate thrives in dynamic startup environments, is adept at setting priorities independently, and is open to making calculated decisions. We are moving swiftly and need individuals who can help navigate our path forward.Your MissionTransform state-of-the-art research models into fast, efficient, and production-ready systems through techniques such as sparsification, distillation, and quantization.Oversee the optimization lifecycle for critical models: establish metrics, conduct experiments, and evaluate trade-offs among latency, cost, and quality.Collaborate closely with researchers and engineers to convert innovative concepts into deployable solutions.RequirementsExtensive experience in deep learning with PyTorch.Practical experience in model optimization and compression, including knowledge distillation, pruning/sparsification, quantization, and mixed precision.Familiarity with efficient architectures such as low-rank adapters.Strong grasp of inference performance and GPU/accelerator fundamentals.Proficient in Python coding and adherence to best practices in research engineering.Experience with large models and datasets in cloud environments.Capability to read ML literature, reproduce results, and modify ideas accordingly.
About TavusTavus is at the forefront of innovation in human computing. Our mission is to develop AI Humans: an advanced interface that bridges the gap between individuals and machines, eliminating the friction found in current technologies. Our state-of-the-art human simulation models empower machines to see, hear, respond, and even exhibit realistic appearances—facilitating genuine, face-to-face interactions. AI Humans integrate the emotional insight of humans with the scalability and dependability of machines, making them reliable agents accessible 24/7, in any language, on our terms.Imagine having access to an affordable therapist, a personal trainer that fits your schedule, or a team of medical assistants dedicated to providing personalized care for every patient. With Tavus, individuals, enterprises, and developers have the tools to create AI Humans that connect, comprehend, and act with empathy on a large scale.We are a Series A company supported by esteemed investors such as Sequoia Capital, Y Combinator, and Scale Venture Partners.Join us in shaping a future where machines and humans genuinely understand one another.The PositionWe are seeking an AI Researcher to join our core AI team and advance the frontiers of multimodal conversational intelligence. If you excel in dynamic environments, enjoy transforming abstract concepts into functional code, and derive motivation from pushing the boundaries of possibility, this role is designed for you.Your Responsibilities Engage in research focusing on Foundational Multimodal Models specifically in the realm of Conversational Avatars (such as Neural Avatars and Talking-Heads).Develop models for video, audio, and language sequences utilizing Autoregressive and Predictive Architectures (e.g., V-JEPA) and/or Diffusion methodologies, with a focus on temporal and sequential data rather than static images.Collaborate closely with the Applied ML team to implement your research into production systems.Remain at the forefront of multimodal learning and assist us in defining what “cutting edge” will mean in the future.Ideal Candidate ProfilePhD (or nearing completion) in a relevant field, or equivalent practical research experience.Experience in multimodal machine learning, particularly focused on conversational interfaces.
Bland Inc. seeks a Machine Learning Researcher specializing in Multimodal Large Language Models (LLMs) to join the team in San Francisco. The focus is on advancing AI systems that integrate language with other types of data. Role overview This position centers on research and development aimed at improving how AI models process and understand information from multiple sources, such as text combined with images or other modalities. What you will do Investigate how language interacts with additional data types within multimodal LLMs Create and evaluate new methods to enhance AI model performance Work closely with colleagues on projects designed to push the boundaries of machine learning Location This role is based in San Francisco.
About UsTavus is an innovative research lab at the forefront of human computing technology. Our mission is to create AI Humans—advanced interfaces that bridge the gap between individuals and machines, eliminating the friction found in current systems. Our real-time human simulation models empower machines to see, hear, respond, and appear realistic, facilitating genuine, face-to-face conversations. With AI Humans, we blend the emotional intelligence inherent in humans with the extensive reach and reliability of machines, enabling them to serve as capable and trusted agents available 24/7, capable of communicating in any language.Envision a therapist accessible to everyone, a personal trainer that tailors sessions to your schedule, or a fleet of medical assistants dedicated to providing personalized attention to every patient. Tavus enables individuals, enterprises, and developers to create AI Humans that connect, empathize, and act with understanding on a large scale.Backed by prestigious investors like Sequoia Capital, Y Combinator, and Scale Venture Partners, we are a Series A company ready to shape the future of human-machine interaction.Join us in transforming a future where humans and machines genuinely comprehend one another.The RoleWe are seeking a passionate AI Researcher to join our core AI team and advance the science of audio-visual avatar generation. If you thrive in dynamic startup environments, enjoy experimenting with generative models, and are excited to see your research translated into production, you will find a welcoming home here.Your Mission Conduct research and develop cutting-edge audio-visual generation models for conversational agents (e.g., Neural Avatars, Talking Heads).Focus on models that intricately align with conversation flows, ensuring seamless integration of verbal and non-verbal cues.Experiment with diffusion models (DDPMs, LDMs, etc.), long-video generation, and audio synthesis.Collaborate closely with the Applied ML team to transition your research into practical applications.Stay updated on the latest breakthroughs in multimodal generation and contribute to the evolution of this field.You Will Excel If You Have:A PhD (or nearing completion) in a relevant discipline, or equivalent hands-on research experience.Proficiency in applying image/video generation techniques and a solid understanding of machine learning principles.
About Our Team:At OpenAI, we are dedicated to ensuring that artificial general intelligence (AGI) serves the greater good of humanity. Our API has emerged as the most widely embraced AI platform in the industry, catering to a diverse clientele ranging from startups and independent developers to Fortune 500 companies. By leveraging our multimodal APIs—which encompass real-time interactions, text-to-speech (TTS), speech generation, and image creation—we empower users to effectively harness the full spectrum of AI capabilities at scale.About the Role:We are on the lookout for an Engineering Manager to spearhead our multimodal API product suite. In this pivotal role, you will guide a talented team focused on delivering cutting-edge APIs for real-time processing, speech transcription, speech generation, and image creation. You will be instrumental in shaping the product roadmap and developing the tools that enable developers to connect with millions of end users through AI-driven audio, video, and imagery.In this role, you will:Lead, mentor, and cultivate a high-performing engineering team dedicated to multimodal API products, including our real-time API, Whisper transcription models, TTS speech generation models, and DALLE image generation APIs.Collaborate with product managers, designers, and various stakeholders to articulate the strategic vision and product roadmap.Work alongside our research teams to enhance our core multimodal models tailored for API customer use cases.Steer technical and architectural decisions with a focus on scalability, robustness, and user experience.Promote a culture of innovation, continuous improvement, and accountability within your team.Qualifications:Demonstrated experience in managing engineering teams that successfully deliver complex, high-quality products at scale.Strong technical expertise with proficiency in modern software engineering practices and system architecture.Exceptional collaboration and communication skills to effectively engage with diverse teams and stakeholders.Familiarity with or a strong passion for multimodal AI, encompassing speech technologies, real-time systems, and image generation.Adept at thriving in a fast-paced, dynamic startup environment.
Thinking Machines Lab
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We envision a future where everyone can harness the knowledge and tools necessary to make AI work for their unique objectives.Comprising a team of scientists, engineers, and innovators, we have developed some of the most widely employed AI products, including ChatGPT and Character.ai, as well as open-weight models such as Mistral and popular open-source projects like PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleAt Thinking Machines, we prioritize a multimodal-first approach. We are seeking new team members to push the boundaries of visual perception and multimodal learning. Our focus is on understanding the interplay between vision and language at scale. We design innovative architectures that integrate pixels and text, create datasets and evaluation methods that assess real-world comprehension, and develop representations that enable models to connect abstract concepts with the physical world. Our aim is to build multimodal systems that seamlessly integrate into real-world applications.Your work will be at the intersection of visual understanding, multimodal reasoning, and large-scale model training. You will contribute to the development of architectures, data, and evaluation tools that teach AI to perceive, comprehend, and collaborate effectively. The ideal candidate is inquisitive about multimodal interfaces, possesses experience in conducting large-scale experiments, and is adept at contributing to complex engineering systems. While we seek individuals with expertise in multimodality, our collaborative environment encourages all new hires to work across modalities as a unified team.This role merges foundational research with practical engineering since we do not differentiate between these roles internally. You will be expected to write high-performance code and analyze technical reports. This position is perfect for someone who enjoys both deep theoretical inquiry and hands-on experimentation and is eager to influence the foundational aspects of AI learning.Note: This is an "evergreen role" that we keep open continuously to express interest in this research area. We receive a high volume of applications, and there may not always be an immediate position that perfectly matches your experience and skills. We encourage you to apply regardless. Applications are reviewed regularly, and we reach out to candidates as new opportunities arise. You are welcome to reapply if your experience increases, but please refrain from applying more than once every six months. Additionally, we may post specific roles for particular project or team needs, where you are also welcome to apply directly in addition to this evergreen role.
The Bot Company
The Bot CompanyAt The Bot Company, we are on a mission to create an innovative robotic assistant for every household.Our dynamic team, composed of talented engineers, designers, and operators, is based in San Francisco. We have a rich background from industry leaders such as Tesla, Cruise, OpenAI, Google, and Pixar, and we have successfully delivered products to hundreds of millions of users, honing our ability to create exceptional products and experiences.We pride ourselves on maintaining a streamlined team structure that fosters swift decision-making and minimizes bureaucracy. Each member is considered an Individual Contributor, granted substantial autonomy, ownership, and accountability. Our culture enables us to work across the technology stack with an emphasis on rapid iteration and execution.What We Seek in CandidatesCandidates for all positions at The Bot Company must exhibit remarkable sharpness and the capacity to thrive in high-pressure environments. We expect candidates to showcase:Exceptional Cognitive Abilities: You possess quick thinking, instant learning capabilities, and the ability to reason across diverse domains.Engineering Curiosity: You demonstrate an innate desire to understand how systems function, even beyond your area of expertise.Performance-Driven Attitude: You excel in fast-paced settings, effectively navigate ambiguity, and thrive under demanding circumstances.Machine Learning: Multimodal Foundation ModelsWe are developing unified foundation models capable of reasoning across text, images, video, and kinematics to inform intelligent robotic behaviors.You will engage with large-scale multimodal networks, overseeing the complete process from data handling to model training and deployment.Your ResponsibilitiesConstruct Native Multimodal Policies: Create architectures where vision, language, and other modalities are represented in a unified manner.Enhance Cross-Modal Reasoning: Explore and implement strategies to ensure that the model not only correlates modalities but also comprehends them (e.g., linking visual physics to kinematic constraints).Manage the Training Loop from Start to Finish: Design, execute, troubleshoot, and refine large-scale training experiments; identify failure points, enhance data mixtures, and tighten evaluations to achieve measurable improvements.Deploy and Refine Real Systems: Integrate models into practical robotic frameworks, enhance robot code for model deployment, and optimize performance for edge inference.
About Hike Medical Hike Medical is building the future of musculoskeletal care by combining advanced technology with practical healthcare solutions. Based in San Francisco’s Rincon Hill, the team develops a platform that spans three core areas: an AI-powered vision system for rapid web-based foot scans that generate custom 3D-printed orthotics, an AI agent platform that manages the entire DME workflow from intake through claims, and SoleForge, a high-scale 3D printing facility for custom medical devices. Hike Medical partners with some of the world’s largest employers and major orthotics and prosthetics organizations. Fortune 50 companies trust the platform to support employee well-being, and a broad network of clinical partners keeps the company connected to real-world needs. Custom insoles are just the starting point. The long-term goal is to reshape the industry with bionic devices: AI-designed, robotically manufactured orthotic and prosthetic products. The company aims to reach this milestone by 2040. Learn more at bionics2040.com. With $22 million raised across Seed and Series A rounds from leading investors, Hike Medical offers a results-oriented culture for those interested in the intersection of AI, manufacturing, and healthcare.
Join VOLT, a trailblazer in crafting advanced AI perception systems that enhance safety and security through real-time risk detection in the physical world.We are on the lookout for a Senior Applied AI & Machine Learning Engineer dedicated to designing, optimizing, and deploying multimodal AI models capable of functioning reliably in diverse real-world scenarios. This is a hands-on role focused on transitioning models from conceptual data to practical production, encompassing both edge devices and cloud infrastructures.In this position, you will engage with vision, video, and language-based models that interpret real-world scenes and events, ensuring their accuracy, latency, robustness, and cost-effectiveness in production systems.Reporting directly to the Head of Engineering, you will play a pivotal role in advancing VOLT AI’s core perception platform.
About Our TeamAt OpenAI, we are dedicated to ensuring that artificial general intelligence (AGI) serves and benefits all of humanity. A vital component of this mission involves developing models that genuinely understand and resonate with human preferences. Our Human Data team is instrumental in making this a reality.The Human Data engineering team is responsible for creating sophisticated systems that facilitate scalable and high-quality human feedback, which is crucial for training and refining OpenAI's most advanced models. Our engineers work in close collaboration with top-tier researchers to implement alignment techniques—from initial experimental concepts to production-ready feedback loops.Position OverviewWe are seeking passionate software engineers to become part of the Human Data team, tasked with developing the platforms, prototypes, tools, and infrastructure essential for training, aligning, and evaluating our AI models. In this role, you will collaborate with researchers and cross-functional teams to actualize alignment concepts, influence the training of future models, and enhance how our models engage with the real world.We are looking for individuals who thrive on technical ownership, enjoy working across the stack, and are eager to tackle complex challenges in a dynamic, impactful environment.This position is based in San Francisco, CA, and follows a hybrid work model of three days in the office each week. We also provide relocation assistance for new hires.Your ResponsibilitiesDevelop and maintain robust full-stack systems for feedback collection, data labeling, and evaluation pipelines while ensuring high levels of security.Convert experimental alignment research into scalable production infrastructure, including inference and model training systems.Design and enhance user-facing tools and backend services to support high-quality data workflows.Collaborate with researchers, engineers, and program leads to refine feedback loops and model interaction strategies.Lead infrastructure improvements that promote faster iterations and scaling across OpenAI’s cutting-edge models, from internal research tools to production-level ChatGPT.QualificationsProven software engineering skills with experience in building scalable production systems.A strong preference for full-stack development with end-to-end ownership—from backend pipelines to user interfaces.Driven by high-impact projects and capable of navigating ambiguous challenges.
Our MissionAt Altos Labs, we are dedicated to restoring cell health and resilience through innovative cell rejuvenation techniques aimed at reversing diseases, injuries, and disabilities that can arise throughout life.For further insights, please visit our website at altoslabs.com.Our ValueOur singular Altos Value is: Everyone Owns Achieving Our Inspiring Mission.Diversity at AltosWe firmly believe that diverse perspectives are crucial for scientific innovation. At Altos, exceptional scientists and industry leaders collaborate globally to further our shared mission. We prioritize Belonging, ensuring all employees feel valued for their unique perspectives, and we hold ourselves accountable for maintaining a diverse and inclusive environment.Your Contributions to AltosAs a member of our team, you will accelerate and enhance our efforts in developing unified, multi-modal generative foundation models tailored for multiscale biology. You will be a key player in multidisciplinary teams that create the computational platforms essential for Altos to fulfill its mission.In this position, you will collaborate with other scientists and engineers across the Institute of Computation to design, develop, and scale cutting-edge foundation models that address biological inquiries and assist in discovering novel interventions for aging and disease. Your focus will be on synthesizing unstructured multimodal signals with structured relational data and knowledge graphs that depict biological realities.The ideal candidate will excel in a dynamic environment that values teamwork, transparency, scientific excellence, originality, and integrity.
About the Role Anthropic is hiring a Software Engineer to support our Research Data Platform. This position is based in San Francisco, CA or New York City, NY. The team works closely with engineers and researchers to create data solutions that strengthen our research efforts. The work directly supports Anthropic’s mission to develop safe and beneficial AI.
About Our Innovative TeamJoin the Workload team at OpenAI, where we are at the forefront of designing and managing the cutting-edge infrastructure that drives the training and inference of large language models (LLMs) at an unprecedented scale. Our systems are engineered to harmonize the complex processes of model training and serving, abstracting performance, parallelism, and execution across extensive GPU and accelerator networks. This robust foundation allows researchers to concentrate on elevating model capabilities, while we take care of the scalability, efficiency, and reliability needed to bring these advanced models to life.Your Role and ResponsibilitiesWe are seeking a talented engineer to design and implement the dataset infrastructure that will fuel OpenAI’s next-generation training stack. Your primary focus will be on creating standardized dataset interfaces, scaling pipelines across thousands of GPUs, and proactively identifying and addressing performance bottlenecks. Collaboration with multimodal researchers and infrastructure teams will be key to ensuring that our datasets are unified, efficient, and user-friendly.Key Responsibilities Include:Design and maintain standardized dataset APIs, including those for multimodal (MM) data that exceeds memory capacity.Develop proactive testing and validation pipelines for dataset loading at GPU scale.Work collaboratively to integrate datasets into training and inference pipelines, ensuring seamless user experiences.Document and maintain dataset interfaces to ensure they are discoverable, consistent, and easily adoptable by other teams.Establish validation systems to assure datasets remain reproducible and unchanged once standardized.Identify and troubleshoot performance bottlenecks in distributed dataset loading, such as stragglers impacting global training speed.Create visualization and inspection tools to highlight errors, bugs, or bottlenecks in datasets.Ideal Candidate ProfilePossess strong engineering fundamentals and experience in distributed systems, data pipelines, or infrastructure.Have a proven track record in building APIs, modular code, and scalable abstractions, with a user-centric approach to design.Be adept at debugging performance issues across large-scale machine fleets.Demonstrate a passion for advancing data infrastructure to enhance research capabilities.
AfterQuery
About AfterQuery AfterQuery partners with leading AI labs to advance training data and evaluation frameworks. The team builds high-signal datasets and runs thorough evaluations that go beyond standard benchmarks. As a post-Series A, early-stage company in San Francisco, AfterQuery gives each team member room to shape the future of AI models. Role Overview: Research Scientist - Frontier Data This role focuses on designing datasets and developing evaluation systems that influence how top AI models are trained and assessed. Working closely with research teams at major AI labs, the scientist explores new data collection techniques, investigates where models fall short, and sets up metrics to track progress. The work is hands-on and experimental, moving quickly from hypothesis to live testing and directly impacting large-scale model training. Key Responsibilities Design data slides and analyze data structures to uncover model weaknesses in areas like finance, software development, and enterprise operations. Build and refine evaluation rubrics and reward signals for RLHF and RLVR training approaches. Study annotator behavior and run experiments to improve model capabilities across different domains. Develop quantitative frameworks to measure dataset quality, diversity, and their effect on model alignment and performance. Work with research teams to turn training objectives into concrete data and evaluation needs. What We Look For Experience as an undergraduate or master’s research student (PhD not required). Background or internships with RL environments or AI safety and benchmarking organizations (e.g., METR, Artificial Analysis) is a strong plus. Genuine interest in how data structure, selection, and quality affect model outcomes. Demonstrated skill in designing experiments, acting quickly, and extracting insights from complex data. Comfort working across sectors such as finance, software engineering, and policy. Strong quantitative background and familiarity with LLM training pipelines, RLHF/RLVR methods, or evaluation frameworks. A hands-on mindset focused on building practical solutions.
About CartesiaAt Cartesia, our vision is to create the future of artificial intelligence—intelligent systems that are seamlessly integrated into daily life. We aim to overcome current limitations by enabling models to continuously understand and analyze vast streams of audio, video, and text data—ranging from 1 billion text tokens to 1 trillion video tokens—right on your device.Our pioneering team, comprised of PhDs from the Stanford AI Lab, has developed State Space Models (SSMs), a groundbreaking approach to training efficient, large-scale foundation models. With a rich blend of expertise in model innovation and systems engineering, alongside a product-focused engineering team, we are committed to developing and delivering cutting-edge AI models and user experiences.Supported by prominent investors including Index Ventures and Lightspeed Venture Partners, as well as many esteemed advisors and over 90 angel investors from diverse industries, we are at the forefront of AI advancements.About The RoleIn our quest to create truly global AI, we must train our models using datasets that represent the vast diversity of languages and cultures around the world. We are looking for a Research Engineer to take charge of the quality and comprehensiveness of the data that drives our models. As our in-house expert in global data, you will ensure that our models excel across multiple languages, leveraging your keen understanding of linguistic subtleties and your enthusiasm for building inclusive, large-scale datasets.Your ImpactDesign and construct extensive datasets for model training, conducting controlled experiments to evaluate their effect on model performance.Develop assessments for speech models through both manual annotation and automated evaluation metrics.Utilize data generation techniques to enhance model intelligence and reduce biases.Create automated quality control systems to validate and filter the generated data.Collaborate with product teams to ensure optimal support for key languages and markets.What You BringProven experience in developing or working with extensive multilingual datasets.Familiarity with generative models, including speech, text, or multimodal systems.Ability to guide human annotation and evaluation across various languages.Strong analytical skills and a passion for data-driven decision-making.
Join fuku as an Applied Research Engineer in San Francisco, CA, where you will be at the forefront of AI video data research. As a crucial member of our team, your mission will involve building robust, high-performance frameworks and extensive pipelines to process and decode video data with exceptional accuracy. You will tackle complex research challenges, refine machine learning models and APIs, and deliver comprehensive solutions across computer vision, audio, and text processing domains. This role is designed for engineers who thrive in both research and production environments and are eager to spearhead the evolution of video understanding from research to deployment.
Our MissionAt Altos Labs, we are dedicated to revitalizing cell health and resilience through innovative cell rejuvenation techniques to reverse disease, injury, and the disabilities that arise throughout life.Discover more about our vision at altoslabs.com.Our ValuesOur core value is simple yet powerful: Everyone Owns Achieving Our Inspiring Mission.Diversity at AltosWe understand that diverse perspectives are crucial for scientific breakthroughs and exploration. At Altos, exceptional scientists and industry leaders collaborate from around the globe to drive our shared mission forward. We prioritize a culture of belonging, ensuring that every employee feels valued for their unique contributions. We are all responsible for maintaining a diverse and inclusive workplace.Your Contributions to AltosJoin Altos Labs in creating a premier AI ecosystem aimed at addressing the most intricate challenges in human biology. You will be instrumental in designing and developing high-performance, scalable solutions that integrate high-dimensional biomedical imaging with molecular and linguistic data.Your role will involve implementing large-scale multimodal data fusion, advancing beyond basic image analysis to develop predictive models that span various biological domains. You will engage directly with data and coding, partnering with our engineering team to ensure these models are scalable, efficiently trainable in distributed cloud environments, and accessible to our global research network.Key ResponsibilitiesModel Development: Create, implement, and train large-scale foundational models (e.g., Vision Transformers, Multimodal LLMs) capable of embedding spatial data and integrating diverse modalities.Innovative Data Fusion: Apply cutting-edge cross-domain mapping and fusion techniques to align heterogeneous biological datasets.Scaling & Training: Develop and oversee high-performance ML pipelines designed to handle petabyte-scale image collections and multi-omics data streams in a cloud infrastructure.Technical Collaboration: Work closely with experimental scientists and software engineers to convert biological complexity into high-performance code and reliable distributed systems.Who You AreWe seek a technical expert who excels at unraveling "unsolvable" challenges through programming and meticulous experimentation. We welcome candidates at the Scientist I, Scientist II, or Senior Scientist levels.
Sign in to browse more jobs
Create account — see all 6,116 results

