Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Experience
Qualifications
The ideal candidate will have proficiency in Python and experience with frameworks such as Torch and Diffusers. You will have the opportunity to utilize our extensive GPU cluster for training and inference, collaborating with a dynamic team that is committed to rapidly deploying groundbreaking AI advancements.
About the job
We are seeking a talented Applied Machine Learning Engineer with a comprehensive understanding of the generative media landscape. You should possess an up-to-date knowledge of emerging methodologies and be capable of identifying gaps in the current market. Your role will involve innovating and developing machine learning models that enhance user experiences, requiring both novel training techniques and the fine-tuning of existing models with fresh datasets.
About fal
fal is at the forefront of AI innovation, dedicated to challenging the status quo and creating transformative solutions in the generative media space. Join us and be part of a team that values creativity, learning, and collaboration.
Similar jobs
1 - 20 of 8,813 Jobs
Search for Senior Applied Ai Engineer Machine Learning For Systems Infrastructure
About UsAt Applied Compute, we specialize in creating Specific Intelligence solutions for enterprises, developing agents that learn continuously from an organization’s processes, data, expertise, and objectives. We recognize a significant gap between the capabilities of AI models in isolation and their practical applications in real-world business contexts. Our systems often fall short because they lack adaptability to feedback. To address this, we are building a continual learning infrastructure that captures context, memory, and decision-making processes throughout the enterprise, enabling specialized agents to effectively execute real tasks.What Excites Us: We operate at a unique intersection where our product team constructs the platform that fuels a new generation of digital coworkers. Our research team pushes the boundaries of post-training and reinforcement learning, creating innovative product experiences. Our applied research engineers collaborate closely with clients to deploy models into production. This blend of strong product focus, deep research, and hands-on customer engagement is crucial for integrating AI into the enterprise. We are product-driven, research-informed, and actively engaged with our clients.Our Team: Our diverse team consists of engineers, researchers, and operators, many of whom are former founders. We have built RL infrastructure at leading organizations like OpenAI and Scale AI, and developed systems at Together, Two Sigma, and Watershed. We proudly serve Fortune 50 clients alongside companies like DoorDash, Mercor, and Cognition. Our work is supported by renowned investors, including Benchmark, Sequoia, and Lux.Who Thrives in Our Environment: We seek individuals eager to apply cutting-edge research and complex systems to tackle real-world challenges. You should be adept at quickly adapting to new environments, whether it’s a fresh codebase, a client’s data architecture, or an unfamiliar problem domain. A genuine enjoyment of customer interactions—listening, empathizing, and understanding how tasks are accomplished within their organizations—is essential. Those with entrepreneurial backgrounds, extensive side projects, or demonstrated end-to-end ownership typically excel in our company.
Full-time|$166K/yr - $210.3K/yr|On-site|San Francisco, California
P-1380 Join Databricks as a Senior Applied AI Engineer, where you will harness the power of machine learning, scheduling, and optimization algorithms to enhance the efficiency and performance of our engineering systems and infrastructure. Our Applied AI team tackles some of the most challenging and fascinating issues in the industry, ensuring that Databricks infrastructure and products operate at peak performance and cost efficiency. This role is critical, as our customers depend on us to deliver the most optimized workloads. Your Impact: Develop comprehensive systems from the ground up within a dynamic team of seasoned professionals. Influence the direction of our applied machine learning investment areas by collaborating with engineering and product teams across the organization. Lead the design and implementation of advanced AI models and systems that enhance the capabilities and performance of Databricks' products, infrastructure, and services. Architect and deploy robust, scalable machine learning infrastructure, including data storage, processing, model training, serving components, and monitoring systems to facilitate seamless integration of AI/ML models into production environments. Explore innovative modeling techniques in the realm of machine learning for systems. Contribute to the wider AI community by publishing research, presenting at conferences, and actively engaging in open-source projects, thereby strengthening Databricks' reputation as an industry leader.
Join Our Team at MacroscopeAt Macroscope, we are dedicated to being the definitive source of truth for any software development company. Our mission is to empower leaders with clarity and provide engineers with the time they need to innovate.We enable leaders to gain insights into the evolution of their products and codebases—tracking changes, understanding team contributions, and identifying progress—all grounded in the ultimate source of truth: the code itself.Founded by experienced entrepreneurs who have successfully built and sold multiple companies, and held executive positions in public tech firms, we are backed by top-tier venture capital firms such as Lightspeed Venture Partners, Thrive Capital, Google Ventures, and Adverb.The RoleWe are seeking a Senior Applied Machine Learning Engineer who will be responsible for designing, developing, and optimizing the ML and AI systems that drive our core offerings. You will have full ownership of the systems, overseeing everything from data collection and evaluation to model experimentation and large-scale production deployment.This cross-functional position entails leading the ML/AI lifecycle for one of our most vital features: AI Code Review. Collaborating closely with our co-founders, you will make pivotal decisions that shape our product's development—ranging from building high-quality datasets to interpreting experimental results and enhancing model performance architecture. Additionally, you will play a significant role in crafting and implementing software that seamlessly integrates our models with our backend applications and user experience, offering a unique opportunity to influence our product's evolution significantly.Technology Stack: Typescript/React (frontend), Golang (backend), Temporal, Google Cloud (GCP), Postgres, Terraform, and custom-built AST "code walkers" in several programming languages including Golang, Typescript, Swift, Python, and Rust.
Join us at Foxglove, where we are revolutionizing the robotics industry by building robust data infrastructure for real-world applications.As robotics transitions from research environments to practical implementations in factories, warehouses, vehicles, and field operations, data becomes essential for engineers to troubleshoot failures, understand unexpected behaviors, and enhance robotic systems.At Foxglove, we provide the observability, visualization, and data infrastructure that enable robotics and autonomous systems teams to efficiently ingest, store, query, replay, and analyze extensive volumes of multimodal sensor data from live systems and production fleets.About the RoleWe are seeking a talented Applied Machine Learning Engineer with strong infrastructure insights to design, deploy, and scale the machine learning systems that power our data platform. In this impactful role, you will be responsible for optimizing production ML infrastructure—from enhancing inference pipeline throughput to establishing training and evaluation workflows. You will focus on high-priority challenges, such as developing retrieval applications for petabyte-scale multimodal robotics data, utilizing cutting-edge models to create high-performance search and data mining products, and fostering an internal ML flywheel for rapid iteration. This is a hands-on, application-driven position rather than a research-focused role.Key ResponsibilitiesDeploy and manage inference infrastructure for production ML workloads, focusing on model serving, scalability, and cost efficiency.Build and oversee vector database integrations and embedding applications to facilitate semantic search across various multimodal robotics data types (image, video, point cloud, and time series).Design and implement evaluation and training infrastructure to enhance model performance rapidly.Lead cloud architecture decisions and tools to optimize inference latency, throughput, cost, and reliability at scale.Collaborate closely with product engineers to deliver application-driven ML features that empower developers at the forefront of robotics and physical AI, steering clear of prototype experiments.Identify appropriate off-the-shelf solutions for production and determine when to build versus buy.
Join Our Innovative Team at ArcadeAt Arcade, we are revolutionizing the way physical products are created with our cutting-edge AI platform. We empower individuals to turn their creative ideas into tangible products seamlessly, utilizing natural language and generative AI. Our mission is to democratize product design, making it as effortless as sharing a post online.Backed by a remarkable $42M in funding from industry-leading investors including Reid Hoffman and Ashton Kutcher, our company is a rising star in the tech landscape. Guided by our founder Mariam Naficy and a team steeped in AI and design expertise, we are at the forefront of a new frontier that merges AI, personal expression, and on-demand manufacturing.Your Role as an Applied AI EngineerWe are on the lookout for an Applied AI Engineer to enhance our generative AI capabilities. This position combines hands-on model development with the integration of advanced AI techniques into our production systems. You will collaborate with diverse teams to conduct research, experiment with models, and implement AI-driven products.
About AbridgeAbridge, established in 2018, is dedicated to enhancing the understanding of healthcare through advanced AI technology. Our platform is specifically designed for medical conversations, streamlining clinical documentation processes and allowing healthcare professionals to prioritize patient care.Our robust technology converts patient-clinician dialogues into structured clinical notes in real-time, integrating seamlessly with electronic medical records (EMR). With our unique Linked Evidence approach and auditable AI framework, we are the sole entity that aligns AI-generated summaries with verified ground truths, fostering trust among healthcare providers. As leaders in generative AI within the healthcare sector, we are committed to setting benchmarks for the ethical implementation of AI across health systems.Our dynamic team comprises practicing MDs, AI researchers, PhDs, creative thinkers, technologists, and engineers, all collaborating to empower individuals and enhance the healthcare experience. We have offices located in San Francisco's Mission District, New York's SoHo, and Pittsburgh's East Liberty.The RoleAs a Senior Machine Learning Infrastructure Engineer at Abridge, you will be essential in constructing and refining the core infrastructure that supports our machine learning models. Your contributions will be crucial in boosting the scalability, efficiency, and performance of our AI solutions. You will collaborate with the Infrastructure and Research teams to build, deploy, optimize, and orchestrate our AI models.What You'll DoDesign, deploy, and maintain scalable Kubernetes clusters for AI model training and inference.Develop, optimize, and maintain high-performance ML serving and training infrastructure, ensuring minimal latency.Work alongside ML and product teams to enhance backend infrastructure for AI-driven applications, focusing on model deployment and efficiency.Improve compute-intensive workflows and maximize GPU utilization for ML tasks.Create a robust orchestration system for model APIs.Partner with leadership to formulate and execute strategies for scaling infrastructure as the company expands, guaranteeing sustained efficiency and performance.
Company Overview At Specter, we are pioneering a software-defined "control plane" designed to enhance the real-world perception of physical assets. Our mission begins with safeguarding American businesses by providing them with comprehensive insights into their physical environments.To achieve this, we are developing a robust hardware-software ecosystem leveraging multi-modal wireless mesh sensing technology. This innovation allows us to significantly reduce the cost and time involved in sensor deployment by a factor of ten. Ultimately, our platform aims to serve as the perception engine for businesses, facilitating real-time visibility and autonomous management of their operational perimeters.Our co-founders, Xerxes and Philip, are deeply committed to empowering our partners in the rapidly evolving landscape of physical AI and robotics. We are a dynamic, rapidly expanding team comprised of talent from Anduril, Tesla, Uber, and the U.S. Special Forces.Position Overview Specter is seeking a dedicated Machine Learning Infrastructure Engineer to construct and optimize the ML systems that drive real-time perception and inference capabilities across our edge-cloud platform. This position will involve overseeing the training, deployment, and enhancement of computer vision and sensor fusion models, aimed at enabling autonomous monitoring and decision-making for our clients' physical assets.Key Responsibilities Include:Design and implement scalable ML training pipelines for computer vision applications, including object detection, tracking, classification, and segmentation.Develop efficient model serving infrastructures to facilitate real-time inference on edge devices with limited computational and power resources.Optimize models for deployment on embedded hardware, employing techniques such as quantization, pruning, TensorRT, ONNX, and CoreML.Create continuous training and evaluation systems to enhance model performance through feedback loops derived from production data.Establish data pipelines for the ingestion, labeling, versioning, and management of extensive multi-modal sensor datasets, including video, radar, lidar, and thermal data.Implement model monitoring frameworks, A/B testing methodologies, and performance analytics for deployed perception systems.Collaborate with perception researchers to transition models from research environments to scalable production across thousands of edge nodes.Construct tools and infrastructure for distributed training, hyperparameter optimization, and experiment tracking.
Role Overview Paraform is hiring an Applied AI Engineer in San Francisco. This role focuses on building and deploying AI systems that directly serve real users. The ideal candidate brings 2-5 years of experience, a strong grasp of modern LLM-based technologies, and a track record of turning advanced models into reliable product features. Success in this role depends on sound product sense and the ability to weigh trade-offs between LLM and traditional machine learning approaches. Experience with LLM-powered applications, retrieval systems, agentic workflows, or automation is valuable. Familiarity with classic ML techniques, such as ranking, recommendation, or classification, will help in designing hybrid systems that balance performance, cost, and reliability. What You Will Do Design and build AI systems to improve matchmaking, ranking, and automation in the Paraform marketplace. Develop LLM-driven features, including retrieval pipelines and agentic workflows, to streamline recruiter and company interactions. Own systems end-to-end: from data pipelines and model design to deployment, monitoring, and iteration in production. Work closely with product managers, ML engineers, and full-stack teams to deliver AI capabilities that shape marketplace outcomes. Create evaluation frameworks to measure real-world performance, reliability, and business impact, not just offline metrics. Set best practices for building and maintaining production AI systems, balancing model quality, cost, latency, and maintainability. Advance the integration of AI into product experiences across the platform. What We Look For 2-5 years of experience at an AI-focused startup (Series A through D). Background working on products with a broad user base, beyond single-enterprise deployments. Proficient in Python and Typescript. Experience developing agentic systems that drive measurable business or user outcomes. Comfort with ambiguity and building in 0 to 1 environments. Ability to communicate technical trade-offs clearly to non-technical stakeholders.
Join David AIAt David AI, we are pioneering the audio data research landscape. Our research and development approach to data ensures that we deliver datasets with the same precision and rigor that leading AI labs apply to their models. Our mission is to seamlessly integrate AI into everyday life, leveraging audio as a key channel. As we witness advancements in audio AI and the emergence of new use cases, we recognize that high-quality training data is the critical component. This is where David AI steps in.Founded in 2024 by a group of former engineers and operators from Scale AI, we have rapidly established partnerships with major FAANG companies and AI labs. Recently, we secured a $50M Series B funding round from prominent investors including Meritech, NVIDIA, Jack Altman (Alt Capital), Amplify Partners, and First Round Capital.Our team is sharp, humble, and ambitious. We are on the lookout for talented individuals in research, engineering, product management, and operations to join us in our mission to redefine the audio AI landscape.About Our Machine Learning TeamOur Machine Learning team operates at the forefront of innovative research and practical application, transforming raw audio into high-quality data for top AI labs and enterprises. We manage the entire machine learning lifecycle—from exploring novel speech processing algorithms to deploying models that handle terabytes of audio data daily.Your RoleAs an Applied ML Engineer at David AI, you will develop state-of-the-art speech and audio models, establish production inference systems, and create robust pipelines that demonstrate the true potential of high-quality data.Key ResponsibilitiesResearch and Design: Create solutions using advanced signal processing algorithms and cutting-edge ML models tailored for speech and audio applications.Development: Build production-grade inference algorithms, pipelines, and APIs in collaboration with cross-functional teams to extract valuable insights for our clients.Collaboration: Work alongside our Operations team to gather valuable training and evaluation datasets to enhance our model quality.Architecture: Design systems that ensure durable and resilient inference and evaluations.
Company OverviewEcho Neurotechnologies is a dynamic startup revolutionizing the Brain-Computer Interface (BCI) sector. We are committed to creating innovative hardware solutions powered by artificial intelligence, with the goal of enhancing the lives of individuals with disabilities and promoting independence through advanced technology.Team CultureBecome part of a close-knit group of passionate professionals in a fast-paced environment. As part of our early-stage team, you will have the chance to influence important decisions that yield substantial, lasting results. We prioritize continuous learning and collaboration, ensuring your contributions are integral to our collective success.Job SummaryWe are on the lookout for a Senior Machine Learning Infrastructure Engineer to join our talented team. In this pivotal role, you will be responsible for designing, constructing, and scaling infrastructure that supports large-scale data processing, modeling, and analysis. You will play an essential role in developing a high-performance, production-ready ML ecosystem that facilitates swift experimentation across diverse datasets, including neural signals and behavioral data. You'll have substantial ownership of our ML R&D platform, collaborating closely with domain experts to develop new cloud infrastructure, data pipelines, and modeling workflows, ultimately leading to the creation of state-of-the-art models for neuroscientific breakthroughs and neural decoding, thereby improving the lives of patients with severe neurological disorders.Key ResponsibilitiesCreate adaptable and efficient ML infrastructure:Design and implement ML cloud infrastructure for extensive modeling and analytics.Facilitate diverse model exploration, hyperparameter tuning, pretraining, fine-tuning, and evaluation.Develop and refine scalable distributed training pipelines, incorporating model sharding, cross-GPU communication, and real-time training monitoring.Manage and sustain robust ML platforms and services throughout the model lifecycle.Make strategic architecture decisions balancing performance, cost, reliability, and scalability.Build flexible and scalable data platforms:Design and optimize large-scale databases and data pipelines to ensure reliable data access.
At Physical Intelligence, we are pioneering general-purpose AI applications for the physical world. Our innovative approach involves orchestrating thousands of accelerators across a diverse ecosystem of GPU and TPU clusters, which encompass various hardware generations, cloud platforms, and cluster configurations.Researchers frequently encounter challenges in identifying the optimal cluster for their tasks, understanding resource availability, and configuring their workloads efficiently. This process is not scalable. To enhance productivity, we require an intelligent scheduling and compute system that can automatically determine the best job placements based on availability, hardware compatibility, cost considerations, and priority levels, allowing researchers to concentrate on their scientific endeavors.This position encompasses the complete ownership of this challenge: the development of scheduling systems, placement logic, cluster management frameworks, and operational tools essential for seamless operations.This role is distinct from traditional cloud DevOps; it focuses on resource allocation intelligence, utilization efficiency, fault tolerance, and ensuring a smooth experience for large-scale distributed training.About the TeamThe ML Infrastructure team is dedicated to bolstering and accelerating Physical Intelligence’s fundamental modeling initiatives by creating systems that ensure large-scale training is reliable, reproducible, and efficient. You will collaborate closely with the ML Infrastructure, data platform, and research teams to eliminate compute scheduling as a bottleneck.Key Responsibilities- Lead Intelligent Job Scheduling and Placement: Design and implement multi-tenant scheduling systems that automatically allocate training jobs to the most suitable cluster based on hardware specifications, topology, availability, cost, and priority. Facilitate equitable resource sharing across teams and projects through quota management, priority tiers, and preemption policies. Simplify cluster discrepancies so researchers can submit jobs without needing detailed knowledge of cluster specifics.- Enhance Multi-cluster Orchestration: Develop the control plane responsible for overseeing the job lifecycle across various clusters (including mixed GPU/TPU setups, multi-generational hardware, both on-premises and cloud-based) and enable effortless job migration, failover, and rescheduling.- Optimize Accelerator Utilization and Performance: Continuously monitor and enhance GPU/TPU usage across the entire fleet. Apply priority, preemption, queuing, and fairness strategies that balance research momentum with cost efficiency.- Guarantee Scalability and Stability: Implement fault detection, automatic recovery mechanisms, and resilience strategies for long-running multi-node training tasks. Oversee health checks, node management, and scaling strategies to ensure optimal performance.
Mach9’s Machine Learning Infrastructure Engineers create and maintain the backbone for production AI models used in civil engineering and surveying. The team manages a machine learning pipeline that processes over 10,000 miles of labeled survey data, supports image segmentation networks, and runs 3D prediction models. These systems deliver real-time inference capabilities directly to surveyors and engineers working in the field. Role overview This position is designed for mid-career engineers with a strong background in both training and inference aspects of machine learning infrastructure. The work involves handling large-scale data and ensuring reliable performance for demanding, real-world applications. What you will do Build and improve training pipelines for deep transformer models using hundreds of terabytes of 3D point cloud and image data. Design and implement inference infrastructure to support both offline detection algorithms and responsive, real-time inference integrated with CAD software. Location Based in San Francisco.
About Nooks.ai:Nooks is a cutting-edge AI Sales Assistant Platform (ASAP) designed to streamline sales processes, allowing representatives to concentrate on building relationships and closing deals. Our innovative platform has empowered thousands of sales professionals to achieve their targets, saving clients countless hours and generating substantial revenue. Trusted by sales teams at industry leaders like Hubspot, Rippling, and Toast, Nooks is transforming the sales landscape.Backed by over $70M in investments from top-tier venture capital firms, including Kleiner Perkins, Nooks has experienced remarkable growth, achieving a 4x and 3x increase in ARR over the past two years. We are on an ambitious trajectory to triple our growth once again this year.For more information, visit Nooks.ai.The RoleNote: Job title will be aligned with candidate experience.We are seeking a passionate Applied Machine Learning Engineer to join our dynamic team, tackling exciting technical challenges in the emerging field of AI-powered real-time collaboration. This role is pivotal in integrating machine learning features into the Nooks platform. The ideal candidate will have hands-on experience in a business where machine learning plays a central role.Key responsibilities will involve training production models to enhance their accuracy for specific sales applications, while aligning our technical strategy with performance, cost, and feasibility factors.Examples of Engineering Challenges You Might EncounterThese examples are illustrative; prior experience in all areas is not required. We hope you find some of these challenges intriguing!Real-time Audio AI & Precision/Recall/Latency Trade-offs (Algorithms & Models)Utilizing audio data, transcription, silence detection, and multiple signals to discern if a live call is a voicemail, a human, or a dial tree. Managing latency alongside precision/recall trade-offs is crucial for prompt human detection, involving advanced techniques like LLM embeddings, few-shot learning, data labeling, and continuous performance monitoring.Intelligent Call Funnels & Playbooks (Data Wrangling, Backend Engineering, GPT-3, UX)Analyzing the conversational flow to optimize call funnels and playbook strategies, focusing on data visibility and user experience.
About UsAt Applied Compute, we are pioneering the development of Specific Intelligence for enterprises, creating agents that continuously learn from a company’s processes, data, expertise, and objectives. Our mission is to bridge the gap between isolated AI capabilities and their effective application within real business environments. Traditional AI systems often fall short as they lack the ability to adapt based on feedback. Our innovative continual learning layer captures context, memory, and decision-making processes across the enterprise, enabling specialized agents to engage in meaningful work.What Excites Us: We operate at the exciting intersection of product development and cutting-edge research. Our product team designs the platform that empowers a new generation of digital coworkers, while our research team drives advancements in post-training and reinforcement learning to enhance user experiences. As an applied research engineer, you will work directly with clients to implement models in production, combining robust product development with deep research insights to facilitate AI integration in enterprises.Meet Our Team: Our diverse team consists of engineers, researchers, and operators, many of whom are former founders. We have previously built reinforcement learning infrastructure at OpenAI, established data foundations at Scale AI, and contributed to significant systems at companies like Together, Two Sigma, and Watershed. We collaborate with Fortune 50 clients, including DoorDash, Mercor, and Cognition, and are proud to be backed by reputable investors such as Benchmark, Sequoia, and Lux.Who Thrives Here: We seek individuals who are passionate about applying innovative research and complex systems to solve real-world challenges. You should feel comfortable navigating new environments rapidly—be it a fresh codebase, a client’s data architecture, or an unfamiliar problem domain. A genuine enjoyment for customer interaction, empathy, and a deep understanding of their operational workflows are essential. Candidates with entrepreneurial backgrounds, extensive side projects, or a proven track record of end-to-end ownership typically excel in our environment.
Full-time|$176K/yr - $220K/yr|On-site|San Francisco, CA; New York, NY
About This Role Join Scale AI's Applied ML team as a Machine Learning Research Engineer, focusing on the development of advanced data infrastructure for leading agentic large language models (LLMs) such as ChatGPT, Gemini, and Llama. You will be responsible for architecting scalable multi-agent systems aimed at validating agentic reasoning and behaviors, enhancing human expertise, and conducting research to address real-world agent reliability failures, even in the face of strong benchmarks. Your contributions will directly impact the deployment of production fixes. This role is ideal for exceptional engineers who possess a deep research rigor and a strong commitment to creating practical, high-impact systems. You will iterate rapidly using data, leverage AI tools for accelerated development, and collaborate closely with engineering, product, and research teams. If you have a knack for transforming cutting-edge agent research into dependable deployed systems, we would love to hear from you.
Full-time|On-site|Denver, CO; New York, NY; San Francisco, CA; Seattle, WA
Quizlet Inc. is looking for an Applied AI Engineer to create AI-driven features that support student learning. This position centers on developing and deploying machine learning solutions aimed at making study experiences more effective and engaging for a global user base. What you will do Design and implement machine learning models to enhance Quizlet’s educational tools Work on features that help students study more efficiently and enjoyably Locations Denver, CO New York, NY San Francisco, CA Seattle, WA
About UsAt Speak, our mission is to revolutionize language learning.Learning a new language can transform lives by unlocking opportunities in diverse cultures, careers, and communities. With over two billion individuals around the globe striving to learn a language, we recognize that traditional one-on-one tutoring remains difficult to access at scale and has seen little innovation over recent decades. Speak is pioneering an AI-driven, human-level tutor accessible right from your pocket, providing a conversation-first experience where learners can practice speaking, receive immediate feedback, and progress through meticulously crafted lessons. Our goal is to facilitate a comprehensive journey from beginner to proficient speaker across various languages.Launched in South Korea in 2019, Speak has quickly become the leading language learning app in the region, now reaching learners across numerous markets and offering instruction in 15+ languages. Supported by over $150 million in venture capital from prestigious investors such as OpenAI, Accel, Founders Fund, and Khosla Ventures, our team is distributed across San Francisco, Seoul, Tokyo, Taipei, and Ljubljana.Role OverviewWe are seeking a skilled Machine Learning Engineer specializing in speech to join our innovative team. In this role, you will take charge of the entire modeling pipeline for speech recognition, encompassing training, experimentation, deployment, and ongoing monitoring. Collaborating closely with Product teams, you will design cutting-edge learning experiences and assess the effectiveness of production models on our users. As part of a nimble and dynamic team, you'll contribute as both a developer and a thought partner on projects related to ASR, assessments, pronunciation improvements, content personalization, and more. This is an exhilarating opportunity to be part of an ML team focused on crafting personalized learning experiences that will transform language education for millions worldwide.
At Causal Labs, we are on a groundbreaking mission to develop general causal intelligence, harnessing AI to (1) forecast future events and (2) pinpoint optimal actions to influence that future.To realize this vision, we are constructing a Large Physics foundation Model (LPM), as the domains governed by physics inherently feature cause-and-effect relationships, which is distinct from visual or textual data.Weather serves as the perfect training environment for our LPM, being the most extensively observed physical system and providing rapid, objective ground truth feedback from sensory data at an unprecedented scale, far exceeding what is utilized for current large language models (LLMs).Our team comprises elite researchers and engineers with backgrounds in self-driving technology, drug discovery, and robotics, including talents from Google DeepMind, Cruise, Waymo, Meta, Nabla Bio, and Apple. We believe that achieving general causal intelligence will be a pivotal technological advancement for humanity.We are searching for infrastructure engineers who are eager to tackle formidable challenges and contribute to our mission.Your expertise in distributed training clusters and performance optimization for large models will be crucial as we address our training and inference challenges. If you possess experience in developing large-scale ML infrastructure within fields like language models, vision systems, robotics, or biology, we invite you to join us.
Full-time|$170K/yr - $250K/yr|On-site|San Francisco
We are seeking a talented Applied Machine Learning Engineer with a comprehensive understanding of the generative media landscape. You should possess an up-to-date knowledge of emerging methodologies and be capable of identifying gaps in the current market. Your role will involve innovating and developing machine learning models that enhance user experiences, requiring both novel training techniques and the fine-tuning of existing models with fresh datasets.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We're dedicated to crafting a future where everyone can harness the power of AI to meet their unique needs and aspirations.Our team comprises scientists, engineers, and innovators who have developed some of the most widely utilized AI products, including ChatGPT and Character.ai, as well as open-weight models like Mistral, in addition to renowned open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleWe are seeking a talented Infrastructure Research Engineer to architect and develop the foundational systems that facilitate the scalable and efficient training of large models using reinforcement learning.This position exists at the crossroads of research and large-scale systems engineering, requiring a professional who not only comprehends the algorithms behind reinforcement learning but also appreciates the practicalities of distributed training and inference at scale. You will have a diverse set of responsibilities, from optimizing rollout and reward pipelines to enhancing the reliability, observability, and orchestration of systems. Collaboration with researchers and infrastructure teams will be essential to ensure reinforcement learning is stable, rapid, and production-ready.Note: This is an evergreen role that we maintain on an ongoing basis to express interest. Due to the high volume of applications we receive, there may not always be an immediate position that aligns perfectly with your skills and experience. We encourage you to apply, as we continuously review applications and reach out to candidates when new opportunities arise. You may reapply after gaining more experience, but please refrain from applying more than once every six months. Additionally, you may notice postings for specific roles that cater to unique project or team needs; in those circumstances, you are welcome to apply directly alongside this evergreen role.What You’ll DoDesign, implement, and optimize the infrastructure that supports large-scale reinforcement learning and post-training workloads.Enhance the reliability and scalability of the RL training pipeline, including distributed RL workloads and training throughput.Create shared monitoring and observability tools to ensure high uptime, debuggability, and reproducibility of RL systems.Work closely with researchers to translate algorithmic concepts into production-quality training pipelines.Develop evaluation and benchmarking infrastructure to assess model performance based on helpfulness, safety, and factual accuracy.Publish and disseminate insights through internal documentation, open-source libraries, or technical reports that contribute to the advancement of scalable AI infrastructure.
Nov 27, 2025
Sign in to browse more jobs
Create account — see all 8,813 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.