Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Experience
Qualifications
Responsibilities:
Architect multi-cloud systems and abstractions to facilitate the SGP platform's seamless operation across existing cloud providers.
Develop custom integrations between Scale AI's platform and client data environments (cloud platforms, data warehouses, internal APIs).
Collaborate directly with platform and product teams, as well as clients, to create and implement scalable infrastructure that meets evolving demands.
Deliver high-quality experiments rapidly to engage and satisfy our customers.
Manage the complete product lifecycle from initial concept through to production.
Possess the ability and willingness to multitask and rapidly acquire new technologies.
Qualifications:
A minimum of 4 years of full-time engineering experience post-graduation.
Proven experience scaling products in hyper-growth startup environments.
Hands-on experience with LLMs, vector databases, and other cutting-edge AI technologies.
Proficiency in Python or JavaScript/TypeScript, and SQL.
Familiarity with Kubernetes.
Experience with major cloud platforms (AWS, Azure, GCP).
Strong communication skills, capable of conveying technical concepts to both technical and non-technical stakeholders.
About the job
Join Scale AI's innovative team as an Infrastructure Software Engineer for our Enterprise Generative AI Platform (SGP). In this dynamic role, you will help design and enhance our enterprise-grade AI platform, which offers robust APIs for knowledge retrieval, inference, evaluation, and more. We're seeking an exceptional engineer who thrives in fast-paced environments and is eager to contribute to the scaling of our core infrastructure.
The ideal candidate will possess a solid foundation in software engineering principles and extensive experience with large-scale distributed systems. Your role will involve implementing solutions across various cloud providers (GCP, Azure, AWS) for clients in highly regulated sectors, including healthcare, telecommunications, finance, and retail.
About Scale AI
Scale AI is at the forefront of artificial intelligence, providing an enterprise-grade generative AI platform that empowers businesses with advanced APIs and tools for effective knowledge management and data analytics. Our mission is to drive innovation across various industries, enabling clients to harness the power of AI effectively.
Similar jobs
1 - 20 of 7,715 Jobs
Search for Ai Software Engineer Cybersecurity Systems
Bishop Fox is a premier authority in offensive security, delivering a wide array of solutions including continuous penetration testing, red teaming, and comprehensive assessments of product, cloud, and application security. Our esteemed clientele includes over a quarter of the Fortune 100, half of the Fortune 10, eight of the top 10 global technology firms, and all leading global media companies. Our managed service platform, innovative service offerings, and commitment to excellence have earned us numerous accolades from industry award programs such as Fast Company, Inc., and SC Media. For more than 16 years, we have actively contributed to the security community, publishing over 16 open-source tools and 50 security advisories in the past five years alone. Discover more at bishopfox.com or connect with us on social media.Who You AreThe Agentic AI Software Engineer in Cybersecurity Systems is tasked with designing, developing, and deploying cutting-edge AI-driven software solutions aimed at enhancing cybersecurity detection, response, analysis, and automation capabilities. In this role, you will focus on constructing and maintaining agent-based AI systems proficient in autonomously generating code, performing security analyses, triaging alerts, identifying vulnerabilities, and proposing remediation strategies, all under structured human oversight and validation.This position demands advanced expertise in computer science, artificial intelligence, machine learning, secure software engineering, distributed systems, and cybersecurity principles. You will be engaged in complex system design, secure AI integration, and the creation of enterprise-grade security automation platforms.
On-site|On-site|San Francisco, CA | New York City, NY | Seattle, WA; Washington, DC
Join Anthropic as a Senior Software Engineer focused on developing cutting-edge AI-driven cybersecurity products. In this dynamic role, you will collaborate with a passionate team to prototype and build innovative solutions that enhance security applications. Your work will bridge research, product development, and customer engagement, allowing you to contribute significantly to the future of cybersecurity.
On-site|On-site|San Francisco, CA | New York City, NY
Join Anthropic as a Research Engineer on our Cybersecurity Reinforcement Learning team, where you'll contribute to the development of AI systems designed for secure coding, vulnerability remediation, and other defensive cybersecurity initiatives. This role blends research with engineering, allowing you to innovate new methodologies while implementing them in code. You will design RL environments, conduct experiments, and collaborate with a diverse team of experts to enhance our AI capabilities while ensuring safety and reliability.
About GranicaGranica is an innovative AI research and infrastructure firm dedicated to creating reliable and steerable representations of enterprise data.We build trust through our product Crunch, a policy-driven health layer that ensures large tabular datasets remain efficient, reliable, and reversible. On this solid foundation, we are developing Large Tabular Models—systems designed to learn cross-column and relational structures in order to provide trustworthy answers and automation with inherent provenance and governance.Our MissionAI is currently hampered not only by the design of models but also by the inefficiencies of the data that supports them. Every redundant byte, poorly organized dataset, and inefficient data pathway contributes to significant costs, latency, and energy waste as we scale.Granica aims to eliminate these inefficiencies. We merge cutting-edge research in information theory, probabilistic modeling, and distributed systems to craft self-optimizing data infrastructures: systems that consistently enhance the representation and utilization of information by AI.Our engineering team collaborates closely with the Granica Research group led by Prof. Andrea Montanari of Stanford University, bridging advancements in information theory and learning efficiency with large-scale distributed systems. Together, we firmly believe that the next major advancement in AI will stem from breakthroughs in efficient systems rather than merely larger models.Your ContributionsGlobal Metadata Substrate: Design a transactional and metadata substrate that facilitates time-travel, schema evolution, and atomic consistency across massive petabyte-scale tabular datasets.Adaptive Engines: Develop systems that autonomously reorganize data, learning from access patterns and workloads to maintain peak efficiency without the need for manual tuning.Intelligent Data Layouts: Optimize bit-level organization (including encoding, compression, and layout) to maximize signal extraction per byte read.Autonomous Compute Pipelines: Create distributed compute systems that scale predictably, adapt to dynamic loads, and ensure reliability under failure conditions.Research to Production: Apply new algorithms in compression, representation, and optimization that emerge from ongoing research. We encourage opportunities to publish and open-source your work.Latency as Intelligence: Design systems that inherently minimize latency as a measure of intelligence.
About GranicaGranica is a pioneering AI research and infrastructure company dedicated to creating reliable and steerable representations of enterprise data.We build trust through Crunch, a policy-driven health layer designed to keep extensive tabular datasets efficient, reliable, and reversible. From this foundation, we are developing Large Tabular Models—systems that learn cross-column and relational structures to provide trustworthy answers and automation, complete with built-in provenance and governance.Our MissionThe current limitations of AI are not solely due to model design but also to the inefficiencies of the data that supports it. At scale, every redundant byte, poorly organized dataset, and inefficient data path contributes to significant costs, latency, and energy waste.Granica’s mission is to eliminate these inefficiencies. We leverage cutting-edge research in information theory, probabilistic modeling, and distributed systems to create self-optimizing data infrastructures that continuously enhance how information is represented and utilized by AI.Our engineering team collaborates closely with the Granica Research group led by Prof. Andrea Montanari from Stanford University, merging advancements in information theory and learning efficiency with large-scale distributed systems. We believe that the next major breakthrough in AI will stem from innovations in efficient systems, rather than simply larger models.What You Will CreateGlobal Metadata Substrate. Design and refine the global metadata and transactional substrate that enables atomic consistency and schema evolution across exabyte-scale data systems.Adaptive Engines. Architect systems that self-optimize, reorganizing and compressing data according to access patterns, achieving unprecedented efficiency improvements.Intelligent Data Layouts. Innovate new encoding and layout strategies that challenge the theoretical limits of signal per byte read.Autonomous Compute Pipelines. Spearhead the development of distributed compute platforms that scale predictively and maintain reliability even under extreme load and failure conditions.Research to Production. Partner with Granica Research to transform advances in compression and probabilistic modeling into production-ready, industry-leading systems.Latency as Intelligence. Propel systems forward by optimizing for latency as a key aspect of intelligence.
The Role As a new graduate Software Engineer focusing on Agentic AI Systems, are you ready to apply your skills at the forefront of AI agents? This position could be the perfect fit for you. Join our dynamic team within the "agent lab" for Natural Language Understanding, where you will significantly contribute to our mission: enhancing the scope of work that can be reliably executed by AI agents at scale. You'll have the chance to shape the Moveworks AI Assistant platform across various domains, including agent orchestration, sandboxed file systems, code execution, latency optimization, agent memory management, LLM self-reflection and improvement, execution environment simulation, enterprise knowledge graphs, and multimodal I/O. Equipped with cutting-edge tools including state-of-the-art LLMs from providers like OpenAI, our team prioritizes rapid development on scalable infrastructure, tackling challenging product and engineering issues, and maximizing value for our clients. If you're eager to make a meaningful impact alongside a talented, dedicated, and mission-driven team, we invite you to connect with us.
Join Saris AI as an AI Systems Engineer, where you will play a pivotal role in designing and implementing innovative solutions that leverage artificial intelligence technologies. You will collaborate closely with cross-functional teams to develop AI systems that enhance our products and services, driving impactful results for our clients.
Join Our Innovative Team at SazabiAs we approach the year 2026, the tech world faces a looming "infinite software crisis." How do we effectively support, maintain, and manage the vast surge in application development?Our solution is Sazabi: the AI-native observability platform designed specifically for dynamic engineering teams.Sazabi empowers teams to inquire about their production systems in straightforward language, visualize operations automatically, and identify root causes up to 10 times faster. Forget about tedious instrumentation, complex dashboard setups, and alert configurations—just get the answers you need.We are proud to be supported by innovators from industry-leading AI companies, including Vercel, Graphite, Daytona, Browserbase, LangChain, and Replit.
Full-time|$144.5K/yr - $180.6K/yr|Hybrid|San Francisco, CA
Welcome to Planet Labs. We are dedicated to leveraging space technology to enhance life on Earth.As the architects of the largest constellation of imaging satellites ever created, Planet Labs provides an unparalleled dataset of empirical information through our innovative cloud-based platform. Our impact spans commercial, environmental, and humanitarian sectors, merging our identity as both a space and data company.Our data empowers customers globally to innovate, drive revenue, conduct research, and tackle the most pressing challenges our world faces.At Planet, we manage every aspect of hardware design, manufacturing, data processing, and software engineering, fostering a dynamic environment filled with experts from diverse fields.We prioritize a people-first approach in our corporate culture and community, striving for continuous improvement to support our team members and prepare for future growth. Join us at Planet Labs and contribute to our mission of transforming global perspectives.With a global workforce, our employees work remotely from locations including San Francisco, Washington DC, Germany, Austria, Slovenia, and The Netherlands.About the Role:Planet Labs aims to capture daily images of the entire world, making significant global changes visible, accessible, and actionable. We are at a pivotal moment, transitioning from extensive AI research to a focused delivery model. To facilitate this, we are establishing a new product group dedicated to launching an AI Geospatial Assistant that will revolutionize how our customers utilize global imagery to address critical challenges in forensics and daily change detection.Our objective is to simplify complex insights through an intuitive interface, requiring no user training. Operating with a startup mentality, our team emphasizes rapid learning and customer-driven milestones to efficiently progress from private alpha to general availability.As a Software Engineer, you will play a key role in developing the backend systems that will bring our AI Geospatial Assistant to fruition. While our research teams focus on creating core models, you will be tasked with the 'last mile' of delivery, designing high-throughput backend services, scaling our systems, and ensuring that our workflows operate swiftly, reliably, and cost-effectively on a global scale.This position is full-time and hybrid, requiring you to work from our San Francisco office three days a week.
The RoleAre you a seasoned software engineer with a passion for innovation in production systems? If you're eager to leverage your expertise in the dynamic field of AI agents, this position is tailored for you.As a key member of our Natural Language Understanding team within the 'agent lab', you will contribute significantly to our mission: pushing the boundaries of what agents can achieve reliably at scale. Your work will enhance the Moveworks AI Assistant platform across various domains, including agent orchestration, sandboxed execution environments, latency reduction, agent memory management, LLM self-reflection, execution environment simulation, enterprise knowledge integration, and multimodal input/output.Equipped with cutting-edge enterprise AI tools, including top-tier LLMs from providers like OpenAI, you'll be part of a fast-paced team focused on scalable infrastructure and solving complex engineering challenges to deliver exceptional value to our customers. If you are ready to make a significant impact alongside a dedicated and driven team, we invite you to explore this opportunity with us.
The Role Are you a skilled software engineer with a proven track record in building and refining production systems? Are you eager to apply your expertise at the forefront of AI technology? If so, this opportunity may be perfect for you. As a Senior Software Engineer on our Natural Language Understanding team within the “agent lab,” you will be pivotal in our mission to enhance the capabilities of AI agents for reliable, scalable performance. You will have the chance to influence the evolution of the Moveworks AI Assistant platform in several key areas: agent orchestration, sandboxed file systems, latency optimization, and multimodal I/O, among others. You will leverage the best tools in enterprise AI, including cutting-edge LLMs from top providers like OpenAI. Our team prioritizes rapid innovation on scalable infrastructure while tackling challenging product and engineering obstacles to deliver exceptional value to our clients. If you are looking to achieve the pinnacle of your career alongside a passionate, dedicated team focused on making an impact, we invite you to connect with us.
About Our TeamThe Platform Systems team at OpenAI is at the forefront of innovation, merging advanced AI technologies with large-scale distributed systems. We are tasked with creating the engineering and research infrastructure essential for training OpenAI's premier models on some of the most powerful, custom-built supercomputers globally.Our team is dedicated to developing the core software for model training, delving deep into the technological stack. This encompasses collective communication, compute efficiency, parallelism strategies, fault tolerance, failure detection, and observability. The systems we design are pivotal to enhancing OpenAI's research capabilities, facilitating reliable and efficient training at the leading edge of technology.We work in close partnership with researchers across the organization, continuously integrating insights from various OpenAI projects to advance our training platform.About the RoleAs a Software Engineer specializing in Platform Systems, you will architect and develop distributed systems that enhance visibility into large-scale training operations, ensuring their dependable operation at scale.Your responsibilities will include designing systems for failure detection, tracing, and observability that pinpoint slow or malfunctioning nodes, identify performance bottlenecks, and assist engineers in optimizing extensive distributed training tasks. This infrastructure is integral to the functionality of OpenAI's training stack and is continuously evolving to accommodate new use cases and increasingly intricate workloads.This position is central to our training infrastructure, merging systems engineering, performance analysis, and large-scale debugging.Key ResponsibilitiesDesign and develop distributed failure detection, tracing, and profiling systems tailored for large-scale AI training jobs.Create tools to identify slow, faulty, or errant nodes and deliver actionable insights into system behavior.Enhance observability, reliability, and performance across OpenAI's training platform.Troubleshoot and resolve issues within complex, high-throughput distributed systems.Collaborate effectively with systems, infrastructure, and research teams to advance platform capabilities.Adapt and expand failure detection and tracing systems to support new training paradigms and workloads.Ideal Candidate ProfilePossesses a deep passion for performance, stability, and observability in distributed systems.Demonstrates proficiency in systems engineering and performance analysis.Has experience in debugging high-throughput distributed systems.Exhibits strong collaboration skills with a track record of working with cross-functional teams.Shows adaptability and eagerness to embrace new technologies and methodologies.
About Our TeamThe Frontier Systems team at OpenAI is at the forefront of technology, responsible for creating, deploying, and maintaining some of the world's largest supercomputers. These supercomputers are pivotal for training our most advanced AI models, pushing the boundaries of innovation.We transform sophisticated data center designs into operational systems and develop the software infrastructure necessary for extensive frontier model training. Our goal is to ensure these hyperscale supercomputers operate reliably and efficiently, supporting groundbreaking AI research.About the RoleAs a key member of the Frontier Systems team, you will be instrumental in designing the critical infrastructure that ensures our supercomputers function seamlessly for pioneering AI research. In this role, you'll address system-level challenges and implement automation solutions that minimize disruptions during large-scale training processes.Your responsibilities will encompass end-to-end ownership of your projects, allowing you to make significant contributions to our mission. This position is ideal for individuals who excel in diagnosing complex system issues and crafting automation strategies to proactively resolve problems across a vast network of machines.Your Responsibilities Include:Enhancing system health checks to maintain the stability of our hyperscale supercomputers during model training.Conducting in-depth investigations into hardware failures and system-level bugs to uncover root causes.Developing automation tools that monitor and resolve issues across thousands of systems, enabling uninterrupted research progress.You May Be a Great Fit If You Possess:7+ years of hands-on experience in software engineering.Strong proficiency in Python and shell scripting.Expertise in analyzing complex data sets using SQL, PromQL, Pandas, or other relevant tools.Experience in creating reproducible analyses.A solid balance of skills in both building and operationalizing systems.Prior experience with hardware is not a prerequisite for this role.Preferred Qualifications:Familiarity with the intricacies of hardware components, protocols, and Linux tools (e.g., PCIe, Infiniband, networking, power management, kernel performance tuning).Experience with system optimization and performance tuning.
Join the Leading Force in Offensive SecurityBishop Fox is at the forefront of offensive security, delivering top-notch solutions encompassing continuous penetration testing, red teaming, attack surface management, and comprehensive assessments for products, cloud, and applications. With a client base that includes over a quarter of the Fortune 100, half of the Fortune 10, and all the leading global media companies, our reputation speaks for itself. Our continuous innovation, managed service platform, and commitment to excellence have earned us accolades from esteemed industry award programs such as Fast Company, Inc., and SC Media. For more than 16 years, we have been dedicated to giving back to the security community, publishing over 16 open source tools and 50 security advisories in just the past five years. Discover more about our mission at bishopfox.com or connect with us on social media.Your Role as a Staff AI Software EngineerThis is not merely another engineering position; you will become part of a dynamic startup-like team within Bishop Fox, combining the agility and innovative spirit of an early-stage venture with the backing of a renowned security leader. As an experienced AI engineer, you will excel in developing robust systems capable of functioning in complex and unpredictable environments. Your focus is on reliability, evaluation, and scalability—ensuring that solutions perform consistently in production environments.Your MissionJoin us in building autonomous AI agents that can identify real vulnerabilities in production applications, equipped to think, adapt, and hack like the world's foremost penetration testers.
The Role Are you a seasoned software engineer with a passion for technical leadership and a track record of enhancing production systems? If you're eager to leverage your expertise at the forefront of AI technology, this opportunity might be perfect for you. As a Staff Software Engineer on our pioneering Natural Language Understanding team within the “agent lab,” you will be instrumental in our mission to expand the boundaries of what AI agents can accomplish reliably and at scale. You'll lead the transformation of the Moveworks AI Assistant platform in key areas such as agent orchestration, sandboxed file systems, code execution, latency optimization, agent memory management, LLM self-reflection and improvement, execution environment simulation, enterprise knowledge graphs, and multimodal I/O. Equipped with cutting-edge enterprise AI tools, including top-tier LLMs from providers like OpenAI, our team focuses on rapid development of scalable infrastructure, tackling complex engineering challenges, and maximizing value for our customers. If you're ready to elevate your career alongside a passionate and impact-driven team, we would be thrilled to engage with you.
About Retell AIAt Retell AI, we are pioneering the future of call centers through innovative voice AI technology. Our cutting-edge solutions are transforming how companies engage with customers.In just 18 months since our inception, we've empowered thousands of businesses with our AI voice agents that efficiently manage sales, support, and logistics calls, significantly reducing the need for large teams of human agents. Supported by industry-leading investors including Y Combinator and Alt Capital, we've grown our annual recurring revenue from $5M to an impressive $36M while expanding our team from 5 to 20 talented individuals since 2025.Our ambitious vision for 2026 is to develop a state-of-the-art customer experience platform where entire contact centers are driven by AI. Unlike basic automation requiring constant human oversight, we’re engineering intelligent AI “workers” capable of serving as frontline agents, quality assurance analysts, and managerial roles, all while optimizing customer interactions continuously.We are rapidly expanding and seeking driven builders who thrive on solving complex technical challenges, act decisively, and wish to make a tangible impact in one of the fastest-growing voice AI startups.Join us in shaping the future!Recognized as a top 50 AI application in the a16z list: https://tinyurl.com/5853dt2xRanked #4 in Brex's Fast-Growing Software Vendors of 2025: https://www.brex.com/journal/brex-benchmark-december-2025Featured among the top startups on: https://leanaileaderboard.com/
Full-time|$240K/yr - $280K/yr|Hybrid|San Francisco, California
About SentryAt Sentry, we are committed to transforming the way developers build software. With a mission to eradicate poor software experiences, we empower developers to create better applications more efficiently, ensuring a seamless encounter with technology.Backed by over $217 million in funding and trusted by more than 100,000 organizations, including industry giants like Disney, Microsoft, and Atlassian, we are at the forefront of performance monitoring and error tracking solutions. Our innovative tools enable companies to focus on product development rather than bug fixes.We embrace a hybrid work environment across our global offices, designating Mondays, Tuesdays, and Thursdays as in-office collaboration days to foster meaningful team interactions. If you are passionate about creating solutions that enhance the digital experience, join us in developing the next wave of software monitoring tools.About the RoleAs a Senior Software Engineer on Sentry’s AI/ML team, you will play a pivotal role in constructing the evaluation infrastructure that assesses the accuracy, reliability, and performance of our AI systems in real-world scenarios. This position is essential for ensuring that our debugging agents and AI-driven features operate correctly, safely, and predictably as they scale. You will design datasets, benchmarks, and test harnesses that convert vague AI behavior into quantifiable metrics, enabling the team to deploy AI solutions with confidence.In This Role You WillDevelop and implement robust evaluation frameworks to assess accuracy, reliability, regressions, and edge cases within AI systems.Generate and manage high-quality datasets, golden test cases, and benchmarks based on real production data.Create automated test harnesses and metrics pipelines to continuously evaluate models, prompts, and workflows.Collaborate closely with applied AI engineers and product leaders to establish clear definitions of success and translate them into measurable criteria.Oversee the evaluation lifecycle for significant AI projects, from initial experimentation to ongoing production monitoring.You'll Love This Job If YouHave a strong commitment to accuracy, rigor, and measurement in AI systems.Enjoy transforming ambiguous product objectives and model behaviors into precise tests and metrics.Take pleasure in building foundational infrastructure that facilitates rapid iteration and boosts team confidence.Thrive in collaborative environments and relish the opportunity to influence model design through effective evaluation.
Full-time|$200K/yr - $280K/yr|On-site|San Francisco Office
About UnifyUnify is revolutionizing the way revenue teams operate by developing the first AI-powered system of action. Our mission is to transform outbound strategies into a robust growth engine, ensuring that go-to-market execution is observable, repeatable, and scalable. Established in 2023 by industry leaders from Ramp and Scale AI, our talented team boasts backgrounds from renowned companies like Airbnb, Meta, Waymo, and Perplexity.In 2024, Unify achieved an impressive 8x revenue growth and serves a diverse clientele including Perplexity, Cursor, SoFi, and Justworks. We are a dynamic team fueled by high energy and intensity, having secured $58M in funding from leading investors such as Thrive, Emergence, and OpenAI. Join us in shaping the future of GTM!About the Role:As a Senior Software Engineer specializing in AI at Unify, you will be at the forefront of innovation, developing new AI products and enhancing our AI platform. Your work will encompass agents, retrieval systems, classification, fine-tuning, reinforcement learning, and LLM inference infrastructure. If you are passionate about creating production-ready AI systems and wish to contribute to cutting-edge applications used by premier GTM teams, this role is for you.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, we are dedicated to empowering humanity by advancing collaborative general intelligence. Our vision is to create a future where everyone can access the knowledge and tools necessary to harness AI for their unique needs.Our diverse team of scientists, engineers, and builders has developed some of the most recognized AI products, including ChatGPT and Character.ai, as well as notable open-weight models like Mistral, and popular open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleWe are currently seeking versatile infrastructure and systems engineers to help construct the foundational systems that support our models and facilitate research and product development. Your contributions will enable teams to create and deliver groundbreaking AI products.As a member of a small, high-impact team, you will be responsible for architecting and scaling the core infrastructure that underpins our operations. This role involves working across the entire technical stack, addressing complex distributed systems challenges, and developing robust, scalable platforms.Infrastructure is vital to our success; it serves as the foundation for every innovation. You will collaborate directly with researchers to expedite experiments, enhance infrastructure efficiency, and derive critical insights from our models, products, and data assets.Note: This is an evergreen role, meaning we are continuously accepting expressions of interest. Due to the volume of applications, there may not always be an immediate match for your skills and experience. However, we encourage you to apply. Applications are reviewed regularly, and we reach out to candidates as new opportunities arise. You may reapply if you gain additional experience, but please wait at least six months between applications. Additionally, we occasionally post specific roles for particular projects or teams, and you are welcome to apply for those as well.What You’ll DoInterviews will be conducted in a general manner, but project selection will consider your interests and experience alongside the needs of the organization. This flexible approach allows us to align talented engineers with the infrastructure teams where they will have the greatest influence and opportunities for growth.Depending on your expertise and interests, you may contribute to various areas such as:Core Infrastructure: Supporting teams that train, research, and ultimately serve AI models by building the infrastructure required for reliable and secure training of frontier models. This may include developing systems and managing large Kubernetes clusters with GPU workloads.
Full-time|$216.2K/yr - $270.3K/yr|On-site|San Francisco, CA; New York, NY
Join Scale AI's innovative team as an Infrastructure Software Engineer for our Enterprise Generative AI Platform (SGP). In this dynamic role, you will help design and enhance our enterprise-grade AI platform, which offers robust APIs for knowledge retrieval, inference, evaluation, and more. We're seeking an exceptional engineer who thrives in fast-paced environments and is eager to contribute to the scaling of our core infrastructure. The ideal candidate will possess a solid foundation in software engineering principles and extensive experience with large-scale distributed systems. Your role will involve implementing solutions across various cloud providers (GCP, Azure, AWS) for clients in highly regulated sectors, including healthcare, telecommunications, finance, and retail.
Mar 26, 2026
Sign in to browse more jobs
Create account — see all 7,715 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.