Senior/Staff Software Engineer, Simulation Platform
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Senior
Qualifications
About Waabi
Waabi is at the forefront of Physical AI, spearheading advancements in autonomous transportation through innovative technology. Our dynamic team is dedicated to transforming the future of mobility with groundbreaking solutions for commercial autonomous vehicles.
Similar jobs
Search for Staff Ai Platform Engineer Inference Agentic Systems
1,200 results
Role Overview Paytm is hiring a Staff AI Platform Engineer in Toronto, Canada, with a focus on Inference and Agentic Systems. This role centers on designing and improving AI-powered platforms that support intelligent, agent-like features for users and business applications. What You Will Do Work with a skilled team to build and refine AI solutions that support agentic behaviors and inference-driven capabilities. Apply experience in machine learning, software engineering, and system architecture to develop and scale reliable AI platforms. Contribute technical leadership and hands-on expertise to projects that advance Paytm’s AI offerings. Key Skills Deep knowledge of machine learning and inference methods Strong background in software engineering Experience designing and maintaining complex system architectures
Cerebras Systems
Cerebras Systems is revolutionizing AI technology with the world's largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture combines the immense computational power of multiple GPUs into a single chip while maintaining unparalleled programming simplicity. This allows us to provide extraordinary training and inference speeds, empowering machine learning users to seamlessly execute large-scale ML applications without the complexities of managing numerous GPUs or TPUs. We proudly serve a diverse clientele, including leading model laboratories, global corporations, and innovative AI-centric startups. Notably, OpenAI recently formed a multi-year partnership with Cerebras, committing to deploy 750 megawatts of scale to enhance critical workloads with ultra-high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, achieving speeds over 10 times faster than GPU-based hyperscale cloud inference services. This remarkable speed transformation enhances user experiences and facilitates real-time iterations while augmenting intelligence through advanced agentic computation. About The RoleThe Inference Core Platform team is integral to Cerebras’ mission of delivering the world’s fastest AI inference. Our engineers develop the core software and hardware infrastructure that enables low-latency, high-speed, and high-throughput deployment on the Cerebras Wafer-Scale Engine (WSE). We oversee the entire stack—from model compilation and scheduling to custom hardware kernels and driver development.The Platform Benchmarking team is crucial in enhancing the performance and scalability of AI inference on one of the most advanced computing systems ever developed. We spearhead the establishment of core inference capabilities and implement performance improvements at every development phase, from initial prototyping to full production deployment.We seek enthusiastic engineers eager to redefine the boundaries of AI inference. If you're passionate about developing systems that measure, analyze, and optimize performance on a large scale, this is your chance to make a transformative impact on the future of AI.
Cerebras Systems
Cerebras Systems is at the forefront of AI technology, having developed the world's largest AI chip, which is 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers unparalleled AI compute power equivalent to dozens of GPUs on a single chip, combined with the ease of programming as if it were a single device. This innovative approach enables us to achieve industry-leading training and inference speeds, allowing machine learning practitioners to run extensive ML applications effortlessly, without the complexities associated with managing numerous GPUs or TPUs. Cerebras is trusted by leading model labs, global enterprises, and pioneering AI-native startups. Notably, OpenAI recently announced a multi-year partnership with Cerebras, aimed at deploying 750 megawatts of scale, revolutionizing critical workloads with ultra high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, exceeding GPU-based hyperscale cloud inference services by more than 10 times. This significant enhancement in speed is redefining the user experience of AI applications, facilitating real-time iterations and amplifying intelligence through enhanced agentic computation.About The RoleAs a member of the inference performance team, you will work at the critical intersection of hardware and software, enhancing end-to-end model inference speed and throughput. Your focus will encompass low-level kernel performance debugging and optimization, system-level performance analysis, performance modeling, and the creation of tools for performance diagnostics and projections.ResponsibilitiesDevelop performance models (kernel-level, end-to-end) to forecast the performance of state-of-the-art and client ML models.Optimize and troubleshoot our kernel micro code and compiler algorithms to enhance ML model inference speed, throughput, and compute utilization on the Cerebras WSE.Analyze and debug runtime performance at the system and cluster level.Create tools and infrastructure to visualize performance data collected from the Wafer Scale Engine and our compute cluster.
Role overview Harvey seeks a Staff Software Engineer in Toronto to drive the evolution of its AI Platform. The focus is on designing and building scalable systems that can adapt to business needs and deliver reliable performance. Key responsibilities Architect and develop essential components for the AI Platform, prioritizing both performance and scalability. Collaborate with engineers and data scientists to address complex technical challenges. Enhance the platform’s stability, speed, and overall functionality. Collaboration and impact This role works alongside a talented group in engineering and data science. Contributions will have a direct effect on the platform’s capabilities and support Harvey’s business outcomes.
Cerebras Systems
Cerebras Systems is at the forefront of AI innovation, creating the world’s largest AI chip, a staggering 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers the computational power of dozens of GPUs within a single chip, paired with the simplicity of a unified programming interface. This unique approach enables us to achieve unparalleled training and inference speeds, empowering machine learning practitioners to execute large-scale ML applications effortlessly, without the complexities associated with hundreds of GPUs or TPUs.Among our esteemed clientele are leading model laboratories, global enterprises, and pioneering AI-native startups. Recently, OpenAI announced a multi-year collaboration with Cerebras, aiming to leverage 750 megawatts of scale to revolutionize key workloads through ultra-high-speed inference.Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference provides the fastest Generative AI inference solution available today, boasting speeds over ten times faster than GPU-based hyperscale cloud services. This extraordinary increase in speed is reshaping the user experience of AI applications, enabling real-time iterations and enhancing intelligence through advanced agentic computation.About The RoleJoin our inference model team, dedicated to advancing state-of-the-art models by numerically validating and accelerating innovative concepts on our wafer-scale hardware. In this role, you will prototype architectural enhancements, construct performance evaluation pipelines, and translate quantitative insights into actionable changes that drive production success.Key ResponsibilitiesPrototype and benchmark innovative concepts such as new attention mechanisms, mixture of experts (MoE), speculative decoding, and other emerging advancements.Create agent-driven automation tools that design experiments, schedule runs, triage regressions, and prepare pull requests.Collaborate closely with compiler, runtime, and silicon teams, gaining a unique perspective on the complete software/hardware innovation stack.Stay current with the latest open- and closed-source models; execute them on wafer scale first to identify new optimization opportunities.
About MagicalMagical is an innovative automation platform that integrates cutting-edge AI technology into the healthcare sector, providing AI agents capable of functioning effectively in real-world settings.Our mission is to create 'AI employees' that streamline tedious and time-consuming processes, specifically in healthcare—a massive $4 trillion industry laden with administrative burdens. We focus on automating critical tasks such as claims processing, prior authorizations, and eligibility checks, allowing healthcare professionals to dedicate more time to patient care.Our AchievementsAs pioneers in the transition to agentic automation in healthcare, we have achieved significant traction:Dramatic revenue growth with clients expanding into new workflows ahead of contract renewals.Rapid proof-of-concepts completed in just 7 days, delivering tangible value in an industry where similar timelines typically span months.Self-repairing automations with production-level reliability at scale, outperforming competitors who struggle to launch functional solutions.Unlike many AI firms that make grand promises, we provide dependable solutions that yield measurable outcomes. Our backing includes notable investors such as Greylock, Coatue, and Lightspeed, with a total funding of $41M. Our founder, Harpaul Sambhi, is a seasoned entrepreneur who previously sold his first company to LinkedIn.Role OverviewIn your role as a Senior Backend Engineer on the Agent Platform team, you will develop the core systems that drive our AI agents—covering aspects from orchestration and state management to model integration and evaluation pipelines. This position is centered around platform engineering, focusing on building the infrastructure that empowers our team to deliver reliable, production-grade agentic automation.You will work at the forefront of multi-agent systems, designing collaborative strategies for agents to tackle complex healthcare workflows. You will take ownership of essential backend services, overseeing them from architecture to deployment while developing verification systems that ensure our agents perform accurately every time.This hybrid position requires you to be onsite in our Toronto office two days each week.
Role Overview Paytm is seeking a Lead Staff Platform Engineer focused on AI Infrastructure in Toronto, Canada. This position centers on building and scaling the systems behind Paytm's AI inference platform, which supports both internal teams and enterprise clients as they develop new customer-facing AI solutions. What You Will Do Own and manage GPU infrastructure, model hosting, and model serving for the AI inference platform. Enable multi-model routing across different modalities, including voice, vision, and fintech workflows. Oversee deployment of proprietary and third-party models on shared GPU and accelerator clusters. Develop self-service platforms that let teams provision compute, deploy, and customize models using APIs and control planes. Establish the AI control plane for Paytm Intelligence (Pi), with a focus on policy-driven routing, quotas, observability, and cost visibility. Impact Areas Accelerate delivery of AI agents and features across payments, risk management, fraud detection, collections, support, and developer tooling. Improve reliability and efficiency of hardware utilization. Give teams the tools to use AI infrastructure without rebuilding core systems each time. Location This role is based in Toronto, Canada.
Affinity
Join Affinity as a Senior AI Engineer on our AI Platform team, where you will leverage your expertise in artificial intelligence to drive innovative solutions. In this role, you will collaborate with cross-functional teams to design, develop, and implement AI models that enhance our product offerings. You will work in a fast-paced environment, utilizing cutting-edge technologies to solve complex challenges and improve user experiences.
Role overview The Staff Engineer for Patient Platform at Prenuvo will play a key part in shaping technology that connects patients and healthcare professionals. The focus is on building systems that enhance patient experience and expand access to healthcare through thoughtful engineering. What you will do Design and implement systems that support both patients and healthcare teams Create solutions to improve how patients interact with healthcare services Use engineering expertise to boost the platform's reliability and reach Location This role is based in Toronto, Ontario, Canada.
Hello, Future Homie!At Homebase, we are on a mission to empower small businesses to flourish. Our team is dynamic, innovative, and dedicated to making a meaningful impact. We prioritize empathy in our development process, act swiftly, and embrace bold initiatives that create substantial real-world results. Every member of our team is committed to elevating our standards, supporting one another, and celebrating our collective achievements.We’re not just developing an application; we’re cultivating unstoppable teams. So, what do you think? Are you ready to join us?Your Impact Begins HereWe are in search of a Staff Engineer (E5) to help shape the architectural future of Homebase’s Core Platform. This pivotal role will span our Identity & Notifications (IN) and Billing & Businesses (BB) teams—two essential domains that facilitate customer access, communication, and transactions within Homebase.As a senior technical leader, you will play a dual role as both a hands-on contributor and a mentor. Reporting directly to the Senior Manager of Core Platform, you will collaborate with engineering managers and individual contributors to design scalable services in line with our service-oriented architecture (SOA) and event-driven architecture (EDA) vision. Your leadership will guide engineers in adopting best practices while identifying opportunities to enhance reliability, performance, and developer velocity across our systems.This is a crucial position for a seasoned platform engineer who is passionate about enabling teams, refining architectures, and tackling complex system challenges that yield impactful results at scale.Act as the technical liaison between the Identity & Notifications and Billing & Businesses teams, ensuring architectural integrity and dependability.Lead the design and implementation of new patterns aligned with SOA and EDA.Mentor engineers from both teams, enhancing their design, debugging, and systems thinking capabilities.Collaborate with Engineering Managers and Product Managers to clarify technical requirements, define project scopes, and sequence foundational tasks.Identify system bottlenecks, advocate for architectural enhancements, and implement practical solutions.Steer the technical direction of high-impact projects, particularly those introducing SOA/EDA patterns into our legacy Rails systems.Promote an AI-first development approach: utilize AI tools to enhance code quality, design exploration, and delivery speed.
Since its inception, Fivetran has been dedicated to simplifying and ensuring reliable access to data, much like the electricity that powers our lives. Our platform enables customer data to flow effortlessly into their warehouses, structured and ready for querying, without the need for engineering or maintenance. We take pride in empowering organizations to harness the full potential of their data every day.About the RoleJoin our dynamic ETL Platform team at Fivetran, where we are building robust data pipelines to support the modern data architecture of thousands of companies.As a Staff Software Engineer, you will play a pivotal role in managing the core platform that underpins Census’ sync engine. Your responsibilities will include ensuring the seamless flow of customer data—on time, every time. You will design and implement significant projects like our recent Sync Tracking overhaul, which enhances our ability to capture, store, and process large datasets. This provides our customers with unparalleled visibility over their data streams, ensuring that every record reaches its destination accurately. Our team is also responsible for managing a scalable and adaptable fleet of machines to meet the evolving demands of our clients.This is a full-time, hybrid position based in our Toronto office.
Join our dynamic team at Harvey as a Software Engineer, where you will play a pivotal role in shaping our cutting-edge AI platform. This is an exciting opportunity to work with innovative technologies and contribute to projects that push the boundaries of artificial intelligence.As a Software Engineer, you will collaborate with cross-functional teams to design, develop, and implement software solutions that enhance our AI capabilities. Your work will directly impact the efficiency and effectiveness of our platform, helping to drive our mission forward.
shyftlabs
Join shyftlabs as a Sales Engineer specializing in AI and Data Platforms. In this role, you will leverage your technical expertise to support the sales team in delivering innovative solutions to our clients. You will work closely with customers to understand their needs and demonstrate how our products can drive their success. Your ability to communicate complex technical concepts in an understandable way will be key to your success.
Rivian and Volkswagen Group Technologies
Rivian and Volkswagen Group Technologies have joined forces to advance the future of electric vehicles. This partnership brings together deep expertise in automotive engineering, software, and cloud connectivity. The focus is on building technology that sets new standards for software-defined vehicles worldwide, with an emphasis on AI, connectivity, and security. Role overview The Senior Software Engineer - AI Platform will lead architectural design and strategic development for the GenAI platform. This position centers on integrating large language models (LLMs) into both internal and customer-facing applications at scale. The aim is to drive cognitive automation, streamline workflows, and support better decision-making across the organization. What you will do Design and develop AI-driven applications using large language models. Integrate LLMs into production systems for both internal teams and external customers. Establish best practices for building distributed systems that are resilient, scalable, and observable. Deliver production-grade tools that are reliable and maintainable. Location This role is based in Toronto, Ontario.
NuBank
About UsNuBank is recognized as one of the largest digital financial platforms globally, serving over 127 million customers across Brazil, Mexico, and Colombia. Guided by our mission to combat complexity and empower individuals, we are transforming financial services in Latin America, marking just the beginning of the vibrant future we are crafting.As a publicly traded company on the New York Stock Exchange (NYSE: NU), we leverage proprietary technology, data intelligence, and an efficient operational model to provide financial products that are straightforward, accessible, and human-centric.Our contributions have been acknowledged in prestigious global rankings, including Time 100 Companies, Fast Company’s Most Innovative Companies, and Forbes World’s Best Bank. Explore more on our career page at NuBank Careers.About the RoleAs a Staff Software Engineer at NuBank, you will play a pivotal role in the AI Private Banker team, our intelligent financial partner designed to assist millions of customers in understanding their finances, making informed decisions, and executing actions effortlessly. You will be responsible for defining the technical architecture that supports conversational AI, recommendation engines, and tailored financial experiences across mobile and web platforms.Your expertise will span the entire stack, from developing Flutter mobile applications that engage millions of users to architecting distributed backend systems that facilitate real-time recommendations and conversational AI interactions. This role is not just about feature development; it’s about establishing a robust engineering foundation that democratizes financial empowerment at scale.Your technical decisions will have a direct impact on how millions of individuals navigate their financial journeys. You will be at the forefront of innovation, merging banking with cutting-edge AI technologies, crafting systems that are both groundbreaking and dependable. Join a team of inquisitive minds reshaping the possibilities in this space.
Tali AI is a rapidly expanding startup in Canada dedicated to enhancing healthcare accessibility through artificial intelligence. We are developing a pioneering AI platform designed to streamline clinical workflows by automating data collection, processing, and sharing. This allows healthcare professionals to concentrate on delivering quality care. Our platform is utilized by thousands of clinicians across Canada and the United States, spanning various specialties and seamlessly integrated with the complex landscape of North American health record systems. Our technical vision is to create a comprehensive operating system for clinical workflows, enabling healthcare teams to customize the platform to automate their data management tasks.We tackle genuinely challenging problems:Establishing real-time audio pipelines where latency impacts outcomes and accuracy is critical.Integrating with a diverse ecosystem of clinical software, including health records, billing, scheduling, referral management, and more.Creating reliable AI workflows and agents in a domain where errors can have significant consequences.Implementing a privacy and compliance framework that is robust by design.We operate in a fast-paced market that values rapid innovation, while building and maintaining clinician trust is paramount.
Extreme Networks, Inc.
Role Overview Extreme Networks is hiring a Staff Linux Systems Engineer for the Platform & Infrastructure team in Toronto, Canada. This role focuses on strengthening and evolving the company's Linux-based systems and core infrastructure. What You Will Do Work closely with teams across the organization to design and implement scalable Linux systems Contribute technical expertise to improve reliability, performance, and security of infrastructure Support the development and deployment of solutions that address client requirements What We’re Looking For Extensive experience with Linux systems engineering Strong background in building and maintaining infrastructure at scale Ability to collaborate effectively with cross-functional teams
Join Waabi, an innovative leader in Physical AI, founded by AI pioneer Raquel Urtasun. We are pioneering the future of autonomous transportation, developing cutting-edge technology for commercial autonomous trucks and robotaxis. Our esteemed team collaborates with top-tier partners across AI, automotive, logistics, and deep technology sectors.With offices located in Toronto, San Francisco, Dallas, and Pittsburgh, Waabi is rapidly expanding and is in search of diverse, innovative, and collaborative individuals eager to make a positive impact on the world. For more information, please visit: www.waabi.aiAt the heart of Waabi’s mission is our groundbreaking simulation technology. Our state-of-the-art closed-loop simulation engine, Waabi World, employs the latest generative AI technologies to create a highly realistic environment for training and evaluating self-driving systems. The Simulation Platform team is tasked with delivering a modular, scalable platform that seamlessly integrates the onboard autonomy system, intelligent actor models, and sensor simulation.Key Responsibilities:- Enhance and contribute to Waabi World, our advanced self-driving simulator, to expedite the development of autonomous vehicles.- Optimize the scalability and performance of the simulation platform to facilitate efficient simulations at a massive scale.- Design and develop orchestration software that interconnects simulation subcomponents, including the autonomy system, data-driven actor models, and sensor simulations powered by generative AI.- Engage in technical and architectural discussions, collaborating with researchers and engineers to share insights and innovative ideas.- Collaborate closely with internal users to grasp their needs and workflows, proposing new features to enhance their experience with the simulator.- Promote best practices and set standards for software development.- Lead complex cross-functional technical projects that have a significant impact across Waabi's systems and the organization as a whole.
Extreme Networks
About Us At Extreme Networks, we are dedicated to crafting seamless networking solutions that enable individuals and organizations to thrive in an increasingly connected world. We are on the lookout for a Director of AI Engineering to spearhead the design, development, and implementation of our cutting-edge AI-native systems.This pivotal role demands a dynamic leader who merges technical expertise with visionary thinking. You will set the strategic direction for AI initiatives, ensuring that innovative concepts transition from research to scalable, production-ready solutions. Your leadership will be crucial in launching enterprise-grade AI applications that revolutionize network design, optimization, security, and support.
Marble AI
Marble AI develops AI-powered tools designed to help accounting firms and their clients manage taxes with greater efficiency. The team’s work centers on simplifying complex tax workflows for mid-sized and large firms. As an early-stage company, Marble AI encourages collaboration and focuses on solving important challenges for its customers. Role overview The Platform Engineer will play a key role in strengthening and expanding the infrastructure that supports Marble AI’s tax platform. Working closely with the engineering team, this position aims to accelerate product development and maintain secure systems capable of handling sensitive financial information. This role combines elements of infrastructure, DevOps, and Site Reliability Engineering. The Platform Engineer will design and maintain the platform, enhance developer tools, and ensure that systems remain secure, observable, and resilient. Input into the direction of the data platform and security practices across the technology stack is also expected. What you will do Build and manage cloud infrastructure using infrastructure-as-code principles. Take responsibility for system reliability, observability (with OTEL), and incident response. Design and refine developer tooling and CI/CD pipelines. Collaborate with product and data teams to address scalable data platform needs. Maintain strong security practices, especially in handling sensitive tax data. Optimize system performance, scalability, and cost across services. Contribute to infrastructure, backend systems, and developer workflows. Requirements Extensive experience with cloud infrastructure such as AWS or similar platforms. Hands-on expertise with Kubernetes, Terraform, and Helm. Background in DevOps, SRE, or platform engineering roles. Experience building and maintaining CI/CD pipelines or developer tools. Strong understanding of distributed systems, networking, and system reliability. Ability to deliver practical solutions while considering long-term system design. Location Toronto
Sign in to browse more jobs
Create account — see all 1,200 results

