Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Mid to Senior
Qualifications
Proven experience in software engineering, preferably with a focus on AI and machine learning. Strong knowledge of inference systems and agentic technologies. Familiarity with programming languages such as Python, Java, or similar. Experience with cloud platforms and deployment of scalable applications. Excellent problem-solving skills and the ability to work in a team-oriented environment.
About the job
Paytm is hiring a Staff AI Platform Engineer in Toronto, Canada, with a focus on Inference and Agentic Systems. This role centers on designing and improving AI-powered platforms that support intelligent, agent-like features for users and business applications.
What You Will Do
Work with a skilled team to build and refine AI solutions that support agentic behaviors and inference-driven capabilities.
Apply experience in machine learning, software engineering, and system architecture to develop and scale reliable AI platforms.
Contribute technical leadership and hands-on expertise to projects that advance Paytm’s AI offerings.
Key Skills
Deep knowledge of machine learning and inference methods
Strong background in software engineering
Experience designing and maintaining complex system architectures
About Paytm
Paytm is a leading digital payments and financial services company in India, known for its innovative approach to technology and customer-centric services. With a mission to bring financial inclusion to millions, we strive to create a seamless and efficient financial ecosystem.
Role Overview Paytm is hiring a Staff AI Platform Engineer in Toronto, Canada, with a focus on Inference and Agentic Systems. This role centers on designing and improving AI-powered platforms that support intelligent, agent-like features for users and business applications. What You Will Do Work with a skilled team to build and refine AI solutions that support agentic behaviors and inference-driven capabilities. Apply experience in machine learning, software engineering, and system architecture to develop and scale reliable AI platforms. Contribute technical leadership and hands-on expertise to projects that advance Paytm’s AI offerings. Key Skills Deep knowledge of machine learning and inference methods Strong background in software engineering Experience designing and maintaining complex system architectures
Cerebras Systems is revolutionizing AI technology with the world's largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture combines the immense computational power of multiple GPUs into a single chip while maintaining unparalleled programming simplicity. This allows us to provide extraordinary training and inference speeds, empowering machine learning users to seamlessly execute large-scale ML applications without the complexities of managing numerous GPUs or TPUs. We proudly serve a diverse clientele, including leading model laboratories, global corporations, and innovative AI-centric startups. Notably, OpenAI recently formed a multi-year partnership with Cerebras, committing to deploy 750 megawatts of scale to enhance critical workloads with ultra-high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, achieving speeds over 10 times faster than GPU-based hyperscale cloud inference services. This remarkable speed transformation enhances user experiences and facilitates real-time iterations while augmenting intelligence through advanced agentic computation. About The RoleThe Inference Core Platform team is integral to Cerebras’ mission of delivering the world’s fastest AI inference. Our engineers develop the core software and hardware infrastructure that enables low-latency, high-speed, and high-throughput deployment on the Cerebras Wafer-Scale Engine (WSE). We oversee the entire stack—from model compilation and scheduling to custom hardware kernels and driver development.The Platform Benchmarking team is crucial in enhancing the performance and scalability of AI inference on one of the most advanced computing systems ever developed. We spearhead the establishment of core inference capabilities and implement performance improvements at every development phase, from initial prototyping to full production deployment.We seek enthusiastic engineers eager to redefine the boundaries of AI inference. If you're passionate about developing systems that measure, analyze, and optimize performance on a large scale, this is your chance to make a transformative impact on the future of AI.
Cerebras Systems is at the forefront of AI technology, having developed the world's largest AI chip, which is 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers unparalleled AI compute power equivalent to dozens of GPUs on a single chip, combined with the ease of programming as if it were a single device. This innovative approach enables us to achieve industry-leading training and inference speeds, allowing machine learning practitioners to run extensive ML applications effortlessly, without the complexities associated with managing numerous GPUs or TPUs. Cerebras is trusted by leading model labs, global enterprises, and pioneering AI-native startups. Notably, OpenAI recently announced a multi-year partnership with Cerebras, aimed at deploying 750 megawatts of scale, revolutionizing critical workloads with ultra high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, exceeding GPU-based hyperscale cloud inference services by more than 10 times. This significant enhancement in speed is redefining the user experience of AI applications, facilitating real-time iterations and amplifying intelligence through enhanced agentic computation.About The RoleAs a member of the inference performance team, you will work at the critical intersection of hardware and software, enhancing end-to-end model inference speed and throughput. Your focus will encompass low-level kernel performance debugging and optimization, system-level performance analysis, performance modeling, and the creation of tools for performance diagnostics and projections.ResponsibilitiesDevelop performance models (kernel-level, end-to-end) to forecast the performance of state-of-the-art and client ML models.Optimize and troubleshoot our kernel micro code and compiler algorithms to enhance ML model inference speed, throughput, and compute utilization on the Cerebras WSE.Analyze and debug runtime performance at the system and cluster level.Create tools and infrastructure to visualize performance data collected from the Wafer Scale Engine and our compute cluster.
Role overview Harvey seeks a Staff Software Engineer in Toronto to drive the evolution of its AI Platform. The focus is on designing and building scalable systems that can adapt to business needs and deliver reliable performance. Key responsibilities Architect and develop essential components for the AI Platform, prioritizing both performance and scalability. Collaborate with engineers and data scientists to address complex technical challenges. Enhance the platform’s stability, speed, and overall functionality. Collaboration and impact This role works alongside a talented group in engineering and data science. Contributions will have a direct effect on the platform’s capabilities and support Harvey’s business outcomes.
Cerebras Systems is at the forefront of AI innovation, creating the world’s largest AI chip, a staggering 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers the computational power of dozens of GPUs within a single chip, paired with the simplicity of a unified programming interface. This unique approach enables us to achieve unparalleled training and inference speeds, empowering machine learning practitioners to execute large-scale ML applications effortlessly, without the complexities associated with hundreds of GPUs or TPUs.Among our esteemed clientele are leading model laboratories, global enterprises, and pioneering AI-native startups. Recently, OpenAI announced a multi-year collaboration with Cerebras, aiming to leverage 750 megawatts of scale to revolutionize key workloads through ultra-high-speed inference.Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference provides the fastest Generative AI inference solution available today, boasting speeds over ten times faster than GPU-based hyperscale cloud services. This extraordinary increase in speed is reshaping the user experience of AI applications, enabling real-time iterations and enhancing intelligence through advanced agentic computation.About The RoleJoin our inference model team, dedicated to advancing state-of-the-art models by numerically validating and accelerating innovative concepts on our wafer-scale hardware. In this role, you will prototype architectural enhancements, construct performance evaluation pipelines, and translate quantitative insights into actionable changes that drive production success.Key ResponsibilitiesPrototype and benchmark innovative concepts such as new attention mechanisms, mixture of experts (MoE), speculative decoding, and other emerging advancements.Create agent-driven automation tools that design experiments, schedule runs, triage regressions, and prepare pull requests.Collaborate closely with compiler, runtime, and silicon teams, gaining a unique perspective on the complete software/hardware innovation stack.Stay current with the latest open- and closed-source models; execute them on wafer scale first to identify new optimization opportunities.
About MagicalMagical is an innovative automation platform that integrates cutting-edge AI technology into the healthcare sector, providing AI agents capable of functioning effectively in real-world settings.Our mission is to create 'AI employees' that streamline tedious and time-consuming processes, specifically in healthcare—a massive $4 trillion industry laden with administrative burdens. We focus on automating critical tasks such as claims processing, prior authorizations, and eligibility checks, allowing healthcare professionals to dedicate more time to patient care.Our AchievementsAs pioneers in the transition to agentic automation in healthcare, we have achieved significant traction:Dramatic revenue growth with clients expanding into new workflows ahead of contract renewals.Rapid proof-of-concepts completed in just 7 days, delivering tangible value in an industry where similar timelines typically span months.Self-repairing automations with production-level reliability at scale, outperforming competitors who struggle to launch functional solutions.Unlike many AI firms that make grand promises, we provide dependable solutions that yield measurable outcomes. Our backing includes notable investors such as Greylock, Coatue, and Lightspeed, with a total funding of $41M. Our founder, Harpaul Sambhi, is a seasoned entrepreneur who previously sold his first company to LinkedIn.Role OverviewIn your role as a Senior Backend Engineer on the Agent Platform team, you will develop the core systems that drive our AI agents—covering aspects from orchestration and state management to model integration and evaluation pipelines. This position is centered around platform engineering, focusing on building the infrastructure that empowers our team to deliver reliable, production-grade agentic automation.You will work at the forefront of multi-agent systems, designing collaborative strategies for agents to tackle complex healthcare workflows. You will take ownership of essential backend services, overseeing them from architecture to deployment while developing verification systems that ensure our agents perform accurately every time.This hybrid position requires you to be onsite in our Toronto office two days each week.
Role Overview Paytm is seeking a Lead Staff Platform Engineer focused on AI Infrastructure in Toronto, Canada. This position centers on building and scaling the systems behind Paytm's AI inference platform, which supports both internal teams and enterprise clients as they develop new customer-facing AI solutions. What You Will Do Own and manage GPU infrastructure, model hosting, and model serving for the AI inference platform. Enable multi-model routing across different modalities, including voice, vision, and fintech workflows. Oversee deployment of proprietary and third-party models on shared GPU and accelerator clusters. Develop self-service platforms that let teams provision compute, deploy, and customize models using APIs and control planes. Establish the AI control plane for Paytm Intelligence (Pi), with a focus on policy-driven routing, quotas, observability, and cost visibility. Impact Areas Accelerate delivery of AI agents and features across payments, risk management, fraud detection, collections, support, and developer tooling. Improve reliability and efficiency of hardware utilization. Give teams the tools to use AI infrastructure without rebuilding core systems each time. Location This role is based in Toronto, Canada.
Join Affinity as a Senior AI Engineer on our AI Platform team, where you will leverage your expertise in artificial intelligence to drive innovative solutions. In this role, you will collaborate with cross-functional teams to design, develop, and implement AI models that enhance our product offerings. You will work in a fast-paced environment, utilizing cutting-edge technologies to solve complex challenges and improve user experiences.
Role overview The Staff Engineer for Patient Platform at Prenuvo will play a key part in shaping technology that connects patients and healthcare professionals. The focus is on building systems that enhance patient experience and expand access to healthcare through thoughtful engineering. What you will do Design and implement systems that support both patients and healthcare teams Create solutions to improve how patients interact with healthcare services Use engineering expertise to boost the platform's reliability and reach Location This role is based in Toronto, Ontario, Canada.
Hello, Future Homie!At Homebase, we are on a mission to empower small businesses to flourish. Our team is dynamic, innovative, and dedicated to making a meaningful impact. We prioritize empathy in our development process, act swiftly, and embrace bold initiatives that create substantial real-world results. Every member of our team is committed to elevating our standards, supporting one another, and celebrating our collective achievements.We’re not just developing an application; we’re cultivating unstoppable teams. So, what do you think? Are you ready to join us?Your Impact Begins HereWe are in search of a Staff Engineer (E5) to help shape the architectural future of Homebase’s Core Platform. This pivotal role will span our Identity & Notifications (IN) and Billing & Businesses (BB) teams—two essential domains that facilitate customer access, communication, and transactions within Homebase.As a senior technical leader, you will play a dual role as both a hands-on contributor and a mentor. Reporting directly to the Senior Manager of Core Platform, you will collaborate with engineering managers and individual contributors to design scalable services in line with our service-oriented architecture (SOA) and event-driven architecture (EDA) vision. Your leadership will guide engineers in adopting best practices while identifying opportunities to enhance reliability, performance, and developer velocity across our systems.This is a crucial position for a seasoned platform engineer who is passionate about enabling teams, refining architectures, and tackling complex system challenges that yield impactful results at scale.Act as the technical liaison between the Identity & Notifications and Billing & Businesses teams, ensuring architectural integrity and dependability.Lead the design and implementation of new patterns aligned with SOA and EDA.Mentor engineers from both teams, enhancing their design, debugging, and systems thinking capabilities.Collaborate with Engineering Managers and Product Managers to clarify technical requirements, define project scopes, and sequence foundational tasks.Identify system bottlenecks, advocate for architectural enhancements, and implement practical solutions.Steer the technical direction of high-impact projects, particularly those introducing SOA/EDA patterns into our legacy Rails systems.Promote an AI-first development approach: utilize AI tools to enhance code quality, design exploration, and delivery speed.
Since its inception, Fivetran has been dedicated to simplifying and ensuring reliable access to data, much like the electricity that powers our lives. Our platform enables customer data to flow effortlessly into their warehouses, structured and ready for querying, without the need for engineering or maintenance. We take pride in empowering organizations to harness the full potential of their data every day.About the RoleJoin our dynamic ETL Platform team at Fivetran, where we are building robust data pipelines to support the modern data architecture of thousands of companies.As a Staff Software Engineer, you will play a pivotal role in managing the core platform that underpins Census’ sync engine. Your responsibilities will include ensuring the seamless flow of customer data—on time, every time. You will design and implement significant projects like our recent Sync Tracking overhaul, which enhances our ability to capture, store, and process large datasets. This provides our customers with unparalleled visibility over their data streams, ensuring that every record reaches its destination accurately. Our team is also responsible for managing a scalable and adaptable fleet of machines to meet the evolving demands of our clients.This is a full-time, hybrid position based in our Toronto office.
Join our dynamic team at Harvey as a Software Engineer, where you will play a pivotal role in shaping our cutting-edge AI platform. This is an exciting opportunity to work with innovative technologies and contribute to projects that push the boundaries of artificial intelligence.As a Software Engineer, you will collaborate with cross-functional teams to design, develop, and implement software solutions that enhance our AI capabilities. Your work will directly impact the efficiency and effectiveness of our platform, helping to drive our mission forward.
Join shyftlabs as a Sales Engineer specializing in AI and Data Platforms. In this role, you will leverage your technical expertise to support the sales team in delivering innovative solutions to our clients. You will work closely with customers to understand their needs and demonstrate how our products can drive their success. Your ability to communicate complex technical concepts in an understandable way will be key to your success.
Full-time|On-site|Canada, Toronto; USA, Durham; USA, Miami; USA, Palo Alto
About UsNuBank is recognized as one of the largest digital financial platforms globally, serving over 127 million customers across Brazil, Mexico, and Colombia. Guided by our mission to combat complexity and empower individuals, we are transforming financial services in Latin America, marking just the beginning of the vibrant future we are crafting.As a publicly traded company on the New York Stock Exchange (NYSE: NU), we leverage proprietary technology, data intelligence, and an efficient operational model to provide financial products that are straightforward, accessible, and human-centric.Our contributions have been acknowledged in prestigious global rankings, including Time 100 Companies, Fast Company’s Most Innovative Companies, and Forbes World’s Best Bank. Explore more on our career page at NuBank Careers.About the RoleAs a Staff Software Engineer at NuBank, you will play a pivotal role in the AI Private Banker team, our intelligent financial partner designed to assist millions of customers in understanding their finances, making informed decisions, and executing actions effortlessly. You will be responsible for defining the technical architecture that supports conversational AI, recommendation engines, and tailored financial experiences across mobile and web platforms.Your expertise will span the entire stack, from developing Flutter mobile applications that engage millions of users to architecting distributed backend systems that facilitate real-time recommendations and conversational AI interactions. This role is not just about feature development; it’s about establishing a robust engineering foundation that democratizes financial empowerment at scale.Your technical decisions will have a direct impact on how millions of individuals navigate their financial journeys. You will be at the forefront of innovation, merging banking with cutting-edge AI technologies, crafting systems that are both groundbreaking and dependable. Join a team of inquisitive minds reshaping the possibilities in this space.
Tali AI is a rapidly expanding startup in Canada dedicated to enhancing healthcare accessibility through artificial intelligence. We are developing a pioneering AI platform designed to streamline clinical workflows by automating data collection, processing, and sharing. This allows healthcare professionals to concentrate on delivering quality care. Our platform is utilized by thousands of clinicians across Canada and the United States, spanning various specialties and seamlessly integrated with the complex landscape of North American health record systems. Our technical vision is to create a comprehensive operating system for clinical workflows, enabling healthcare teams to customize the platform to automate their data management tasks.We tackle genuinely challenging problems:Establishing real-time audio pipelines where latency impacts outcomes and accuracy is critical.Integrating with a diverse ecosystem of clinical software, including health records, billing, scheduling, referral management, and more.Creating reliable AI workflows and agents in a domain where errors can have significant consequences.Implementing a privacy and compliance framework that is robust by design.We operate in a fast-paced market that values rapid innovation, while building and maintaining clinician trust is paramount.
Role Overview Extreme Networks is hiring a Staff Linux Systems Engineer for the Platform & Infrastructure team in Toronto, Canada. This role focuses on strengthening and evolving the company's Linux-based systems and core infrastructure. What You Will Do Work closely with teams across the organization to design and implement scalable Linux systems Contribute technical expertise to improve reliability, performance, and security of infrastructure Support the development and deployment of solutions that address client requirements What We’re Looking For Extensive experience with Linux systems engineering Strong background in building and maintaining infrastructure at scale Ability to collaborate effectively with cross-functional teams
Join Waabi, an innovative leader in Physical AI, founded by AI pioneer Raquel Urtasun. We are pioneering the future of autonomous transportation, developing cutting-edge technology for commercial autonomous trucks and robotaxis. Our esteemed team collaborates with top-tier partners across AI, automotive, logistics, and deep technology sectors.With offices located in Toronto, San Francisco, Dallas, and Pittsburgh, Waabi is rapidly expanding and is in search of diverse, innovative, and collaborative individuals eager to make a positive impact on the world. For more information, please visit: www.waabi.aiAt the heart of Waabi’s mission is our groundbreaking simulation technology. Our state-of-the-art closed-loop simulation engine, Waabi World, employs the latest generative AI technologies to create a highly realistic environment for training and evaluating self-driving systems. The Simulation Platform team is tasked with delivering a modular, scalable platform that seamlessly integrates the onboard autonomy system, intelligent actor models, and sensor simulation.Key Responsibilities:- Enhance and contribute to Waabi World, our advanced self-driving simulator, to expedite the development of autonomous vehicles.- Optimize the scalability and performance of the simulation platform to facilitate efficient simulations at a massive scale.- Design and develop orchestration software that interconnects simulation subcomponents, including the autonomy system, data-driven actor models, and sensor simulations powered by generative AI.- Engage in technical and architectural discussions, collaborating with researchers and engineers to share insights and innovative ideas.- Collaborate closely with internal users to grasp their needs and workflows, proposing new features to enhance their experience with the simulator.- Promote best practices and set standards for software development.- Lead complex cross-functional technical projects that have a significant impact across Waabi's systems and the organization as a whole.
About Us At Extreme Networks, we are dedicated to crafting seamless networking solutions that enable individuals and organizations to thrive in an increasingly connected world. We are on the lookout for a Director of AI Engineering to spearhead the design, development, and implementation of our cutting-edge AI-native systems.This pivotal role demands a dynamic leader who merges technical expertise with visionary thinking. You will set the strategic direction for AI initiatives, ensuring that innovative concepts transition from research to scalable, production-ready solutions. Your leadership will be crucial in launching enterprise-grade AI applications that revolutionize network design, optimization, security, and support.
About Marble AI Marble AI builds AI-powered tools that help accounting firms and their clients manage taxes more efficiently. The team focuses on simplifying complex tax workflows for mid-to-large-sized firms. As an early-stage company, Marble AI values collaboration and aims to solve high-impact problems for its customers. Role Overview The Platform Engineer will help strengthen and expand the infrastructure behind Marble AI’s tax platform. This position works closely with the engineering team to speed up product development and ensure systems can securely handle sensitive financial data. The role blends infrastructure, DevOps, and Site Reliability Engineering (SRE). Responsibilities include designing and maintaining the platform, improving developer tools, and making sure systems are secure, observable, and resilient. The Platform Engineer will also have a say in the direction of Marble AI’s data platform and security practices across the stack. What You’ll Do Build and manage cloud infrastructure, treating infrastructure as code. Own system reliability, observability (using OTEL), and incident response processes. Design and improve developer tooling and CI/CD pipelines. Work with product and data teams to support scalable data platform needs. Apply and maintain strong security practices, especially for sensitive tax data. Optimize system performance, scalability, and cost across services. Contribute across infrastructure, backend systems, and developer workflows. What We’re Looking For Deep experience with cloud infrastructure (AWS or similar platforms). Hands-on experience with Kubernetes, Terraform, and Helm. Previous work in DevOps, SRE, or platform engineering roles. Experience building and maintaining CI/CD pipelines or developer tooling. Solid understanding of distributed systems, networking, and system reliability. Ability to balance practical solutions with long-term system design. Location Toronto
Cresta is dedicated to transforming customer interactions into a strategic advantage by harnessing the full capabilities of the contact center. Our innovative platform merges cutting-edge AI with human intelligence to provide contact centers with valuable customer insights, streamline automated processes, and empower every team member to enhance productivity and efficiency. Originating from the renowned Stanford AI lab, Cresta boasts co-founder and chairman Sebastian Thrun, the visionary behind Google X, Waymo, and Udacity, alongside CEO Ping Wu, co-founder of Google Contact Center AI, and Tim Shi, an early member of OpenAI.Embark with us on an exciting journey to revolutionize the workforce through AI. The future of work is here, and it is at Cresta.About the Role:We are on the lookout for passionate UX interns to contribute to the evolution of Cresta’s human + AI agents platform, which facilitates AI-driven business workflows and automation for contact center operators. Your role will involve designing exceptional end-user experiences and the tools that enable these interactions.As a product design intern at Cresta, you will be at the forefront of a significant technological advancement with LLM and Generative AI. This unique opportunity allows you to make a meaningful impact on how individuals interact with AI. Under the guidance of a design supervisor, you will engage in the complete UX process, including product design, user research, and visual design components, to deliver outstanding user experiences.Responsibilities:Generate innovative UX concepts for Conversation Intelligence and AI Agents, taking into account diverse user personas.Analyze user and business challenges, propose viable solutions, and collaborate with engineers to implement designs.Conduct user research, including usability testing, competitive analysis, customer interviews, behavioral insights, and metrics analysis to inform design decisions.Rapidly prototype (using tools like Figma, Cursor, Claude, Voice API, etc.) to explore various UI paradigms.Work closely with Engineers, PMs, ML experts, Customer Success, and Sales teams throughout the organization.Effectively present and communicate design concepts to supervisors and cross-functional teams.
Role Overview Paytm is hiring a Staff AI Platform Engineer in Toronto, Canada, with a focus on Inference and Agentic Systems. This role centers on designing and improving AI-powered platforms that support intelligent, agent-like features for users and business applications. What You Will Do Work with a skilled team to build and refine AI solutions that support agentic behaviors and inference-driven capabilities. Apply experience in machine learning, software engineering, and system architecture to develop and scale reliable AI platforms. Contribute technical leadership and hands-on expertise to projects that advance Paytm’s AI offerings. Key Skills Deep knowledge of machine learning and inference methods Strong background in software engineering Experience designing and maintaining complex system architectures
Cerebras Systems is revolutionizing AI technology with the world's largest AI chip, which is 56 times larger than traditional GPUs. Our innovative wafer-scale architecture combines the immense computational power of multiple GPUs into a single chip while maintaining unparalleled programming simplicity. This allows us to provide extraordinary training and inference speeds, empowering machine learning users to seamlessly execute large-scale ML applications without the complexities of managing numerous GPUs or TPUs. We proudly serve a diverse clientele, including leading model laboratories, global corporations, and innovative AI-centric startups. Notably, OpenAI recently formed a multi-year partnership with Cerebras, committing to deploy 750 megawatts of scale to enhance critical workloads with ultra-high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, achieving speeds over 10 times faster than GPU-based hyperscale cloud inference services. This remarkable speed transformation enhances user experiences and facilitates real-time iterations while augmenting intelligence through advanced agentic computation. About The RoleThe Inference Core Platform team is integral to Cerebras’ mission of delivering the world’s fastest AI inference. Our engineers develop the core software and hardware infrastructure that enables low-latency, high-speed, and high-throughput deployment on the Cerebras Wafer-Scale Engine (WSE). We oversee the entire stack—from model compilation and scheduling to custom hardware kernels and driver development.The Platform Benchmarking team is crucial in enhancing the performance and scalability of AI inference on one of the most advanced computing systems ever developed. We spearhead the establishment of core inference capabilities and implement performance improvements at every development phase, from initial prototyping to full production deployment.We seek enthusiastic engineers eager to redefine the boundaries of AI inference. If you're passionate about developing systems that measure, analyze, and optimize performance on a large scale, this is your chance to make a transformative impact on the future of AI.
Cerebras Systems is at the forefront of AI technology, having developed the world's largest AI chip, which is 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers unparalleled AI compute power equivalent to dozens of GPUs on a single chip, combined with the ease of programming as if it were a single device. This innovative approach enables us to achieve industry-leading training and inference speeds, allowing machine learning practitioners to run extensive ML applications effortlessly, without the complexities associated with managing numerous GPUs or TPUs. Cerebras is trusted by leading model labs, global enterprises, and pioneering AI-native startups. Notably, OpenAI recently announced a multi-year partnership with Cerebras, aimed at deploying 750 megawatts of scale, revolutionizing critical workloads with ultra high-speed inference. Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference offers the fastest Generative AI inference solution globally, exceeding GPU-based hyperscale cloud inference services by more than 10 times. This significant enhancement in speed is redefining the user experience of AI applications, facilitating real-time iterations and amplifying intelligence through enhanced agentic computation.About The RoleAs a member of the inference performance team, you will work at the critical intersection of hardware and software, enhancing end-to-end model inference speed and throughput. Your focus will encompass low-level kernel performance debugging and optimization, system-level performance analysis, performance modeling, and the creation of tools for performance diagnostics and projections.ResponsibilitiesDevelop performance models (kernel-level, end-to-end) to forecast the performance of state-of-the-art and client ML models.Optimize and troubleshoot our kernel micro code and compiler algorithms to enhance ML model inference speed, throughput, and compute utilization on the Cerebras WSE.Analyze and debug runtime performance at the system and cluster level.Create tools and infrastructure to visualize performance data collected from the Wafer Scale Engine and our compute cluster.
Role overview Harvey seeks a Staff Software Engineer in Toronto to drive the evolution of its AI Platform. The focus is on designing and building scalable systems that can adapt to business needs and deliver reliable performance. Key responsibilities Architect and develop essential components for the AI Platform, prioritizing both performance and scalability. Collaborate with engineers and data scientists to address complex technical challenges. Enhance the platform’s stability, speed, and overall functionality. Collaboration and impact This role works alongside a talented group in engineering and data science. Contributions will have a direct effect on the platform’s capabilities and support Harvey’s business outcomes.
Cerebras Systems is at the forefront of AI innovation, creating the world’s largest AI chip, a staggering 56 times larger than traditional GPUs. Our revolutionary wafer-scale architecture delivers the computational power of dozens of GPUs within a single chip, paired with the simplicity of a unified programming interface. This unique approach enables us to achieve unparalleled training and inference speeds, empowering machine learning practitioners to execute large-scale ML applications effortlessly, without the complexities associated with hundreds of GPUs or TPUs.Among our esteemed clientele are leading model laboratories, global enterprises, and pioneering AI-native startups. Recently, OpenAI announced a multi-year collaboration with Cerebras, aiming to leverage 750 megawatts of scale to revolutionize key workloads through ultra-high-speed inference.Thanks to our groundbreaking wafer-scale architecture, Cerebras Inference provides the fastest Generative AI inference solution available today, boasting speeds over ten times faster than GPU-based hyperscale cloud services. This extraordinary increase in speed is reshaping the user experience of AI applications, enabling real-time iterations and enhancing intelligence through advanced agentic computation.About The RoleJoin our inference model team, dedicated to advancing state-of-the-art models by numerically validating and accelerating innovative concepts on our wafer-scale hardware. In this role, you will prototype architectural enhancements, construct performance evaluation pipelines, and translate quantitative insights into actionable changes that drive production success.Key ResponsibilitiesPrototype and benchmark innovative concepts such as new attention mechanisms, mixture of experts (MoE), speculative decoding, and other emerging advancements.Create agent-driven automation tools that design experiments, schedule runs, triage regressions, and prepare pull requests.Collaborate closely with compiler, runtime, and silicon teams, gaining a unique perspective on the complete software/hardware innovation stack.Stay current with the latest open- and closed-source models; execute them on wafer scale first to identify new optimization opportunities.
About MagicalMagical is an innovative automation platform that integrates cutting-edge AI technology into the healthcare sector, providing AI agents capable of functioning effectively in real-world settings.Our mission is to create 'AI employees' that streamline tedious and time-consuming processes, specifically in healthcare—a massive $4 trillion industry laden with administrative burdens. We focus on automating critical tasks such as claims processing, prior authorizations, and eligibility checks, allowing healthcare professionals to dedicate more time to patient care.Our AchievementsAs pioneers in the transition to agentic automation in healthcare, we have achieved significant traction:Dramatic revenue growth with clients expanding into new workflows ahead of contract renewals.Rapid proof-of-concepts completed in just 7 days, delivering tangible value in an industry where similar timelines typically span months.Self-repairing automations with production-level reliability at scale, outperforming competitors who struggle to launch functional solutions.Unlike many AI firms that make grand promises, we provide dependable solutions that yield measurable outcomes. Our backing includes notable investors such as Greylock, Coatue, and Lightspeed, with a total funding of $41M. Our founder, Harpaul Sambhi, is a seasoned entrepreneur who previously sold his first company to LinkedIn.Role OverviewIn your role as a Senior Backend Engineer on the Agent Platform team, you will develop the core systems that drive our AI agents—covering aspects from orchestration and state management to model integration and evaluation pipelines. This position is centered around platform engineering, focusing on building the infrastructure that empowers our team to deliver reliable, production-grade agentic automation.You will work at the forefront of multi-agent systems, designing collaborative strategies for agents to tackle complex healthcare workflows. You will take ownership of essential backend services, overseeing them from architecture to deployment while developing verification systems that ensure our agents perform accurately every time.This hybrid position requires you to be onsite in our Toronto office two days each week.
Role Overview Paytm is seeking a Lead Staff Platform Engineer focused on AI Infrastructure in Toronto, Canada. This position centers on building and scaling the systems behind Paytm's AI inference platform, which supports both internal teams and enterprise clients as they develop new customer-facing AI solutions. What You Will Do Own and manage GPU infrastructure, model hosting, and model serving for the AI inference platform. Enable multi-model routing across different modalities, including voice, vision, and fintech workflows. Oversee deployment of proprietary and third-party models on shared GPU and accelerator clusters. Develop self-service platforms that let teams provision compute, deploy, and customize models using APIs and control planes. Establish the AI control plane for Paytm Intelligence (Pi), with a focus on policy-driven routing, quotas, observability, and cost visibility. Impact Areas Accelerate delivery of AI agents and features across payments, risk management, fraud detection, collections, support, and developer tooling. Improve reliability and efficiency of hardware utilization. Give teams the tools to use AI infrastructure without rebuilding core systems each time. Location This role is based in Toronto, Canada.
Join Affinity as a Senior AI Engineer on our AI Platform team, where you will leverage your expertise in artificial intelligence to drive innovative solutions. In this role, you will collaborate with cross-functional teams to design, develop, and implement AI models that enhance our product offerings. You will work in a fast-paced environment, utilizing cutting-edge technologies to solve complex challenges and improve user experiences.
Role overview The Staff Engineer for Patient Platform at Prenuvo will play a key part in shaping technology that connects patients and healthcare professionals. The focus is on building systems that enhance patient experience and expand access to healthcare through thoughtful engineering. What you will do Design and implement systems that support both patients and healthcare teams Create solutions to improve how patients interact with healthcare services Use engineering expertise to boost the platform's reliability and reach Location This role is based in Toronto, Ontario, Canada.
Hello, Future Homie!At Homebase, we are on a mission to empower small businesses to flourish. Our team is dynamic, innovative, and dedicated to making a meaningful impact. We prioritize empathy in our development process, act swiftly, and embrace bold initiatives that create substantial real-world results. Every member of our team is committed to elevating our standards, supporting one another, and celebrating our collective achievements.We’re not just developing an application; we’re cultivating unstoppable teams. So, what do you think? Are you ready to join us?Your Impact Begins HereWe are in search of a Staff Engineer (E5) to help shape the architectural future of Homebase’s Core Platform. This pivotal role will span our Identity & Notifications (IN) and Billing & Businesses (BB) teams—two essential domains that facilitate customer access, communication, and transactions within Homebase.As a senior technical leader, you will play a dual role as both a hands-on contributor and a mentor. Reporting directly to the Senior Manager of Core Platform, you will collaborate with engineering managers and individual contributors to design scalable services in line with our service-oriented architecture (SOA) and event-driven architecture (EDA) vision. Your leadership will guide engineers in adopting best practices while identifying opportunities to enhance reliability, performance, and developer velocity across our systems.This is a crucial position for a seasoned platform engineer who is passionate about enabling teams, refining architectures, and tackling complex system challenges that yield impactful results at scale.Act as the technical liaison between the Identity & Notifications and Billing & Businesses teams, ensuring architectural integrity and dependability.Lead the design and implementation of new patterns aligned with SOA and EDA.Mentor engineers from both teams, enhancing their design, debugging, and systems thinking capabilities.Collaborate with Engineering Managers and Product Managers to clarify technical requirements, define project scopes, and sequence foundational tasks.Identify system bottlenecks, advocate for architectural enhancements, and implement practical solutions.Steer the technical direction of high-impact projects, particularly those introducing SOA/EDA patterns into our legacy Rails systems.Promote an AI-first development approach: utilize AI tools to enhance code quality, design exploration, and delivery speed.
Since its inception, Fivetran has been dedicated to simplifying and ensuring reliable access to data, much like the electricity that powers our lives. Our platform enables customer data to flow effortlessly into their warehouses, structured and ready for querying, without the need for engineering or maintenance. We take pride in empowering organizations to harness the full potential of their data every day.About the RoleJoin our dynamic ETL Platform team at Fivetran, where we are building robust data pipelines to support the modern data architecture of thousands of companies.As a Staff Software Engineer, you will play a pivotal role in managing the core platform that underpins Census’ sync engine. Your responsibilities will include ensuring the seamless flow of customer data—on time, every time. You will design and implement significant projects like our recent Sync Tracking overhaul, which enhances our ability to capture, store, and process large datasets. This provides our customers with unparalleled visibility over their data streams, ensuring that every record reaches its destination accurately. Our team is also responsible for managing a scalable and adaptable fleet of machines to meet the evolving demands of our clients.This is a full-time, hybrid position based in our Toronto office.
Join our dynamic team at Harvey as a Software Engineer, where you will play a pivotal role in shaping our cutting-edge AI platform. This is an exciting opportunity to work with innovative technologies and contribute to projects that push the boundaries of artificial intelligence.As a Software Engineer, you will collaborate with cross-functional teams to design, develop, and implement software solutions that enhance our AI capabilities. Your work will directly impact the efficiency and effectiveness of our platform, helping to drive our mission forward.
Join shyftlabs as a Sales Engineer specializing in AI and Data Platforms. In this role, you will leverage your technical expertise to support the sales team in delivering innovative solutions to our clients. You will work closely with customers to understand their needs and demonstrate how our products can drive their success. Your ability to communicate complex technical concepts in an understandable way will be key to your success.
Full-time|On-site|Canada, Toronto; USA, Durham; USA, Miami; USA, Palo Alto
About UsNuBank is recognized as one of the largest digital financial platforms globally, serving over 127 million customers across Brazil, Mexico, and Colombia. Guided by our mission to combat complexity and empower individuals, we are transforming financial services in Latin America, marking just the beginning of the vibrant future we are crafting.As a publicly traded company on the New York Stock Exchange (NYSE: NU), we leverage proprietary technology, data intelligence, and an efficient operational model to provide financial products that are straightforward, accessible, and human-centric.Our contributions have been acknowledged in prestigious global rankings, including Time 100 Companies, Fast Company’s Most Innovative Companies, and Forbes World’s Best Bank. Explore more on our career page at NuBank Careers.About the RoleAs a Staff Software Engineer at NuBank, you will play a pivotal role in the AI Private Banker team, our intelligent financial partner designed to assist millions of customers in understanding their finances, making informed decisions, and executing actions effortlessly. You will be responsible for defining the technical architecture that supports conversational AI, recommendation engines, and tailored financial experiences across mobile and web platforms.Your expertise will span the entire stack, from developing Flutter mobile applications that engage millions of users to architecting distributed backend systems that facilitate real-time recommendations and conversational AI interactions. This role is not just about feature development; it’s about establishing a robust engineering foundation that democratizes financial empowerment at scale.Your technical decisions will have a direct impact on how millions of individuals navigate their financial journeys. You will be at the forefront of innovation, merging banking with cutting-edge AI technologies, crafting systems that are both groundbreaking and dependable. Join a team of inquisitive minds reshaping the possibilities in this space.
Tali AI is a rapidly expanding startup in Canada dedicated to enhancing healthcare accessibility through artificial intelligence. We are developing a pioneering AI platform designed to streamline clinical workflows by automating data collection, processing, and sharing. This allows healthcare professionals to concentrate on delivering quality care. Our platform is utilized by thousands of clinicians across Canada and the United States, spanning various specialties and seamlessly integrated with the complex landscape of North American health record systems. Our technical vision is to create a comprehensive operating system for clinical workflows, enabling healthcare teams to customize the platform to automate their data management tasks.We tackle genuinely challenging problems:Establishing real-time audio pipelines where latency impacts outcomes and accuracy is critical.Integrating with a diverse ecosystem of clinical software, including health records, billing, scheduling, referral management, and more.Creating reliable AI workflows and agents in a domain where errors can have significant consequences.Implementing a privacy and compliance framework that is robust by design.We operate in a fast-paced market that values rapid innovation, while building and maintaining clinician trust is paramount.
Role Overview Extreme Networks is hiring a Staff Linux Systems Engineer for the Platform & Infrastructure team in Toronto, Canada. This role focuses on strengthening and evolving the company's Linux-based systems and core infrastructure. What You Will Do Work closely with teams across the organization to design and implement scalable Linux systems Contribute technical expertise to improve reliability, performance, and security of infrastructure Support the development and deployment of solutions that address client requirements What We’re Looking For Extensive experience with Linux systems engineering Strong background in building and maintaining infrastructure at scale Ability to collaborate effectively with cross-functional teams
Join Waabi, an innovative leader in Physical AI, founded by AI pioneer Raquel Urtasun. We are pioneering the future of autonomous transportation, developing cutting-edge technology for commercial autonomous trucks and robotaxis. Our esteemed team collaborates with top-tier partners across AI, automotive, logistics, and deep technology sectors.With offices located in Toronto, San Francisco, Dallas, and Pittsburgh, Waabi is rapidly expanding and is in search of diverse, innovative, and collaborative individuals eager to make a positive impact on the world. For more information, please visit: www.waabi.aiAt the heart of Waabi’s mission is our groundbreaking simulation technology. Our state-of-the-art closed-loop simulation engine, Waabi World, employs the latest generative AI technologies to create a highly realistic environment for training and evaluating self-driving systems. The Simulation Platform team is tasked with delivering a modular, scalable platform that seamlessly integrates the onboard autonomy system, intelligent actor models, and sensor simulation.Key Responsibilities:- Enhance and contribute to Waabi World, our advanced self-driving simulator, to expedite the development of autonomous vehicles.- Optimize the scalability and performance of the simulation platform to facilitate efficient simulations at a massive scale.- Design and develop orchestration software that interconnects simulation subcomponents, including the autonomy system, data-driven actor models, and sensor simulations powered by generative AI.- Engage in technical and architectural discussions, collaborating with researchers and engineers to share insights and innovative ideas.- Collaborate closely with internal users to grasp their needs and workflows, proposing new features to enhance their experience with the simulator.- Promote best practices and set standards for software development.- Lead complex cross-functional technical projects that have a significant impact across Waabi's systems and the organization as a whole.
About Us At Extreme Networks, we are dedicated to crafting seamless networking solutions that enable individuals and organizations to thrive in an increasingly connected world. We are on the lookout for a Director of AI Engineering to spearhead the design, development, and implementation of our cutting-edge AI-native systems.This pivotal role demands a dynamic leader who merges technical expertise with visionary thinking. You will set the strategic direction for AI initiatives, ensuring that innovative concepts transition from research to scalable, production-ready solutions. Your leadership will be crucial in launching enterprise-grade AI applications that revolutionize network design, optimization, security, and support.
About Marble AI Marble AI builds AI-powered tools that help accounting firms and their clients manage taxes more efficiently. The team focuses on simplifying complex tax workflows for mid-to-large-sized firms. As an early-stage company, Marble AI values collaboration and aims to solve high-impact problems for its customers. Role Overview The Platform Engineer will help strengthen and expand the infrastructure behind Marble AI’s tax platform. This position works closely with the engineering team to speed up product development and ensure systems can securely handle sensitive financial data. The role blends infrastructure, DevOps, and Site Reliability Engineering (SRE). Responsibilities include designing and maintaining the platform, improving developer tools, and making sure systems are secure, observable, and resilient. The Platform Engineer will also have a say in the direction of Marble AI’s data platform and security practices across the stack. What You’ll Do Build and manage cloud infrastructure, treating infrastructure as code. Own system reliability, observability (using OTEL), and incident response processes. Design and improve developer tooling and CI/CD pipelines. Work with product and data teams to support scalable data platform needs. Apply and maintain strong security practices, especially for sensitive tax data. Optimize system performance, scalability, and cost across services. Contribute across infrastructure, backend systems, and developer workflows. What We’re Looking For Deep experience with cloud infrastructure (AWS or similar platforms). Hands-on experience with Kubernetes, Terraform, and Helm. Previous work in DevOps, SRE, or platform engineering roles. Experience building and maintaining CI/CD pipelines or developer tooling. Solid understanding of distributed systems, networking, and system reliability. Ability to balance practical solutions with long-term system design. Location Toronto
Cresta is dedicated to transforming customer interactions into a strategic advantage by harnessing the full capabilities of the contact center. Our innovative platform merges cutting-edge AI with human intelligence to provide contact centers with valuable customer insights, streamline automated processes, and empower every team member to enhance productivity and efficiency. Originating from the renowned Stanford AI lab, Cresta boasts co-founder and chairman Sebastian Thrun, the visionary behind Google X, Waymo, and Udacity, alongside CEO Ping Wu, co-founder of Google Contact Center AI, and Tim Shi, an early member of OpenAI.Embark with us on an exciting journey to revolutionize the workforce through AI. The future of work is here, and it is at Cresta.About the Role:We are on the lookout for passionate UX interns to contribute to the evolution of Cresta’s human + AI agents platform, which facilitates AI-driven business workflows and automation for contact center operators. Your role will involve designing exceptional end-user experiences and the tools that enable these interactions.As a product design intern at Cresta, you will be at the forefront of a significant technological advancement with LLM and Generative AI. This unique opportunity allows you to make a meaningful impact on how individuals interact with AI. Under the guidance of a design supervisor, you will engage in the complete UX process, including product design, user research, and visual design components, to deliver outstanding user experiences.Responsibilities:Generate innovative UX concepts for Conversation Intelligence and AI Agents, taking into account diverse user personas.Analyze user and business challenges, propose viable solutions, and collaborate with engineers to implement designs.Conduct user research, including usability testing, competitive analysis, customer interviews, behavioral insights, and metrics analysis to inform design decisions.Rapidly prototype (using tools like Figma, Cursor, Claude, Voice API, etc.) to explore various UI paradigms.Work closely with Engineers, PMs, ML experts, Customer Success, and Sales teams throughout the organization.Effectively present and communicate design concepts to supervisors and cross-functional teams.
Dec 24, 2025
Sign in to browse more jobs
Create account — see all 1,195 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.