Technical Staff Member It Engineering jobs in New York – Browse 4,415 openings on RoboApply Jobs

Technical Staff Member It Engineering jobs in New York

Open roles matching “Technical Staff Member It Engineering” with location signals for New York. 4,415 active listings on RoboApply Jobs.

4,415 jobs found

1 - 20 of 4,415 Jobs
Apply
Vibecode logoVibecode logo
Full-time|On-site|New York City

Vibecode is on a mission to make artificial intelligence accessible to 100 million more people. The team is dedicated to using advanced technology to inspire and educate, with a focus on the possibilities of AI-assisted coding. Role overview This New York City-based role centers on exploring how AI tools can influence the future of coding and design. The wor…

Apr 21, 2026
Apply
Valkai logoValkai logo
Full-time|On-site|New York City

THE ROLE:As a pivotal member of our team, you will transform the interaction dynamics between life sciences organizations and agents. Your creativity will help craft innovative experiences that seamlessly integrate into daily workflows. You will engage with cutting-edge models and develop interfaces that empower users to harness their capabilities effectively.WHAT WE ARE LOOKING FOR:Analytical Thinkers: You possess a comprehensive approach to problem-solving, effectively balancing business and technical considerations.Technology Agnostic: You appreciate technology for its practicalities and are open to learning and adapting to new tools and ideas.Self-Motivated and Organized: You thrive in situations of ambiguity and take initiative to create meaningful impact.Product Ownership: You take responsibility for the quality and usability of the products you create, demonstrating pride in your work.Determined: You are relentless in pursuing goals, overcoming obstacles with perseverance.YOUR KEY RESPONSIBILITIES:Manage retrieval of vast document scales with diverse formats (graphs, tables, diagrams, etc.)Develop organizational interfaces to enhance collaboration with AI, integrating into everyday workflowsDesign and implement strategic planning tools involving human oversightWork directly with enterprise clients to create product experiences that automate routine tasksAssess large language models across various life science tasksEnsure government-level data security for sensitive informationConduct internet-scale data collectionDESIRABLE QUALIFICATIONS:Experience in a rapidly growing early-stage startup environment.Background in building agents or fine-tuning language models for complex tasks.Experience in developing large-scale content indexing systems.Familiarity with API development and integrations.

Feb 7, 2026
Apply
reflectionai logoreflectionai logo
Full-time|On-site|New York

Role Overview reflectionai is seeking a Technical Staff Member for the IT Engineering team in New York. This role focuses on designing, building, and maintaining IT infrastructure that supports the company’s work in artificial intelligence. What You Will Do Work closely with team members to develop and support IT systems Contribute to the design and implementation of infrastructure projects Help maintain reliable, scalable technology to back AI initiatives About the Team Join a group of engineers committed to advancing reflectionai’s technology. Collaboration and hands-on problem-solving are at the core of this team’s approach.

Apr 16, 2026
Apply
Anterior logoAnterior logo
Full-time|On-site|New York

About AnteriorAt Anterior, we are dedicated to revolutionizing the landscape of healthcare administration. Our mission is to ensure that healthcare operations are seamless and virtually invisible, enabling clinicians to concentrate on what truly matters - patient care. Our AI-driven platform, meticulously crafted by clinicians for clinicians, streamlines administrative tasks and enhances patient outcomes. By merging clinical knowledge with innovative technology, we are at the forefront of transforming healthcare operations through responsible AI.We are experiencing rapid growth, backed by prestigious investors such as Sequoia, New Enterprise Associates (NEA), and Neo, alongside notable angel investors including the founders of DeepMind, Google AI, and Inflection AI.At Anterior, we are committed to enhancing healthcare experiences for everyone. We strive to leverage technology in ways that make clinical decision-making as frictionless as possible.Our innovative technology is designed for large-scale clients, built on the principles of simplicity and robust foundations, ensuring that all aspects follow suit.The TeamOur engineering team at Anterior is characterized by a strong sense of product ownership and a commitment to solving significant challenges that simplify healthcare workflows. We work collaboratively in multidisciplinary teams to deliver secure, scalable, and user-centric solutions, adhering to best practices and modular design principles.We value diverse experiences and skill sets, continuously expanding our team to creatively and effectively address the challenges posed by modern healthcare.About YouYou possess a deep commitment to excellence, advocating for doing things The Right Way. You believe in the principle of Less is More and focus on the foundational elements of system design and architecture, such as pacelc and consistency guarantees. Your passion extends beyond the code you write; you are weary of over-engineered solutions and prioritize the fundamentals. With experience from dealing with complex and ineffective systems, you strive to create resilient, enduring code.We have openings across the stack, from infrastructure to frontend development. We invite you to share your proudest firefighting story during the interview, bring along your GitHub portfolio, and let's explore if we are a good match. If you are a junior developer with a passion for coding and innovation, we encourage you to apply. Share some of your work with us, and let's connect!About the RoleAs a Member of the Technical Staff at Anterior, you will take ownership of problems end-to-end, from system design to production. Your role will span across the stack (with a focus on backend), developing the core platform that facilitates clinical decision-making at scale. Daily tasks may involve:Collaborating with multidisciplinary teams to develop and deploy innovative solutions.Designing scalable systems that enhance operational efficiency in healthcare.Ensuring code quality through best practices and extensive testing.

Jan 28, 2025
Apply
Modal logoModal logo
FullTime|On-site|New York

About Us:At Modal, we empower AI teams with an unparalleled infrastructure foundation. Our platform offers instant access to GPUs, rapid container startups, and integrated storage solutions, making it seamless to train models, execute batch jobs, and deliver low-latency inference. Leading companies like Suno, Lovable, and Substack trust Modal to transition from prototypes to full-scale productions without the complexity of infrastructure management.As a dynamic and rapidly expanding team based in NYC, San Francisco, and Stockholm, we have achieved a remarkable 9-figure Annual Recurring Revenue (ARR) and recently secured a Series B funding at a valuation of $1.1 billion. Our growing customer base includes thousands who rely on us for their production AI workloads.Joining Modal means becoming part of one of the most innovative AI infrastructure companies at a pivotal moment in its growth, with abundant opportunities for personal and professional development. Our talented team includes creators of renowned open-source projects (like Seaborn and Luigi), accomplished researchers, competition medalists, and seasoned engineering and product leaders.The Role:In your role at Modal, you will play a crucial part in enhancing our cloud services, upon which our customers depend for their essential production systems. As we rapidly expand our cloud infrastructure offerings, we aim to significantly bolster our reliability while scaling our platform and customer base.We are looking for individuals who are deep systems thinkers with a passion for achieving high reliability. As our inaugural reliability-focused team member, you will have the unique opportunity to shape our reliability systems and practices and collaborate closely with our development teams.Identify and implement architectural enhancements to boost reliability, performance, and availability.Cultivate a culture of reliability throughout Modal’s engineering teams.Design and establish essential operational processes including deployments, upgrades, rollbacks, and postmortem reviews.Become an integral part of our core engineering team by participating in on-call rotations, addressing production incidents as they arise.Develop monitoring systems that guarantee superior service quality for our customers.Troubleshoot and resolve production issues across all service layers.

Jan 19, 2026
Apply
Sciemo logoSciemo logo
Full-time|Hybrid|New York City

About SciemoSciemo is revolutionizing the consumer goods industry through advanced AI technology that empowers businesses to make quicker, smarter, and more human decisions throughout the Integrated Business Planning (IBP) process. Our platform turns chaotic, fragmented data into actionable insights, aiding decision-makers in real-time by simplifying complexity.OverviewAs a pioneering startup, Sciemo specializes in AI solutions tailored for consumer brands. We harness machine learning, generative AI, agent-based systems, and graph technologies to deliver insights in seconds and tangible business outcomes in minutes.Your RoleWe are on the lookout for a Founding Member of Technical Staff, who will serve as both a Data Scientist and a Machine Learning Engineer. In this critical position, you will be instrumental in the design, development, and deployment of the intelligence that powers our AI products. You’ll engage in a wide range of applied AI efforts, from data science and machine learning to large-scale production engineering. This dual role demands not only advanced model development expertise but also the engineering acumen to implement and sustain robust, scalable systems.You will work closely with data engineers, product leads, backend engineers, and customer-facing teams to ensure our AI systems provide measurable value in practical scenarios. As one of our initial technical team members, you will help shape our AI strategy, establish technical standards, and set best practices for scalable applied AI.Key ResponsibilitiesDevelop and Deploy AI Systems:Architect, build, and deploy ML/GenAI products on cloud infrastructure (AWS or comparable).Design and implement comprehensive AI workflows: data ingestion, feature engineering, modeling, evaluation, and deployment.Establish automated pipelines for continuous learning, model promotion, and performance monitoring.System Architecture & Reliability:Lead the design of ML orchestration frameworks (such as Airflow, Kedro, ZenML, Flyte) to guarantee reproducibility and system reliability.

Oct 10, 2025
Apply
Deeptune logoDeeptune logo
Full-time|$200K/yr - $300K/yr|On-site|New York City

Join Deeptune’s Innovative TeamAt Deeptune, we are pioneering the development of high-fidelity training environments for AI agents, enabling them to learn real-world tasks through advanced reinforcement learning techniques. Collaborating with leading AI laboratories, we are at the forefront of training the next generation of agentic models, contributing to significant advancements in computer operations, code generation, and complex task execution.Our close-knit team of around 20 talented engineers and operators, hailing from prestigious organizations such as Anthropic, Scale AI, Palantir, Hebbia, Glean, and Retool, is based in New York City. Recently, we successfully secured $43 million in Series A funding, led by Andreessen Horowitz, with contributions from 776, Abstract Ventures, and Inspired Capital.If you are eager to tackle some of the toughest challenges at the intersection of reinforcement learning and practical AI implementation, we would love to connect with you!Learn more: Fortune Exclusive on Our Series A | a16z: Why We’re Investing in Deeptune

Mar 22, 2026
Apply
Reflection AI logoReflection AI logo
Full-time|On-site|New York

Our MissionAt Reflection AI, our mission is to develop open superintelligence and make it available to everyone.We are creating open weight models that cater to individuals, agents, enterprises, and even nations. Our skilled team of AI researchers and innovators hails from leading organizations such as DeepMind, OpenAI, Google Brain, Meta, Character.AI, and Anthropic.About the RoleThe Compute Platform team at Reflection AI focuses on ensuring our compute layer is robust and highly available. Our K8s-based platform spans multiple neo-clouds, tackling complex systems challenges related to multi-cloud scheduling, node health, and performance debugging. You will collaborate closely with our training teams to design strategies for fault tolerance, health checks, and remediation processes.Key ResponsibilitiesCluster Management: Develop and maintain tools for automatic remediation, topology-aware scheduling, capacity planning, and expedited hardware debugging.Platform Engineering: Design and refine our cluster management stack to efficiently handle workloads across extensive multi-GPU fleets.Monitoring & Observability: Establish an all-encompassing monitoring system for the cluster, emphasizing durability and active performance benchmarking.Roadmap Execution: Prepare the infrastructure for next-gen GPU deployments and larger cluster sizes. In the long run, you will contribute to managing multi-cloud storage, petabyte-scale data replication, and optimizing GPU-to-GPU network performance.

Mar 20, 2026
Apply
Reflection AI logoReflection AI logo
Full-time|On-site|NYC

Our MissionAt Reflection AI, our mission is to create open superintelligence and ensure its accessibility for everyone.We are crafting open weight models for individuals, organizations, and even nations. Our talented team comprises AI researchers and entrepreneurs hailing from renowned institutions such as DeepMind, OpenAI, Google Brain, Meta, Character.AI, Anthropic, and more.FoundationsVision:Establish and maintain a comprehensive company-wide foundations platform that empowers every team by delivering dependable, scalable developer infrastructure, Site Reliability Engineering (SRE) capabilities, and high-throughput data ingestion tools, enabling Reflection to accelerate as we grow.What This Team DoesWe are responsible for developing and managing the essential data systems and pipelines that fuel our research, training, and production environments. This platform facilitates rapid experimentation, reliable model development, and scalable production workflows by integrating ingestion, processing, and orchestration throughout the data lifecycle.Design ingestion and orchestration patterns for both batch and streaming data workloads.Construct scalable compute and storage foundations (formats, engines, runtimes) that support extensive data processing.Guarantee reproducible pipelines through versioning, backfills, and isolated execution environments.Deliver trusted data quality, lineage, and governance signals to empower teams in making informed production decisions.Sustain predictable cost and performance through established guardrails, budgets, and ongoing system optimization.Facilitate a unified data layer that supports research, training, and production across the model development lifecycle.About the RoleYou will play a pivotal role in constructing the core data systems and pipelines that drive our research, training, and production environments. Your responsibilities will include designing and implementing reliable, scalable ingestion and orchestration patterns for batch and streaming workloads, developing storage and compute foundations that enable reproducible experimentation and rapid iteration, and establishing data quality and governance standards that teams can rely on for production decisions. You will also provide the foundational data layer that unifies ingestion, processing, and workflow management throughout model development.

Mar 12, 2026
Apply
Reflection AI logoReflection AI logo
Full-time|On-site|NYC

Our MissionAt Reflection AI, our mission is to develop open superintelligence accessible to everyone.We are creating open weight models for individuals, enterprises, agents, and nation-states. Our diverse team includes AI researchers and industry veterans from leading organizations such as DeepMind, OpenAI, Google Brain, Meta, Character.AI, and Anthropic.FoundationsVision: We aim to establish and maintain a company-wide foundational platform designed to enhance every team by offering dependable, scalable development infrastructure, Site Reliability Engineering (SRE) capabilities, and high-throughput data ingestion tools. This enables Reflection to expedite its growth as we scale.What This Team DoesOur team builds and manages the essential platform layer that every engineering team relies on. We craft clear, opinionated golden paths for cloud infrastructure, networking, and access patterns, ensuring that the safest and most scalable routes are also the easiest to navigate. We work across various domains including cloud infrastructure, network primitives, security protocols, and developer tools to empower engineers to deploy rapidly while ensuring reliability, security, and cost predictability.Define and uphold golden paths for cloud projects and access patterns.Implement secure defaults and guardrails, making the safe path the most straightforward option.Construct and manage core networking components such as routing, DNS, load balancing, and connectivity.Establish baseline observability and operational best practices across services and infrastructure.Oversee developer infrastructure including build systems, CI/CD, caching, and testing infrastructure to facilitate rapid shipping.Enhance cloud cost visibility and reduction through effective tooling, guardrails, and optimization strategies without hindering team progress.What You'll Work WithCloud & Multi-Cloud Architecture: GCP, AWS, multi-cloud configurations, VPC architecture, private DNS, load balancers, advanced routing (BGP), and large-scale Kubernetes clusters.Identity & Security Foundations: IAM, OIDC, RBAC, KMS, secrets management, policy-as-code (OPA, Gatekeeper), secure defaults, and cloud guardrails.Infrastructure & Automation: Terraform, Pulumi, foundational Kubernetes, CI/CD systems, Bazel, build caching, and hermetic builds.Observability & Cost Governance: Prometheus, Grafana, OpenTelemetry, and more.

Mar 12, 2026
Apply
anterior logoanterior logo
Full-time|Remote|New York

Role overview anterior seeks a Senior Member of Technical Staff with a strong technical background and a practical mindset. This is a fully remote role. The team emphasizes collaboration and creative approaches to overcome challenges and keep projects on track. What you will do Collaborate with engineers and other colleagues to design and develop new technologies Help implement systems that shape the future of anterior’s products Use technical expertise to tackle complex problems and deliver outcomes Remote work anterior is headquartered in New York, but this position is open to candidates working from any location.

Apr 28, 2026
Apply
Reflection AI logoReflection AI logo
Full-time|On-site|NYC

Our VisionAt Reflection AI, we are dedicated to the mission of creating open superintelligence and ensuring it is accessible to everyone. Our team is committed to developing open weight models that cater to individuals, agents, enterprises, and even nation-states. Our diverse group of AI researchers and innovators hail from esteemed organizations such as DeepMind, OpenAI, Google Brain, Meta, Character.AI, and Anthropic.Foundational ObjectivesVision:We aim to establish and manage a company-wide foundational platform that enhances every team’s productivity by delivering dependable, scalable developer infrastructure, site reliability engineering capabilities, and high-throughput data ingestion tools—empowering Reflection AI to progress swiftly as we expand.Team ResponsibilitiesOur team is responsible for constructing and managing essential shared services that fuel our research, training, and production environments. These systems form the backbone that supports various teams in model development, deployment, and evaluation, integrating data, compute, and workflow management while facilitating rapid experimentation and robust production systems.Design and manage shared services that multiple teams utilize across research and production workflows.Establish and maintain reliability targets through Service Level Indicators (SLIs), Service Level Objectives (SLOs), and effective on-call practices.Ensure operational readiness through comprehensive runbooks, incident playbooks, and capacity planning.Guarantee correctness and performance under load, addressing issues like consistency, tail latency, and failure modes.Create APIs, SDKs, and internal platforms that support high-velocity experimentation and iteration.Minimize operational burden through enhanced tooling, standardization, and scalable platform patterns across teams.Technologies You'll Engage WithContainer Abstractions: Containers-as-a-Service, Kubernetes abstraction layers, container orchestration, reproducible environments, multi-tenant isolation.Distributed Systems Architecture: Sharding, replication, coordination services, high-concurrency systems, concurrency control.Service Development Stack: gRPC, Protobuf, Go, Rust, C++.Reliability & Performance: Idempotency, retries, backpressure, SLI/SLO design, tail latency optimization, service reliability engineering.Your ProfileWe are looking for a talented individual with a solid background in distributed systems and a passion for building scalable solutions.

Mar 12, 2026
Apply
Modal logoModal logo
FullTime|On-site|New York

About Us:At Modal, we empower AI teams with robust infrastructure solutions. Our platform offers instant access to GPUs, rapid container startups, and integrated storage, enabling seamless model training, batch processing, and low-latency inference. Leading companies such as Suno, Lovable, and Substack trust Modal to transition from prototype to production effortlessly, without the complexities of infrastructure management.As a rapidly expanding organization based in New York, San Francisco, and Stockholm, we have achieved a remarkable nine-figure ARR and recently secured a Series B funding round at a valuation of $1.1 billion. Our extensive customer base, which includes industry leaders like Lovable, Scale AI, Substack, and Suno, depends on us for their production AI workloads.Joining Modal means becoming part of one of the fastest-growing AI infrastructure companies at a pivotal stage, with numerous opportunities for personal and professional growth. Our team is composed of the creators of renowned open-source projects (e.g., Seaborn, Luigi), academic scholars, international competition medalists, and seasoned engineering and product leaders with extensive experience.Your Role:We are seeking talented engineers who have a proven track record in optimizing machine learning systems for performance at scale. If you are passionate about contributing to open-source initiatives and enhancing Modal’s container runtime to improve language and diffusion models for higher throughput and reduced latency, we want to hear from you!Key Qualifications:A minimum of 5 years of experience in crafting high-quality, high-performance code.Proficient in working with PyTorch, high-level ML frameworks, and inference engines (e.g., vLLM or TensorRT).Strong understanding of Nvidia GPU architecture and CUDA programming.Hands-on experience in machine learning performance engineering, including optimizing GPU performance through techniques such as debugging SM occupancy, rewriting algorithms for compute efficiency, and minimizing host overhead.Familiarity with low-level operating system principles (Linux kernel, file systems, containers) is a plus.Willingness to work in-person at our offices in New York, San Francisco, or Stockholm.

Dec 18, 2024
Apply
Cohere logoCohere logo
FullTime|On-site|New York

About UsAt Cohere, we're on a mission to harness and scale intelligence for the betterment of humanity. We specialize in training and deploying cutting-edge models for developers and enterprises, enabling them to create transformative AI experiences such as content generation, semantic search, retrieval-augmented generation (RAG), and intelligent agents. We are committed to fostering the widespread adoption of AI technologies.Our passion for excellence drives us to continuously enhance our models and the value they deliver to our clients. We thrive in a fast-paced environment where hard work and innovation are paramount to achieving the best outcomes for our users.Cohere is made up of a diverse team of researchers, engineers, designers, and more, all of whom are leaders in their fields. We believe that a variety of perspectives is essential to creating outstanding products.Join us in our quest to shape the future!Role OverviewOur rapidly expanding team of researchers and engineers is dedicated to building robust machine learning systems and enhancing the efficiency of large language model (LLM) inference. We focus on developing innovative techniques that improve production execution of models, resulting in reduced latency, increased throughput, and consistent performance across various workloads.As a member of this team, you will engage with the inference stack to enhance key performance metrics by analyzing model execution, identifying performance bottlenecks, and crafting pioneering optimizations. You will work closely with both modeling and systems teams to test, measure, and implement enhancements that significantly improve inference speed. As the team progresses, you will have the chance to deepen your knowledge in advanced performance techniques, including GPU/CUDA optimizations, kernel-level enhancements, and execution strategies for mixture of experts (MoE) and large-scale architectures.Note: Cohere has offices in Toronto, Montreal, San Francisco, New York, Paris, Seoul, and London. We embrace a remote-friendly approach, strategically distributing teams based on interests, expertise, and time zones to enhance collaboration and flexibility. The Model Efficiency team is primarily based in the EST and PST time zones.Who You Are5+ years of experience coding high-performance, production-level softwareProficient in C++ or Python (experience in Rust/Go is also a plus)Strong understanding of machine learning concepts and frameworksExperience in optimizing ML systems for production environmentsExcellent problem-solving skills and ability to collaborate effectively in a team setting

Nov 7, 2025
Apply
Fluidstack logoFluidstack logo
Full-time|$90K/yr - $150K/yr|On-site|New York, NY

About FluidstackAt Fluidstack, we are at the forefront of creating infrastructure for advanced intelligence. We collaborate with leading AI laboratories, governments, and enterprises, including Mistral, Poolside, Black Forest Labs, Meta, and others to facilitate computing capabilities at unprecedented speeds.Our commitment to realizing Artificial General Intelligence (AGI) is driven by urgency and excellence. Our dedicated team is passionate about delivering top-tier infrastructure, considering our customers' success as our own. If you are driven by purpose, have a relentless pursuit of excellence, and are prepared to work diligently to propel the future of intelligence, we invite you to help us shape what comes next.About the RoleAs a Member of Technical Staff in Corporate IT, you will engage in managing identities, devices, and SaaS platforms while providing support to end users. The ideal candidate will be organized, collaborative, and eager to learn within established frameworks.Deliver hands-on IT support to our in-office and remote employees by effectively resolving Help Desk tickets with a focus on exceptional customer service.Manage Fluidstack’s Mobile Device Management (MDM) fleet, which includes macOS, iOS, and Windows devices.Provide in-office support through troubleshooting basic network issues along with A/V systems.Assist in managing and optimizing Okta and our key SaaS tools by overseeing integrations and role-based access controls.Support daily procurement and asset management by coordinating hardware purchases, tracking assets, and maintaining accurate inventory throughout their lifecycle from acquisition to decommissioning.Help keep IT documentation clear, organized, and accessible by developing user-friendly technical documents for both technical and non-technical audiences.Contribute to IT projects by assisting in execution, prioritizing tasks, and ensuring transparent communication with stakeholders through regular updates.Maintain a secure environment by supporting essential security practices, including access reviews and basic identity security tasks.Assist in the onboarding and offboarding processes for Fluidstack employees and contractors.Support the IT team by creating straightforward, reliable automations that minimize manual tasks and enhance operational efficiency.

Nov 24, 2025
Apply
Eagle logoEagle logo
Full-time|$150K/yr - $300K/yr|On-site|New York City

Join Eagle as a Member of Technical Staff!Eagle is revolutionizing the built environment, akin to Pixar Studios, by integrating advanced AI into civil, structural, and MEP engineering. With the support of Lightspeed Venture Partners, we acquire and transform engineering firms to tackle some of the world’s most pressing infrastructure, energy, and climate challenges.Our mission is to equip engineers with cutting-edge tools that enhance their capabilities and streamline their processes. By leveraging AI, we aim to close the automation gap in engineering, where currently less than 5% of tasks have been automated.As a Member of Technical Staff, you will collaborate closely with our founders to develop Cortex, our innovative data platform. You will work alongside experienced engineers to create systems that make institutional knowledge accessible, automate quality assurance, and produce engineering deliverables efficiently.Your Responsibilities:Develop and implement core components of our platform, including data connectors and document intelligence for AEC artifacts.Lead the creation of systems across LLM pipelines, document parsing, and backend infrastructure.Collaborate with design partners, delivering your work for immediate use by engineers.Research and apply emerging AI frameworks and methodologies.Contribute to fostering a positive engineering culture.

Apr 2, 2026
Apply
Chakra Labs logo
Full-time|On-site|Brooklyn

About UsAt Chakra Labs, we are dedicated to transforming human taste into a sophisticated intelligence framework. Our focus is on creating high-fidelity trajectories and environments that push the boundaries of AI research, collaborating with leading laboratories in the field.Why Join Us?Engage with cutting-edge technologies spanning data, AI, and infrastructure.Receive mentorship from a highly qualified and experienced team.Enjoy a competitive salary package, comprehensive benefits, and a daily meal stipend of $30, alongside an attractive equity offering.Considerations Before ApplyingThis is our passion. While we value fun, we are dedicated to hard work. If you prefer a standard 9-5 role, this may not be the right environment for you.Uncertainty is part of the journey. The landscapes of data and AI are ever-evolving. While we have a defined mission, those uncomfortable with ambiguity may find this role challenging.Your ProfileAdvanced Knowledge: You hold a Master's or PhD in machine learning, with a solid grasp of linear algebra and stochastic gradient descent, capable of deriving linear regression manually.Post-Training & Environments: You’re adept at executing SFT on distributed GPU clusters utilizing FSDP. You possess a keen interest in enhancing existing models with high-quality data and can conceptualize new environments and online reinforcement learning (GRPO, PPO).Proficient in Python: Familiarity with FastAPI and PyTorch is essential, including REST APIs and backend development, along with the ability to abstract high-level concepts in Torch using lower-level primitives.Experience: We seek approximately 3-5 years of relevant experience, though we value skills and alignment with our expectations over strict qualifications.

May 6, 2025
Apply
Fireworks AI logoFireworks AI logo
Full-time|On-site|New York, NY; San Mateo, CA

Join Fireworks AI as a Technical Staff Member specializing in Cloud Infrastructure. In this pivotal role, you will be at the forefront of our innovative cloud solutions, collaborating with a dynamic team of professionals dedicated to advancing cloud technologies. Your expertise will contribute to building and maintaining robust infrastructure that supports our evolving business needs.

May 1, 2026
Apply
xai logoxai logo
Full-time|On-site|New York, NY; Palo Alto, CA

About the Role X Money, part of xai, is hiring a Technical Staff Member in New York, NY or Palo Alto, CA. This position works closely with engineers and researchers to build new technology for the financial sector. The team focuses on improving user experiences and developing new solutions for financial products.

Apr 14, 2026
Apply
Vibecode logoVibecode logo
Full-time|On-site|New York City

Vibecode aims to make AI-assisted coding accessible to millions more people. The team, based in New York City, explores ways to inspire a broad audience through technology and creative outreach. Role overview This Technical Staff Member position centers on Growth & Content. The role brings together research, content creation, and practical work with the latest AI technologies. Projects often involve both technical and creative problem-solving, with a focus on sharing insights and fostering community engagement. What you will do Research new technologies and evaluate their real-world applications Test and refine early-access AI models and runtimes to enhance performance Create in-depth educational and inspirational content for platforms like YouTube, X, Instagram, and LinkedIn Develop original projects that blend multiple AI tools in inventive ways Use a strong visual approach to ensure work is distinctive and polished Location This role is based in New York City.

Apr 21, 2026

Sign in to browse more jobs

Create account — see all 4,415 results

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.