Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Unlock Your Potential
Generate Job-Optimized Resume
One Click And Our AI Optimizes Your Resume to Match The Job Description.
Is Your Resume Optimized For This Role?
Find Out If You're Highlighting The Right Skills And Fix What's Missing
Experience Level
Experience
Qualifications
About the RoleAs a Software Engineer, you will play a crucial role in constructing the infrastructure that facilitates high-quality and rapid research. You will be responsible for developing and maintaining pipelines that handle extensive EEG datasets, as well as creating the infrastructure that supports our experimental setups and data analyses. This position offers the opportunity to contribute to groundbreaking neural decoding technologies while collaborating with leading researchers in the field of brain-computer interfaces. You might be a good fit if youPossess over 5 years of hands-on software engineering experience, with a strong command of Python and systems-level architecture. Have experience in building and sustaining high-performance data pipelines for the processing and storage of large datasets. Are adept at creating the infrastructure and front-end tools necessary for real-time visualization of experimental stimuli. Have a solid background in managing comprehensive software stacks, cloud infrastructure, and dependency management. Are skilled in debugging low-level system performance issues or architecting new foundational services across the full tech stack. Take pride in owning the complete technical lifecycle of a project, from its initial design to production deployment and ongoing maintenance. Strong candidates may haveExperience in developing internal libraries, SDKs, or tools that enhance machine learning and computational neuroscience workflows. A history of building core Python infrastructure such as interpreters, package ecosystems, typing tools, or maintaining widely-adopted open-source Python development tools. Expertise in optimizing Python performance, including familiarity with Cython, profiling tools, or enhancing Python's computational efficiency.
About the job
About alljoined
At alljoined, we are revolutionizing the way humans interact with technology. Our mission is to eliminate the communication barriers between the brain and machines through innovative, non-invasive methods. By leveraging cutting-edge deep learning techniques on large-scale EEG datasets collected from cost-effective devices, we aim to decode visual information, text, and eventually internal thoughts. We pride ourselves on our advanced capabilities and complete vertical integration, striving to create a transformative consumer interface that will redefine experiences at home and in the workplace.
We are in search of passionate and talented individuals to expand our elite team of researchers dedicated to building the next generation of interfaces that enhance individual lives and promote societal well-being.
About alljoined
At alljoined, we are at the forefront of bridging the gap between human cognition and technology. Our non-invasive techniques utilize deep learning to interpret brain signals, aiming to unlock new dimensions of human-machine interaction. Join us as we build a future where technology seamlessly integrates with our thoughts and experiences.
Similar jobs
1 - 20 of 10,323 Jobs
Search for Machine Learning Researcher At Alljoined San Francisco
Full-time|$140K/yr - $250K/yr|On-site|San Francisco
About AlljoinedAt Alljoined, we are pioneering the future of communication between humans and technology by developing non-invasive methods to decode brain activity. By leveraging cutting-edge deep learning techniques on extensive EEG datasets collected through cost-effective hardware, we aim to interpret images, text, and video, with a long-term vision of understanding internal thoughts. Our capabilities are industry-leading, and we are fully vertically integrated. Our mission is to create a universal consumer interface that revolutionizes daily interactions both at home and in the workplace.We are on the lookout for exceptional researchers to expand our elite team, dedicated to creating the next transformative interface that enhances individual lives and contributes positively to society.About the RoleWe invite you to apply for the position of Machine Learning Researcher within our core R&D team. In this role, you will be responsible for conceptualizing and executing advanced machine learning models for EEG-based neural decoding, disseminating impactful research, and establishing the foundational infrastructure for our brain decoding systems. You'll collaborate with top-tier experts in neural decoding and AI, driving innovation in brain-computer interfaces.Key ResponsibilitiesResearch & Model Development:Craft, train, and enhance state-of-the-art deep learning models for neural decoding, utilizing the latest advancements in machine learning architectures such as transformers and diffusion models.Investigate innovative methodologies for modeling high-frequency time-series EEG datasets alongside various other data modalities.Convert research findings into production-ready code that seamlessly integrates with our proprietary brain-computer interface stack.Collaboration & Publication:Work in tandem with a multidisciplinary team of neuroscientists and ML engineers to develop scalable, end-to-end neural decoding solutions.Publish research outcomes in leading ML and AI conferences such as NeurIPS, ICML, ICLR, and CVPR, and actively engage in open-source communities as appropriate.
Full-time|$120K/yr - $160K/yr|On-site|San Francisco
About alljoinedAt alljoined, we are revolutionizing the way humans interact with technology. Our mission is to eliminate the communication barriers between the brain and machines through innovative, non-invasive methods. By leveraging cutting-edge deep learning techniques on large-scale EEG datasets collected from cost-effective devices, we aim to decode visual information, text, and eventually internal thoughts. We pride ourselves on our advanced capabilities and complete vertical integration, striving to create a transformative consumer interface that will redefine experiences at home and in the workplace.We are in search of passionate and talented individuals to expand our elite team of researchers dedicated to building the next generation of interfaces that enhance individual lives and promote societal well-being.
Full-time|$120K/yr - $160K/yr|On-site|San Francisco
About AlljoinedAlljoined is dedicated to revolutionizing the interaction between humans and technology by developing non-invasive methods to decode thoughts from the brain. Utilizing cutting-edge deep learning techniques on extensive EEG datasets gathered using cost-effective hardware, we initially decode images, text, and video, with aspirations to eventually interpret internal thoughts. Our capabilities are at the forefront of technology, and we are fully vertically integrated. Our mission is to create a comprehensive consumer interface that will radically enhance our daily interactions at home and in the workplace.We are in the process of expanding our elite team of researchers, aiming to create the next interface that will not only improve individual lives but also contribute positively to society.About the RoleIn the role of Computational Neuroscientist, you will be instrumental in propelling our research initiatives forward. Through the application of advanced computational and neuroimaging methodologies, you will design and refine stimulus paradigms in collaboration with our research team, conducting in-depth analyses of collected data to extract insights across various paradigms and architectures. You will collaborate with a dedicated team of research coordinators who will facilitate large-scale data collection, alongside engineers and researchers to devise experiments aimed at optimizing next-generation neural decoding techniques.About YouPhD or postdoctoral experience in neuroscience or a related discipline.Substantial experience in the design and execution of neuroscience studies, with a focus on functional brain imaging data analysis (EEG, fMRI).Strong proficiency in Python, particularly with libraries such as MNE, Numpy, PsychoPy, and machine learning frameworks including PyTorch, SciPy, and Scikit-Learn.Extensive background in utilizing advanced computational, statistical, and modeling techniques to analyze and interpret neuroimaging data.A solid understanding of psychophysics and human behavior.Proven ability to conduct independent research while collaborating effectively in an interdisciplinary environment.
At Causal Labs, we are on a groundbreaking mission to develop general causal intelligence—artificial intelligence that not only predicts future events but also determines the most effective actions to influence those outcomes.To achieve this monumental goal, we are constructing a Large Physics Foundation Model (LPM). Our focus is on domains governed by physical laws, which inherently exhibit cause-and-effect relationships, setting them apart from traditional visual or textual data.Weather serves as the ideal training environment for our LPM, being one of the most extensively observed physical systems available. It provides immediate, objective feedback from sensory observations and boasts data scales significantly larger than those currently employed to train existing language models.Our team at Causal Labs includes leading researchers and engineers with backgrounds in self-driving technology, drug discovery, and robotics, hailing from prestigious organizations such as Google DeepMind, Cruise, Waymo, Meta, Nabla Bio, and Apple. We firmly believe that achieving general causal intelligence will represent one of the most critical technological advancements for our civilization.We are seeking innovative researchers eager to confront unsolved challenges in the field.This role presents an opportunity to create powerful models rooted in observable feedback and verifiable ground truths. If you possess experience in pioneering research and training large-scale models from the ground up in areas such as language and vision models, robotics, or biology, we invite you to join our mission.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our ambition is to enhance human potential by advancing collaborative general intelligence. We envision a future where individuals have the tools and knowledge to harness AI for their distinct requirements and aspirations.Our team comprises dedicated scientists, engineers, and innovators who have contributed to some of the most renowned AI products, including ChatGPT and Character.ai, along with open-weight models like Mistral, and influential open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleWe are seeking an Infrastructure Research Engineer to architect, optimize, and sustain the computational frameworks that facilitate large-scale language model training. You will create high-performance machine learning kernels (e.g., CUDA, CuTe, Triton), enable effective low-precision arithmetic operations, and enhance the distributed computing infrastructure essential for training expansive models.This position is ideal for an engineer who thrives in close collaboration with hardware and research disciplines. You will partner with researchers and systems architects to merge algorithmic design with hardware efficiency. Your responsibilities will include prototyping new kernel implementations, evaluating performance across various hardware generations, and helping to establish the numerical and parallelism strategies crucial for scaling next-generation AI systems.Note: This is an evergreen role that remains open continuously for expressions of interest. We receive numerous applications, and there may not always be an immediate opportunity that aligns with your qualifications. However, we encourage you to apply, as we regularly assess applications and will reach out as new positions become available. You are also welcome to reapply after gaining additional experience, but please refrain from applying more than once every six months. Additionally, you may notice postings for specific roles catering to particular projects or team needs. In such cases, you are encouraged to apply directly alongside this evergreen listing.What You’ll DoDesign and develop custom ML kernels (e.g., CUDA, CuTe, Triton) for key LLM operations such as attention, matrix multiplication, gating, and normalization, optimized for contemporary GPU and accelerator architectures.Conceptualize compute primitives aimed at alleviating memory bandwidth bottlenecks and enhancing kernel compute efficiency.Collaborate with research teams to synchronize kernel-level optimizations with model architecture and algorithmic objectives.Create and maintain a library of reusable kernels and performance benchmarks that serve as the foundation for internal model training.Contribute to the stability and scalability of our infrastructure, ensuring it meets the growing demands of AI development.
About Liquid AIFounded as a spin-off from MIT CSAIL, Liquid AI specializes in creating versatile AI systems designed for optimal performance across various deployment platforms, including data center accelerators and on-device hardware. Our technology emphasizes low latency, minimal memory consumption, privacy, and dependability. We collaborate with leading enterprises in sectors such as consumer electronics, automotive, life sciences, and financial services. As we experience rapid growth, we are on the lookout for exceptional talent to join our team.The OpportunityThe Data team at Liquid AI drives the development of our Liquid Foundation Models, focusing on pre-training, vision, audio, and emerging modalities. With the stagnation of public data sources, the effectiveness of our models increasingly relies on specially curated datasets. We are seeking engineers with a machine learning mindset who can efficiently gather, filter, and synthesize high-quality data at scale.At Liquid AI, we regard data as a research challenge rather than an infrastructural issue. Our engineers conduct experiments, design ablations, and assess how data-related decisions impact model quality. We will align you with a team where you can experience rapid growth and make a significant impact, be it in pre-training, post-training reinforcement learning, vision-language, audio, or multimodal applications.While we prefer candidates in San Francisco and Boston, we are open to considering other locations.What We're Looking ForWe are in search of a candidate who:Thinks like a researcher and executes like an engineer: You should be able to formulate hypotheses, conduct experiments, and evaluate results. Our engineers produce research-level code while our researchers implement production systems.Learns quickly and adapts: You will be working in rapidly evolving modalities, so the ability to quickly grasp new domains and thrive in ambiguity is essential.Prioritizes data quality: We hold data quality in high regard; tasks such as filtering, deduplication, augmentation, and evaluation are key responsibilities, not afterthoughts.Solves problems autonomously: Data engineers operate within training groups (pre-training and multimodal). While collaboration is crucial, we expect ownership and self-direction.The WorkDevelop and maintain data processing, filtering, and selection pipelines at scale.Establish pipelines for pretraining, midtraining, supervised fine-tuning, and preference optimization datasets.Design synthetic data generation systems utilizing large language models (LLMs), structured prompting, and domain-specific generative techniques.
Join Handshake as a Machine Learning Research Scientist and contribute to groundbreaking projects that leverage advanced algorithms and data analysis to drive innovation. In this role, you will collaborate with a dynamic team to design, implement, and evaluate machine learning models that enhance our products and services. Your expertise will be pivotal in unlocking new insights from data, improving user experiences, and shaping the future of our technology.
Company Overview:At Specter, we are pioneering a software-defined control plane for the physical realm, beginning with safeguarding American enterprises through comprehensive monitoring of their physical assets.Our innovative approach leverages a connected hardware-software ecosystem built on advanced multi-modal wireless mesh sensing technology. This breakthrough enables us to reduce the deployment costs and time for sensors by a factor of 10. Our ultimate goal is to establish a perception engine that provides real-time visibility of a company’s physical environment and facilitates autonomous operations management.Co-founders Xerxes and Philip are dedicated to empowering our partners in the rapidly evolving landscape of physical AI and robotics. Join our dynamic and rapidly expanding team comprised of talents from Anduril, Tesla, Uber, and the U.S. Special Forces.Position Overview:We are seeking a Perception AI Engineer who will be instrumental in transforming sensor data pipelines into actionable insights for our clients.Key Responsibilities:Implement and deploy a range of deep-learning models, including vision, vision-language, and large language models, within our sophisticated distributed perception system.Design and scale a production-ready data collection, labeling, and model retraining platform.Lead the design of a multimodal software user interface.
Join the Center for AI Safety (CAIS), a pioneering research and advocacy organization dedicated to addressing the societal-scale risks posed by artificial intelligence. We tackle the most pressing challenges in AI through rigorous technical research, innovative field-building initiatives, and proactive policy engagement, in collaboration with our sister organization, the Center for AI Safety Action Fund.As a Research Scientist, you will spearhead and conduct transformative research aimed at enhancing the safety and dependability of cutting-edge AI systems. Your responsibilities will include designing and executing experiments on large language models, developing the necessary tools for training and evaluating models at scale, and converting your findings into publishable research. You will work closely with CAIS researchers and external partners from academia and industry, utilizing our compute cluster for large-scale model training and evaluation. Your research will focus on critical areas such as AI honesty, robustness, transparency, and the detection of trojan/backdoor behaviors, all aimed at mitigating real-world risks associated with advanced AI technologies.
At Exa, we are revolutionizing the way AI applications access information by building a cutting-edge search engine from the ground up. Our team is dedicated to developing a robust infrastructure capable of crawling the web, training advanced embedding models, and creating high-performance vector databases using Rust to facilitate seamless searches.As part of our ML team, you'll be instrumental in training foundational models that refine search capabilities. Our mission? To deliver precise answers to even the most complex queries, effectively transforming the web into an incredibly powerful knowledge database.We are seeking a talented Machine Learning Research Engineer who is passionate about crafting embedding models that enhance web search efficiency. Your responsibilities will include innovating novel transformer-based architectures, curating extensive datasets, conducting evaluations, and continuously improving our state-of-the-art models.
bland is looking for a Machine Learning Researcher with a focus on audio. This position is based in San Francisco and centers on advancing how machines process and understand sound. The team works on pushing the boundaries of audio technology for a range of platforms. Responsibilities Research and develop new machine learning techniques for audio applications Contribute to projects that improve audio processing and analysis Collaborate with colleagues to bring research ideas into real-world audio products Location This role requires working onsite in San Francisco.
About MercorMercor sits at the forefront of labor markets and artificial intelligence research, collaborating with premier AI laboratories and enterprises to harness the human intelligence crucial for AI evolution.Our expansive talent network empowers the training of cutting-edge AI models, akin to how educators impart knowledge to students—sharing insights, experiences, and contexts that transcend mere code. Currently, our network comprises over 30,000 experts, generating collective earnings exceeding $2 million daily.At Mercor, we are pioneering a unique category of work where expertise fuels AI progress. Realizing this vision necessitates a bold, fast-paced, and deeply dedicated team. You will collaborate with researchers, operators, and AI firms that are at the vanguard of transforming systems that redefine society.As a profitable Series C company, Mercor is valued at $10 billion and maintains an in-office presence five days a week at our new headquarters in San Francisco.About the RoleIn your capacity as a Research Engineer at Mercor, you will operate at the intersection of engineering and applied AI research. You will play a pivotal role in post-training and Reinforcement Learning from Human Feedback (RLVR), synthetic data generation, and large-scale evaluation workflows essential for advancing frontier language models.Your contributions will help train large language models to adeptly utilize tools, exhibit agentic behavior, and engage in real-world reasoning within production environments. You will be instrumental in shaping rewards, conducting post-training experiments, and constructing scalable systems to enhance model performance. Your responsibilities will also include designing and evaluating datasets, creating scalable data augmentation pipelines, and developing rubrics and evaluators that expand the learning potential of LLMs.
Join us at Physical Intelligence as a Research Scientist, where you will be at the forefront of innovation in machine learning and robotics. We are in search of exceptional researchers across all experience levels who demonstrate a strong track record of impactful research results. Ideal candidates will possess a solid foundation in both practical implementation and theoretical frameworks, showcasing a blend of system-building capabilities and significant conceptual, algorithmic, or theoretical advancements. We value diverse backgrounds and encourage applications from both traditional academic researchers and those with unique, unconventional experiences.We are committed to fostering a diverse and inclusive workplace. In accordance with the San Francisco Fair Chance Ordinance, we welcome applications from qualified individuals with arrest and conviction records.
OverviewPluralis Research is at the forefront of Protocol Learning, innovating a decentralized approach to train and deploy AI models that democratizes access beyond just well-funded corporations. By aggregating computational resources from diverse participants, we incentivize collaboration while safeguarding against centralized control of model weights, paving the way for a truly open and cooperative environment for advanced AI.We are seeking a talented Machine Learning Training Platform Engineer to design, develop, and scale the core infrastructure that powers our decentralized ML training platform. In this role, you will have ownership over essential systems including infrastructure orchestration, distributed computing, and service integration, facilitating ongoing experimentation and large-scale model training.ResponsibilitiesMulti-Cloud Infrastructure: Create resource management systems that provision and orchestrate computing resources across AWS, GCP, and Azure using infrastructure-as-code tools like Pulumi or Terraform. Manage dynamic scaling, state synchronization, and concurrent operations across hundreds of diverse nodes.Distributed Training Systems: Design fault-tolerant infrastructure for distributed machine learning, including GPU clusters, NVIDIA runtime, S3 checkpointing, large dataset management and streaming, health monitoring, and resilient retry strategies.Real-World Networking: Develop systems that simulate and manage real-world network conditions—such as bandwidth shaping, latency injection, and packet loss—while accommodating dynamic node churn and ensuring efficient data flow across workers with varying connectivity, as our training occurs on consumer nodes and non-co-located infrastructure.
Achira is seeking a Machine Learning Research Engineer to help improve workflows and systems for artificial intelligence projects. This position is based in the San Francisco office. Role overview This role centers on developing and refining machine learning pipelines. The focus is on efficient deployment and scaling of AI models in production environments. Collaboration with colleagues from different disciplines is a key part of the work, aiming to bring forward new ideas and solid practices in machine learning systems. What you will do Design and optimize machine learning workflows for better performance and scalability Work closely with cross-functional teams to implement improvements in AI systems Support the deployment process, helping ensure models run efficiently in real-world settings Location This position is based at Achira's San Francisco office.
Full-time|$148K/yr - $200K/yr|Hybrid|San Francisco, California, United States
About Taskrabbit:Taskrabbit is an innovative marketplace platform that seamlessly connects individuals with Taskers to manage everyday home tasks, including furniture assembly, handyman services, moving assistance, and much more.At Taskrabbit, we aim to transform lives one task at a time. We celebrate innovation, inclusion, and hard work, fostering a collaborative, pragmatic, and fast-paced culture. We seek talented, entrepreneurially minded, data-driven individuals who possess a passion for empowering others to pursue their passions. In partnership with IKEA, we are creating more opportunities for individuals to earn a consistent, meaningful income on their terms by establishing enduring relationships with clients in communities globally.Taskrabbit operates as a hybrid company, with team members located across the US and EU, and has been recognized as a Built In — Best Places to Work for 2022, 2023, and 2024, receiving accolades across various national and regional categories. Join us at Taskrabbit, where your contributions will be significant, your ideas appreciated, and your potential maximized!This position operates on a hybrid schedule, requiring two days of in-office collaboration per week. It can be based in our San Francisco office or our new New York City office (opening March 2026).About the RoleMachine Learning is a foundational element at Taskrabbit, and we are in search of an experienced Senior Machine Learning Engineer to join our team and help mold the future of ML/AI at Taskrabbit. This distinct, full-stack role is designed for someone who is enthusiastic about the entire machine learning lifecycle—from initial research and model development to constructing the robust infrastructure necessary for deploying and scaling your innovations.As a Senior Machine Learning Engineer, you will engage with exciting challenges that directly influence how users discover and interact with home services on the Taskrabbit platform. You will play a vital role in enhancing our capabilities in areas such as search ranking, content discovery, and recommendation systems. Collaborating closely with data scientists and fellow engineers, you will design and implement cutting-edge algorithms, ensuring the scalability, reliability, and optimization of our models in production alongside software engineers.
Full-time|$240K/yr - $260K/yr|On-site|San Francisco, CA
About VSCO At VSCO, we empower photographers with an innovative platform that provides essential tools, a vibrant community, and the visibility needed for creative and professional growth. We cultivate an authentic creative environment that welcomes photographers of all skill levels, offering a space that inspires opportunity, collaboration, and connection. Our mission is to support photographers in their journeys, enabling them to thrive and connect with fellow creatives and businesses through our comprehensive suite of tools, available on both mobile and desktop. We seek individuals who are passionate and proactive in advancing our mission. Our team members have the opportunity to make a significant impact, and we believe that collaborative efforts yield stronger results. Our core values are essential to our team culture and guide our hiring process. Learn more about what you can expect when joining VSCO on our Careers Page. About The Role As a Senior Machine Learning Engineer, you will harness the power of AI and machine learning to create innovative, reliable user-facing product features. You will leverage your extensive technical background and hands-on experience in deploying machine learning models to deliver impactful solutions based on real-world feedback. Your focus on measurable outcomes and customer satisfaction drives your work, blending innovation with practical implementation. You will be highly skilled in Python and adept across the data and machine learning stack, enabling you to develop and launch models efficiently while ensuring scalability and maintainability. Whether working with traditional algorithms or cutting-edge deep learning and generative AI, you will expertly navigate the complexity of each problem, managing every phase from defining the challenge to deployment and iterative improvement. Your dedication to software engineering excellence will inform your thoughtful approach to system design for machine learning, encompassing data quality, pipeline design, feature workflows, model serving, and ongoing monitoring and enhancement. By integrating machine learning deeply within our cohesive product experiences, you will collaborate effectively with cross-functional teams, aligning on objectives, defining success metrics, and driving meaningful outcomes. You will stay informed about the rapidly evolving AI landscape, maintaining a discerning perspective that allows your team to focus on significant advancements while avoiding distractions. The Day to Day Design and implement ML-powered features for search, discovery, personalization, and more.
Job OverviewJoin Eragon as a Machine Learning Engineer and lead the charge in transforming innovative AI models into scalable, production-grade systems. This position is pivotal in bridging research and real-world applications by designing and optimizing systems that enhance vital workflows throughout the enterprise.In collaboration with our research, product, and engineering teams, you will convert cutting-edge capabilities into dependable, high-performance systems ready for production.Key ResponsibilitiesModel Development & Deployment: Craft, refine, and deploy machine learning models within production settings.Systems Engineering: Architect scalable pipelines for training, inference, evaluation, and comprehensive monitoring.Performance Optimization: Enhance the latency, throughput, cost-efficiency, and reliability of ML systems.Data & Infrastructure: Manipulate large datasets and ensure seamless integration of models with internal systems and APIs.Cross-Functional Collaboration: Collaborate with product and engineering teams to provide end-to-end AI functionalities.Evaluation & Monitoring: Develop robust evaluation frameworks and feedback loops to ensure system effectiveness.
About the RoleJoin Thorin as an AI Researcher and play a pivotal role in shaping the core research initiatives that drive our AI innovations. In this position, you will operate at the crossroads of machine learning research, practical model development, and product application, enhancing our understanding and automation of enterprise workflows.This position merges theoretical research with hands-on implementation, transitioning ideas from conceptual stages through experimentation into functional components that enrich Thorin’s offerings.Your ResponsibilitiesResearch & InnovationConduct innovative machine learning research aligned with real-world product demands.Investigate new model architectures, training methods, and evaluation techniques specifically designed for understanding and automating organizational workflows.Model Development & EvaluationCreate, implement, and assess ML/AI methodologies that enhance model efficacy for essential tasks.Collaborate closely with cross-functional teams to integrate research outcomes into tangible products that meet user needs.
Full-time|$225K/yr - $550K/yr|On-site|San Francisco
At magic.dev, we are committed to advancing humanity by developing safe artificial general intelligence (AGI) that tackles the world's most pressing challenges. Our unique approach focuses on automating research and code generation to enhance model performance and alignment more effectively than traditional methods. By leveraging cutting-edge pre-training, domain-specific reinforcement learning, ultra-long context processing, and efficient inference-time computation, we aim to redefine the capabilities of AGI.Role OverviewAs a Research Engineer, you will play a pivotal role in training, evaluating, and deploying large-scale AI models alongside innovative inference-time computing methods. You will contribute to the creation of extensive internet-scale datasets and support the prototyping of groundbreaking research and product initiatives.Key ResponsibilitiesEnhance inference throughput for cutting-edge model architecturesDevelop and refine frameworks that underpin our research and production processesTrain trillion-parameter models using large GPU clustersCurate post-training datasets to bolster specific capabilitiesConstruct internet-scale data pipelines and web crawlersDesign, prototype, and optimize innovative model architecturesContribute to cutting-edge research in long-context, inference-time computation, reinforcement learning, and additional domainsQualificationsProven software engineering expertiseIn-depth understanding of deep learning literatureExperience with both pre-training and post-training of large language models (LLMs)Strong capability to generate and assess research ideasFamiliarity with large distributed systemsProficient in managing substantial ETL workloadsCompensation and BenefitsAnnual salary ranging from $225,000 to $550,000 based on experienceEquity is a significant component of total compensation401(k) plan with a 6% salary matchComprehensive health, dental, and vision insurance for you and your dependentsUnlimited paid time offVisa sponsorship and relocation assistance availableBe part of a small, dynamic, and focused team
Jan 24, 2024
Sign in to browse more jobs
Create account — see all 10,323 results
Tailoring 0 resumes…
Tailoring 0 resumes…
We'll move completed jobs to Ready to Apply automatically.