<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>f28927b0-573</externalid>
      <Title>Machine Learning Systems Research Engineer, Agent Post-training - Enterprise GenAI</Title>
      <Description><![CDATA[<p>At Scale, our mission is to accelerate the development of AI applications. We are working on an arsenal of proprietary research and resources that serve all of our enterprise clients. As an ML Sys Research Engineer, you&#39;ll work on building out the algorithms for our next-gen Agent RL training platform, support large scale training, and research and integrate state-of-the-art technologies to optimize our ML system.</p>
<p>Your customer will be other MLREs and AAIs on the Enterprise AI team who are taking the training algorithms and applying them to client use-cases ranging from next-generation AI cybersecurity firewall LLMs to training foundation healthtech search models.</p>
<p>If you are excited about shaping the future of the modern AI movement, we would love to hear from you!</p>
<p>Key Responsibilities:</p>
<ul>
<li>Build, profile and optimize our training and inference framework.</li>
<li>Post-train state of the art models, developed both internally and from the community, to define stable post-training recipes for our enterprise engagements.</li>
<li>Collaborate with ML teams to accelerate their research and development, and enable them to develop the next generation of models and data curation.</li>
<li>Create a next-gen agent training algorithm for multi-agent/multi-tool rollouts.</li>
</ul>
<p>Ideal Candidate:</p>
<ul>
<li>At least 1-3 years of LLM training in a production environment.</li>
<li>Passionate about system optimization.</li>
<li>Experience with post-training methods like RLHF/RLVR and related algorithms like PPO/GRPO etc.</li>
<li>Ability to demonstrate know-how on how to operate the architecture of the modern GPU cluster.</li>
<li>Experience with multi-node LLM training and inference.</li>
<li>Strong software engineering skills, proficient in frameworks and tools such as CUDA, Pytorch, transformers, flash attention, etc.</li>
<li>Strong written and verbal communication skills to operate in a cross functional team environment.</li>
<li>PhD or Masters in Computer Science or a related field.</li>
</ul>
<p>Compensation:</p>
<p>We offer competitive compensation packages, including base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>
<p>Benefits:</p>
<ul>
<li>Comprehensive health, dental and vision coverage.</li>
<li>Retirement benefits.</li>
<li>A learning and development stipend.</li>
<li>Generous PTO.</li>
<li>Commuter stipend.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$189,600-$237,000 USD</Salaryrange>
      <Skills>LLM training, System optimization, Post-training methods, GPU cluster operation, Multi-node LLM training, Inference, CUDA, Pytorch, Transformers, Flash attention</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Scale</Employername>
      <Employerlogo>https://logos.yubhub.co/scale.com.png</Employerlogo>
      <Employerdescription>Scale is a leading AI data foundry that helps fuel advancements in AI, including generative AI, defense applications, and autonomous vehicles.</Employerdescription>
      <Employerwebsite>https://www.scale.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/scaleai/jobs/4625341005</Applyto>
      <Location>San Francisco, CA; New York, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>539e2a23-ddf</externalid>
      <Title>Tech Lead Manager- MLRE, ML Systems</Title>
      <Description><![CDATA[<p>You will lead the development of our internal distributed framework for large language model training. The platform powers MLEs, researchers, data scientists, and operators for fast and automatic training and evaluation of LLMs. It also serves as the underlying training framework for the data quality evaluation pipeline.</p>
<p>You will work closely with Scale’s ML teams and researchers to build the foundation platform which supports all our ML research and development works. You will be building and optimising the platform to enable our next generation LLM training, inference and data curation.</p>
<p>Key responsibilities include:</p>
<ul>
<li>Building, profiling and optimising our training and inference framework.</li>
<li>Collaborating with ML and research teams to accelerate their research and development, and enable them to develop the next generation of models and data curation.</li>
<li>Researching and integrating state-of-the-art technologies to optimise our ML system.</li>
</ul>
<p>The ideal candidate will have:</p>
<ul>
<li>Passionate about system optimisation.</li>
<li>Experience with multi-node LLM training and inference.</li>
<li>Experience with developing large-scale distributed ML systems.</li>
<li>Experience with post-training methods like RLHF/RLVR and related algorithms like PPO/GRPO etc.</li>
<li>Strong software engineering skills, proficient in frameworks and tools such as CUDA, PyTorch, transformers, flash attention, etc.</li>
</ul>
<p>Nice to haves include demonstrated expertise in post-training methods and/or next generation use cases for large language models including instruction tuning, RLHF, tool use, reasoning, agents, and multimodal, etc.</p>
<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$264,800-$331,000 USD</Salaryrange>
      <Skills>system optimisation, multi-node LLM training and inference, large-scale distributed ML systems, post-training methods, software engineering skills, CUDA, PyTorch, transformers, flash attention, next generation use cases for large language models, instruction tuning, RLHF, tool use, reasoning, agents, multimodal</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Scale</Employername>
      <Employerlogo>https://logos.yubhub.co/scale.com.png</Employerlogo>
      <Employerdescription>Scale provides training and evaluation data and end-to-end solutions for the ML lifecycle.</Employerdescription>
      <Employerwebsite>https://scale.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/scaleai/jobs/4618046005</Applyto>
      <Location>San Francisco, CA; New York, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>57a8aa85-77e</externalid>
      <Title>Staff Machine Learning Research Engineer, Agent Post-training - Enterprise GenAI</Title>
      <Description><![CDATA[<p>We are seeking a Staff Machine Learning Research Engineer to join our Enterprise ML Research Lab. As a key member of our team, you will build out our next-gen Agent RL training platform, integrating cutting-edge research into our training stack. You will train state-of-the-art models, design solutions for complex multi-agent systems, and collaborate with our team to deploy use-cases ranging from next-generation AI cybersecurity firewall LLMs to training foundation healthtech search models.</p>
<p>The ideal candidate will have 5+ years of LLM training in a production environment, experience with post-training methods like RLHF/RLVR and related algorithms like PPO/GRPO, and publications in top conferences such as NEURIPS, ICLR, or ICML within the last two years. A PhD or Masters in Computer Science or a related field is required.</p>
<p>In addition to a competitive salary, you will receive equity-based compensation, comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. This role may also be eligible for additional benefits such as a commuter stipend.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$189,600-$237,000 USD</Salaryrange>
      <Skills>LLM training, Post-training methods, RLHF/RLVR, PPO/GRPO, NEURIPS, ICLR, ICML, Computer Science, PhD, Masters</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Scale</Employername>
      <Employerlogo>https://logos.yubhub.co/scale.com.png</Employerlogo>
      <Employerdescription>Scale is a leading AI data foundry, helping fuel the most exciting advancements in AI, including generative AI, defense applications, and autonomous vehicles.</Employerdescription>
      <Employerwebsite>https://www.scale.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/scaleai/jobs/4625337005</Applyto>
      <Location>San Francisco, CA; New York, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>840bab06-7be</externalid>
      <Title>ML Research Engineer, ML Systems</Title>
      <Description><![CDATA[<p>Job Description:</p>
<p>Scale&#39;s ML platform (RLXF) team builds our internal distributed framework for large language model training and inference. The platform has been powering MLEs, researchers, data scientists and operators for fast and automatic training and evaluation of LLM&#39;s, as well as evaluation of data quality.</p>
<p>At Scale, we&#39;re uniquely positioned at the heart of the field of AI as an indispensable provider of training and evaluation data and end-to-end solutions for the ML lifecycle. You will work closely across Scale&#39;s ML teams and researchers to build the foundation platform that supports all our ML research and development. You will be building and optimizing the platform to enable our next generation of LLM training, inference and data curation.</p>
<p>Responsibilities:</p>
<ul>
<li>Build, profile and optimize our training and inference framework</li>
<li>Collaborate with ML teams to accelerate their research and development and enable them to develop the next generation of models and data curation</li>
<li>Research and integrate state-of-the-art technologies to optimize our ML system</li>
</ul>
<p>Ideal Candidate:</p>
<ul>
<li>Strong excitement about system optimization</li>
<li>Experience with multi-node LLM training and inference</li>
<li>Experience with developing large-scale distributed ML systems</li>
<li>Strong software engineering skills, proficient in frameworks and tools such as CUDA, Pytorch, transformers, flash attention, etc.</li>
<li>Strong written and verbal communication skills and the ability to operate in a cross functional team environment</li>
</ul>
<p>Nice to Have:</p>
<ul>
<li>Demonstrated expertise in post-training methods &amp;/or next generation use cases for large language models including instruction tuning, RLHF, tool use, reasoning, agents, and multimodal, etc.</li>
</ul>
<p>Compensation Packages:</p>
<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You&#39;ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>
<p>Please note that our policy requires a 90-day waiting period before reconsidering candidates for the same role. This allows us to ensure a fair and thorough evaluation of all applicants.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$189,600-$237,000 USD</Salaryrange>
      <Skills>System Optimization, Multi-node LLM Training and Inference, Large-Scale Distributed ML Systems, CUDA, Pytorch, Transformers, Flash Attention, Post-Training Methods, Next Generation Use Cases for Large Language Models, Instruction Tuning, RLHF, Tool Use, Reasoning, Agents, Multimodal</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Scale</Employername>
      <Employerlogo>https://logos.yubhub.co/scale.com.png</Employerlogo>
      <Employerdescription>Scale develops reliable AI systems for the world&apos;s most important decisions, providing high-quality data and full-stack technologies for leading models.</Employerdescription>
      <Employerwebsite>https://scale.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/scaleai/jobs/4534631005</Applyto>
      <Location>San Francisco, CA; Seattle, WA; New York, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>b2637f59-e14</externalid>
      <Title>Full-Stack Software Engineer, Reinforcement Learning</Title>
      <Description><![CDATA[<p>As a Full-Stack Software Engineer in RL, you&#39;ll build the platforms, tools, and interfaces that power environment creation, data collection, and training observability. The quality of Claude&#39;s next generation depends on the quality of the data we train it on , and the systems you build are what make that data possible. You&#39;ll own product surfaces end-to-end , from backend services and APIs to the web UIs that researchers, external vendors, and thousands of data labelers use every day.\n\nYou don&#39;t need a background in ML research. What matters is that you can take an ambiguous, high-stakes problem and ship a polished, reliable product against it, fast. This team moves very quickly. Claude writes a lot of the code we commit, which means the bottleneck isn&#39;t typing , it&#39;s judgment, taste, and the ability to react to what researchers need next.\n\nYou&#39;ll iterate on data collection strategies to distill the knowledge of thousands of human experts around the world into our models, and you&#39;ll do it in a loop that closes in hours and days, not quarters or months.\n\nAnthropic&#39;s Reinforcement Learning organization leads the research and development that trains Claude to be capable, reliable, and safe. We&#39;ve contributed to every Claude model, with significant impact on the autonomy and coding capabilities of our most advanced models.\n\nOur work spans teaching models to use computers effectively, advancing code generation through RL, pioneering fundamental RL research for large language models, and building the scalable training methodologies behind our frontier production models.\n\nThe RL org is organized around four goals: solving the science of long-horizon tasks and continual learning, scaling RL data and environments to be comprehensive and diverse, automating software engineering end-to-end, and training the frontier production model.\n\nOur engineering teams build the environments, evaluation systems, data pipelines, and tooling that make all of this possible , from realistic agentic training environments and scalable code data generation to human data collection platforms and production training operations.\n\n### Responsibilities\n\n<em>   Build and extend web platforms for RL environment creation, management, and quality review , including environment configuration, versioning, and validation workflows\n</em>   Develop vendor-facing interfaces and tooling that let external partners create, submit, and iterate on training environments with minimal friction\n<em>   Design and implement platforms for human data collection at scale, including labeling workflows, quality assurance systems, and feedback mechanisms that surface reward signal integrity issues early\n</em>   Build evaluation dashboards and observability UIs that give researchers real-time insight into environment quality, training run health, and reward hacking\n<em>   Create backend services and APIs that connect environment authoring tools, data collection systems, and RL training infrastructure\n</em>   Build and expand scalable code data generation pipelines, producing diverse programming tasks with robust reward signals across languages and difficulty levels\n<em>   Develop onboarding automation and documentation tooling so new vendors and internal users ramp up in hours, not weeks\n</em>   Partner closely with RL researchers, data operations, and vendor management to translate ambiguous requirements into well-scoped, well-designed products\n\n### Requirements\n\n<em>   Strong software engineering fundamentals and real full-stack range , you&#39;re comfortable owning a surface from database schema to frontend\n</em>   Proficient in Python and a modern web stack (React, TypeScript, or similar)\n<em>   Track record of shipping systems that solved a hard problem, not just shipped on time , e.g. you built the thing that made your team 10x faster, or the internal tool nobody thought was possible\n</em>   Operate with high agency: you identify what needs to be done and drive it forward without waiting for a ticket\n<em>   Found yourself wondering &quot;why isn&#39;t this moving faster?&quot; in previous roles , and then have done something about it\n</em>   Care about UX and can build interfaces that are intuitive for both technical researchers and non-technical labelers\n<em>   Communicate clearly with researchers, operations teams, and engineers, and can turn vague asks into well-scoped work\n</em>   Thrive in a fast-moving environment where priorities shift, Claude is your pair programmer, and the next problem is often one nobody has solved before\n<em>   Care about Anthropic&#39;s mission to build safe, beneficial AI and want your work to contribute directly to it\n\n### Nice to Have\n\n</em>   Built data collection, labeling, or annotation platforms , ideally ones that had to scale across many vendors or many task types\n<em>   Background building multi-tenant platforms with role-based access, audit trails, and vendor management workflows\n</em>   Experience with cloud infrastructure (GCP or AWS), Docker, and CI/CD pipelines\n<em>   Familiarity with LLM training, fine-tuning, or evaluation workflows\n</em>   Experience with async Python (Trio, asyncio) or high-throughput API design\n<em>   Background in dashboards, monitoring, or observability tooling\n</em>   Experience working directly with external vendors or partners on technical integrations\n<em>   A background that isn&#39;t a straight line , e.g. math or physics into SWE, competitive programming, research into engineering, or a side project that outgrew its scope\n\n### Representative Projects\n\n</em>   Building a unified platform for human data collection that integrates labeling workflows, vendor management, and QA for complex agentic tasks\n<em>   Developing vendor onboarding automation that handles Docker registry access, API token management, and environment validation\n</em>   Creating evaluation and observability dashboards that catch reward hacks, measure environment difficulty, and give real-time feedback during production training\n<em>   Building environment quality review workflows that let researchers browse, grade, and provide feedback on training environments\n</em>   Developing automated environment quality pipelines that validate correctness and difficulty calibration before environments hit production training\n*   Building internal tools for browsing and analyzing training run results, environment statistics, and data collection progress</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$300,000-$405,000 USD</Salaryrange>
      <Skills>Python, Modern web stack, React, TypeScript, Strong software engineering fundamentals, Full-stack range, Database schema, Frontend, Cloud infrastructure, Docker, CI/CD pipelines, LLM training, Fine-tuning, Evaluation workflows, Async Python, High-throughput API design, Dashboards, Monitoring, Observability tooling, Data collection, Labeling, Annotation platforms, Multi-tenant platforms, Role-based access, Audit trails, Vendor management workflows</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a company working on developing artificial intelligence systems. It has a quickly growing team of researchers, engineers, and business leaders.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5186067008</Applyto>
      <Location>San Francisco, CA | New York City, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>c9ab5cbc-dd6</externalid>
      <Title>Research Engineer, Performance RL</Title>
      <Description><![CDATA[<p>We&#39;re hiring a Research Engineer to join our Code RL team within the RL organization. As a Research Engineer, you&#39;ll advance our models&#39; ability to safely write correct, fast code for accelerators.</p>
<p>You&#39;ll need to know accelerator performance well to turn it into tasks and signals models can learn from. Specifically, you will:</p>
<ul>
<li>Invent, design and implement RL environments and evaluations.</li>
<li>Conduct experiments and shape our research roadmap.</li>
<li>Deliver your work into training runs.</li>
<li>Collaborate with other researchers, engineers, and performance engineering specialists across and outside Anthropic.</li>
</ul>
<p>We&#39;re looking for someone with expertise in accelerators (CUDA, ROCm, Triton, Pallas), ML framework programming (JAX or PyTorch), and experience with balancing research exploration with engineering implementation.</p>
<p>Strong candidates may also have experience with reinforcement learning, porting ML workloads between different types of accelerators, and familiarity with LLM training methodologies.</p>
<p>The annual compensation range for this role is $350,000-$850,000 USD.</p>
<p>Please note that we&#39;re an extremely collaborative group, and we value communication skills. The easiest way to understand our research directions is to read our recent research.</p>
<p>We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$350,000-$850,000 USD</Salaryrange>
      <Skills>accelerator performance, ML framework programming, reinforcement learning, RL environments and evaluations, experiments and research roadmap, training runs, collaboration with researchers and engineers, CUDA, ROCm, Triton, Pallas, JAX, PyTorch, LLM training methodologies</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that focuses on creating reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5160330008</Applyto>
      <Location>San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>faffcca4-e94</externalid>
      <Title>Research Engineer, Cybersecurity Reinforcement Learning</Title>
      <Description><![CDATA[<p>About the role</p>
<p>We&#39;re hiring for the Cybersecurity RL team within Horizons. As a Research Engineer, you&#39;ll help to safely advance the capabilities of our models in secure coding, vulnerability remediation, and other areas of defensive cybersecurity.</p>
<p>This role blends research and engineering, requiring you to both develop novel approaches and realize them in code. Your work will include designing and implementing RL environments, conducting experiments and evaluations, delivering your work into production training runs, and collaborating with other researchers, engineers, and cybersecurity specialists across and outside Anthropic.</p>
<p>The role requires domain expertise in cybersecurity paired with interest or experience in training safe AI models. For example, you might be a white hat hacker who&#39;s curious about how LLMs could augment or transform your work, a security engineer interested in how AI could help harden systems at scale, or a detection and response professional wondering how models could enhance defensive workflows.</p>
<p>Responsibilities</p>
<ul>
<li>Design and implement RL environments for secure coding and vulnerability remediation</li>
<li>Conduct experiments and evaluations to assess the effectiveness of our models</li>
<li>Deliver your work into production training runs to advance the capabilities of our models</li>
<li>Collaborate with other researchers, engineers, and cybersecurity specialists across and outside Anthropic</li>
</ul>
<p>Requirements</p>
<ul>
<li>Experience in cybersecurity research</li>
<li>Experience with machine learning</li>
<li>Strong software engineering skills</li>
<li>Ability to balance research exploration with engineering implementation</li>
<li>Passion for AI&#39;s potential and commitment to developing safe and beneficial systems</li>
</ul>
<p>Strong candidates may also have:</p>
<ul>
<li>Professional experience in security engineering, fuzzing, detection and response, or other applied defensive work</li>
<li>Experience participating in or building CTF competitions and cyber ranges</li>
<li>Academic research experience in cybersecurity</li>
<li>Familiarity with RL techniques and environments</li>
<li>Familiarity with LLM training methodologies</li>
</ul>
<p>Logistics</p>
<ul>
<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>
<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>
<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>
<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>
<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>
</ul>
<p>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>
<p>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links,visit anthropic.com/careers directly for confirmed position openings.</p>
<p>How we&#39;re different</p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>
<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>
<p>Come work with us!</p>
<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$300,000-$405,000 USD</Salaryrange>
      <Skills>cybersecurity research, machine learning, software engineering, research exploration, engineering implementation, security engineering, fuzzing, detection and response, RL techniques, LLM training methodologies</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5025624008</Applyto>
      <Location>San Francisco, CA | New York City, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>157be224-49f</externalid>
      <Title>Machine Learning Systems Engineer, RL Engineering</Title>
      <Description><![CDATA[<p>About the role:</p>
<p>You want to build the cutting-edge systems that train AI models like Claude. You&#39;re excited to work at the frontier of machine learning, implementing and improving advanced techniques to create ever more capable, reliable and steerable AI.</p>
<p>As an ML Systems Engineer on our Reinforcement Learning Engineering team, you&#39;ll be responsible for the critical algorithms and infrastructure that our researchers depend on to train models. Your work will directly enable breakthroughs in AI capabilities and safety.</p>
<p>You&#39;ll focus obsessively on improving the performance, robustness, and usability of these systems so our research can progress as quickly as possible.</p>
<p>Our finetuning researchers train our production Claude models, and internal research models, using RLHF and other related methods. Your job will be to build, maintain, and improve the algorithms and systems that these researchers use to train models.</p>
<p>You&#39;ll be responsible for improving the speed, reliability, and ease-of-use of these systems.</p>
<p>Strong candidates may also have experience with:</p>
<p>High performance, large scale distributed systems Large scale LLM training Python Implementing LLM finetuning algorithms, such as RLHF</p>
<p>Representative projects:</p>
<p>Profiling our reinforcement learning pipeline to find opportunities for improvement Building a system that regularly launches training jobs in a test environment so that we can quickly detect problems in the training pipeline Making changes to our finetuning systems so they work on new model architectures Building instrumentation to detect and eliminate Python GIL contention in our training code Diagnosing why training runs have started slowing down after some number of steps, and fixing it Implementing a stable, fast version of a new training algorithm proposed by a researcher</p>
<p>Deadline to apply: None. Applications will be reviewed on a rolling basis.</p>
<p>The annual compensation range for this role is $500,000-$850,000 USD.</p>
<p>Logistics:</p>
<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices. Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed.</p>
<p>Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>
<p>We think AI systems like the ones we&#39;re building have enormous social and ethical implications. We think this makes representation even more important, and we strive to include a range of diverse perspectives on our team.</p>
<p>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links,visit anthropic.com/careers directly for confirmed position openings.</p>
<p>How we&#39;re different:</p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles.</p>
<p>We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time.</p>
<p>As such, we greatly value communication skills.</p>
<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>
<p>Come work with us!</p>
<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>
<p>Guidance on Candidates&#39; AI Usage: Learn about our policy for using AI in our application process</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$500,000-$850,000 USD</Salaryrange>
      <Skills>High performance, large scale distributed systems, Large scale LLM training, Python, Implementing LLM finetuning algorithms, such as RLHF</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/4952051008</Applyto>
      <Location>San Francisco, CA | New York City, NY | Seattle, WA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>1507524b-770</externalid>
      <Title>Research Engineer, Performance RL</Title>
      <Description><![CDATA[<p>We&#39;re hiring a Research Engineer to join our Code RL team within the RL organization. As a Research Engineer, you&#39;ll advance our models&#39; ability to safely write correct, fast code for accelerators.</p>
<p>You&#39;ll need to know accelerator performance well to turn it into tasks and signals models can learn from. Specifically, you will:</p>
<ul>
<li>Invent, design and implement RL environments and evaluations.</li>
<li>Conduct experiments and shape our research roadmap.</li>
<li>Deliver your work into training runs.</li>
<li>Collaborate with other researchers, engineers, and performance engineering specialists across and outside Anthropic.</li>
</ul>
<p>You may be a good fit if you:</p>
<ul>
<li>Have expertise with accelerators (CUDA, ROCm, Triton, Pallas), ML framework programming (JAX or PyTorch).</li>
<li>Have worked across the stack – kernels, model code, distributed systems.</li>
<li>Know how to balance research exploration with engineering implementation.</li>
<li>Are passionate about AI&#39;s potential and committed to developing safe and beneficial systems.</li>
</ul>
<p>Strong candidates may also have:</p>
<ul>
<li>Experience with reinforcement learning.</li>
<li>Experience porting ML workloads between different types of accelerators.</li>
<li>Familiarity with LLM training methodologies.</li>
</ul>
<p>The annual compensation range for this role is $350,000-$850,000 USD.</p>
<p>We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science.</p>
<p>We kitchen is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$350,000-$850,000 USD</Salaryrange>
      <Skills>accelerators, ML framework programming, distributed systems, reinforcement learning, LLM training methodologies, CUDA, ROCm, Triton, Pallas, JAX, PyTorch</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5160330008</Applyto>
      <Location>San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>76fd624c-e23</externalid>
      <Title>Full-Stack Software Engineer, Reinforcement Learning</Title>
      <Description><![CDATA[<p>As a Full-Stack Software Engineer in RL, you&#39;ll build the platforms, tools, and interfaces that power environment creation, data collection, and training observability. The quality of Claude&#39;s next generation depends on the quality of the data we train it on , and the systems you build are what make that data possible. You&#39;ll own product surfaces end-to-end , from backend services and APIs to the web UIs that researchers, external vendors, and thousands of data labelers use every day. You don&#39;t need a background in ML research. What matters is that you can take an ambiguous, high-stakes problem and ship a polished, reliable product against it, fast.</p>
<p>This team moves very quickly. Claude writes a lot of the code we commit, which means the bottleneck isn&#39;t typing , it&#39;s judgment, taste, and the ability to react to what researchers need next. You&#39;ll iterate on data collection strategies to distill the knowledge of thousands of human experts around the world into our models, and you&#39;ll do it in a loop that closes in hours and days, not quarters or months.</p>
<p>Our work spans teaching models to use computers effectively, advancing code generation through RL, pioneering fundamental RL research for large language models, and building the scalable training methodologies behind our frontier production models. The RL org is organized around four goals: solving the science of long-horizon tasks and continual learning, scaling RL data and environments to be comprehensive and diverse, automating software engineering end-to-end, and training the frontier production model.</p>
<p>Our engineering teams build the environments, evaluation systems, data pipelines, and tooling that make all of this possible , from realistic agentic training environments and scalable code data generation to human data collection platforms and production training operations.</p>
<p>Responsibilities:</p>
<ul>
<li>Build and extend web platforms for RL environment creation, management, and quality review , including environment configuration, versioning, and validation workflows</li>
<li>Develop vendor-facing interfaces and tooling that let external partners create, submit, and iterate on training environments with minimal friction</li>
<li>Design and implement platforms for human data collection at scale, including labeling workflows, quality assurance systems, and feedback mechanisms that surface reward signal integrity issues early</li>
<li>Build evaluation dashboards and observability UIs that give researchers real-time insight into environment quality, training run health, and reward hacking</li>
<li>Create backend services and APIs that connect environment authoring tools, data collection systems, and RL training infrastructure</li>
<li>Build and expand scalable code data generation pipelines, producing diverse programming tasks with robust reward signals across languages and difficulty levels</li>
<li>Develop onboarding automation and documentation tooling so new vendors and internal users ramp up in hours, not weeks</li>
<li>Partner closely with RL researchers, data operations, and vendor management to translate ambiguous requirements into well-scoped, well-designed products</li>
</ul>
<p>You May Be a Good Fit If You:</p>
<ul>
<li>Have strong software engineering fundamentals and real full-stack range , you&#39;re comfortable owning a surface from database schema to frontend</li>
<li>Are proficient in Python and a modern web stack (React, TypeScript, or similar)</li>
<li>Have a track record of shipping systems that solved a hard problem, not just shipped on time , e.g. you built the thing that made your team 10x faster, or the internal tool nobody thought was possible</li>
<li>Operate with high agency: you identify what needs to be done and drive it forward without waiting for a ticket</li>
<li>Have found yourself wondering &quot;why isn&#39;t this moving faster?&quot; in previous roles , and then have done something about it</li>
<li>Care about UX and can build interfaces that are intuitive for both technical researchers and non-technical labelers</li>
<li>Communicate clearly with researchers, operations teams, and engineers, and can turn vague asks into well-scoped work</li>
<li>Thrive in a fast-moving environment where priorities shift, Claude is your pair programmer, and the next problem is often one nobody has solved before</li>
<li>Care about Anthropic&#39;s mission to build safe, beneficial AI and want your work to contribute directly to it</li>
</ul>
<p>Strong Candidates May Also Have:</p>
<ul>
<li>Built data collection, labeling, or annotation platforms , ideally ones that had to scale across many vendors or many task types</li>
<li>Background building multi-tenant platforms with role-based access, audit trails, and vendor management workflows</li>
<li>Experience with cloud infrastructure (GCP or AWS), Docker, and CI/CD pipelines</li>
<li>Familiarity with LLM training, fine-tuning, or evaluation workflows</li>
<li>Experience with async Python (Trio, asyncio) or high-throughput API design</li>
<li>Background in dashboards, monitoring, or observability tooling</li>
<li>Experience working directly with external vendors or partners on technical integrations</li>
<li>A background that isn&#39;t a straight line , e.g. math or physics into SWE, competitive programming, research into engineering, or a side project that outgrew its scope</li>
</ul>
<p>Representative Projects:</p>
<ul>
<li>Building a unified platform for human data collection that integrates labeling workflows, vendor management, and QA for complex agentic tasks</li>
<li>Developing vendor onboarding automation that handles Docker registry access, API token management, and environment validation</li>
<li>Creating evaluation and observability dashboards that catch reward hacks, measure environment difficulty, and give real-time feedback during production training</li>
<li>Building environment quality review workflows that let researchers browse, grade, and provide feedback on training environments</li>
<li>Developing automated environment quality pipelines that validate correctness and difficulty calibration before environments hit production training</li>
<li>Building internal tools for browsing and analyzing training run results, environment statistics, and data collection progress</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$300,000-$405,000 USD</Salaryrange>
      <Skills>Python, Modern web stack, React, TypeScript, Cloud infrastructure, Docker, CI/CD pipelines, LLM training, Fine-tuning, Evaluation workflows, Async Python, High-throughput API design, Dashboards, Monitoring, Observability tooling, Data collection, Labeling, Annotation, Multi-tenant platforms, Role-based access, Audit trails, Vendor management workflows</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a company that creates reliable, interpretable, and steerable AI systems. It has a quickly growing team of researchers, engineers, policy experts, and business leaders.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5186067008</Applyto>
      <Location>San Francisco, CA | New York City, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>2907e75d-d4e</externalid>
      <Title>Research Engineer, Frontier Safety Risk Assessment</Title>
      <Description><![CDATA[<p>Job Title: Research Engineer, Frontier Safety Risk Assessment</p>
<p>We are seeking 2 Research Engineers for the Frontier Safety Risk Assessment team within the AGI Safety and Alignment Team.</p>
<p>As a Research Engineer, you will contribute novel research towards our ability to measure and assess risk from frontier models. This might include:</p>
<ul>
<li>Identifying new risk pathways within current areas (loss of control, ML R&amp;D, cyber, CBRN, harmful manipulation) or in new ones;</li>
<li>Conceiving of, designing, and developing new ways to measure pre-mitigation and post-mitigation risk;</li>
<li>Forecasting and scenario planning for future risks which are not yet material.</li>
</ul>
<p>Your work will involve complex conceptual thinking as well as engineering. You should be comfortable with research that is uncertain, under-constrained, and which does not have an achievable “right answer”. You should also be skilled at engineering, especially using Python, and able to rapidly familiarise yourself with internal and external codebases. Lastly, you should be able to adapt to pragmatic constraints around compute and researcher time that require us to prioritise effort based on the value of information.</p>
<p>Although this job description is written for a Research Engineer, all members of this team are better thought of as members of technical staff. We expect everyone to contribute to the research as well as the engineering and to be strong in both areas.</p>
<p>The role will mostly depend on your general ability to assess and manage future risks, rather than from specialist knowledge within the risk domains, but insofar as specialist knowledge is helpful, knowledge in ML R&amp;D and loss of control as risk domains are likely the most valuable.</p>
<p>About You</p>
<p>In order to set you up for success as a Research Engineer at Google DeepMind, we look for the following skills and experience:</p>
<ul>
<li>You have extensive research experience with deep learning and/or foundation models (for example, but not necessarily, a PhD in machine learning).</li>
<li>You are adept at generating ideas and designing experiments, and implementing these in Python with real AI systems.</li>
<li>You are keen to address risks from foundation models, and have thought about how to do so. You plan for your research to impact production systems on a timescale between “immediately” and “a few years”.</li>
<li>You are excited to work with strong contributors to make progress towards a shared ambitious goal.</li>
<li>With strong, clear communication skills, you are confident engaging technical stakeholders to share research insights tailored to their background.</li>
</ul>
<p>In addition, any of the following would be an advantage:</p>
<ul>
<li>Experience in areas such as frontier risk assessment and/or mitigations, safety, and alignment.</li>
<li>Engineering experience with LLM training and inference.</li>
<li>PhD in Computer Science or Machine Learning related field.</li>
<li>A track record of publications at venues such as NeurIPS, ICLR, ICML, RL/DL, EMNLP, AAAI and UAI.</li>
<li>Experience with collaborating or leading an applied research project.</li>
</ul>
<p>At Google DeepMind, we value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact. We are committed to equal employment opportunity regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition (including breastfeeding) or any other basis as protected by applicable law. If you have a disability or additional need that requires accommodation, please do not hesitate to let us know.</p>
<p>At Google DeepMind, we want employees and their families to live happier and healthier lives, both in and out of work, and our benefits reflect that. Some select benefits we offer: enhanced maternity, paternity, adoption, and shared parental leave, private medical and dental insurance for yourself and any dependents, and flexible working options. We strive to continually improve our working environment, and provide you with excellent facilities such as healthy food, an on-site gym, faith rooms, terraces etc.</p>
<p>We are also open to relocating candidates and offer a bespoke service and immigration support to make it as easy as possible (depending on eligibility).</p>
<p>The US base salary range for this full-time position is between $136,000 - $245,000 + bonus + equity + benefits. Your recruiter can share more about the specific salary range for your targeted location during the hiring process.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$136,000 - $245,000 + bonus + equity + benefits</Salaryrange>
      <Skills>Python, Deep learning, Foundation models, Risk assessment, Mitigation, Forecasting, Scenario planning, LLM training and inference, PhD in Computer Science or Machine Learning related field, Track record of publications at venues such as NeurIPS, ICLR, ICML, RL/DL, EMNLP, AAAI and UAI, Experience with collaborating or leading an applied research project</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Google DeepMind</Employername>
      <Employerlogo>https://logos.yubhub.co/deepmind.com.png</Employerlogo>
      <Employerdescription>Google DeepMind is a subsidiary of Alphabet Inc., a multinational conglomerate headquartered in Mountain View, California.</Employerdescription>
      <Employerwebsite>https://deepmind.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/deepmind/jobs/7493360</Applyto>
      <Location>London, UK; New York City, New York, US; San Francisco, California, US</Location>
      <Country></Country>
      <Postedate>2026-03-16</Postedate>
    </job>
    <job>
      <externalid>b151fcc2-2fb</externalid>
      <Title>Member of Technical Staff, High Performance Computing Engineer</Title>
      <Description><![CDATA[<p>We are looking for experienced Member of Technical Staff, High Performance Computing Engineers to help build and scale the infrastructure that trains our frontier models and powers the next evolution of our personal AI, Copilot.</p>
<p>This role offers the unique opportunity to work on some of the largest scale supercomputers in the world – a rare chance to operate at such a significant scale.</p>
<p><strong>Responsibilities</strong></p>
<p>Design, operate, and maintain large-scale HPC environments, drawing on hands-on engineering experience in production settings.</p>
<p>Own the deployment, configuration, and day-to-day operation of HPC schedulers (e.g., SLURM, Kubernetes), ensuring reliable and efficient job scheduling at scale.</p>
<p>Serve as a technical owner for at least one core HPC domain (GPU compute, high-performance storage, networking, or similar), including ongoing maintenance, performance tuning, and troubleshooting of massive clusters.</p>
<p>Develop and maintain automation and tooling using Bash and/or Python to improve cluster reliability, observability, and operational efficiency.</p>
<p>Partner closely with researchers and engineers to support their workloads, troubleshoot cluster usage issues, and triage failed or underperforming jobs to resolution.</p>
<p>Drive work forward independently by navigating ambiguity and technical roadblocks, delivering incremental improvements that get capabilities into users’ hands quickly.</p>
<p><strong>Qualifications</strong></p>
<p>Do you have a Bachelor’s degree in computer science, or related technical field AND 4+ years technical engineering experience with deploying or operating on-premise or cloud high-performance clusters, AND 4+ years experience working with high-scale training clusters (ex. working with frameworks/tools such as nvidia InfiniBand clusters, SLURM, Kubernetes, Ray, etc.), AND 4+ years experience building scalable services on top of public cloud infrastructure like Azure, AWS, or GCP, OR equivalent experience?</p>
<p><strong>Preferred Qualifications</strong></p>
<p>Master’s Degree in Computer Science or related technical field AND 6+ years technical engineering experience with deploying or operating on-premise or cloud high-performance clusters, AND 6+ years experience working with high-scale training clusters (ex. working with frameworks/tools such as nvidia InfiniBand clusters, SLURM, Kubernetes, Ray, etc.), AND 6+ years experience building scalable services on top of public cloud infrastructure like Azure, AWS, or GCP, OR equivalent experience.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>HPC, SLURM, Kubernetes, GPU compute, high-performance storage, networking, Bash, Python, nvidia InfiniBand clusters, Ray, LLM training clusters, AI platforms, Machine Learning frameworks, large-scale HPC or GPU systems</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that develops and markets software, services, and solutions for personal and business use. It is one of the largest and most influential technology companies in the world.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-high-performance-computing-engineer-mai-superintelligence-team-3/</Applyto>
      <Location>Zürich</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>34f86990-9e0</externalid>
      <Title>Machine Learning Systems Engineer, RL Engineering</Title>
      <Description><![CDATA[<p><strong>About the role:</strong></p>
<p>You want to build the cutting-edge systems that train AI models like Claude. You&#39;re excited to work at the frontier of machine learning, implementing and improving advanced techniques to create ever more capable, reliable and steerable AI. As an ML Systems Engineer on our Reinforcement Learning Engineering team, you&#39;ll be responsible for the critical algorithms and infrastructure that our researchers depend on to train models. Your work will directly enable breakthroughs in AI capabilities and safety. You&#39;ll focus obsessively on improving the performance, robustness, and usability of these systems so our research can progress as quickly as possible. You&#39;re energized by the challenge of supporting and empowering our research team in the mission to build beneficial AI systems.</p>
<p>Our finetuning researchers train our production Claude models, and internal research models, using RLHF and other related methods. Your job will be to build, maintain, and improve the algorithms and systems that these researchers use to train models. You’ll be responsible for improving the speed, reliability, and ease-of-use of these systems.</p>
<p><strong>You may be a good fit if you:</strong></p>
<ul>
<li>Have 4+ years of software engineering experience</li>
<li>Like working on systems and tools that make other people more productive</li>
<li>Are results-oriented, with a bias towards flexibility and impact</li>
<li>Pick up slack, even if it goes outside your job description</li>
<li>Enjoy pair programming (we love to pair!)</li>
<li>Want to learn more about machine learning research</li>
<li>Care about the societal impacts of your work</li>
</ul>
<p><strong>Strong candidates may also have experience with:</strong></p>
<ul>
<li>High performance, large scale distributed systems</li>
<li>Large scale LLM training</li>
<li>Python</li>
<li>Implementing LLM finetuning algorithms, such as RLHF</li>
</ul>
<p><strong>Representative projects:</strong></p>
<ul>
<li>Profiling our reinforcement learning pipeline to find opportunities for improvement</li>
<li>Building a system that regularly launches training jobs in a test environment so that we can quickly detect problems in the training pipeline</li>
<li>Making changes to our finetuning systems so they work on new model architectures</li>
<li>Building instrumentation to detect and eliminate Python GIL contention in our training code</li>
<li>Diagnosing why training runs have started slowing down after some number of steps, and fixing it</li>
<li>Implementing a stable, fast version of a new training algorithm proposed by a researcher</li>
</ul>
<p><strong>Logistics</strong></p>
<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>
<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work. We think AI systems like the ones we&#39;re building have enormous social and ethical implications. We think this makes representation even more important, and we strive to include a range of diverse perspectives on our team.</p>
<p><strong>Your safety matters to us.</strong> To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</p>
<p><strong>How we&#39;re different</strong></p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>
<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$500,000 - $850,000USD</Salaryrange>
      <Skills>High performance, large scale distributed systems, Large scale LLM training, Python, Implementing LLM finetuning algorithms, such as RLHF, High performance, large scale distributed systems, Large scale LLM training, Python, Implementing LLM finetuning algorithms, such as RLHF</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a company that aims to create reliable, interpretable, and steerable AI systems. It has a team of researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</Employerdescription>
      <Employerwebsite>https://job-boards.greenhouse.io</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/4952051008</Applyto>
      <Location>San Francisco, CA | New York City, NY | Seattle, WA</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>b0188062-45f</externalid>
      <Title>Research Engineer, Cybersecurity Reinforcement Learning</Title>
      <Description><![CDATA[<p><strong>About the role</strong></p>
<p>We&#39;re hiring for the Cybersecurity RL team within Horizons. As a Research Engineer, you&#39;ll help to safely advance the capabilities of our models in secure coding, vulnerability remediation, and other areas of defensive cybersecurity.</p>
<p>This role blends research and engineering, requiring you to both develop novel approaches and realise them in code. Your work will include designing and implementing RL environments, conducting experiments and evaluations, delivering your work into production training runs, and collaborating with other researchers, engineers, and cybersecurity specialists across and outside Anthropic.</p>
<p><strong>You may be a good fit if you:</strong></p>
<ul>
<li>Have experience in cybersecurity research.</li>
<li>Have experience with machine learning.</li>
<li>Have strong software engineering skills.</li>
<li>Can balance research exploration with engineering implementation.</li>
<li>Are passionate about AI&#39;s potential and committed to developing safe and beneficial systems.</li>
</ul>
<p><strong>Strong candidates may also have:</strong></p>
<ul>
<li>Professional experience in security engineering, fuzzing, detection and response, or other applied defensive work.</li>
<li>Experience participating in or building CTF competitions and cyber ranges.</li>
<li>Academic research experience in cybersecurity.</li>
<li>Familiarity with RL techniques and environments.</li>
<li>Familiarity with LLM training methodologies.</li>
</ul>
<p><strong>Logistics</strong></p>
<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>
<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p><strong>How we&#39;re different</strong></p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time.</p>
<p><strong>Come work with us!</strong></p>
<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lot more.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$300,000 - $405,000 USD</Salaryrange>
      <Skills>cybersecurity research, machine learning, software engineering, RL techniques and environments, LLM training methodologies, security engineering, fuzzing, detection and response, CTF competitions and cyber ranges, academic research in cybersecurity</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation headquartered in San Francisco, focused on creating reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://job-boards.greenhouse.io</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5025624008</Applyto>
      <Location>San Francisco, CA, New York City, NY</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>c1d20281-7ee</externalid>
      <Title>Member of Technical Staff, High Performance Computing Engineer</Title>
      <Description><![CDATA[<p><strong>Summary</strong></p>
<p>Microsoft AI are looking for a talented Member of Technical Staff, High Performance Computing Engineer at their London office. This role sits at the heart of building and scaling the infrastructure that trains their frontier models and powers the next evolution of their personal AI, Copilot. You&#39;ll work directly with researchers and engineers to support their workloads, troubleshoot cluster usage issues, and triage failed or underperforming jobs to resolution.</p>
<p><strong>About the Role</strong></p>
<p>As a Member of Technical Staff, High Performance Computing Engineer, you will design, operate, and maintain large-scale HPC environments, drawing on hands-on engineering experience in production settings. You will own the deployment, configuration, and day-to-day operation of HPC schedulers (e.g., SLURM, Kubernetes), ensuring reliable and efficient job scheduling at scale. You will serve as a technical owner for at least one core HPC domain (GPU compute, high-performance storage, networking, or similar), including ongoing maintenance, performance tuning, and troubleshooting of massive clusters.</p>
<p><strong>Accountabilities</strong></p>
<ul>
<li>Design, operate, and maintain large-scale HPC environments, drawing on hands-on engineering experience in production settings.</li>
<li>Own the deployment, configuration, and day-to-day operation of HPC schedulers (e.g., SLURM, Kubernetes), ensuring reliable and efficient job scheduling at scale.</li>
<li>Serve as a technical owner for at least one core HPC domain (GPU compute, high-performance storage, networking, or similar), including ongoing maintenance, performance tuning, and troubleshooting of massive clusters.</li>
</ul>
<p><strong>The Candidate we&#39;re looking for</strong></p>
<p><strong>Experience:</strong></p>
<ul>
<li>4+ years technical engineering experience with deploying or operating on-premise or cloud high-performance clusters.</li>
<li>4+ years experience working with high-scale training clusters (ex. working with frameworks/tools such as nvidia InfiniBand clusters, SLURM, Kubernetes, Ray, etc.).</li>
<li>4+ years experience building scalable services on top of public cloud infrastructure like Azure, AWS, or GCP.</li>
</ul>
<p><strong>Technical skills:</strong></p>
<ul>
<li>Experience with LLM training clusters.</li>
<li>Experience working with AI platforms, frameworks, and APIs.</li>
<li>Experience using Machine Learning frameworks, including experience using, deploying, and scaling language learning models, either personally or professionally.</li>
</ul>
<p><strong>Personal attributes:</strong></p>
<ul>
<li>Ability to identify, analyze, and resolve complex technical issues, ensuring optimal performance, scalability, and user experience.</li>
<li>Dedication to writing clean, maintainable, and well-documented code with a focus on application quality, performance, and security.</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Competitive salary and benefits package.</li>
<li>Opportunity to work with a leading technology company and contribute to HERE&#39;s mission.</li>
<li>Collaborative and dynamic work environment.</li>
<li>Professional development opportunities.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>Competitive salary and benefits package</Salaryrange>
      <Skills>High Performance Computing, Cloud Infrastructure, Machine Learning, AI Platforms, Frameworks and APIs, LLM Training Clusters, AI Platforms, Frameworks and APIs</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that empowers every person and every organization on the planet to achieve more. They come together with a growth mindset, innovate to empower others, and collaborate to realize their shared goals.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-high-performance-computing-engineer-mai-superintelligence-team-2/</Applyto>
      <Location>London</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
    <job>
      <externalid>7abfb827-590</externalid>
      <Title>Member of Technical Staff, High Performance Computing Engineer</Title>
      <Description><![CDATA[<p><strong>Summary</strong></p>
<p>Microsoft AI are looking for experienced Member of Technical Staff, High Performance Computing Engineers to help build and scale the infrastructure that trains their frontier models and powers the next evolution of their personal AI, Copilot.</p>
<p><strong>About the Role</strong></p>
<p>This role offers the unique opportunity to work on some of the largest scale supercomputers in the world – a rare chance to operate at such a significant scale. As a Member of Technical Staff, High Performance Computing Engineer, you will design, operate, and maintain large-scale HPC environments, drawing on hands-on engineering experience in production settings. You will own the deployment, configuration, and day-to-day operation of HPC schedulers (e.g., SLURM, Kubernetes), ensuring reliable and efficient job scheduling at scale.</p>
<p><strong>Accountabilities</strong></p>
<ul>
<li>Design, operate, and maintain large-scale HPC environments, drawing on hands-on engineering experience in production settings.</li>
<li>Own the deployment, configuration, and day-to-day operation of HPC schedulers (e.g., SLURM, Kubernetes), ensuring reliable and efficient job scheduling at scale.</li>
</ul>
<p><strong>The Candidate we&#39;re looking for</strong></p>
<p><strong>Experience:</strong></p>
<ul>
<li>4+ years technical engineering experience with deploying or operating on-premise or cloud high-performance clusters.</li>
<li>4+ years experience working with high-scale training clusters (ex. working with frameworks/tools such as nvidia InfiniBand clusters, SLURM, Kubernetes, Ray, etc.).</li>
<li>4+ years experience building scalable services on top of public cloud infrastructure like Azure, AWS, or GCP.</li>
</ul>
<p><strong>Technical skills:</strong></p>
<ul>
<li>Experience with LLM training clusters.</li>
<li>Experience working with AI platforms, frameworks, and APIs.</li>
<li>Experience using Machine Learning frameworks, including experience using, deploying, and scaling language learning models, either personally or professionally.</li>
</ul>
<p><strong>Personal attributes:</strong></p>
<ul>
<li>Ability to identify, analyze, and resolve complex technical issues, ensuring optimal performance, scalability, and user experience.</li>
<li>Dedication to writing clean, maintainable, and well-documented code with a focus on application quality, performance, and security.</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Competitive salary.</li>
<li>Comprehensive benefits package.</li>
<li>Opportunities for professional growth and development.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>Competitive salary</Salaryrange>
      <Skills>High Performance Computing, Cloud Infrastructure, Machine Learning, AI Platforms, Frameworks and APIs, LLM Training Clusters, AI Platforms, Frameworks and APIs</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that empowers every person and every organization on the planet to achieve more. They come together with a growth mindset, innovate to empower others, and collaborate to realize their shared goals.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-high-performance-computing-engineer-mai-superintelligence-team/</Applyto>
      <Location>Multiple Locations, United States</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
    <job>
      <externalid>ba52acc3-4fd</externalid>
      <Title>Engineering Site Lead</Title>
      <Description><![CDATA[<p>We&#39;re seeking an exceptional Site Lead to establish and scale our London office. This is a unique opportunity to shape Perplexity&#39;s presence in one of the world&#39;s leading tech hubs, building teams and culture from the ground up while driving technical excellence in infrastructure and AI systems.</p>
<p><strong>What you&#39;ll do</strong></p>
<p>As Site Lead, you&#39;ll serve as the face of Perplexity in London, responsible for building our technical organization, fostering a world-class engineering culture, and directly managing one or more infrastructure teams. You&#39;ll report to senior leadership and work cross-functionally with teams across our global footprint.</p>
<p><strong>What you need</strong></p>
<ul>
<li>10+ years of experience in software engineering with 5+ years in infrastructure, cloud infrastructure, or AI infrastructure roles</li>
<li>3+ years of people management experience, including building and scaling teams</li>
<li>Proven track record of establishing or significantly growing an engineering site or office</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>distributed systems, cloud platforms, infrastructure automation, GPU infrastructure and orchestration, ML training and inference pipelines, Model serving and deployment at scale, Kubernetes, Terraform, container orchestration, CI/CD systems, experience at companies focused on AI/ML, search, or large-scale consumer applications, previous experience as a site lead, office lead, or similar multi-team leadership role, background in building infrastructure for LLM training or inference, contributions to open-source infrastructure or AI infrastructure projects, experience scaling teams from 0 to 20+ engineers, active involvement in the London or European tech community</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Perplexity</Employername>
      <Employerlogo>https://logos.yubhub.co/perplexity.com.png</Employerlogo>
      <Employerdescription>Perplexity is revolutionizing how people discover and interact with information through AI-powered search and knowledge tools. As we expand our global footprint, we&apos;re establishing a strategic presence in London to drive innovation and growth across Europe.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/perplexity/638e6823-be7f-46c6-9675-7b1197fc9b8c</Applyto>
      <Location>London</Location>
      <Country></Country>
      <Postedate>2026-03-04</Postedate>
    </job>
    <job>
      <externalid>46711770-4ab</externalid>
      <Title>AI Researcher</Title>
      <Description><![CDATA[<p>Perplexity is seeking top-tier AI Research Scientists and Engineers to advance our AI products and capabilities. We&#39;re building the future of AI-powered search and agent experiences through our Sonar models, Deep Research Agent, Comet Agent, and Search products. Join us in creating SOTA experiences that handle hundreds of millions of queries and continue to scale rapidly.</p>
<p><strong>What you&#39;ll do</strong></p>
<p>Research &amp; Development</p>
<ul>
<li>Post-train SOTA LLMs using the latest supervised and reinforcement learning techniques (SFT/DPO/GRPO)</li>
<li>Leverage our rich query/answer dataset to scale model performance across Sonar, Deep Research, Comet, and Search products</li>
</ul>
<p><strong>What you need</strong></p>
<ul>
<li>Proven experience with large-scale LLMs and Deep Learning systems</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$220K – $485K</Salaryrange>
      <Skills>large-scale LLMs, Deep Learning systems, Python/PyTorch, post-training techniques, reinforcement learning, PhD in Machine Learning, AI, Systems, or related areas, C++/CUDA programming skills, experience building LLM training frameworks, academic publications and research impact, experience with agent systems and multi-step reasoning, background in personalization and preference learning</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Perplexity</Employername>
      <Employerlogo>https://logos.yubhub.co/perplexity.com.png</Employerlogo>
      <Employerdescription>Perplexity is a company seeking top-tier AI Research Scientists and Engineers to advance their AI products and capabilities. They&apos;re building the future of AI-powered search and agent experiences through their Sonar models, Deep Research Agent, Comet Agent, and Search products.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/perplexity/8fe61c73-0daf-4432-a47d-44714c1ef764</Applyto>
      <Location>San Francisco, Palo Alto</Location>
      <Country></Country>
      <Postedate>2026-03-04</Postedate>
    </job>
  </jobs>
</source>