<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>cd02d1a1-0e8</externalid>
      <Title>Communications Lead, Claude Code</Title>
      <Description><![CDATA[<p>We&#39;re looking for a Communications Lead to own comms for Claude Code. You&#39;ll sit on the Product Communications team, working day-to-day with the Claude Code product team, developer relations, and marketing.</p>
<p>The media landscape for developer tools doesn&#39;t look like it did five years ago. We need someone who understands both traditional press and the channels where developers form opinions. You might have come up through an in-house comms team, or you might have run launches inside product marketing, handled press from a DevRel role, or found your way to this work from somewhere adjacent.</p>
<p>You should be a Claude Code user yourself and know the product well.</p>
<p>Responsibilities:</p>
<ul>
<li>Own communications for Claude Code, from the big launches to the steady rhythm of updates, community moments, and everything in between</li>
<li>Build and maintain strong relationships with journalists, newsletter writers, podcasters, and creators covering dev tools and the AI ecosystem</li>
<li>Lead cross-functional product launch communications, coordinating messaging across comms, marketing, developer relations, and product</li>
<li>Advise leadership and DevRel when things move fast or catch fire, whether it’s an incident or a community thread</li>
<li>Translate complex technical work into stories that land with developers and still make sense to broader audiences</li>
<li>Develop messaging frameworks and content strategies that work across technical and non-technical audiences</li>
<li>Prepare Claude Code engineers and product leads for external moments: podcasts, talks, press, etc.</li>
<li>Think across channels (press, social, community, owned) and know which lever to pull for each moment</li>
<li>Pay attention to what&#39;s actually working and build the program from there</li>
</ul>
<p>You may be a good fit if you:</p>
<ul>
<li>Have 8–12 years of experience in communications, PR, or developer marketing, with meaningful time focused on technical products or developer audiences</li>
<li>Use Claude Code heavily and can talk specifically about how you use it in your day-to-day</li>
<li>Are high-agency and low-ego, with a bias to action</li>
<li>Write clearly and concisely, whether it&#39;s a launch post or a cross-functional update, a lot of context moves through this role and people need to be able to follow it</li>
<li>Have a deep understanding of both traditional media channels and the emerging platforms where technical communities engage</li>
<li>Are very online, follow the right people, know what&#39;s moving through Hacker News and developer social chatter, and catch things early</li>
<li>Have real fluency in developer culture and know how trust gets earned there</li>
</ul>
<p>Strong candidates may also</p>
<ul>
<li>Have experience at developer tools companies, infrastructure products, or open source projects</li>
<li>Have an existing network in developer media, technical journalism, or the creator space</li>
<li>Have experience managing communications for AI or ML products</li>
</ul>
<p>The annual compensation range for this role is $185,000-$255,000 USD.</p>
<p>Logistics</p>
<ul>
<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>
<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>
<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>
<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>
<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>
</ul>
<p>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>
<p>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links,visit anthropic.com/careers directly for confirmed position openings.</p>
<p>How we&#39;re different</p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>
<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>
<p>Come work with us!</p>
<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$185,000-$255,000 USD</Salaryrange>
      <Skills>communications, PR, developer marketing, technical products, developer audiences, AI, ML, GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, Learning from Human Preferences</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5153586008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>540ce49c-271</externalid>
      <Title>Member of Technical Staff - Multimodal Understanding</Title>
      <Description><![CDATA[<p><strong>About the Role</strong></p>
<p>You will join the multimodal team to push toward superhuman multimodal intelligence. Advance understanding and generation across modalities,image, video, audio, and text,spanning the full stack: data curation/acquisition, tokenizer training, large-scale pre-training, post-training/alignment, infrastructure/scaling, evaluation, tooling/demos, and end-to-end product experiences.</p>
<p>Collaborate cross-functionally with pre-training, post-training, reasoning, data, applied, and product teams to deliver frontier capabilities in multimodal reasoning, world modeling, tool use, agentic behaviors, and interactive human-AI collaboration. Contribute to building models that can see, hear, reason about, and interact with the world in real time at unprecedented levels.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Design, build, and optimize large-scale distributed systems for multimodal pre-training, post-training, inference, data processing, and tokenization at web/petabyte scale.</li>
<li>Develop high-throughput pipelines for data acquisition, preprocessing, filtering, generation, decoding, loading, crawling, visualization, and management (images, videos, audio + text).</li>
<li>Advance multimodal capabilities including spatial-temporal compression, cross-modal alignment, world modeling, reasoning, emergent abilities, audio/image/video understanding &amp; generation, real-time video processing, and noisy data handling.</li>
<li>Drive data quality and studies: curation (human/synthetic), filtering techniques, analysis, and scalable pipelines to support trillion-parameter models.</li>
<li>Create evaluation frameworks, internal benchmarks, reward models, and metrics that capture real-world usage, failure modes, interactive dynamics, and human-AI synergy.</li>
<li>Innovate on algorithms, modeling approaches, hardware/software/algorithm co-design, and scaling paradigms for state-of-the-art performance.</li>
<li>Build research tooling, user-friendly interfaces, prototypes/demos, full-stack applications, and enable rapid iteration based on feedback.</li>
<li>Work across the stack (pre-training → SFT/RL/post-training) to enable reasoning, tool calling, agentic behaviors, orchestration, and seamless real-time interactions.</li>
</ul>
<p><strong>Basic Qualifications</strong></p>
<ul>
<li>Hands-on experience with multimodal pre-training, post-training, or fine-tuning (vision, audio, video, or cross-modal).</li>
<li>Expert-level proficiency in Python (core language), with strong experience in at least one of: JAX / PyTorch / XLA.</li>
<li>Proven track record building or optimizing large-scale distributed ML systems (training/inference optimization, GPU utilization, multi-GPU/TPU setups, hardware co-design).</li>
<li>Deep experience designing and running data pipelines at scale: curation, filtering, generation, quality studies, especially for noisy/real-world multimodal data.</li>
<li>Strong fundamentals in evaluation design, benchmarks, reward modeling, or RL techniques (particularly for interactive/agentic behaviors).</li>
<li>Proactive self-starter who thrives in high-intensity environments and is passionate about pushing multimodal AI frontiers.</li>
<li>Willingness to own end-to-end initiatives and do whatever it takes to deliver breakthrough user experiences.</li>
</ul>
<p><strong>Preferred Skills and Experience</strong></p>
<ul>
<li>Experience leading major improvements in model capabilities through better data, modeling, algorithms, or scaling.</li>
<li>Familiarity with state-of-the-art in multimodal LLMs, scaling laws, tokenizers, compression techniques, reasoning, or agentic systems.</li>
<li>Proficiency in Rust and/or C++ for performance-critical components.</li>
<li>Hands-on work with large-scale orchestration tools such as Spark, Ray, or Kubernetes.</li>
<li>Background building full-stack tooling: performant interfaces, real-time research demos/apps, or end-to-end product ownership.</li>
<li>Passion for end-to-end user experience in interactive, real-time multimodal AI systems.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$180,000 - $440,000 USD</Salaryrange>
      <Skills>Multimodal pre-training, Post-training, Fine-tuning, Python, JAX, PyTorch, XLA, Large-scale distributed ML systems, Data pipelines, Evaluation design, Benchmarks, Reward modeling, RL techniques, State-of-the-art in multimodal LLMs, Scaling laws, Tokenizers, Compression techniques, Reasoning, Agentic systems, Rust, C++, Spark, Ray, Kubernetes, Full-stack tooling</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>xAI</Employername>
      <Employerlogo>https://logos.yubhub.co/xai.com.png</Employerlogo>
      <Employerdescription>xAI creates AI systems that can accurately understand the universe and aid humanity in its pursuit of knowledge.</Employerdescription>
      <Employerwebsite>https://www.xai.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/xai/jobs/5111374007?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Palo Alto, CA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>6aa46bac-783</externalid>
      <Title>Software Engineer, Cybersecurity Products</Title>
      <Description><![CDATA[<p><strong>About Anthropic</strong></p>
<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>
<p><strong>About the Role</strong></p>
<p>We&#39;re looking for engineers to join a new effort building AI-powered products and capabilities for cybersecurity. You&#39;ll work across the stack to prototype new ideas and build from the ground up.</p>
<p>This role sits at the intersection of research, product, and go-to-market. You&#39;ll work closely with research teams to develop new model capabilities for security applications, prototype and iterate quickly to validate ideas, and engage directly with customers and partners to inform what we build. The right candidate has the technical depth to engage with research, the product instincts to know what&#39;s worth building, and the drive to move fast.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Prototype and build new AI-powered products for cybersecurity</li>
</ul>
<ul>
<li>Iterate quickly based on customer feedback and what you learn</li>
</ul>
<ul>
<li>Collaborate with research teams to identify and develop new model capabilities for security applications</li>
</ul>
<ul>
<li>Engage directly with customers and partners to understand workflows and inform product direction</li>
</ul>
<p><strong>You may be a good fit if you:</strong></p>
<ul>
<li>Have 7+ years of experience as a software engineer</li>
</ul>
<ul>
<li>Experience developing cybersecurity products</li>
</ul>
<ul>
<li>Enjoy fast iteration and are energized by prototyping new ideas</li>
</ul>
<ul>
<li>Have strong product instincts and enjoy defining what to build, not just how to build it</li>
</ul>
<ul>
<li>Are comfortable working closely with research and go-to-market teams</li>
</ul>
<ul>
<li>Have strong communication skills and can work effectively across functions</li>
</ul>
<p><strong>Strong candidates may also have:</strong></p>
<ul>
<li>Experience in incident response, reverse engineering, network analysis, penetration testing, or similar fields</li>
</ul>
<ul>
<li>Experience working with AI/ML models and building products on top of them</li>
</ul>
<ul>
<li>Experience building agentic applications</li>
</ul>
<p><strong>Deadline to apply:</strong></p>
<p>None. Applications will be reviewed on a rolling basis.</p>
<p><strong>Logistics</strong></p>
<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>
<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p><strong>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</strong></p>
<p><strong>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</strong></p>
<p><strong>How we&#39;re different</strong></p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>
<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>
<p><strong>Come work with us!</strong></p>
<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues. <strong>Guidance on Candidates&#39; AI Usage:</strong> Learn about our policy for using AI in our application process</p>
<p>Interested in building your career at Anthropic? Get future opportunities by following us on LinkedIn and Twitter.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$320,000 - $405,000 USD</Salaryrange>
      <Skills>software engineer, cybersecurity products, AI/ML models, incident response, reverse engineering, network analysis, penetration testing, agentic applications, circuit-based interpretability, multimodal neurons, scaling laws, AI &amp; compute, concrete problems in AI safety, learning from human preferences</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation headquartered in San Francisco, with a mission to create reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://job-boards.greenhouse.io</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5063007008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco, CA | New York City, NY | Seattle, WA; Washington, DC</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>5d38ab71-400</externalid>
      <Title>Research Engineer, Pretraining Scaling</Title>
      <Description><![CDATA[<p><strong>About Anthropic</strong></p>
<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>
<p><strong>About the Role:</strong></p>
<p>Anthropic&#39;s ML Performance and Scaling team trains our production pretrained models, work that directly shapes the company&#39;s future and our mission to build safe, beneficial AI systems. As a Research Engineer on this team, you&#39;ll ensure our frontier models train reliably, efficiently, and at scale. This is demanding, high-impact work that requires both deep technical expertise and a genuine passion for the craft of large-scale ML systems.</p>
<p>This role lives at the boundary between research and engineering. You&#39;ll work across our entire production training stack: performance optimisation, hardware debugging, experimental design, and launch coordination. During launches, the team works in tight lockstep, responding to production issues that can&#39;t wait for tomorrow.</p>
<p><strong>Responsibilities:</strong></p>
<ul>
<li>Own critical aspects of our production pretraining pipeline, including model operations, performance optimisation, observability, and reliability</li>
<li>Debug and resolve complex issues across the full stack—from hardware errors and networking to training dynamics and evaluation infrastructure</li>
<li>Design and run experiments to improve training efficiency, reduce step time, increase uptime, and enhance model performance</li>
<li>Respond to on-call incidents during model launches, diagnosing problems quickly and coordinating solutions across teams</li>
<li>Build and maintain production logging, monitoring dashboards, and evaluation infrastructure</li>
<li>Add new capabilities to the training codebase, such as long context support or novel architectures</li>
<li>Collaborate closely with teammates across SF and London, as well as with Tokens, Architectures, and Systems teams</li>
<li>Contribute to the team&#39;s institutional knowledge by documenting systems, debugging approaches, and lessons learned</li>
</ul>
<p><strong>You May Be a Good Fit If You:</strong></p>
<ul>
<li>Have hands-on experience training large language models, or deep expertise with JAX, TPU, PyTorch, or large-scale distributed systems</li>
<li>Genuinely enjoy both research and engineering work—you&#39;d describe your ideal split as roughly 50/50 rather than heavily weighted toward one or the other</li>
<li>Are excited about being on-call for production systems, working long days during launches, and solving hard problems under pressure</li>
<li>Thrive when working on whatever is most impactful, even if that changes day-to-day based on what the production model needs</li>
<li>Excel at debugging complex, ambiguous problems across multiple layers of the stack</li>
<li>Communicate clearly and collaborate effectively, especially when coordinating across time zones or during high-stress incidents</li>
<li>Are passionate about the work itself and want to refine your craft as a research engineer</li>
<li>Care about the societal impacts of AI and responsible scaling</li>
</ul>
<p><strong>Strong Candidates May Also Have:</strong></p>
<ul>
<li>Previous experience training LLM’s or working extensively with JAX/TPU, PyTorch, or other ML frameworks at scale</li>
<li>Contributed to open-source LLM frameworks (e.g., open\_lm, llm-foundry, mesh-transformer-jax)</li>
<li>Published research on model training, scaling laws, or ML systems</li>
<li>Experience with production ML systems, observability tools, or evaluation infrastructure</li>
<li>Background as a systems engineer, quant, or in other roles requiring both technical depth and operational excellence</li>
</ul>
<p><strong>What Makes This Role Unique:</strong></p>
<p>This is not a typical research engineering role. The work is highly operational—you&#39;ll be deeply involved in keeping our production models training smoothly, which means being responsive to incidents, flexible about priorities, and comfortable with uncertainty. During launches, the team often works extended hours and may need to respond to issues on evenings and weekends.</p>
<p>However, this operational intensity comes with extraordinary learning opportunities. You&#39;ll gain hands-on experience with some of the largest, most sophisticated training runs in the industry. You&#39;ll work alongside world-class researchers and engineers, and the institutional knowledge you build will compound in ways that can&#39;t be easily transferred. For people who thrive on this type of work, it&#39;s uniquely rewarding.</p>
<p>We&#39;re building a close-knit team of people who genuinely care about doing excellent work together. If you&#39;re someone who wants to be part of training the models that will define the future of AI—and you&#39;re excited about the full reality of what that entails—we&#39;d love to hear from you.</p>
<p><strong>Logistics</strong></p>
<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>
<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong></p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $850,000USD</Salaryrange>
      <Skills>JAX, TPU, PyTorch, large-scale distributed systems, model operations, performance optimisation, observability, reliability, model training, scaling laws, ML systems, open-source LLM frameworks, production ML systems, observability tools, evaluation infrastructure, systems engineer, quant</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a growing organisation working on creating reliable, interpretable, and steerable AI systems. Their mission is to build safe and beneficial AI systems for users and society.</Employerdescription>
      <Employerwebsite>https://job-boards.greenhouse.io</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/4938432008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>a05bfa1a-d23</externalid>
      <Title>Research Engineer, Pretraining Scaling</Title>
      <Description><![CDATA[<p><strong>About Anthropic</strong></p>
<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>
<p><strong>About the Role:</strong></p>
<p>Anthropic&#39;s ML Performance and Scaling team trains our production pretrained models, work that directly shapes the company&#39;s future and our mission to build safe, beneficial AI systems. As a Research Engineer on this team, you&#39;ll ensure our frontier models train reliably, efficiently, and at scale. This is demanding, high-impact work that requires both deep technical expertise and a genuine passion for the craft of large-scale ML systems.</p>
<p>This role lives at the boundary between research and engineering. You&#39;ll work across our entire production training stack: performance optimization, hardware debugging, experimental design, and launch coordination. During launches, the team works in tight lockstep, responding to production issues that can&#39;t wait for tomorrow.</p>
<p><strong>Responsibilities:</strong></p>
<ul>
<li>Own critical aspects of our production pretraining pipeline, including model operations, performance optimization, observability, and reliability</li>
<li>Debug and resolve complex issues across the full stack—from hardware errors and networking to training dynamics and evaluation infrastructure</li>
<li>Design and run experiments to improve training efficiency, reduce step time, increase uptime, and enhance model performance</li>
<li>Respond to on-call incidents during model launches, diagnosing problems quickly and coordinating solutions across teams</li>
<li>Build and maintain production logging, monitoring dashboards, and evaluation infrastructure</li>
<li>Add new capabilities to the training codebase, such as long context support or novel architectures</li>
<li>Collaborate closely with teammates across SF and London, as well as with Tokens, Architectures, and Systems teams</li>
<li>Contribute to the team&#39;s institutional knowledge by documenting systems, debugging approaches, and lessons learned</li>
</ul>
<p><strong>You May Be a Good Fit If You:</strong></p>
<ul>
<li>Have hands-on experience training large language models, or deep expertise with JAX, TPU, PyTorch, or large-scale distributed systems</li>
<li>Genuinely enjoy both research and engineering work—you&#39;d describe your ideal split as roughly 50/50 rather than heavily weighted toward one or the other</li>
<li>Are excited about being on-call for production systems, working long days during launches, and solving hard problems under pressure</li>
<li>Thrive when working on whatever is most impactful, even if that changes day-to-day based on what the production model needs</li>
<li>Excel at debugging complex, ambiguous problems across multiple layers of the stack</li>
<li>Communicate clearly and collaborate effectively, especially when coordinating across time zones or during high-stress incidents</li>
<li>Are passionate about the work itself and want to refine your craft as a research engineer</li>
<li>Care about the societal impacts of AI and responsible scaling</li>
</ul>
<p><strong>Strong Candidates May Also Have:</strong></p>
<ul>
<li>Previous experience training LLM’s or working extensively with JAX/TPU, PyTorch, or other ML frameworks at scale</li>
<li>Contributed to open-source LLM frameworks (e.g., open\_lm, llm-foundry, mesh-transformer-jax)</li>
<li>Published research on model training, scaling laws, or ML systems</li>
<li>Experience with production ML systems, observability tools, or evaluation infrastructure</li>
<li>Background as a systems engineer, quant, or in other roles requiring both technical depth and operational excellence</li>
</ul>
<p><strong>What Makes This Role Unique:</strong></p>
<p>This is not a typical research engineering role. The work is highly operational—you&#39;ll be deeply involved in keeping our production models training smoothly, which means being responsive to incidents, flexible about priorities, and comfortable with uncertainty. During launches, the team often works extended hours and may need to respond to issues on evenings and weekends.</p>
<p>However, this operational intensity comes with extraordinary learning opportunities. You&#39;ll gain hands-on experience with some of the largest, most sophisticated training runs in the industry. You&#39;ll work alongside world-class researchers and engineers, and the institutional knowledge you build will compound in ways that can&#39;t be easily transferred. For people who thrive on this type of work, it&#39;s uniquely rewarding.</p>
<p>We&#39;re building a close-knit team of people who genuinely care about doing excellent work together. If you&#39;re someone who wants to be part of training the models that will define the future of AI—and you&#39;re excited about the full reality of what that entails—we&#39;d love to hear from you.</p>
<p><strong>Logistics</strong></p>
<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>
<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>£260,000 - £630,000GBP</Salaryrange>
      <Skills>JAX, TPU, PyTorch, large-scale distributed systems, model operations, performance optimization, observability, reliability, debugging, experimental design, launch coordination, production logging, monitoring dashboards, evaluation infrastructure, collaboration, communication, open-source LLM frameworks, research on model training, scaling laws, ML systems, production ML systems, observability tools, evaluation infrastructure, systems engineering, quant, operational excellence</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a company that creates reliable, interpretable, and steerable AI systems. It has a quickly growing team of researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/4938436008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>London</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
  </jobs>
</source>