<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>46fbbe1a-6b9</externalid>
      <Title>Software Engineer, Cloud Engineering</Title>
      <Description><![CDATA[<p>Join us in building the future of finance.</p>
<p>Our mission is to democratize finance for all. An estimated $124 trillion of assets will be inherited by younger generations in the next two decades. The largest transfer of wealth in human history. If you’re ready to be at the epicenter of this historic cultural and financial shift, keep reading.</p>
<p>We are building an elite team, applying frontier technologies to the world’s biggest financial problems. We’re looking for bold thinkers. Sharp problem-solvers. Builders who are wired to make an impact. Robinhood isn’t a place for complacency, it’s where ambitious people do the best work of their careers. We’re a high-performing, fast-moving team with ethics at the center of everything we do. Expectations are high, and so are the rewards.</p>
<p>The Bitstamp Cloud Engineering team is responsible for designing, maintaining, and scaling the AWS infrastructure that powers our global crypto exchange. As part of the Robinhood family, we are aligning our systems to support global expansion while maintaining the stability and reliability our institutional customers depend on. The team partners closely with Security, Data, Product Engineering, and Platform teams to ensure our infrastructure supports secure trading, high availability, and regulatory requirements. We value practical solutions, operational rigor, and steady progress toward automation and long-term scalability!</p>
<p>As a Senior Cloud Engineer, you will be an established individual contributor responsible for the reliability, scalability, and modernization of our AWS environment. You will manage core compute and database infrastructure while contributing to projects that align Bitstamp’s systems with Robinhood’s broader platform architecture. This role balances hands-on operational support with forward-looking automation work, reducing manual effort and improving system resilience. Your work will directly support high-volume trading systems and ensure our platform performs consistently as we scale globally!</p>
<p>This role is based in our Ljubljana office(s), with in-person attendance expected at least 3 days per week. At Robinhood, we believe in the power of in-person work to accelerate progress, spark innovation, and strengthen community. Our office experience is intentional, energizing, and designed to fully support high-performing teams.</p>
<p>Applications for this role will be accepted through April 27th, 2026 Requires participation in an on-call rotation to support business needs.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Manage and troubleshoot AWS services including EC2, ECS, Aurora, and DynamoDB to ensure high availability and performance.</li>
<li>Contribute to infrastructure projects that align Bitstamp systems with Robinhood’s global platform architecture and shared services.</li>
<li>Identify manual operational processes and implement automation using infrastructure-as-code and workflow tooling.</li>
<li>Monitor database and compute capacity, adjusting configurations to support platform growth and transaction volume.</li>
<li>Participate in the on-call rotation, diagnosing and resolving production issues to maintain 24/7 system stability.</li>
</ul>
<p><strong>Requirements</strong></p>
<ul>
<li>2–3+ years of hands-on experience in Cloud Engineering, DevOps, or Infrastructure Engineering within AWS environments.</li>
<li>Practical experience with EC2, ECS, and database technologies such as Aurora and DynamoDB.</li>
<li>Use structured troubleshooting methods and data analysis to resolve moderately complex infrastructure issues.</li>
<li>Comfortable balancing operational responsibilities with incremental automation improvements.</li>
<li>Communicate technical concepts clearly and collaborate effectively with engineering and platform teams.</li>
</ul>
<p><strong>What we offer</strong></p>
<ul>
<li>Challenging, high-impact work to grow your career</li>
<li>Performance driven compensation with multipliers for outsized impact and bonus programs</li>
<li>Top tier benefits to fuel your work, including supplemental health insurance, ancillary insurance, and mental health support programs</li>
<li>Lifestyle wallet - a highly flexible employer-paid benefits spending account expenses beyond traditional benefits such as wellness, childcare, learning, and more.</li>
<li>Time off to recharge including company holidays, paid time off, sick time, paid volunteer time off, parental leave, and more!</li>
<li>Exceptional office experience with catered meals, events, and comfortable workspaces.</li>
<li>Monthly commuter stipend to help offset in-office commuting costs</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>AWS, EC2, ECS, Aurora, DynamoDB, Cloud Engineering, DevOps, Infrastructure Engineering</Skills>
      <Category>Engineering</Category>
      <Industry>Finance</Industry>
      <Employername>Bitstamp</Employername>
      <Employerlogo>https://logos.yubhub.co/bitstamp.net.png</Employerlogo>
      <Employerdescription>Bitstamp is a cryptocurrency exchange that operates globally. It was founded in 2011 and is headquartered in Luxembourg.</Employerdescription>
      <Employerwebsite>https://www.bitstamp.net/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/robinhood/jobs/7589432?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Ljubljana, Slovenia</Location>
      <Country></Country>
      <Postedate>2026-04-25</Postedate>
    </job>
    <job>
      <externalid>5bc76aca-281</externalid>
      <Title>Research Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>
<p>We are a dynamic team passionate about AI and its potential to transform society. Our diverse workforce thrives in competitive environments and is committed to driving innovation.</p>
<p>Our teams are distributed between France, USA, UK, Germany and Singapore. We are creative, low-ego and team-spirited.</p>
<p>Join us to be part of a pioneering company shaping the future of AI.</p>
<p>Together, we can make a meaningful impact.</p>
<p>Role Summary</p>
<p>Research Engineer, Data Infrastructure</p>
<p>The Data Infrastructure team at Mistral AI is architecting the backbone of our frontier model training and fine-tuning ecosystem. We are building the specialized compute and data fabrics required to power the development of world-class AI.</p>
<p>Our vision is to operate some of the largest compute fleets in production and build data lakes and metadata systems with a roadmap toward exabyte-scale architecture.</p>
<p>We are currently in the process of building a high-performance training platform designed for massive scale across both on-premise and cloud-native Kubernetes environments.</p>
<p>We are leading a strategic transition from legacy scheduling to modern orchestration.</p>
<p>With numerous clusters distributed across various regions, we are focussed on implementing sophisticated multi-cluster orchestration and cloud-bursting capabilities to better utilize our global resources and ensure our researchers have seamless access to compute wherever it resides.</p>
<p>Our mission is to evolve our current systems into a platform that is as durable as it is flexible.</p>
<p>Location: Paris / London (hybrid) or remote EU/UK with one hub day per month.</p>
<p>About the Role</p>
<p>This role focuses on building and operating the next generation of data infrastructure at Mistral AI.</p>
<p>You will be a core contributor to our evolution, helping us design and scale massive compute fleets and storage systems designed for high performance and scalability.</p>
<p>You will help us move toward a future of decoupled control and data planes, scaling big data compute and storage platforms while ensuring secure and governed data access for MLOps and research.</p>
<p>You will take full lifecycle ownership: from architecting the migration away from legacy orchestrators to implementing production-grade pipelines and participating in on-call rotations for critical training jobs.</p>
<p>In this role, you will:</p>
<ul>
<li>Build &amp; Scale: Help us reach our goal of operating massive distributed compute and storage systems</li>
</ul>
<ul>
<li>Global Orchestration: Architect and maintain multi-cluster orchestration layers to optimize workload placement across diverse hardware and regions.</li>
</ul>
<ul>
<li>Design Future-Proof Storage: Architect our transition to modern storage formats to handle fine-tuning datasets at a scale that anticipates exabyte growth.</li>
</ul>
<ul>
<li>Platform Engineering: Contribute to the development of our internal training platform, ensuring seamless model training and fine-tuning capabilities across Kubernetes and SLURM based environments.</li>
</ul>
<ul>
<li>Metadata &amp; Lineage: Implement and manage systems to provide clear visibility and lineage as our data and model pipelines grow in complexity.</li>
</ul>
<ul>
<li>Operational Excellence: Use modern deployment workflows to manage cloud-native deployments, ensuring our data platform can scale by orders of magnitude while remaining reliable and efficient.</li>
</ul>
<p><strong>You might thrive in this role if you:</strong></p>
<ul>
<li>Have 4+ years of experience in Data Infrastructure, MLOps, or Infrastructure Engineering.</li>
</ul>
<ul>
<li>Have experience or a strong interest in supporting foundational compute and storage platforms.</li>
</ul>
<ul>
<li>Are proficient in Python and enjoy solving the &quot;brittle data lake&quot; problem with modern, columnar storage standards.</li>
</ul>
<ul>
<li>Are well-versed in Kubernetes-native tooling and excited to debug large-scale distributed systems across multi-cluster environments.</li>
</ul>
<ul>
<li>Take pride in building and operating scalable, reliable, and secure systems from the ground up.</li>
</ul>
<ul>
<li>Are comfortable with ambiguity and the challenges of building high-scale infrastructure in a rapid-growth AI environment.</li>
</ul>
<p>Benefits</p>
<p>France</p>
<ul>
<li>Competitive cash salary and equity</li>
</ul>
<ul>
<li>Food: Daily lunch vouchers</li>
</ul>
<ul>
<li>Sport: Monthly contribution to a Gympass subscription</li>
</ul>
<ul>
<li>Transportation: Monthly contribution to a mobility pass</li>
</ul>
<ul>
<li>Health: Full health insurance for you and your family</li>
</ul>
<ul>
<li>Parental: Generous parental leave policy</li>
</ul>
<ul>
<li>Visa sponsorship</li>
</ul>
<p>UK</p>
<ul>
<li>Competitive cash salary and equity</li>
</ul>
<ul>
<li>Insurance</li>
</ul>
<ul>
<li>Transportation: Reimburse office parking charges, or £90 per month for public transport</li>
</ul>
<ul>
<li>Sport: £90 per month reimbursement for gym membership</li>
</ul>
<ul>
<li>Meal voucher: £200 monthly allowance for meals</li>
</ul>
<ul>
<li>Pension plan: SmartPension (percentages are 5% Employee &amp; 3% Employer)</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Kubernetes, Data Infrastructure, MLOps, Infrastructure Engineering, Cloud-Native Deployments, Modern Deployment Workflows, Columnar Storage Standards, Distributed Systems, Multi-Cluster Environments</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo>https://logos.yubhub.co/mistral.ai.png</Employerlogo>
      <Employerdescription>Mistral AI is a company that develops high-performance, optimized, open-source and cutting-edge AI models, products and solutions.</Employerdescription>
      <Employerwebsite>https://mistral.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/071a5491-ea01-413f-ad78-f85b5e4c2215?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Paris</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>eea394f9-dab</externalid>
      <Title>Research Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>
<p>Role Summary</p>
<p>This role focuses on building and operating the next generation of data infrastructure at Mistral AI. You will be a core contributor to our evolution, helping us design and scale massive compute fleets and storage systems designed for high performance and scalability. You will help us move toward a future of decoupled control and data planes, scaling big data compute and storage platforms while ensuring secure and governed data access for MLOps and research. You will take full lifecycle ownership: from architecting the migration away from legacy orchestrators to implementing production-grade pipelines and participating in on-call rotations for critical training jobs.</p>
<p>Responsibilities</p>
<ul>
<li>Build &amp; Scale: Help us reach our goal of operating massive distributed compute and storage systems</li>
<li>Global Orchestration: Architect and maintain multi-cluster orchestration layers to optimize workload placement across diverse hardware and regions.</li>
<li>Design Future-Proof Storage: Architect our transition to modern storage formats to handle fine-tuning datasets at a scale that anticipates exabyte growth.</li>
<li>Platform Engineering: Contribute to the development of our internal training platform, ensuring seamless model training and fine-tuning capabilities across Kubernetes and SLURM based environments.</li>
<li>Metadata &amp; Lineage: Implement and manage systems to provide clear visibility and lineage as our data and model pipelines grow in complexity.</li>
<li>Operational Excellence: Use modern deployment workflows to manage cloud-native deployments, ensuring our data platform can scale by</li>
</ul>
<p>About you</p>
<ul>
<li>Have 4+ years of experience in Data Infrastructure, MLOps, or Infrastructure Engineering.</li>
<li>Have experience or a strong interest in supporting foundational compute and storage platforms.</li>
<li>Are proficient in Python and enjoy solving the &quot;brittle data lake&quot; problem with modern, columnar storage standards.</li>
<li>Are well-versed in Kubernetes-native tooling and excited to debug large-scale distributed systems across multi-cluster environments.</li>
<li>Take pride in building and operating scalable, reliable, and secure systems from the ground up.</li>
<li>Are comfortable with ambiguity and the challenges of building high-scale infrastructure in a rapid-growth AI environment.</li>
</ul>
<p>What we offer</p>
<ul>
<li>Competitive salary and equity.</li>
<li>Healthcare: Medical/Dental/Vision covered for you and your family.</li>
<li>Pension: 401K (6% matching)</li>
<li>PTO: 18 days</li>
<li>Transportation: Reimburse office parking charges, or $120/month for public transport</li>
<li>Sport: $120/month reimbursement for gym membership</li>
<li>Meal stipend: $400 monthly allowance for meals (solution might evolve as we grow bigger)</li>
<li>Visa sponsorship</li>
<li>Coaching: we offer BetterUp coaching on a voluntary basis</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Kubernetes, SLURM, Data Infrastructure, MLOps, Infrastructure Engineering</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo>https://logos.yubhub.co/mistral.ai.png</Employerlogo>
      <Employerdescription>Mistral AI is a company that develops and provides high-performance, optimized, open-source and cutting-edge AI models, products and solutions.</Employerdescription>
      <Employerwebsite>https://mistral.ai/careers</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/37f53ee5-dd88-43e3-be6a-70e3db159c8f?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Palo Alto</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>d0203959-b2e</externalid>
      <Title>Research Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>Mistral AI is a pioneering company shaping the future of AI. We believe in the power of AI to simplify tasks, save time, and enhance learning and creativity.</p>
<p>Role Summary</p>
<p>This role focuses on building and operating the next generation of data infrastructure at Mistral AI. You will be a core contributor to our evolution, helping us design and scale massive compute fleets and storage systems designed for high performance and scalability. You will help us move toward a future of decoupled control and data planes, scaling big data compute and storage platforms while ensuring secure and governed data access for MLOps and research. You will take full lifecycle ownership: from architecting the migration away from legacy orchestrators to implementing production-grade pipelines and participating in on-call rotations for critical training jobs.</p>
<p>Responsibilities</p>
<ul>
<li>Build &amp; Scale: Help us reach our goal of operating massive distributed compute and storage systems</li>
<li>Global Orchestration: Architect and maintain multi-cluster orchestration layers to optimize workload placement across diverse hardware and regions.</li>
<li>Design Future-Proof Storage: Architect our transition to modern storage formats to handle fine-tuning datasets at a scale that anticipates exabyte growth.</li>
<li>Platform Engineering: Contribute to the development of our internal training platform, ensuring seamless model training and fine-tuning capabilities across Kubernetes and SLURM based environments.</li>
<li>Metadata &amp; Lineage: Implement and manage systems to provide clear visibility and lineage as our data and model pipelines grow in complexity.</li>
<li>Operational Excellence: Use modern deployment workflows to manage cloud-native deployments, ensuring our data platform can scale by</li>
</ul>
<p>About You</p>
<ul>
<li>Have 4+ years of experience in Data Infrastructure, MLOps, or Infrastructure Engineering.</li>
<li>Have experience or a strong interest in supporting foundational compute and storage platforms.</li>
<li>Are proficient in Python and enjoy solving the &quot;brittle data lake&quot; problem with modern, columnar storage standards.</li>
<li>Are well-versed in Kubernetes-native tooling and excited to debug large-scale distributed systems across multi-cluster environments.</li>
<li>Take pride in building and operating scalable, reliable, and secure systems from the ground up.</li>
<li>Are comfortable with ambiguity and the challenges of building high-scale infrastructure in a rapid-growth AI environment.</li>
</ul>
<p>What We Offer</p>
<ul>
<li>Competitive salary and equity.</li>
<li>Healthcare: Medical/Dental/Vision covered for you and your family.</li>
<li>Pension: 401K (6% matching).</li>
<li>PTO: 18 days.</li>
<li>Transportation: Reimburse office parking charges, or $120/month for public transport.</li>
<li>Sport: $120/month reimbursement for gym membership.</li>
<li>Meal stipend: $400 monthly allowance for meals (solution might evolve as we grow bigger).</li>
<li>Visa sponsorship.</li>
<li>Coaching: we offer BetterUp coaching on a voluntary basis.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Kubernetes, SLURM, Data Infrastructure, MLOps, Infrastructure Engineering</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo>https://logos.yubhub.co/mistral.ai.png</Employerlogo>
      <Employerdescription>Mistral AI provides high-performance, optimized, open-source and cutting-edge AI models, products and solutions for enterprise and personal needs.</Employerdescription>
      <Employerwebsite>https://mistral.ai/careers</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/37f53ee5-dd88-43e3-be6a-70e3db159c8f?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Palo Alto</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>dbfbd1d2-0a3</externalid>
      <Title>Research Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>Mistral AI is a pioneering company shaping the future of AI. We believe in the power of AI to simplify tasks, save time, and enhance learning and creativity.</p>
<p>Role Summary</p>
<p>The Data Infrastructure team at Mistral AI is architecting the backbone of our frontier model training and fine-tuning ecosystem. We are building the specialized compute and data fabrics required to power the development of world-class AI.</p>
<p>In this role, you will be a core contributor to our evolution, helping us design and scale massive compute fleets and storage systems designed for high performance and scalability. You will help us move toward a future of decoupled control and data planes, scaling big data compute and storage platforms while ensuring secure and governed data access for MLOps and research.</p>
<p>Responsibilities</p>
<ul>
<li>Build &amp; Scale: Help us reach our goal of operating massive distributed compute and storage systems</li>
<li>Global Orchestration: Architect and maintain multi-cluster orchestration layers to optimize workload placement across diverse hardware and regions.</li>
<li>Design Future-Proof Storage: Architect our transition to modern storage formats to handle fine-tuning datasets at a scale that anticipates exabyte growth.</li>
<li>Platform Engineering: Contribute to the development of our internal training platform, ensuring seamless model training and fine-tuning capabilities across Kubernetes and SLURM based environments.</li>
<li>Metadata &amp; Lineage: Implement and manage systems to provide clear visibility and lineage as our data and model pipelines grow in complexity.</li>
<li>Operational Excellence: Use modern deployment workflows to manage cloud-native deployments, ensuring our data platform can scale by orders of magnitude while remaining reliable and efficient.</li>
</ul>
<p>You might thrive in this role if you:</p>
<ul>
<li>Have 4+ years of experience in Data Infrastructure, MLOps, or Infrastructure Engineering.</li>
<li>Have experience or a strong interest in supporting foundational compute and storage platforms.</li>
<li>Are proficient in Python and enjoy solving the &quot;brittle data lake&quot; problem with modern, columnar storage standards.</li>
<li>Are well-versed in Kubernetes-native tooling and excited to debug large-scale distributed systems across multi-cluster environments.</li>
<li>Take pride in building and operating scalable, reliable, and secure systems from the ground up.</li>
<li>Are comfortable with ambiguity and the challenges of building high-scale infrastructure in a rapid-growth AI environment.</li>
</ul>
<p>Benefits</p>
<p>France</p>
<ul>
<li>Competitive cash salary and equity</li>
<li>Food: Daily lunch vouchers</li>
<li>Sport: Monthly contribution to a Gympass subscription</li>
<li>Transportation: Monthly contribution to a mobility pass</li>
<li>Health: Full health insurance for you and your family</li>
<li>Parental: Generous parental leave policy</li>
</ul>
<p>UK</p>
<ul>
<li>Competitive cash salary and equity</li>
<li>Insurance</li>
<li>Transportation: Reimburse office parking charges, or £90 per month for public transport</li>
<li>Sport: £90 per month reimbursement for gym membership</li>
<li>Meal voucher: £200 monthly allowance for meals</li>
<li>Pension plan: SmartPension (percentages are 5% Employee &amp; 3% Employer)</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Kubernetes, Data Infrastructure, MLOps, Infrastructure Engineering, Columnar Storage Standards</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo>https://logos.yubhub.co/mistral.ai.png</Employerlogo>
      <Employerdescription>Mistral AI is a company that develops and provides artificial intelligence (AI) technology and solutions. It has a diverse workforce and operates globally.</Employerdescription>
      <Employerwebsite>https://mistral.ai/careers</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/071a5491-ea01-413f-ad78-f85b5e4c2215?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Paris</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>26a321e1-136</externalid>
      <Title>Software Engineer, Codex Core Agents</Title>
      <Description><![CDATA[<p><strong>Compensation</strong></p>
<p>$230K – $385K • Offers Equity</p>
<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>
<ul>
<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>
</ul>
<ul>
<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>
</ul>
<ul>
<li>401(k) retirement plan with employer match</li>
</ul>
<ul>
<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>
</ul>
<ul>
<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>
</ul>
<ul>
<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>
</ul>
<ul>
<li>Mental health and wellness support</li>
</ul>
<ul>
<li>Employer-paid basic life and disability coverage</li>
</ul>
<ul>
<li>Annual learning and development stipend to fuel your professional growth</li>
</ul>
<ul>
<li>Daily meals in our offices, and meal delivery credits as eligible</li>
</ul>
<ul>
<li>Relocation support for eligible employees</li>
</ul>
<ul>
<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>
</ul>
<p><strong>About the Team</strong></p>
<p>The Codex Core Agent team builds the kernel of Codex. We own making the agent better, accelerating research, and making those improvements real in production for our users.</p>
<p><strong>About the Role</strong></p>
<p>We’re looking for engineers to build the infrastructure that powers Codex agents in production. This role focuses on the systems that let models safely execute code, interact with tools, complete long-running tasks, and operate reliably and efficiently at scale.</p>
<p><strong>What You’ll Do</strong></p>
<ul>
<li>Design and build execution environments for AI agents, including sandboxing, isolation, and reproducibility.</li>
</ul>
<ul>
<li>Develop systems for agent orchestration across multi-step, tool-using workflows.</li>
</ul>
<ul>
<li>Build infrastructure for running, testing, and debugging code generated by models.</li>
</ul>
<ul>
<li>Create state and memory systems that allow agents to persist context across long-running tasks.</li>
</ul>
<ul>
<li>Optimize tokens, latency, reliability, and cost across Codex’s production fleet.</li>
</ul>
<ul>
<li>Support model rollouts, capacity planning, and the core tradeoffs between quality, speed, and economics to manage a fleet of frontier agents at scale.</li>
</ul>
<ul>
<li>Build shared platform capabilities that unblock product teams, partner teams, and open source Codex.</li>
</ul>
<p><strong>You Might Be a Good Fit If You</strong></p>
<ul>
<li>Have strong experience in distributed systems or infrastructure engineering.</li>
</ul>
<ul>
<li>Have built systems involving containers, sandboxing, or virtualization.</li>
</ul>
<ul>
<li>Are comfortable working across backend systems, APIs, and developer tooling.</li>
</ul>
<ul>
<li>Care deeply about system reliability, performance, and security.</li>
</ul>
<ul>
<li>Enjoy working on ambiguous, zero-to-one problems.</li>
</ul>
<ul>
<li>Want to help build the systems that turn model capability into a dependable software engineering agent.</li>
</ul>
<p><strong>Bonus Points</strong></p>
<ul>
<li>Experience with code execution platforms, CI/CD systems, or build systems.</li>
</ul>
<ul>
<li>Familiarity with LLMs, agents, or tool-use frameworks.</li>
</ul>
<ul>
<li>Background in security engineering or isolation systems.</li>
</ul>
<ul>
<li>Experience building developer platforms, IDE tooling, or open source infrastructure.</li>
</ul>
<p><strong>About OpenAI</strong></p>
<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>
<p>We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic.</p>
<p>For additional information, please see [OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement](https://cdn.openai.com/policies/eeo-policy-statement.pdf).</p>
<p>Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.</p>
<p>To notify OpenAI that you believe this job posting is non-compliant, please submit a report through [this form](https://form.asana.com/?d=57018692298241&amp;k=5MqR40fZd7jlxVUh5J-UeA). No response will be provided to inquiries unrelated to job posting compliance.</p>
<p>We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this [link](https://form.asana.com/?k=bQ7w9h3iexRlicUdWRiwvg&amp;d=57018692298241).</p>
<p>[OpenAI Global Applicant Privacy Policy](https://cdn.openai.com/policies/global-employee-and-contractor-privacy-policy.pdf)</p>
<p>At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>Full time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$230K – $385K</Salaryrange>
      <Skills>distributed systems, infrastructure engineering, containers, sandboxing, virtualization, backend systems, APIs, developer tooling, security, code execution platforms, CI/CD systems, build systems, LLMs, agents, tool-use frameworks, security engineering, isolation systems, developer platforms, IDE tooling, open source infrastructure</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>OpenAI</Employername>
      <Employerlogo>https://logos.yubhub.co/openai.com.png</Employerlogo>
      <Employerdescription>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity.</Employerdescription>
      <Employerwebsite>https://openai.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/openai/7ade7a12-845c-4e3a-af23-c028420bd181?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco; London, UK; New York City; Seattle</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>bdf4e05a-b8c</externalid>
      <Title>MTS - Site Reliability Engineer</Title>
      <Description><![CDATA[<p>As Microsoft continues to push the boundaries of AI, we are on the lookout for individuals to work with us on the most interesting and challenging AI questions of our time. Our vision is bold and broad , to build systems that have true artificial intelligence across agents, applications, services, and infrastructure. It’s also inclusive: we aim to make AI accessible to all , consumers, businesses, developers , so that everyone can realize its benefits.</p>
<p>We’re looking for an experienced Site Reliability Engineer (SRE) to join our infrastructure team. In this role, you’ll blend software engineering and systems engineering to keep our large-scale distributed AI infrastructure reliable and efficient. You’ll work closely with ML researchers, data engineers, and product developers to design and operate the platforms that power training, fine-tuning, and serving generative AI models.</p>
<p>Microsoft’s mission is to empower every person and every organization on the planet to achieve more. As employees we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.</p>
<p>Responsibilities:</p>
<p>Reliability &amp; Availability: Ensure uptime, resiliency, and fault tolerance of AI model training and inference systems.</p>
<p>Observability: Design and maintain monitoring, alerting, and logging systems to provide real-time visibility into model serving pipelines and infra.</p>
<p>Performance Optimization: Analyze system performance and scalability, optimize resource utilization (compute, GPU clusters, storage, networking).</p>
<p>Automation &amp; Tooling: Build automation for deployments, incident response, scaling, and failover in hybrid cloud/on-prem CPU+GPU environments.</p>
<p>Incident Management: Lead on-call rotations, troubleshoot production issues, conduct blameless postmortems, and drive continuous improvements.</p>
<p>Security &amp; Compliance: Ensure data privacy, compliance, and secure operations across model training and serving environments.</p>
<p>Collaboration: Partner with ML engineers and platform teams to improve developer experience and accelerate research-to-production workflows.</p>
<p>Qualifications:</p>
<p>Required Qualifications: 4+ years of experience in Site Reliability Engineering, DevOps, or Infrastructure Engineering roles.</p>
<p>Preferred Qualifications: Strong proficiency in Kubernetes, Docker, and container orchestration. Knowledge of CI/CD pipelines for Inference and ML model deployment. Hands-on experience with public cloud platforms like Azure/AWS/GCP and infrastructure-as-code. Expertise in monitoring &amp; observability tools (Grafana, Datadog, OpenTelemetry, etc.). Strong programming/scripting skills in Python, Go, or Bash. Solid knowledge of distributed systems, networking, and storage. Experience running large-scale GPU clusters for ML/AI workloads (preferred). Familiarity with ML training/inference pipelines. Experience with high-performance computing (HPC) and workload schedulers (Kubernetes operators). Background in capacity planning &amp; cost optimization for GPU-heavy environments.</p>
<p>Work on cutting-edge infrastructure that powers the future of Generative AI. Collaborate with world-class researchers and engineers. Impact millions of users through reliable and responsible AI deployments. Competitive compensation, equity options, and comprehensive benefits.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$119,800 - $234,700 per year</Salaryrange>
      <Skills>Site Reliability Engineering, DevOps, Infrastructure Engineering, Kubernetes, Docker, container orchestration, CI/CD pipelines, ML model deployment, public cloud platforms, Azure, AWS, GCP, infrastructure-as-code, monitoring &amp; observability tools, Grafana, Datadog, OpenTelemetry, Python, Go, Bash, distributed systems, networking, storage, GPU clusters, ML training/inference pipelines, high-performance computing, workload schedulers, capacity planning, cost optimization, cloud architecture, containerization, microservices, API design, security, compliance, agile development, scrum, kanban</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft is a multinational technology company that develops, manufactures, licenses, and supports a wide range of software products, services, and devices.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/mts-site-reliability-engineer/?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Redmond</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>5ff592ac-9d8</externalid>
      <Title>Sr. Software Engineer, Inference</Title>
      <Description><![CDATA[<p>We are seeking a Senior Software Engineer to join our Inference team, responsible for building and maintaining critical systems that serve Claude to millions of users worldwide. The team has a dual mandate: maximizing compute efficiency to serve our explosive customer growth, while enabling breakthrough research by giving our scientists the high-performance inference infrastructure they need to develop next-generation models.</p>
<p>As a Senior Software Engineer, you will be responsible for designing, implementing, and deploying large-scale distributed systems, including intelligent request routing, fleet-wide orchestration, and load balancing. You will work closely with our research team to develop new inference features and integrate new AI accelerator platforms.</p>
<p>To succeed in this role, you should have significant software engineering experience, particularly with distributed systems, and be results-oriented with a bias towards flexibility and impact. You should also be able to pick up slack, even if it goes outside your job description, and thrive in environments where technical excellence directly drives both business results and research breakthroughs.</p>
<p>Responsibilities:</p>
<ul>
<li>Design and implement large-scale distributed systems, including intelligent request routing, fleet-wide orchestration, and load balancing</li>
<li>Work closely with our research team to develop new inference features and integrate new AI accelerator platforms</li>
<li>Collaborate with cross-functional teams to ensure seamless deployment and operation of our systems</li>
<li>Analyze observability data to tune performance based on real-world production workloads</li>
<li>Manage multi-region deployments and geographic routing for global customers</li>
</ul>
<p>Requirements:</p>
<ul>
<li>Bachelor&#39;s degree or equivalent combination of education, training, and/or experience</li>
<li>Significant software engineering experience, particularly with distributed systems</li>
<li>Results-oriented with a bias towards flexibility and impact</li>
<li>Ability to pick up slack, even if it goes outside your job description</li>
<li>Thrives in environments where technical excellence directly drives both business results and research breakthroughs</li>
</ul>
<p>Preferred Qualifications:</p>
<ul>
<li>Experience with Kubernetes and cloud infrastructure (AWS, GCP)</li>
<li>Familiarity with machine learning systems and infrastructure</li>
<li>Strong communication and collaboration skills</li>
</ul>
<p>Benefits:</p>
<ul>
<li>Competitive compensation and benefits</li>
<li>Optional equity donation matching</li>
<li>Generous vacation and parental leave</li>
<li>Flexible working hours</li>
<li>Lovely office space in which to collaborate with colleagues</li>
</ul>
<p>Guidance on Candidates&#39; AI Usage: Learn about our policy for using AI in our application process</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>£225,000-£325,000 GBP</Salaryrange>
      <Skills>Distributed systems, Kubernetes, Cloud infrastructure, Machine learning systems, Infrastructure engineering, Python, Rust, Java, C++, Go</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5152348008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>London, UK</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>b0e99a49-d99</externalid>
      <Title>Senior Engineering Manager - Infrastructure</Title>
      <Description><![CDATA[<p>About Us</p>
<p>We&#39;re looking for an Infrastructure Senior Engineering Manager to help us build a seamless, reliable platform for the dbt platform across AWS, Azure, and GCP.</p>
<p>Our team&#39;s mission is to create a seamless developer experience by providing a stable, observable, and easy-to-use infrastructure platform. Over the past year, we&#39;ve designed and operationalized a next-gen cell-based architecture, scaling the dbt platform across all three cloud providers. Now, we&#39;re focused on automation, self-service, and improving developer velocity through better tooling, processes, and infrastructure design.</p>
<p>As a Senior Engineering Manager, you&#39;ll lead your team on infrastructure projects to refine our platform while ensuring performance, reliability, and an excellent developer experience. You&#39;ll collaborate across teams, tackle real infrastructure challenges, and help shape the future of the multi-cloud dbt platform.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Build, lead, and coach a team of 8-12 engineers to manage the infrastructure for the dbt platform and report to the Director of Infrastructure</li>
<li>Empower your team to achieve big goals by giving them product and business context and supporting team ownership of the roadmap, product development lifecycle, and technical excellence</li>
<li>Dive deep into our product to frame tradeoffs and make decisions about what, how, and when we build</li>
<li>Partner with Product Marketing, Solutions Architecture, and Customer Support to build delightful migration experiences, helping our customers seamlessly move off legacy deployments</li>
<li>Coach engineers in product thinking, quality, and software engineering. Build individualized growth plans and match interests and capabilities to team goals</li>
<li>Work with peer managers to evolve organizational processes like product training, technical decision making, project execution, and planning</li>
</ul>
<p><strong>Requirements</strong></p>
<ul>
<li>5+ years in people management with a software or infrastructure engineering team</li>
<li>Experience managing senior individual contributors (Staff+ level)</li>
<li>Experience supporting a cloud-based infrastructure with complex resource requirements and global deployment strategy</li>
<li>Deep understanding of Terraform and cloud infrastructure state management</li>
</ul>
<p><strong>Nice to Have</strong></p>
<ul>
<li>Experience leading teams through all parts of the product development lifecycle</li>
<li>Have successfully partnered across teams and departments to coordinate cross-cutting initiatives</li>
<li>You are interested in our mission and values. You are inspired to drive progress in the data and analytics ecosystem</li>
</ul>
<p><strong>Compensation &amp; Benefits</strong></p>
<p>Salary: We offer competitive compensation packages commensurate with experience, including salary, equity, and where applicable, performance-based pay. Our Talent Acquisition Team can answer questions around dbt Labs&#39; total rewards during your interview process.</p>
<p>In select locations (including Boston, Chicago, Denver, Los Angeles, Philadelphia, New York Metro, San Francisco, DC Metro, Seattle, Austin), an alternate range may apply, as specified below.</p>
<p>The typical starting salary range for this role is: $223,000 - $270,000 USD</p>
<p>The typical starting salary range for this role in the select locations listed is: $248,000 - $300,000 US</p>
<p>Equity Stake Benefits</p>
<ul>
<li>dbt Labs offers: unlimited vacation, 401k w/3% guaranteed contribution, excellent healthcare, paid parental leave, wellness stipend, home office stipend, and more!</li>
</ul>
<p><strong>Our Hiring Process</strong></p>
<ul>
<li>Interview with a Talent Acquisition Partner (30 Mins)</li>
<li>Technical Interview with Hiring Manager (60 Mins)</li>
<li>Team Interviews ( 3 rounds, 45 Mins each)</li>
<li>Final Values Interview (30 Mins)</li>
</ul>
<p>If you’re passionate about building well-designed, high-impact software, we’d love to hear from you!</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange>$223,000 - $270,000 USD</Salaryrange>
      <Skills>Terraform, Cloud infrastructure state management, People management, Software engineering, Infrastructure engineering, Product development lifecycle, Technical decision making, Project execution, Process improvement</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>dbt Labs</Employername>
      <Employerlogo>https://logos.yubhub.co/getdbt.com.png</Employerlogo>
      <Employerdescription>dbt Labs is a leading analytics engineering platform, used by over 90,000 teams every week, with over $100 million in annual recurring revenue.</Employerdescription>
      <Employerwebsite>https://www.getdbt.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/dbtlabsinc/jobs/4686309005?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>US - Remote</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>67b4ccd7-51d</externalid>
      <Title>Senior Software Engineer, Observability Insights</Title>
      <Description><![CDATA[<p>Join CoreWeave&#39;s Observability team, where we are building the next-generation insights layer for AI systems.</p>
<p>Our team empowers internal and external users to understand, troubleshoot, and optimize complex AI workloads by transforming telemetry into actionable insights.</p>
<p>As a Senior Software Engineer on the Observability Insights team, you will lead the development of agentic interfaces and product experiences that sit atop CoreWeave&#39;s telemetry layer.</p>
<p>You&#39;ll design multi-tenant APIs, managed Grafana experiences, and MCP-based tool servers to help customers and internal teams interact with data in innovative ways.</p>
<p>Collaborating closely with PMs and engineering leadership, your work will shape the end-to-end observability experience and influence how people engage with cutting-edge AI infrastructure.</p>
<p><strong>About the role</strong></p>
<ul>
<li>6+ years of experience in software or infrastructure engineering building production-grade backend systems and distributed APIs.</li>
</ul>
<ul>
<li>Strong focus on developer-facing infrastructure, with a customer-obsessed approach to SDKs, CLIs, and APIs.</li>
</ul>
<ul>
<li>Proficient in reliability engineering, including fault-tolerant design, SLOs, error budgets, and multi-tenant system resilience.</li>
</ul>
<ul>
<li>Familiar with observability systems such as ClickHouse, Loki, VictoriaMetrics, Prometheus, and Grafana.</li>
</ul>
<ul>
<li>Experienced in agentic applications or LLM-based features, including grounding, tool calling, and operational safety.</li>
</ul>
<ul>
<li>Comfortable writing production code primarily in Go, with the ability to integrate Python components when needed.</li>
</ul>
<ul>
<li>Collaborative experience in agile teams delivering end-to-end telemetry-to-insights pipelines.</li>
</ul>
<p><strong>Preferred</strong></p>
<ul>
<li>Experience operating Kubernetes clusters at scale, especially for AI workloads.</li>
</ul>
<ul>
<li>Hands-on experience with logging, tracing, and metrics platforms in production, with deep knowledge of cardinality, indexing, and query optimization.</li>
</ul>
<ul>
<li>Experienced in running distributed systems or API services at cloud scale, including event streaming and data pipeline management.</li>
</ul>
<ul>
<li>Familiarity with LLM frameworks, MCP, and agentic tooling (e.g., Langchain, AgentCore).</li>
</ul>
<p><strong>Why CoreWeave?</strong></p>
<p>At CoreWeave, we work hard, have fun, and move fast!</p>
<p>We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on.</p>
<p>We&#39;re not afraid of a little chaos, and we&#39;re constantly learning.</p>
<p>Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>
<ul>
<li>Be Curious at Your Core</li>
</ul>
<ul>
<li>Act Like an Owner</li>
</ul>
<ul>
<li>Empower Employees</li>
</ul>
<ul>
<li>Deliver Best-in-Class Client Experiences</li>
</ul>
<ul>
<li>Achieve More Together</li>
</ul>
<p>We support and encourage an entrepreneurial outlook and independent thinking.</p>
<p>We foster an environment that encourages collaboration and enables the development of innovative solutions to complex problems.</p>
<p>As we get set for takeoff, the organization&#39;s growth opportunities are constantly expanding.</p>
<p>You will be surrounded by some of the best talent in the industry, who will want to learn from you, too.</p>
<p>Come join us!</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$165,000 to $242,000</Salaryrange>
      <Skills>software engineering, infrastructure engineering, backend systems, distributed APIs, reliability engineering, fault-tolerant design, SLOs, error budgets, multi-tenant system resilience, observability systems, ClickHouse, Loki, VictoriaMetrics, Prometheus, Grafana, agentic applications, LLM-based features, grounding, tool calling, operational safety, Go, Python, Kubernetes, logging, tracing, metrics platforms, cardinality, indexing, query optimization, event streaming, data pipeline management, LLM frameworks, MCP, agent tooling, operating Kubernetes clusters</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>CoreWeave</Employername>
      <Employerlogo>https://logos.yubhub.co/coreweave.com.png</Employerlogo>
      <Employerdescription>CoreWeave is a cloud computing company that provides a platform for building and scaling AI.</Employerdescription>
      <Employerwebsite>https://www.coreweave.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/coreweave/jobs/4650163006?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>New York, NY / Sunnyvale, CA</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>bd3983d2-d82</externalid>
      <Title>Staff Technical Program Manager, Reliability &amp; Observability</Title>
      <Description><![CDATA[<p>We&#39;re looking for a technical, hands-on, and mission-driven Staff Technical Program Manager (TPM) to lead Reliability &amp; Observability initiatives. In this role, you will collaborate closely with Machine Learning (ML) engineers, infrastructure engineers, and product managers across Airbnb to develop holistic solutions that ensure the Airbnb platform is robust, highly available, and transparent in its operations.</p>
<p>The Reliability &amp; Observability team enables safe, resilient, and transparent operation of critical systems powering Airbnb. This team creates and maintains the frameworks and platforms for proactive monitoring, alerting, logging, tracing, and incident management, helping teams across Airbnb maintain service health and quickly remediate issues.</p>
<p>As a TPM, you will play a crucial role in tackling projects that power search and recommendations across the entire Airbnb platform,directly influencing how guests and hosts connect in meaningful ways. Success in this role requires a strong sense of ownership, coupled with the ability to think critically and strategically while managing relationships with cross-functional stakeholders.</p>
<p>Responsibilities:</p>
<ul>
<li>Shape and influence the technical direction of projects, ensuring we meet stakeholder needs while maintaining high-quality standards.</li>
</ul>
<ul>
<li>Rapidly prototype and validate project ideas through iterative development cycles, adapting as necessary based on insights and data.</li>
</ul>
<ul>
<li>Balance broad, outcome-driven thinking with attention to critical details, exercising sound judgment to prioritize where deep focus is necessary to ensure successful execution.</li>
</ul>
<ul>
<li>Define and secure stakeholder alignment on clear, measurable success criteria to accelerate AI initiatives.</li>
</ul>
<ul>
<li>Regularly assess risks and opportunities, and devise proactive mitigation strategies to maintain momentum and project success.</li>
</ul>
<ul>
<li>Maintain transparent and effective communication channels to keep stakeholders informed of progress, developments, and challenges.</li>
</ul>
<ul>
<li>Expertly present outcomes and updates to senior leadership, clearly and comprehensively articulating trade-offs, risks, and emerging opportunities.</li>
</ul>
<p>Your Expertise:</p>
<ul>
<li>At least 10 years of work experience, with at least 8 years as a TPM or relevant experience.</li>
</ul>
<ul>
<li>Demonstrated ability to work through ambiguity to detailed solutions.</li>
</ul>
<ul>
<li>Self-motivated, proactive and proven ability to adapt well and work with teams having different operating cadences.</li>
</ul>
<ul>
<li>Sound business judgment, a proven ability to influence others, strong analytical skills, and a track record of taking ownership, leading data driven analyses, and influencing results.</li>
</ul>
<ul>
<li>Experience with ML models, LLMs, LRMs, feature development, model testing and resource management to support the development of AI-powered product experiences.</li>
</ul>
<ul>
<li>Familiar with A/B testing, incremental delivery and deployment.</li>
</ul>
<ul>
<li>Ability to ramp up quickly and learn new technologies with minimal lag time.</li>
</ul>
<ul>
<li>Excellent written and oral business communication and people skills, with the ability to influence stakeholders.</li>
</ul>
<p>Our Commitment To Inclusion &amp; Belonging:</p>
<p>Airbnb is committed to working with the broadest talent pool possible. We believe diverse ideas foster innovation and engagement, and allow us to attract creatively-led people, and to develop the best products, services and solutions.</p>
<p>All qualified individuals are encouraged to apply. We strive to also provide a disability inclusive application and interview process. If you are a candidate with a disability and require reasonable accommodation in order to submit an application, please contact us at: reasonableaccommodations@airbnb.com.</p>
<p>How We&#39;ll Take Care of You:</p>
<p>Our job titles may span more than one career level. The actual base pay is dependent upon many factors, such as: training, transferable skills, work experience, business needs and market demands. The base pay range is subject to change and may be modified in the future. This role may also be eligible for bonus, equity, benefits, and Employee Travel Credits.</p>
<p>Pay Range $194,000-$242,000 USD</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange>$194,000-$242,000 USD</Salaryrange>
      <Skills>Technical Program Management, Reliability &amp; Observability, Machine Learning, Infrastructure Engineering, Product Management, A/B Testing, Incremental Delivery, Deployment</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Airbnb</Employername>
      <Employerlogo>https://logos.yubhub.co/airbnb.com.png</Employerlogo>
      <Employerdescription>Airbnb is a global online marketplace for short-term vacation rentals. It was founded in 2007 and has since grown to become one of the largest and most popular travel platforms in the world.</Employerdescription>
      <Employerwebsite>https://www.airbnb.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/airbnb/jobs/7558202?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>United States</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>51758515-c12</externalid>
      <Title>Member of Technical Staff</Title>
      <Description><![CDATA[<p>We are seeking a highly skilled Member of Technical Staff to join our team in managing and enhancing reliability across a multi-data center environment.</p>
<p>This role focuses on automating processes, building and implementing robust observability solutions, and ensuring seamless operations for mission-critical AI infrastructure.</p>
<p>The ideal candidate will combine strong coding abilities with hands-on data center experience to build scalable reliability services, optimize system performance, and minimize downtime,including close partnership with facility operations to address physical infrastructure impacts.</p>
<p>In an era where AI workloads demand near-zero downtime, this position plays a pivotal role in bridging software engineering principles with physical data center realities.</p>
<p>By prioritizing automation and observability, team members in this role can reduce mean time to recovery (MTTR) by up to 50% through proactive monitoring and automated remediation, based on industry benchmarks from high-scale environments like those at hyperscale cloud providers.</p>
<p>Responsibilities:</p>
<ul>
<li>Design, develop, and deploy scalable code and services (primarily in Python and Rust, with flexibility for emerging languages) to automate reliability workflows, including monitoring, alerting, incident response, and infrastructure provisioning.</li>
</ul>
<ul>
<li>Implement and maintain observability tools and practices, such as metrics collection, logging, tracing, and dashboards, to provide real-time insights into system health across multiple data centers,open to innovative stacks beyond traditional ones like ELK.</li>
</ul>
<ul>
<li>Collaborate with cross-functional teams,including software development, network engineering, site operations, and facility operations (critical facilities, mechanical/electrical teams, and data center infrastructure management),to identify reliability bottlenecks, automate solutions for fault tolerance, disaster recovery, capacity planning, and physical/environmental risk mitigation (e.g., power redundancy, cooling efficiency, and environmental monitoring integration).</li>
</ul>
<ul>
<li>Troubleshoot and resolve complex issues in data center environments, including hardware failures, environmental anomalies, software bugs, and network-related problems, while adhering to reliability principles like error budgets and SLAs.</li>
</ul>
<ul>
<li>Optimize Linux-based systems for performance, security, and reliability, including kernel tuning, container orchestration (e.g., Kubernetes or emerging alternatives), and scripting for automation.</li>
</ul>
<ul>
<li>Understand network topologies and concepts in large-scale, multi-data center environments to effectively troubleshoot connectivity, routing, redundancy, and performance issues; integrate observability into data center interconnects and facility-level controls for rapid diagnosis and automation.</li>
</ul>
<ul>
<li>Participate in on-call rotations, post-incident reviews (blameless postmortems), and continuous improvement initiatives to enhance overall site reliability, including joint exercises with facility teams for physical failover and recovery scenarios.</li>
</ul>
<ul>
<li>Mentor junior team members and document processes to foster a culture of automation, knowledge sharing, and adaptability to new technologies.</li>
</ul>
<p>Basic Qualifications:</p>
<ul>
<li>Bachelor&#39;s degree in Computer Science, Computer Engineering, Electrical Engineering, or a closely related technical field (or equivalent professional experience).</li>
</ul>
<ul>
<li>5+ years of hands-on experience in site reliability engineering (SRE), infrastructure engineering, DevOps, or systems engineering, preferably supporting large-scale, distributed, or production environments.</li>
</ul>
<ul>
<li>Strong programming skills with proven production experience in Python (required for automation and tooling); experience with Rust or willingness to work in Rust is a plus, but strong coding fundamentals in at least one systems-level language (e.g., Python, Go, C++) are essential.</li>
</ul>
<ul>
<li>Solid experience with Linux systems administration, performance tuning, kernel-level understanding, and scripting/automation in production environments.</li>
</ul>
<ul>
<li>Practical knowledge of containerization and orchestration technologies, such as Docker and Kubernetes (or similar systems).</li>
</ul>
<ul>
<li>Experience implementing observability solutions, including metrics, logging, tracing, monitoring tools (e.g., Prometheus, Grafana, or alternatives), alerting, and dashboards.</li>
</ul>
<ul>
<li>Familiarity with troubleshooting complex issues in distributed systems, including software bugs, hardware failures, network problems, and environmental factors.</li>
</ul>
<ul>
<li>Understanding of networking fundamentals (TCP/IP, routing, redundancy, DNS) in large-scale or multi-site environments.</li>
</ul>
<ul>
<li>Experience participating in on-call rotations, incident response, post-incident reviews (blameless postmortems), and reliability practices such as error budgets or SLAs.</li>
</ul>
<ul>
<li>Ability to collaborate effectively with cross-functional teams (software engineers, network teams, site/facility operations, mechanical/electrical teams).</li>
</ul>
<p>Preferred Skills and Experience:</p>
<ul>
<li>7+ years of experience in SRE or infrastructure roles, ideally in hyperscale, cloud, or AI/ML training infrastructure environments with multi-data center setups.</li>
</ul>
<ul>
<li>Hands-on experience operating or scaling Kubernetes clusters (or equivalent orchestration) at large scale, including automation for provisioning, lifecycle management, and high-availability.</li>
</ul>
<ul>
<li>Proficiency in Rust for systems programming and performance-critical components.</li>
</ul>
<ul>
<li>Direct experience integrating software reliability tools with physical data center infrastructure.</li>
</ul>
<ul>
<li>Experience with observability tools and practices, such as metrics collection, logging, tracing, and dashboards.</li>
</ul>
<ul>
<li>Familiarity with containerization and orchestration technologies, such as Docker and Kubernetes (or similar systems).</li>
</ul>
<ul>
<li>Experience with Linux systems administration, performance tuning, kernel-level understanding, and scripting/automation in production environments.</li>
</ul>
<ul>
<li>Understanding of networking fundamentals (TCP/IP, routing, redundancy, DNS) in large-scale or multi-site environments.</li>
</ul>
<ul>
<li>Experience participating in on-call rotations, incident response, post-incident reviews (blameless postmortems), and reliability practices such as error budgets or SLAs.</li>
</ul>
<ul>
<li>Ability to collaborate effectively with cross-functional teams (software engineers, network teams, site/facility operations, mechanical/electrical teams).</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Rust, Linux systems administration, performance tuning, kernel-level understanding, scripting/automation, containerization, orchestration, observability, metrics collection, logging, tracing, dashboards, networking fundamentals, TCP/IP, routing, redundancy, DNS, Kubernetes, Docker, Grafana, Prometheus, ELK, DevOps, SRE, infrastructure engineering, systems engineering</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>xAI</Employername>
      <Employerlogo>https://logos.yubhub.co/xai.com.png</Employerlogo>
      <Employerdescription>xAI creates AI systems to understand the universe and aid humanity in its pursuit of knowledge.</Employerdescription>
      <Employerwebsite>https://www.xai.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/xai/jobs/5044403007?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Memphis, TN</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>eebf21c4-d1f</externalid>
      <Title>Staff Site Reliability Engineer</Title>
      <Description><![CDATA[<p>Join our Site Reliability Engineering (SRE) team and help ensure the reliability, scalability, and performance of Replit&#39;s infrastructure that serves millions of developers worldwide.</p>
<p>As a Staff Site Reliability Engineer, you will bridge the gap between development and operations, implementing automation and establishing best practices that enable our platform to scale efficiently while maintaining high availability.</p>
<p>We are seeking Staff SREs who are passionate about building and maintaining resilient systems at scale. Your mission will be to proactively find and analyze reliability problems across our stack, then design and implement software and systems to create step-function improvements.</p>
<p>You will design robust observability solutions, lead incident response, automate operational tasks, and continuously improve our infrastructure&#39;s reliability, all while mentoring and educating the broader engineering team to make reliability a core value at Replit.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Architect and Implement Observability: Design, build, and lead the implementation of comprehensive monitoring, logging, and tracing solutions. Create dashboards and metrics that provide real-time visibility into system health and performance, enabling proactive issue detection.</li>
</ul>
<ul>
<li>Define and Drive Reliability Standards: Work with product and engineering teams to define, implement, and track Service Level Objectives (SLOs) and Service Level Indicators (SLIs). Build systems to monitor and report on these metrics, holding teams accountable and ensuring we maintain high reliability standards while balancing innovation speed.</li>
</ul>
<ul>
<li>Lead Incident Management and Response: Act as a senior leader during high-impact incidents, guiding the team to rapid resolution. Conduct thorough, blameless post-mortems and drive the implementation of preventative measures. Develop and refine runbooks and build automation to reduce Mean Time To Recovery (MTTR).</li>
</ul>
<ul>
<li>Drive Automation and Infrastructure as Code: Architect, build, and improve automation to eliminate toil and operational work. Design and maintain CI/CD pipelines and infrastructure automation using tools like Terraform or Pulumi. Create self-healing systems that can automatically respond to common failure scenarios.</li>
</ul>
<ul>
<li>Optimize Performance on Kubernetes: Collaborate with core infrastructure and product teams to performance-tune and optimize our large-scale cloud deployments, with a deep focus on Kubernetes, Docker, and GCP. Identify and resolve performance bottlenecks, implement capacity planning strategies, and reduce latency across global regions.</li>
</ul>
<ul>
<li>Debug and Harden Distributed Systems: Dive deep into debugging extremely difficult technical problems across the stack. Use your findings to design and implement long-term fixes that make our systems and products more robust, operable, and easier to diagnose.</li>
</ul>
<ul>
<li>Provide Staff-Level Guidance: Review feature and system designs from across the company, acting as a key owner for the reliability, scalability, security, and operational integrity of those designs.</li>
</ul>
<ul>
<li>Educate and Mentor: Educate, mentor, and hold accountable the broader engineering team to improve the reliability of our systems, making reliability a core value of the Replit engineering culture.</li>
</ul>
<ul>
<li>Build and Integrate: Write high-quality, well-tested code in Python or Go to meet the needs of your customers, whether it&#39;s building new internal tools or integrating with third-party vendors.</li>
</ul>
<p><strong>Required Skills and Experience</strong></p>
<ul>
<li>8-10 years of experience in Site Reliability Engineering or similar roles (e.g., DevOps, Systems Engineering, Infrastructure Engineering).</li>
</ul>
<ul>
<li>Strong programming skills in languages like Python or Go. You write high-quality, well-tested code.</li>
</ul>
<ul>
<li>Deep understanding of distributed systems. You’ve designed, built, scaled, and maintained production services and know how to compose a service-oriented architecture.</li>
</ul>
<ul>
<li>Deep experience with container orchestration platforms, specifically Kubernetes, and cloud-native technologies.</li>
</ul>
<ul>
<li>Proven track record of designing, implementing, and maintaining sophisticated monitoring and observability solutions (e.g., metrics, logging, tracing).</li>
</ul>
<ul>
<li>Strong incident management skills with extensive experience leading incident response for complex systems and demonstrated critical thinking under pressure.</li>
</ul>
<ul>
<li>Experience with infrastructure as code (e.g., Terraform, Pulumi) and configuration management tools.</li>
</ul>
<ul>
<li>Excellent written and verbal communication skills, with an ability to explain complex technical concepts clearly and simply and a bias toward open, transparent cultural practices.</li>
</ul>
<ul>
<li>Strong interpersonal skills, with experience working with and mentoring engineers from junior to principal levels.</li>
</ul>
<ul>
<li>A willingness to dive into understanding, debugging, and improving any layer of the stack.</li>
</ul>
<ul>
<li>You&#39;re passionate about making software creation accessible and empowering the next generation of builders.</li>
</ul>
<p><strong>Bonus Points</strong></p>
<ul>
<li>Deep experience with Google Cloud Platform (GCP) services and tools.</li>
</ul>
<ul>
<li>Expert-level knowledge of modern observability platforms (e.g., Prometheus, Grafana, Datadog, OpenTelemetry).</li>
</ul>
<ul>
<li>Experience designing and building reliable systems capable of handling high throughput and low latency.</li>
</ul>
<ul>
<li>Significant experience with Go and Terraform.</li>
</ul>
<ul>
<li>Familiarity with working in rapid-growth, startup environments.</li>
</ul>
<ul>
<li>Experience writing company-facing blog posts and training materials.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>Full time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange>$220K - $325K</Salaryrange>
      <Skills>Site Reliability Engineering, DevOps, Systems Engineering, Infrastructure Engineering, Python, Go, Distributed Systems, Container Orchestration, Kubernetes, Cloud-Native Technologies, Monitoring and Observability, Incident Management, Infrastructure as Code, Terraform, Pulumi, Configuration Management, Google Cloud Platform, Prometheus, Grafana, Datadog, OpenTelemetry, Go, Terraform</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Replit</Employername>
      <Employerlogo>https://logos.yubhub.co/replit.com.png</Employerlogo>
      <Employerdescription>Replit is an agentic software creation platform that enables anyone to build applications using natural language, with millions of users worldwide.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/replit/d50ad15b-82d4-452f-b4ea-2a7f5e796170?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Remote (United States)</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>da726093-b19</externalid>
      <Title>Research Engineer, Discovery</Title>
      <Description><![CDATA[<p><strong>About the Role</strong></p>
<p>As a Research Engineer on our team, you will work end to end across the whole model stack, identifying and addressing key infra blockers on the path to scientific AGI. Strong candidates should have familiarity with elements of language model training, evaluation, and inference and eagerness to quickly dive and get up to speed in areas they are not yet an expert on. This may include performance optimization, distributed systems, VM/sandboxing/container deployment, and large scale data pipelines.</p>
<p><strong>Responsibilities:</strong></p>
<ul>
<li>Design and implement large-scale infrastructure systems to support AI scientist training, evaluation, and deployment across distributed environments</li>
<li>Identify and resolve infrastructure bottlenecks impeding progress toward scientific capabilities</li>
<li>Develop robust and reliable evaluation frameworks for measuring progress towards scientific AGI.</li>
<li>Build scalable and performant VM/sandboxing/container architectures to safely execute long-horizon AI tasks and scientific workflows</li>
<li>Collaborate to translate experimental requirements into production-ready infrastructure</li>
<li>Develop large scale data pipelines to handle advanced language model training requirements</li>
<li>Optimize large scale training and inference pipelines for stable and efficient reinforcement learning</li>
</ul>
<p><strong>You may be a good fit if you:</strong></p>
<ul>
<li>Have 6+ years of highly-relevant experience in infrastructure engineering with demonstrated expertise in large-scale distributed systems</li>
<li>Are a strong communicator and enjoy working collaboratively</li>
<li>Possess deep knowledge of performance optimization techniques and system architectures for high-throughput ML workloads</li>
<li>Have experience with containerization technologies (Docker, Kubernetes) and orchestration at scale</li>
<li>Have proven track record of building large-scale data pipelines and distributed storage systems</li>
<li>Excel at diagnosing and resolving complex infrastructure challenges in production environments</li>
<li>Can work effectively across the full ML stack from data pipelines to performance optimization</li>
<li>Have experience collaborating with other researchers to scale experimental ideas</li>
<li>Thrive in fast-paced environments and can rapidly iterate from experimentation to production</li>
</ul>
<p><strong>Strong candidates may also have:</strong></p>
<ul>
<li>Experience with language model training infrastructure and distributed ML frameworks (PyTorch, JAX, etc.)</li>
<li>Background in building infrastructure for AI research labs or large-scale ML organizations</li>
<li>Knowledge of GPU/TPU architectures and language model inference optimization</li>
<li>Experience with cloud platforms (AWS, GCP) at enterprise scale</li>
<li>Familiarity with VM and container orchestration.</li>
<li>Experience with workflow orchestration tools and experiment management systems</li>
<li>History working with large scale reinforcement learning</li>
<li>Comfort with large scale data pipelines (Beam, Spark, Dask, …)</li>
</ul>
<p><strong>Logistics</strong></p>
<ul>
<li>Education requirements: We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</li>
<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>
<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>
</ul>
<p><strong>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</strong></p>
<p><strong>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</strong></p>
<p><strong>How we&#39;re different</strong></p>
<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale projects, and we&#39;re committed to making a positive impact on the world.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$350,000 - $850,000 USD</Salaryrange>
      <Skills>infrastructure engineering, large-scale distributed systems, performance optimization, containerization technologies, orchestration at scale, data pipelines, distributed storage systems, complex infrastructure challenges, ML stack, workflow orchestration tools, experiment management systems, reinforcement learning, large scale data pipelines, language model training infrastructure, distributed ML frameworks, GPU/TPU architectures, language model inference optimization, cloud platforms, VM and container orchestration, workflow orchestration tools, experiment management systems, large scale reinforcement learning, large scale data pipelines</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a company that aims to create reliable, interpretable, and steerable AI systems. It has a team of researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</Employerdescription>
      <Employerwebsite>https://job-boards.greenhouse.io</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/4669581008?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>8c164f95-f8d</externalid>
      <Title>Senior Infrastructure Engineer</Title>
      <Description><![CDATA[<p>Join our Infrastructure Engineering team and help ensure the reliability, scalability, and performance of Replit&#39;s infrastructure that serves millions of developers worldwide. As a Senior Infrastructure Engineer, you will bridge the gap between development and operations, implementing automation and establishing best practices that enable our platform to scale efficiently while maintaining high availability.</p>
<p>We are seeking Senior Infrastructure Engineers who are passionate about building and maintaining resilient systems at scale. Your mission will be to proactively find and analyse reliability problems across our stack, then design and implement software and systems to address them. You will build robust monitoring solutions, automate operational tasks, and continuously improve our infrastructure&#39;s reliability.</p>
<p><strong>You Will:</strong></p>
<ul>
<li>Drive Automation and Infrastructure as Code: Build and improve automation to eliminate toil and operational work. Maintain CI/CD pipelines and infrastructure automation using tools like Terraform or Pulumi. Create self-healing systems that can automatically respond to common failure scenarios.</li>
<li>Optimise Performance and Infrastructure: Collaborate with core infrastructure and product teams to performance tune and optimise our cloud deployments (Kubernetes, Docker, GCP). Identify and resolve performance bottlenecks and implement capacity planning strategies.</li>
<li>Elevate Developer Experience: Design and implement improvements to our build, test, and deployment systems to make software delivery faster, safer, and more reliable for all engineers.</li>
<li>Drive Cross-Team Improvements: Partner with service owners across Replit to understand their pain points, and collaborate on implementing build/test/deploy enhancements within their specific services.</li>
<li>Build Shared Tooling: Create and maintain centralized tooling and automation that improves the engineering lifecycle, from local development to production monitoring.</li>
<li>Debug and Harden Systems: Dive deep into debugging difficult technical problems, making our systems and products more robust, operable, and easier to diagnose.</li>
<li>Collaborate on Design Reviews: Participate in feature and system design reviews, contributing expertise on security, scale, and operational considerations.</li>
<li>Build and Integrate: Write high-quality, well-tested code to meet the needs of your customers, including building pipelines to integrate with 3rd party vendors.</li>
</ul>
<p><strong>Required Skills and Experience:</strong></p>
<ul>
<li>4+ years of experience in Site Reliability Engineering or similar roles (DevOps, Systems Engineering, Infrastructure Engineering).</li>
<li>Strong programming skills in languages like Python or Go.</li>
<li>You write high-quality, well-tested code.</li>
<li>Solid understanding of distributed systems. You&#39;ve built, scaled, and maintained production services and understand service-oriented architecture.</li>
<li>Experience with container orchestration platforms (Kubernetes) and cloud-native technologies.</li>
<li>Experience implementing and maintaining monitoring/observability solutions, with strong skills in debugging and performance tuning.</li>
<li>Strong incident management skills with experience participating in incident response and demonstrated critical thinking under pressure.</li>
<li>Experience with infrastructure as code (e.g., Terraform) and configuration management tools.</li>
<li>Excellent written and verbal communication skills, with an ability to explain technical concepts clearly.</li>
<li>A willingness to dive into understanding, debugging, and improving any layer of the stack.</li>
<li>You&#39;re passionate about making software creation accessible and empowering the next generation of builders.</li>
</ul>
<p><strong>Bonus Points:</strong></p>
<ul>
<li>Experience with Google Cloud Platform (GCP) services and tools.</li>
<li>Knowledge of modern observability platforms (Prometheus, Grafana, Datadog, etc.).</li>
<li>Experience building reliable systems capable of handling high throughput and low latency.</li>
<li>Experience with Go and Terraform.</li>
<li>Familiarity with working in rapid-growth environments.</li>
</ul>
<p>_This is a full-time role that can be held from our Foster City, CA office. The role has an in-office requirement of Monday, Wednesday, and Friday._</p>
<p><strong>Full-Time Employee Benefits Include:</strong></p>
<ul>
<li>Competitive Salary &amp; Equity</li>
<li>401(k) Program with a 4% match</li>
<li>Health, Dental, Vision and Life Insurance</li>
<li>Short Term and Long Term Disability</li>
<li>Paid Parental, Medical, Caregiver Leave</li>
<li>Commuter Benefits</li>
<li>Monthly Wellness Stipend</li>
<li>Autonomous Work Environment</li>
<li>In Office Set-Up Reimbursement</li>
<li>Flexible Time Off (FTO) + Holidays</li>
<li>Quarterly Team Gatherings</li>
<li>In Office Amenities</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$190K - $240K</Salaryrange>
      <Skills>Site Reliability Engineering, DevOps, Systems Engineering, Infrastructure Engineering, Python, Go, Terraform, Kubernetes, Docker, GCP, Monitoring/observability solutions, Debugging and performance tuning, Incident management, Infrastructure as code, Configuration management tools, Google Cloud Platform (GCP) services and tools, Modern observability platforms (Prometheus, Grafana, Datadog, etc.), Building reliable systems capable of handling high throughput and low latency, Go and Terraform, Familiarity with working in rapid-growth environments</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Replit</Employername>
      <Employerlogo>https://logos.yubhub.co/replit.com.png</Employerlogo>
      <Employerdescription>Replit is a software creation platform that enables anyone to build applications using natural language. With millions of users worldwide, Replit is a leading platform in the software development industry.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/replit/16c85abc-763c-4f36-ab67-64f416343384?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Foster City, CA</Location>
      <Country></Country>
      <Postedate>2026-03-07</Postedate>
    </job>
    <job>
      <externalid>b7de618e-5e1</externalid>
      <Title>Site Reliability Engineer</Title>
      <Description><![CDATA[<p>Join our Site Reliability Engineering team and help ensure the reliability, scalability, and performance of Replit&#39;s infrastructure that serves millions of developers worldwide. As a Site Reliability Engineer, you will bridge the gap between development and operations, implementing automation and establishing best practices that enable our platform to scale efficiently while maintaining high availability.</p>
<p>We are seeking SREs who are passionate about building and maintaining resilient systems at scale. Your mission will be to design and implement robust monitoring solutions, automate operational tasks, and continuously improve our infrastructure&#39;s reliability and performance.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Design and Implement Observability Solutions: Develop comprehensive monitoring and alerting systems using modern observability tools. Create dashboards and metrics that provide real-time visibility into system health and performance. Implement logging strategies that enable quick problem identification and resolution.</li>
</ul>
<ul>
<li>Drive Automation and Infrastructure as Code: Architect and implement infrastructure automation solutions using tools like Terraform, Ansible, or Pulumi. Design and maintain CI/CD pipelines that enable reliable and consistent deployments. Create self-healing systems that can automatically respond to common failure scenarios.</li>
</ul>
<ul>
<li>Establish SLOs and SLIs: Work with product and engineering teams to define and implement Service Level Objectives (SLOs) and Service Level Indicators (SLIs). Build systems to track and report on these metrics, ensuring we maintain high reliability standards while balancing innovation speed.</li>
</ul>
<ul>
<li>Incident Management and Response: Lead incident response efforts, conducting thorough post-mortems, and implementing improvements to prevent future occurrences. Develop and maintain runbooks for critical services. Build tools and processes that reduce Mean Time To Recovery (MTTR).</li>
</ul>
<ul>
<li>Performance Optimization: Identify and resolve performance bottlenecks across our infrastructure. Implement capacity planning strategies and optimize resource utilization. Work on reducing latency and improving system efficiency across global regions.</li>
</ul>
<p><strong>Requirements</strong></p>
<ul>
<li>4-8 years of experience in Site Reliability Engineering or similar roles (DevOps, Systems Engineering, Infrastructure Engineering)</li>
</ul>
<ul>
<li>Strong programming skills in languages commonly used for automation (Python, Go, or similar)</li>
</ul>
<ul>
<li>Deep understanding of distributed systems</li>
</ul>
<ul>
<li>Experience with container orchestration platforms (Kubernetes) and cloud-native technologies</li>
</ul>
<ul>
<li>Proven track record of implementing and maintaining monitoring/observability solutions</li>
</ul>
<ul>
<li>Strong incident management skills with experience leading incident response</li>
</ul>
<ul>
<li>Experience with infrastructure as code and configuration management tools</li>
</ul>
<p><strong>Bonus Points</strong></p>
<ul>
<li>Experience with Google Cloud Platform (GCP) services and tools</li>
</ul>
<ul>
<li>Knowledge of modern observability platforms (Prometheus, Grafana, Datadog, etc.)</li>
</ul>
<p><strong>What We Value</strong></p>
<ul>
<li>Problem-solving mindset: Ability to approach complex operational challenges systematically and devise effective solutions</li>
</ul>
<ul>
<li>Self-directed and autonomous: Capable of working independently while collaborating effectively with cross-functional teams</li>
</ul>
<ul>
<li>Strong communication skills: Ability to explain complex technical concepts to both technical and non-technical audiences</li>
</ul>
<ul>
<li>Continuous learning: Passion for staying current with industry best practices and new technologies</li>
</ul>
<ul>
<li>Focus on automation: Strong belief in automating repetitive tasks and building self-healing systems</li>
</ul>
<p><strong>Full-Time Employee Benefits Include</strong></p>
<ul>
<li>Competitive Salary &amp; Equity</li>
</ul>
<ul>
<li>401(k) Program with a 4% match</li>
</ul>
<ul>
<li>Health, Dental, Vision and Life Insurance</li>
</ul>
<ul>
<li>Short Term and Long Term Disability</li>
</ul>
<ul>
<li>Paid Parental, Medical, Caregiver Leave</li>
</ul>
<ul>
<li>Commuter Benefits</li>
</ul>
<ul>
<li>Monthly Wellness Stipend</li>
</ul>
<ul>
<li>Autonomous Work Environment</li>
</ul>
<ul>
<li>In Office Set-Up Reimbursement</li>
</ul>
<ul>
<li>Flexible Time Off (FTO) + Holidays</li>
</ul>
<ul>
<li>Quarterly Team Gatherings</li>
</ul>
<ul>
<li>In Office Amenities</li>
</ul>
<p><strong>Want to Learn More About What We Are Up To?</strong></p>
<ul>
<li>Meet the Replit Agent</li>
</ul>
<ul>
<li>Replit: Make an app for that</li>
</ul>
<ul>
<li>Replit Blog</li>
</ul>
<ul>
<li>Amjad TED Talk</li>
</ul>
<p><strong>Interviewing + Culture at Replit</strong></p>
<ul>
<li>Operating Principles</li>
</ul>
<ul>
<li>Reasons not to work at Replit</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange>$160K - $250K</Salaryrange>
      <Skills>Site Reliability Engineering, DevOps, Systems Engineering, Infrastructure Engineering, Python, Go, Distributed systems, Container orchestration platforms, Cloud-native technologies, Monitoring/observability solutions, Incident management, Infrastructure as code, Configuration management tools, Google Cloud Platform, Prometheus, Grafana, Datadog</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Replit</Employername>
      <Employerlogo>https://logos.yubhub.co/replit.com.png</Employerlogo>
      <Employerdescription>Replit is a software creation platform that enables anyone to build applications using natural language. With millions of users worldwide, Replit is a leading provider of software development tools.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/replit/f6e6158e-eb89-4008-81ea-1b7512bc509d?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>United States</Location>
      <Country></Country>
      <Postedate>2026-03-07</Postedate>
    </job>
    <job>
      <externalid>323bc85d-b69</externalid>
      <Title>Staff Infrastructure Engineer</Title>
      <Description><![CDATA[<p><strong>About the Role:</strong></p>
<p>Join our Infrastructure Engineering team and help ensure the reliability, scalability, and performance of Replit&#39;s infrastructure that serves millions of developers worldwide. As a Staff Infrastructure Engineer, you will bridge the gap between development and operations, implementing automation and establishing best practices that enable our platform to scale efficiently while maintaining high availability.</p>
<p><strong>Responsibilities:</strong></p>
<ul>
<li>Drive Automation and Infrastructure as Code: Architect, build, and improve automation to eliminate toil and operational work. Design and maintain CI/CD pipelines and infrastructure automation using tools like Terraform or Pulumi. Create self-healing systems that can automatically respond to common failure scenarios.</li>
</ul>
<ul>
<li>Optimise Performance and Infrastructure: Collaborate with core infrastructure and product teams to performance tune and optimise our cloud deployments (Kubernetes, Docker, GCP). Identify and resolve performance bottlenecks, implement capacity planning strategies, and reduce latency across global regions.</li>
</ul>
<ul>
<li>Elevate Developer Experience: Design and implement improvements to our build, test, and deployment systems to make software delivery faster, safer, and more reliable for all engineers.</li>
</ul>
<ul>
<li>Drive Cross-Company Improvements: Partner directly with service owners across Replit to understand their pain points, and collaborate on implementing build/test/deploy enhancements within their specific services.</li>
</ul>
<ul>
<li>Build Shared Tooling: Create and maintain centralized tooling and automation that improves the entire engineering lifecycle, from local development to production monitoring.</li>
</ul>
<ul>
<li>Debug and Harden Systems: Dive deep into debugging extremely difficult technical problems, making our systems and products more robust, operable, and easier to diagnose.</li>
</ul>
<ul>
<li>Provide Staff-Level Guidance: Review feature and system designs, acting as an owner for the security, scale, and operational integrity of those designs.</li>
</ul>
<ul>
<li>Educate and Mentor: Educate, mentor, and hold accountable the engineering team to improve the reliability of our systems, making reliability a core value of the Replit engineering culture.</li>
</ul>
<ul>
<li>Build and Integrate: Write high-quality, well-tested code to meet the needs of your customers, including building pipelines to integrate with 3rd party vendors.</li>
</ul>
<p><strong>Required Skills and Experience:</strong></p>
<ul>
<li>8-10 years of experience in Infrastructure Engineering or similar roles (DevOps, Systems Engineering, Site Reliability Engineering).</li>
</ul>
<ul>
<li>Strong programming skills in languages like Python or Go.</li>
</ul>
<ul>
<li>You write high-quality, well-tested code.</li>
</ul>
<ul>
<li>Deep understanding of distributed systems. You&#39;ve designed, built, scaled, and maintained production services and know how to compose a service-oriented architecture.</li>
</ul>
<ul>
<li>Experience with container orchestration platforms (Kubernetes) and cloud-native technologies.</li>
</ul>
<ul>
<li>Proven track record of implementing and maintaining monitoring/observability solutions, with strong skills in debugging and performance tuning.</li>
</ul>
<ul>
<li>Strong incident management skills with experience leading incident response and demonstrated critical thinking under pressure.</li>
</ul>
<ul>
<li>Experience with infrastructure as code (e.g., Terraform) and configuration management tools.</li>
</ul>
<ul>
<li>Excellent written and verbal communication skills, with an ability to explain technical concepts clearly and simply and a bias toward open, transparent cultural practices.</li>
</ul>
<ul>
<li>Strong interpersonal skills, with experience working with engineers from junior to principal levels.</li>
</ul>
<ul>
<li>A willingness to dive into understanding, debugging, and improving any layer of the stack.</li>
</ul>
<ul>
<li>You&#39;re passionate about making software creation accessible and empowering the next generation of builders.</li>
</ul>
<p><strong>Bonus Points:</strong></p>
<ul>
<li>Deep experience with Google Cloud Platform (GCP) services and tools.</li>
</ul>
<ul>
<li>Knowledge of modern observability platforms (Prometheus, Grafana, Datadog, etc.).</li>
</ul>
<ul>
<li>Experience designing and building reliable systems capable of handling high throughput and low latency.</li>
</ul>
<ul>
<li>Experience with Go and Terraform.</li>
</ul>
<ul>
<li>Familiarity with working in rapid-growth environments.</li>
</ul>
<ul>
<li>Experience writing company-facing blog posts and training materials.</li>
</ul>
<p><strong>Full-Time Employee Benefits Include:</strong></p>
<ul>
<li>Competitive Salary &amp; Equity</li>
</ul>
<ul>
<li>401(k) Program with a 4% match</li>
</ul>
<ul>
<li>Health, Dental, Vision and Life Insurance</li>
</ul>
<ul>
<li>Short Term and Long Term Disability</li>
</ul>
<ul>
<li>Paid Parental, Medical, Caregiver Leave</li>
</ul>
<ul>
<li>Commuter Benefits</li>
</ul>
<ul>
<li>Monthly Wellness Stipend</li>
</ul>
<ul>
<li>Autonomous Work Environment</li>
</ul>
<ul>
<li>In Office Set-Up Reimbursement</li>
</ul>
<ul>
<li>Flexible Time Off (FTO) + Holidays</li>
</ul>
<ul>
<li>Quarterly Team Gatherings</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$220K – $325K</Salaryrange>
      <Skills>Infrastructure Engineering, DevOps, Systems Engineering, Site Reliability Engineering, Python, Go, Distributed systems, Container orchestration platforms, Cloud-native technologies, Monitoring/observability solutions, Infrastructure as code, Configuration management tools, Google Cloud Platform, Prometheus, Grafana, Datadog, Go, Terraform, Rapid-growth environments, Company-facing blog posts, Training materials</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Replit</Employername>
      <Employerlogo>https://logos.yubhub.co/replit.com.png</Employerlogo>
      <Employerdescription>Replit is a software creation platform that enables anyone to build applications using natural language. With millions of users worldwide, Replit is democratizing software development by removing traditional barriers to application creation.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/replit/6481ec1e-527c-4c1f-a041-2fb5021e7bd5?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Foster City, CA</Location>
      <Country></Country>
      <Postedate>2026-03-07</Postedate>
    </job>
    <job>
      <externalid>4b563c21-dd0</externalid>
      <Title>Software Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p><strong>Software Engineer, Data Infrastructure</strong></p>
<p><strong>Location</strong></p>
<p>San Francisco</p>
<p><strong>Employment Type</strong></p>
<p>Full time</p>
<p><strong>Department</strong></p>
<p>Applied AI</p>
<p><strong>Compensation</strong></p>
<ul>
<li>$185K – $385K • Offers Equity</li>
</ul>
<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>
<p><strong>Benefits</strong></p>
<ul>
<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>
</ul>
<ul>
<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>
</ul>
<ul>
<li>401(k) retirement plan with employer match</li>
</ul>
<ul>
<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>
</ul>
<ul>
<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>
</ul>
<ul>
<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>
</ul>
<ul>
<li>Mental health and wellness support</li>
</ul>
<ul>
<li>Employer-paid basic life and disability coverage</li>
</ul>
<ul>
<li>Annual learning and development stipend to fuel your professional growth</li>
</ul>
<ul>
<li>Daily meals in our offices, and meal delivery credits as eligible</li>
</ul>
<ul>
<li>Relocation support for eligible employees</li>
</ul>
<ul>
<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>
</ul>
<p><strong>About the Team</strong></p>
<p>Data Platform at OpenAI owns the foundational data stack powering critical product, research, and analytics workflows. We operate some of the largest Spark compute fleets in production; design, and build data lakes and metadata systems on Iceberg and Delta with a vision toward exabyte-scale architecture; run high throughput streaming platforms on Kafka and Flink; provide orchestration with Airflow; and support ML feature engineering tooling such as Chronon. Our mission is to deliver reliable, secure, and efficient data access at scale and accelerate intelligent, AI assisted data workflows.</p>
<p><strong>About the Role</strong></p>
<p>This role focuses on building and operating data infrastructure that supports massive compute fleets and storage systems, designed for high performance and scalability. You’ll help design, build, and operate the next generation of data infrastructure at OpenAI. You will scale and harden big data compute and storage platforms, build and support high-throughput streaming systems, build and operate low latency data ingestions, enable secure and governed data access for ML and analytics, and design for reliability and performance at extreme scale.</p>
<p>You will take full lifecycle ownership: architecture, implementation, production operations, and on-call participation.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Design, build, and maintain data infrastructure systems such as distributed compute, data orchestration, distributed storage, streaming infrastructure, machine learning infrastructure while ensuring scalability, reliability, and security</li>
</ul>
<ul>
<li>Ensure our data platform can scale by orders of magnitude while remaining reliable and efficient</li>
</ul>
<ul>
<li>Accelerate company productivity by empowering your fellow engineers &amp; teammates with excellent data tooling and systems</li>
</ul>
<ul>
<li>Collaborate with product, research and analytics teams to build the technical foundations capabilities that unlock new features and experiences</li>
</ul>
<ul>
<li>Own the reliability of the systems you build, including participation in an on-call rotation for critical incidents</li>
</ul>
<p><strong>Requirements</strong></p>
<ul>
<li>4+ years in data infrastructure engineering OR</li>
</ul>
<ul>
<li>4+ years in infrastructure engineering with a strong interest in data</li>
</ul>
<ul>
<li>Take pride in building and operating scalable, reliable, secure systems</li>
</ul>
<ul>
<li>Are comfortable with ambiguity and rapid change</li>
</ul>
<ul>
<li>Have an intrinsic desire to learn and fill in missing skills, and an equally strong talent for sharing learnings clearly and concisely with others</li>
</ul>
<p><strong>About OpenAI</strong></p>
<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of human diversity.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$185K – $385K • Offers Equity</Salaryrange>
      <Skills>data infrastructure engineering, infrastructure engineering, Spark, Kafka, Flink, Airflow, Chronon, Iceberg, Delta, Terraform, distributed systems, machine learning, data science, cloud computing, containerization, DevOps</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>OpenAI</Employername>
      <Employerlogo>https://logos.yubhub.co/openai.com.png</Employerlogo>
      <Employerdescription>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/openai/f763c6b3-5167-4a67-b691-4c3fa2c44156?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
    <job>
      <externalid>af70e58f-a16</externalid>
      <Title>Technical Program Manager - Compute</Title>
      <Description><![CDATA[<p><strong>Summary</strong></p>
<p>Microsoft AI are looking for a talented Technical Program Manager - Compute at their Redmond office. This role sits at the heart of strategic decision-making, turning market data into actionable insights for a company that&#39;s revolutionising AI technology. You&#39;ll work directly with leadership to shape the company&#39;s direction in the AI market.</p>
<p><strong>About the Role</strong></p>
<p>As a Technical Program Manager - Compute, you will drive projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network. You will collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation. You will leverage data and analytics to define metrics, set baselines and targets for fleet efficiency &amp; optimize. You will advocate for AI team&#39;s resource needs with exec and working level partners across Microsoft. You will foster a culture of collaboration, continuous improvement, and growth. You will own the status of key compute projects, proactively identifying risks and proposing solutions to ensure timely delivery. You will communicate program strategies, progress, and results to executive leadership and key stakeholders, advocating for quality and efficiency within the team.</p>
<p><strong>Accountabilities</strong></p>
<ul>
<li>Drive projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network.</li>
<li>Collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation.</li>
</ul>
<p><strong>The Candidate we&#39;re looking for</strong></p>
<p><strong>Experience:</strong></p>
<ul>
<li>6+ years&#39; experience in technical program management, infrastructure engineering, AI/ML, or product development OR equivalent experience.</li>
</ul>
<p><strong>Technical skills:</strong></p>
<ul>
<li>Strong technical curiosity and judgment.</li>
</ul>
<p><strong>Personal attributes:</strong></p>
<ul>
<li>Proactive attitude and enthusiasm for exploring new methods and technologies in compute and infrastructure.</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Competitive salary.</li>
<li>Comprehensive benefits package.</li>
<li>Opportunities for professional growth and development.</li>
<li>Collaborative and dynamic work environment.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>USD $139,900 - $274,800 per year</Salaryrange>
      <Skills>technical program management, infrastructure engineering, AI/ML, product development, data analytics, cloud computing, containerization</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that is on a mission to train the world&apos;s most capable AI frontier models, pushing the boundaries of scale, performance, and product deployment.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/technical-program-manager-compute-2/?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Redmond</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
    <job>
      <externalid>d19375cb-532</externalid>
      <Title>Technical Program Manager - Compute</Title>
      <Description><![CDATA[<p><strong>Summary</strong></p>
<p>Microsoft AI are looking for a talented Technical Program Manager to join their team in Mountain View. This role will be responsible for driving projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network. The ideal candidate will have experience collaborating with AI researchers, engineers, and infrastructure teams to deliver robust, scalable solutions.</p>
<p><strong>About the Role</strong></p>
<p>As a Technical Program Manager at Microsoft AI, you will be responsible for driving projects and programs related to compute infrastructure. This will include forecasting and allocation resource needs like compute, storage, network. You will collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation. You will leverage data and analytics to define metrics, set baselines and targets for fleet efficiency &amp; optimize. You will also advocate for AI team&#39;s resource needs with exec and working level partners across Microsoft.</p>
<p><strong>Accountabilities</strong></p>
<ul>
<li>Drive projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network.</li>
<li>Collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation.</li>
</ul>
<p><strong>The Candidate we&#39;re looking for</strong></p>
<p><strong>Experience:</strong></p>
<ul>
<li>6+ years&#39; experience in technical program management, infrastructure engineering, AI/ML, or product development.</li>
</ul>
<p><strong>Technical skills:</strong></p>
<ul>
<li>Experience with cloud computing platforms, such as Azure or AWS.</li>
<li>Knowledge of containerization and orchestration tools, such as Docker and Kubernetes.</li>
</ul>
<p><strong>Personal attributes:</strong></p>
<ul>
<li>Strong technical curiosity and judgment.</li>
<li>Proactive attitude and enthusiasm for exploring new methods and technologies in compute and infrastructure.</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Competitive salary and benefits package.</li>
<li>Opportunities for professional growth and development.</li>
<li>Collaborative and dynamic work environment.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>USD $139,900 – $274,800 per year</Salaryrange>
      <Skills>technical program management, infrastructure engineering, AI/ML, product development, cloud computing, containerization, orchestration, cloud computing, containerization, orchestration</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that specializes in artificial intelligence and machine learning. They are known for their innovative products and services that empower individuals and organizations to achieve more. Microsoft AI is committed to pushing the boundaries of AI and making it more accessible to everyone.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/technical-program-manager-compute/?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>Mountain View</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
    <job>
      <externalid>50e40039-fc8</externalid>
      <Title>Technical Program Manager - Compute</Title>
      <Description><![CDATA[<p><strong>Summary</strong></p>
<p>Microsoft AI are looking for a talented Technical Program Manager - Compute at their New York office. This role sits at the heart of strategic decision-making, turning market data into actionable insights for a company that&#39;s revolutionising AI technology. You&#39;ll work directly with leadership to shape the company&#39;s direction in the AI market.</p>
<p><strong>About the Role</strong></p>
<p>As a Technical Program Manager - Compute, you will drive projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network. You will collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation. You will leverage data and analytics to define metrics, set baselines and targets for fleet efficiency &amp; optimize. You will advocate for AI team&#39;s resource needs with exec and working level partners across Microsoft. You will foster a culture of collaboration, continuous improvement, and growth. You will own the status of key compute projects, proactively identifying risks and proposing solutions to ensure timely delivery. You will communicate program strategies, progress, and results to executive leadership and key stakeholders, advocating for quality and efficiency within the team.</p>
<p><strong>Accountabilities</strong></p>
<ul>
<li>Drive projects and programs related to compute infrastructure, including forecasting and allocation resource needs like compute, storage, network.</li>
<li>Collaborate with product teams, engineers, researchers, and external partners to identify gaps and drive timelines toward resolution and mitigation.</li>
</ul>
<p><strong>The Candidate we&#39;re looking for</strong></p>
<p><strong>Experience:</strong></p>
<ul>
<li>6+ years&#39; experience in technical program management, infrastructure engineering, AI/ML, or product development OR equivalent experience.</li>
</ul>
<p><strong>Technical skills:</strong></p>
<ul>
<li>Strong technical curiosity and judgment.</li>
</ul>
<p><strong>Personal attributes:</strong></p>
<ul>
<li>Proactive attitude and enthusiasm for exploring new methods and technologies in compute and infrastructure.</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Competitive salary.</li>
<li>Comprehensive benefits package.</li>
<li>Opportunities for professional growth and development.</li>
<li>Collaborative and dynamic work environment.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>USD $139,900 – $274,800 per year</Salaryrange>
      <Skills>technical program management, infrastructure engineering, AI/ML, product development, strong technical curiosity and judgment, proactive attitude and enthusiasm for exploring new methods and technologies in compute and infrastructure</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a leading technology company that is on a mission to train the world&apos;s most capable AI frontier models, pushing the boundaries of scale, performance, and product deployment. They are a startup-like team inside Microsoft AI, created to push the boundaries of AI toward Humanist Superintelligence—ultra-capable systems that remain controllable, safety-aligned, and anchored to human values.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/technical-program-manager-compute-3/?utm_source=yubhub.co&amp;utm_medium=jobs_feed&amp;utm_campaign=apply</Applyto>
      <Location>New York</Location>
      <Country></Country>
      <Postedate>2026-03-06</Postedate>
    </job>
  </jobs>
</source>