<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>0e51b45e-f97</externalid>
      <Title>Executive Business Partner</Title>
      <Description><![CDATA[<p>We&#39;re hiring an Executive Business Partner to support several technical leaders out of our New York office. This is a non-traditional EA role, requiring creativity in adapting to different people&#39;s work styles and the new challenges that emerge at a fast-moving startup.</p>
<p>You will help our team stay focused and organised, managing personal logistics and any tasks that might fall through the cracks. You&#39;ll also be the operational lead for our NYC office , coordinating with other locations, planning team events, and managing travel and schedules.</p>
<p>Key responsibilities include:</p>
<ul>
<li>Managing calendars, scheduling meetings, and coordinating travel for 3-4 technical leaders</li>
<li>Serving as the primary point of contact between your supported leaders and the rest of the company</li>
<li>Owning NYC office operations: coordinating with SF, managing local logistics including swag and supplies, planning team events and offsites</li>
<li>Supporting recruiting coordination efforts</li>
<li>Tracking projects and commitments so nothing falls through the cracks</li>
</ul>
<p>This role entails real autonomy in making decisions without tight supervision. As the senior operations person in NYC, you will have the opportunity to make a significant impact on the team&#39;s success.</p>
<p>We offer a competitive salary range of $200,000 - $250,000 USD per year, depending on background, skills, and experience. We also provide generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>executive</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$200,000 - $250,000 USD per year</Salaryrange>
      <Skills>Executive or administrative support experience, Track record of supporting technical leaders, Adaptability and flexibility, Discretion and professionalism, Coordinating travel and logistics, Experience managing a satellite office, Coordinating across time zones, Adapting to a fast-changing role, Proven professionalism and discretion, Having followed an executive to a new company</Skills>
      <Category>Operations</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a technology company that specialises in developing artificial intelligence products.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5142474008</Applyto>
      <Location>New York, New York</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>b79d9627-55a</externalid>
      <Title>Research Engineer, Infrastructure, Training Systems</Title>
      <Description><![CDATA[<p>We&#39;re seeking an infrastructure research engineer to design and build scalable, efficient training systems for large models. As a key member of our team, you&#39;ll take ownership of the training stack end-to-end, ensuring every GPU cycle drives scientific progress. Your goal is to make experimentation and training at Thinking Machines fast and reliable, allowing our research teams to focus on science, not system bottlenecks.</p>
<p>Key responsibilities include designing, implementing, and optimizing distributed training systems, developing high-performance optimizations, and establishing standards for reliability, maintainability, and security. You&#39;ll collaborate with researchers and engineers to build scalable infrastructure and publish learnings through internal documentation, open-source libraries, or technical reports.</p>
<p>We&#39;re looking for someone who blends deep systems and performance expertise with a curiosity for machine learning at scale. A strong understanding of deep learning frameworks, such as PyTorch, and experience working on distributed training for large models are preferred. If you have a track record of improving research productivity through infrastructure design or process improvements, that&#39;s a plus.</p>
<p>This role is based in San Francisco, California, and offers a competitive salary range of $350,000 - $475,000 USD per year, depending on background, skills, and experience. We sponsor visas and offer generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD per year</Salaryrange>
      <Skills>deep learning frameworks, distributed training, high-performance optimizations, reliability, maintainability, and security, scalable infrastructure, past experience working on distributed training for large models, track record of improving research productivity through infrastructure design or process improvements, contributions to open-source ML infrastructure</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab develops AI products, including ChatGPT and Character.ai, and contributes to open-source projects like PyTorch.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5013932008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>7e28478b-c37</externalid>
      <Title>Research, Audio Expertise</Title>
      <Description><![CDATA[<p>We&#39;re seeking a researcher to advance the frontier of audio capabilities. You&#39;ll explore how audio models enable more natural and efficient communication/collaboration, preserving more information and capturing user intent.</p>
<p>This is a highly collaborative role. You&#39;ll work closely across pre-training, post-training, and product with world-class researchers, infrastructure engineers, and designers.</p>
<p>As a researcher in this role, you&#39;ll be expected to:</p>
<ul>
<li>Own research projects on audio training, low-latency inference, and conversational responsiveness.</li>
<li>Design and train large-scale models that natively support audio input and output.</li>
<li>Investigate scaling behaviour such as how data, model size, and compute affect capability and efficiency.</li>
<li>Build and maintain audio data pipelines, including preprocessing, filtering, segmentation, and alignment for training and evaluation.</li>
<li>Collaborate with data and infrastructure teams to scale audio training efficiently across distributed systems.</li>
<li>Publish and present research that moves the entire community forward.</li>
</ul>
<p>Share code, datasets, and insights that accelerate progress across industry and academia.</p>
<p>This role blends fundamental research and practical engineering, as we do not distinguish between the two roles internally. You will be expected to write high-performance code and read technical reports.</p>
<p>It&#39;s an excellent fit for someone who enjoys both deep theoretical exploration and hands-on experimentation, and who wants to shape the foundations of how AI learns.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid|senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD</Salaryrange>
      <Skills>Python, PyTorch, TensorFlow, JAX, Machine Learning, Deep Learning, Distributed Compute Environments, Probability, Statistics, Real-time Inference, Streaming Architectures, Optimization for Low Latency, Large-Scale Audio or Multimodal Models, Speech, Audio, Voice, or Similar Areas</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a research organisation that focuses on advancing collaborative general intelligence through AI products and open-source projects.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5002212008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>0a2ea62c-943</externalid>
      <Title>Research Engineer, Infrastructure, RL Systems</Title>
      <Description><![CDATA[<p>We&#39;re looking for an infrastructure research engineer to design and build the core systems that enable scalable, efficient training of large models through reinforcement learning.</p>
<p>This role sits at the intersection of research and large-scale systems engineering: a builder who understands both the algorithms behind RL and the realities of distributed training and inference at scale. You&#39;ll wear many hats, from optimising rollout and reward pipelines to enhancing reliability, observability, and orchestration, collaborating closely with researchers and infra teams to make reinforcement learning stable, fast, and production-ready.</p>
<p>Responsibilities:</p>
<ul>
<li>Design, build, and optimise the infrastructure that powers large-scale reinforcement learning and post-training workloads.</li>
</ul>
<ul>
<li>Improve the reliability and scalability of RL training pipeline, distributed RL workloads, and training throughput.</li>
</ul>
<ul>
<li>Develop shared monitoring and observability tools to ensure high uptime, debuggability, and reproducibility for RL systems.</li>
</ul>
<ul>
<li>Collaborate with researchers to translate algorithmic ideas into production-grade training pipelines.</li>
</ul>
<ul>
<li>Build evaluation and benchmarking infrastructure that measures model progress on helpfulness, safety, and factuality.</li>
</ul>
<ul>
<li>Publish and share learnings through internal documentation, open-source libraries, or technical reports that advance the field of scalable AI infrastructure.</li>
</ul>
<p>We&#39;re looking for someone with strong engineering skills, ability to contribute performant, maintainable code and debug in complex codebases. You should have a good understanding of deep learning frameworks (e.g., PyTorch, JAX) and their underlying system architectures.</p>
<p>Experience training or supporting large-scale language models with tens of billions of parameters or more is a plus. Familiarity with monitoring and observability tools (Prometheus, Grafana, OpenTelemetry) is also a plus.</p>
<p>Logistics:</p>
<ul>
<li>Location: This role is based in San Francisco, California.</li>
</ul>
<ul>
<li>Compensation: Depending on background, skills and experience, the expected annual salary range for this position is $350,000 - $475,000 USD.</li>
</ul>
<ul>
<li>Visa sponsorship: We sponsor visas. While we can&#39;t guarantee success for every candidate or role, if you&#39;re the right fit, we&#39;re committed to working through the visa process together.</li>
</ul>
<ul>
<li>Benefits: Thinking Machines offers generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD</Salaryrange>
      <Skills>deep learning frameworks, PyTorch, JAX, complex codebases, scalable AI infrastructure, large-scale language models, monitoring and observability tools, experience training or supporting large-scale language models, familiarity with monitoring and observability tools</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachineslab.com.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a research organisation that focuses on developing collaborative general intelligence.</Employerdescription>
      <Employerwebsite>https://thinkingmachineslab.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5013930008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>07a3c83e-51e</externalid>
      <Title>Research Engineer, Infrastructure, Numerics</Title>
      <Description><![CDATA[<p>We&#39;re looking for an infrastructure research engineer to design and build the core systems that enable efficient large-scale model training with a focus on numerics. You will focus on improving the numerical foundations of our distributed training stack, from precision formats and kernel optimizations to communication frameworks that make training trillion-parameter models stable, scalable, and fast.</p>
<p>This role is ideal for someone who thrives at the intersection of research and systems engineering: a builder who understands both the math of optimization and the realities of distributed compute.</p>
<p>Responsibilities:</p>
<ul>
<li>Design and optimize distributed training infrastructure for large-scale LLMs, focusing on performance, stability, and reproducibility across multi-GPU and multi-node setups.</li>
<li>Implement and evaluate low-precision numerics (for example, BF16, MXFP8, NVFP4) to improve efficiency without sacrificing model quality.</li>
<li>Develop kernels and communication primitives that use hardware-level support for mixed and low-precision arithmetic.</li>
<li>Collaborate with research teams to co-design model architectures and training recipes that align with emerging numeric formats and stability constraints.</li>
<li>Prototype and benchmark scaling strategies such as data, tensor, and pipeline parallelism that integrate precision-adaptive computation and quantized communication.</li>
<li>Contribute to the design of our internal orchestration and monitoring systems to ensure that thousands of distributed experiments can run efficiently and reproducibly.</li>
<li>Publish and share learnings through internal documentation, open-source libraries, or technical reports that advance the field of scalable AI infrastructure.</li>
</ul>
<p>Skills and Qualifications:</p>
<p>Minimum qualifications:</p>
<ul>
<li>Bachelor’s degree or equivalent experience in computer science, electrical engineering, statistics, machine learning, physics, robotics, or similar.</li>
<li>Understanding of deep learning frameworks (e.g., PyTorch, JAX) and their underlying system architectures.</li>
<li>Thrive in a highly collaborative environment involving many, different cross-functional partners and subject matter experts.</li>
<li>A bias for action with a mindset to take initiative to work across different stacks and different teams where you spot the opportunity to make sure something ships.</li>
<li>Strong engineering skills, ability to contribute performant, maintainable code and debug in complex codebases in areas such as floating-point numerics, low-precision arithmetic, and distributed systems.</li>
</ul>
<p>Preferred qualifications , we encourage you to apply if you meet some but not all of these:</p>
<ul>
<li>Familiarity with distributed frameworks such as PyTorch/XLA, DeepSpeed, Megatron-LM.</li>
<li>Experience implementing FP8, INT8, or block-floating point (MX) formats and understanding their numerical trade-offs.</li>
<li>Prior contributions to open-source deep learning infrastructure such as PyTorch, DeepSpeed, or XLA.</li>
<li>Publications, patents, or projects related to numerical optimization, communication-efficient training, or systems for large models.</li>
<li>Experience training and supporting large-scale AI models.</li>
<li>Track record of improving research productivity through infrastructure design or process improvements.</li>
</ul>
<p>Logistics:</p>
<ul>
<li>Location: This role is based in San Francisco, California.</li>
<li>Compensation: Depending on background, skills and experience, the expected annual salary range for this position is $350,000 - $475,000 USD.</li>
<li>Visa sponsorship: We sponsor visas. While we can&#39;t guarantee success for every candidate or role, if you&#39;re the right fit, we&#39;re committed to working through the visa process together.</li>
<li>Benefits: Thinking Machines offers generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD</Salaryrange>
      <Skills>Bachelor’s degree or equivalent experience in computer science, electrical engineering, statistics, machine learning, physics, robotics, or similar, Understanding of deep learning frameworks (e.g., PyTorch, JAX) and their underlying system architectures, Thriving in a highly collaborative environment involving many, different cross-functional partners and subject matter experts, Strong engineering skills, ability to contribute performant, maintainable code and debug in complex codebases in areas such as floating-point numerics, low-precision arithmetic, and distributed systems, Familiarity with distributed frameworks such as PyTorch/XLA, DeepSpeed, Megatron-LM, Experience implementing FP8, INT8, or block-floating point (MX) formats and understanding their numerical trade-offs, Prior contributions to open-source deep learning infrastructure such as PyTorch, DeepSpeed, or XLA, Publications, patents, or projects related to numerical optimization, communication-efficient training, or systems for large models, Experience training and supporting large-scale AI models, Track record of improving research productivity through infrastructure design or process improvements</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a company that creates AI products, including ChatGPT and Character.ai, and contributes to open-source projects like PyTorch.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5013937008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>ea436883-1a8</externalid>
      <Title>Recruiting Coordination and Operations Specialist</Title>
      <Description><![CDATA[<p>We&#39;re hiring a Recruiting Coordinator to help us build a best-in-class recruiting engine as we scale. This is a coordination-heavy role, but the scope goes beyond scheduling - you&#39;ll own the systems and processes that keep the team organized and moving, and you&#39;ll be expected to improve them over time.</p>
<p>This means helping to build the processes, documentation, and systems that grow with us, not just executing the ones that already exist. Beyond scheduling, you&#39;ll plug into broader recruiting operations such as talent mapping, reporting, referral processing, and events - wherever the team needs support.</p>
<p>Your responsibilities will include:</p>
<p>Maintaining data integrity across our ATS and recruiting tools - auditing regularly and keeping things accurate. Supporting talent mapping, referral processing, inbound application triage, and recruiting events. Tracking interview feedback completion and following up to make sure nothing stalls. Identifying process gaps and proposing improvements - this role is expected to make the function better, not just execute. Collecting, organizing, and auditing interview feedback to ensure nothing gets lost or delayed. Maintaining up-to-date recruiting information in documentation and in Slack, supporting data integrity across recruiting systems.</p>
<p>You&#39;ll also manage high-volume interview coordination across multiple time zones, calendars, and stakeholders while maintaining a world-class candidate experience. Coordinate all interview logistics, from travel and expenses to room setup and technical support. Serve as the main point of contact for candidates, communicating in a warm, professional, and informative manner from initial screening through offer. Anticipate scheduling bottlenecks before they happen: monitor interviewer availability, flag conflicts early, and keep hiring velocity high. Scale interviewer pools by partnering with teams to identify, train, and prepare new interviewers as hiring volume grows. Use modern scheduling tools (e.g., Greenhouse, ModernLoop) to automate scheduling, reduce manual work, and improve response speed. Manage high-volume scheduling across time zones, calendars, and multiple stakeholders.</p>
<p>The ideal candidate will have experience coordinating schedules, managing calendars, or supporting operations in a fast-paced environment. They&#39;ll be excellent at written and verbal communication, able to adapt quickly to different teams and roles, and comfortable with Google Workspace, scheduling tools, and Slack.</p>
<p>Logistics: Location: San Francisco, California. Compensation: $140,000 - $200,000 USD per year, depending on background, skills, and experience. Visa sponsorship: Yes. Benefits: Generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$140,000 - $200,000 USD</Salaryrange>
      <Skills>Experience coordinating schedules, managing calendars, or supporting operations in a fast-paced environment, Excellent written and verbal communication skills, Ability to adapt quickly to different teams and roles, Comfortable with Google Workspace, scheduling tools, and Slack, Do-what-it-takes mindset: proactive, detail-oriented, and able to prioritize tasks effectively, 1+ years in a high-volume scheduling or recruiting coordination role, Experience with ATS, HRIS, and scheduling tools like Greenhouse, GoodTime, Workday, and ModernLoop, Experience partnering closely with recruiters and hiring managers, Proven ability to be proactive in ambiguous situations and propose multiple solutions, Experience supporting technical or leadership-level hiring</Skills>
      <Category>Operations</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a technology company that creates AI products, including ChatGPT and Character.ai.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5156656008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>cba88898-896</externalid>
      <Title>Research Engineer, Infrastructure, Kernels</Title>
      <Description><![CDATA[<p>We&#39;re looking for an infrastructure research engineer to design, optimize, and maintain the compute foundations that power large-scale language model training. You will develop high-performance ML kernels (e.g., CUDA, CuTe, Triton), enable efficient low-precision arithmetic, and improve the distributed compute stack that makes training large models possible.</p>
<p>This role is perfect for an engineer who enjoys working close to the metal and across the research boundary. You&#39;ll collaborate with researchers and systems architects to bridge algorithmic design with hardware efficiency. You&#39;ll prototype new kernel implementations, profile performance across hardware generations, and help define the numerical and parallelism strategies that determine how we scale next-generation AI systems.</p>
<p><strong>Responsibilities</strong></p>
<ul>
<li>Design and implement custom ML kernels (e.g., CUDA, CuTe, Triton) for core LLM operations such as attention, matrix multiplication, gating, and normalization, optimized for modern GPU and accelerator architectures.</li>
<li>Design and think through compute primitives to reduce memory bandwidth bottlenecks and improve kernel compute efficiency.</li>
<li>Collaborate with research teams to align kernel-level optimizations with model architecture and algorithmic goals.</li>
<li>Develop and maintain a library of reusable kernels and performance benchmarks that serve as the foundation for internal model training.</li>
<li>Contribute to infrastructure stability and scalability, ensuring reproducibility, consistency across precision formats, and high utilization of compute resources.</li>
<li>Document and share insights through internal talks, technical papers, or open-source contributions to strengthen the broader ML systems community.</li>
</ul>
<p><strong>Skills and Qualifications</strong></p>
<p>Minimum qualifications:</p>
<ul>
<li>Bachelor’s degree or equivalent experience in computer science, electrical engineering, statistics, machine learning, physics, robotics, or similar.</li>
<li>Strong engineering skills, ability to contribute performant, maintainable code and debug in complex codebases</li>
<li>Understanding of deep learning frameworks (e.g., PyTorch, JAX) and their underlying system architectures.</li>
<li>Thrive in a highly collaborative environment involving many, different cross-functional partners and subject matter experts.</li>
<li>A bias for action with a mindset to take initiative to work across different stacks and different teams where you spot the opportunity to make sure something ships.</li>
<li>Proficiency in CUDA, CuTe, Triton, or other GPU programming frameworks.</li>
<li>Demonstrated ability to analyze, profile, and optimize compute-intensive workloads.</li>
</ul>
<p>Preferred qualifications:</p>
<ul>
<li>Experience training or supporting large-scale language models with tens of billions of parameters or more.</li>
<li>Track record of improving research productivity through infrastructure design or process improvements.</li>
<li>Experience developing or tuning kernels for deep learning frameworks such as PyTorch, JAX, or custom accelerators.</li>
<li>Familiarity with tensor parallelism, pipeline parallelism, or distributed data processing frameworks.</li>
<li>Experience implementing low-precision formats (FP8, INT8, block floating point) or contributing to related compiler stacks (e.g., XLA, TVM).</li>
<li>Contributions to open-source GPU, ML systems, or compiler optimization projects.</li>
<li>Prior research or engineering experience in numerical optimization, communication-efficient training, or scalable AI infrastructure.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD</Salaryrange>
      <Skills>CUDA, CuTe, Triton, GPU programming frameworks, Deep learning frameworks (e.g., PyTorch, JAX), Computer science, Electrical engineering, Statistics, Machine learning, Physics, Robotics, Experience training or supporting large-scale language models with tens of billions of parameters or more, Track record of improving research productivity through infrastructure design or process improvements, Experience developing or tuning kernels for deep learning frameworks such as PyTorch, JAX, or custom accelerators, Familiarity with tensor parallelism, pipeline parallelism, or distributed data processing frameworks, Experience implementing low-precision formats (FP8, INT8, block floating point) or contributing to related compiler stacks (e.g., XLA, TVM), Contributions to open-source GPU, ML systems, or compiler optimization projects, Prior research or engineering experience in numerical optimization, communication-efficient training, or scalable AI infrastructure</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a technology company that has created widely used AI products, including ChatGPT and Character.ai, and open-source projects like PyTorch.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5013934008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>9be280f4-cbc</externalid>
      <Title>Software Engineer, Data Infrastructure</Title>
      <Description><![CDATA[<p>We&#39;re looking for an engineer to join our small, high-impact team responsible for architecting and scaling the core infrastructure behind distributed training pipelines, multimodal data catalogs, and intelligent processing systems that operate over petabytes of data.</p>
<p>As a software engineer on our data infrastructure team, you&#39;ll design, build, and operate scalable, fault-tolerant infrastructure for LLM Research: distributed compute, data orchestration, and storage across modalities. You&#39;ll develop high-throughput systems for data ingestion, processing, and transformation , including training data catalogs, deduplication, quality checks, and search. You&#39;ll also build systems for traceability, reproducibility, and robust quality control at every stage of the data lifecycle.</p>
<p>You&#39;ll collaborate with research teams to unlock new features, improve data quality, and accelerate training cycles. You&#39;ll implement and maintain monitoring and alerting to support platform reliability and performance.</p>
<p>If you&#39;re excited by distributed systems, large-scale data mining, open-source tools like Spark, Kafka, Beam, Ray, and Delta Lake, and enjoy building from the ground up, we&#39;d love to hear from you.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>entry|mid|senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD</Salaryrange>
      <Skills>backend language (Python or Rust), distributed compute frameworks (Apache Spark or Ray), cloud infrastructure, data lake architectures, batch and streaming pipelines, Kafka, dbt, Terraform, Airflow, web crawler, deduplication, data mining, search, file formats and storage systems</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a research organisation that focuses on developing collaborative general intelligence.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5013919008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>4ced2159-802</externalid>
      <Title>Research, Vision Expertise</Title>
      <Description><![CDATA[<p>Thinking Machines Lab is seeking a researcher to join their team in San Francisco. The successful candidate will work on advancing the science of visual perception and multimodal learning. They will design architectures that fuse pixels and text, build datasets and evaluation methods that test real-world comprehension, and develop representations that let models ground abstract concepts in the physical world.</p>
<p>The ideal candidate will have expertise in multimodality and experience running large-scale experiments. They will be comfortable contributing to complex engineering systems and have a strong grasp of probability, statistics, and machine learning fundamentals.</p>
<p>This is an evergreen role, meaning that the position is open on an ongoing basis. The company receives many applications, and there may not always be an immediate role that aligns perfectly with the candidate&#39;s experience and skills. However, they encourage candidates to apply and continuously review applications.</p>
<p>Responsibilities:</p>
<ul>
<li>Own research projects on training and performance analysis of multimodal AI models.</li>
<li>Curate and build large-scale datasets and evaluation benchmarks to advance vision capabilities.</li>
<li>Work with data infrastructure engineers, pretraining researchers and engineers, and product teams to create frontier multimodal models and the products that leverage them.</li>
<li>Publish and present research that moves the entire community forward.</li>
</ul>
<p>Skills and Qualifications:</p>
<ul>
<li>Ability to design, run, and analyze experiments thoughtfully, with demonstrated research judgment and empirical rigor.</li>
<li>Understanding of machine learning fundamentals, large-scale training, and distributed compute environments.</li>
<li>Proficiency in Python and familiarity with at least one deep learning framework (e.g., PyTorch, TensorFlow, or JAX).</li>
<li>Comfortable with debugging distributed training and writing code that scales.</li>
<li>Bachelor&#39;s degree or equivalent experience in Computer Science, Machine Learning, Physics, Mathematics, or a related discipline with strong theoretical and empirical grounding.</li>
</ul>
<p>Preferred qualifications include research or engineering contributions in visual reasoning, spatial understanding, or multimodal architecture design, experience developing evaluation frameworks for multimodal tasks, publications or open-source contributions in vision-language modeling, video understanding, or multimodal AI, and a strong grasp of probability, statistics, and ML fundamentals.</p>
<p>Logistics:</p>
<ul>
<li>Location: San Francisco, California.</li>
<li>Compensation: $350,000 - $475,000 USD per year, depending on background, skills, and experience.</li>
<li>Visa sponsorship: Yes.</li>
<li>Benefits: Generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$350,000 - $475,000 USD per year</Salaryrange>
      <Skills>Python, Deep learning framework (e.g., PyTorch, TensorFlow, or JAX), Machine learning fundamentals, Large-scale training, Distributed compute environments, Visual reasoning, Spatial understanding, Multimodal architecture design, Evaluation frameworks for multimodal tasks, Vision-language modeling, Video understanding, Multimodal AI</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachines.ai.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a research organisation that focuses on advancing collaborative general intelligence. They have developed several widely used AI products, including ChatGPT and Character.ai.</Employerdescription>
      <Employerwebsite>https://thinkingmachines.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5002288008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>a0fe4cba-5d3</externalid>
      <Title>Engineering Manager</Title>
      <Description><![CDATA[<p>We&#39;re hiring an Engineering Manager to lead a team of senior and staff-level engineers across ML infrastructure and product. You will help the team build and scale systems that are reliable, performant, and easy to operate.</p>
<p>This role combines collaboration with hand-on work. You’ll partner with tech leads to set the technical direction for your team and own its execution. You should also be ready to go deep on system design and contribute directly when needed.</p>
<p>Responsibilities:</p>
<ul>
<li>Lead and grow a team of senior and staff-level engineers, setting clear expectations and maintaining a high bar for execution.</li>
</ul>
<ul>
<li>Own architecture, system design, and long-term technical direction for your team&#39;s systems, with emphasis on reliability and performance.</li>
</ul>
<ul>
<li>Contribute directly to design reviews, prototyping, and debugging critical issues.</li>
</ul>
<ul>
<li>Partner with researchers and product teams to define roadmaps and prioritize work.</li>
</ul>
<ul>
<li>Hire and close senior engineering talent. Mentor engineers into technical leaders.</li>
</ul>
<p>Skills and Qualifications:</p>
<ul>
<li>Minimum qualifications:</li>
</ul>
<ul>
<li>Bachelor’s degree or equivalent industry experience in computer science, engineering, or similar.</li>
</ul>
<ul>
<li>8+ years of experience building and scaling production systems, including system design and distributed systems.</li>
</ul>
<ul>
<li>3+ years of engineering management experience in high-growth environments.</li>
</ul>
<ul>
<li>Preferred qualifications , we encourage you to apply even if you don’t meet all preferred qualifications, but at least some:</li>
</ul>
<ul>
<li>Experience managing teams of senior or staff-level engineers.</li>
</ul>
<ul>
<li>Background in infrastructure, systems engineering, or developer productivity.</li>
</ul>
<ul>
<li>Familiarity with AI/ML systems, data infrastructure, or high-performance computing.</li>
</ul>
<ul>
<li>Track record of building or contributing to widely used systems, platforms, or tools.</li>
</ul>
<p>Logistics:</p>
<ul>
<li>Compensation: Depending on background, skills and experience, the expected annual salary range for this position is $400,000 - $500,000 USD.</li>
</ul>
<ul>
<li>Visa sponsorship: We sponsor visas. While we can&#39;t guarantee success for every candidate or role, if you&#39;re the right fit, we&#39;re committed to working through the visa process together.</li>
</ul>
<ul>
<li>Benefits: Thinking Machines offers generous health, dental, and vision benefits, unlimited PTO, paid parental leave, and relocation support as needed.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$400,000 - $500,000 USD</Salaryrange>
      <Skills>computer science, engineering, system design, distributed systems, engineering management, infrastructure, systems engineering, developer productivity, AI/ML systems, data infrastructure, high-performance computing</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachineslab.com.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is a company that empowers humanity through advancing collaborative general intelligence. It has created some of the most widely used AI products.</Employerdescription>
      <Employerwebsite>https://thinkingmachineslab.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5165725008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>594b20c4-c28</externalid>
      <Title>Infrastructure Engineer, Security</Title>
      <Description><![CDATA[<p>We&#39;re looking for an infrastructure engineer to own and evolve the security infrastructure that underpins our foundation models. In this role, you&#39;ll work across compute, storage, networking, and data platforms, making sure our systems are secure, reliable, and built to scale.</p>
<p>You&#39;ll shape controls, architecture, and tooling so that security is part of how the platform works by default. You&#39;ll partner closely with research and product teams, enabling them to move quickly while keeping our models, data, and environments protected.</p>
<p>Key responsibilities include:</p>
<p>Architecting security patterns for platforms and services, including network segmentation, service-to-service authentication, RBAC, and policy enforcement in Kubernetes and cloud environments.</p>
<p>Managing identity, access, and secrets for humans and services: workload and cross-cloud identity, least-privilege IAM, and secrets management.</p>
<p>Building secure platforms for data ingestion, processing, and curation: classification, encryption, access controls, and safe sharing patterns across teams.</p>
<p>Writing threat models and reviewing designs with researchers and engineers to help them ship features and experiments in a safe, scalable way.</p>
<p>Automating security checks and building guardrails: policy-as-code, secure infrastructure baselines, validation in CI/CD, and tools that make the secure path the easiest one.</p>
<p>Requirements include:</p>
<p>Bachelor&#39;s degree or equivalent experience in engineering, or similar.</p>
<p>Strong background with containers and orchestration (e.g., Kubernetes) and how to secure them (namespaces, network policies, pod security, admission controls, etc.).</p>
<p>Practical experience with Infrastructure as Code (Terraform or similar), including secure patterns for provisioning networks, IAM, and shared services.</p>
<p>Solid understanding of cloud networking and security: VPCs, load balancers, service discovery, mTLS, firewalls, and zero-trust-style architectures.</p>
<p>Proficiency with a systems language such as Rust and scripting in Python for building platform components and internal tools.</p>
<p>Evidence of owning complex, production-critical systems, including debugging issues that span infra, security, and application layers.</p>
<p>Preferred qualifications include experience with ML infrastructure, GPU clusters, or large-scale training environments, as well as background in AI labs, HPC environments, or ML-heavy organizations.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$200,000 - $475,000 USD</Salaryrange>
      <Skills>Kubernetes, Infrastructure as Code, Cloud Networking and Security, Systems Language (Rust), Scripting (Python), ML Infrastructure, GPU Clusters, Large-Scale Training Environments, AI Labs, HPC Environments</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Thinking Machines Lab</Employername>
      <Employerlogo>https://logos.yubhub.co/thinkingmachineslab.com.png</Employerlogo>
      <Employerdescription>Thinking Machines Lab is building a future where everyone has access to the knowledge and tools to make AI work for their unique needs and goals.</Employerdescription>
      <Employerwebsite>https://thinkingmachineslab.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/thinkingmachines/jobs/5015964008</Applyto>
      <Location>San Francisco</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
  </jobs>
</source>