<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>64b848a0-7f0</externalid>
      <Title>5-6 month software development internship - GPU Simulation Data Processing Internship</Title>
      <Description><![CDATA[<p>Our internship programs offer real-world projects, hands-on experience, and opportunities to collaborate with passionate teams globally. Explore your interests, share your ideas, and bring them to life while shaping your career path within our inclusive culture that fosters innovation and collaboration.</p>
<p>As a GPU Simulation Data Processing Intern, you will extend a simulation data processing toolbox on GPU architectures to enhance performance and scalability for Synopsys solvers and applications. You will collaborate with the DPF group to design and implement generic tools for pre- and post-processing in numerical simulation domains. You will work closely with team members to identify and solve challenges based on GPU technologies, contributing innovative solutions. You will participate in code reviews, documentation, and optimization of existing tools to ensure high-quality and robust deliverables. You will share ideas and actively contribute to a collaborative, creative environment.</p>
<p>Key responsibilities include:</p>
<ul>
<li>Extending a simulation data processing toolbox on GPU architectures to enhance performance and scalability for Synopsys solvers and applications</li>
<li>Collaborating with the DPF group to design and implement generic tools for pre- and post-processing in numerical simulation domains</li>
<li>Working closely with team members to identify and solve challenges based on GPU technologies, contributing innovative solutions</li>
<li>Participating in code reviews, documentation, and optimization of existing tools to ensure high-quality and robust deliverables</li>
<li>Sharing ideas and actively contributing to a collaborative, creative environment</li>
</ul>
<p>Requirements include:</p>
<ul>
<li>Currently pursuing a master’s or engineering degree in Computer Science or a related field</li>
<li>Strong programming skills in C++ and Python; familiarity with Github for version control</li>
<li>Experience with Cuda and understanding of GPU architectures</li>
<li>Ability to work effectively as part of a team, take initiative, and propose innovative solutions to technical challenges</li>
</ul>
<p>This is a 6-month internship with a full-time schedule (35 hours/week). The internship is located in Lyon, France, and requires the ability to work in-office.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>internship</Jobtype>
      <Experiencelevel>entry</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>C++, Python, CUDA, GPU architectures, Github</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Synopsys</Employername>
      <Employerlogo>https://logos.yubhub.co/careers.synopsys.com.png</Employerlogo>
      <Employerdescription>Synopsys is a leading provider of electronic design automation (EDA) software and services for designing, verifying, and manufacturing electronic systems and microelectronic components.</Employerdescription>
      <Employerwebsite>https://careers.synopsys.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://careers.synopsys.com/job/villeurbanne/5-6-month-software-development-internship-gpu-simulation-data-processing-internship-toolbox-cuda/44408/93816738928</Applyto>
      <Location>Lyon</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>d8ab9aea-841</externalid>
      <Title>Senior Client Engineer - Rendering</Title>
      <Description><![CDATA[<p>Electronic Arts creates next-level entertainment experiences that inspire players and fans around the world. As one of the largest sports entertainment platforms in the world, EA SPORTS FC is redefining football with genre-leading interactive experiences, connecting a global community of fans to The World&#39;s Game through innovation and unrivalled authenticity.</p>
<p>We invite you to join our passionate and dynamic team as we pioneer the future of football fandom. As a Senior Client Engineer - Rendering, you will be responsible for using our existing game engine for game development, building upon the foundation of our current engine to enhance its functionality and performance. You will implement and fine-tune real-time shaders, dynamic lighting, and post-processing effects to enhance visual fidelity while maintaining target frame rates.</p>
<p>Responsibilities:</p>
<ul>
<li>Use our existing game engine for game development, building upon the foundation of our current engine to enhance its functionality and performance.</li>
<li>Implement and fine-tune real-time shaders, dynamic lighting, and post-processing effects to enhance visual fidelity while maintaining target frame rates.</li>
<li>Develop new tools and improve existing ones based on our existing toolchain.</li>
<li>Identify performance bottlenecks within games and improve them.</li>
<li>Communicate with game designers and artists, ensuring that program functionality aligns with design requirements.</li>
<li>Help create technical specifications and software architecture documents.</li>
<li>Communicate project progress and risks to superiors promptly.</li>
</ul>
<p>Qualifications:</p>
<ul>
<li>5+ years of professional experience in real-time graphics or rendering development for games.</li>
<li>Experience with 3D mathematics and algorithms, GPU architectures, and graphics APIs (such as Metal, Vulkan, or DirectX).</li>
<li>Experience in profiling and optimizing rendering performance on iOS and Android platforms.</li>
<li>Expert in C++ and master at least one shader language (such as HLSL or GLSL).</li>
<li>Hands-on experience developing in-house game engines or modifying at least one commercial game engine (such as Unreal Engine or Unity).</li>
<li>A collaborative mindset for working with art, design, and QA teams.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>C++, Real-time graphics, Rendering development, Game engine development, Shader languages włsl, Metal, Vulkan, DirectX, 3D mathematics, GPU architectures</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Electronic Arts</Employername>
      <Employerlogo>https://logos.yubhub.co/jobs.ea.com.png</Employerlogo>
      <Employerdescription>Electronic Arts is a leading video game developer and publisher with a portfolio of popular titles such as EA SPORTS FC.</Employerdescription>
      <Employerwebsite>https://jobs.ea.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ea.com/en_US/careers/JobDetail/213520-Rendering-Engineer-III/213520</Applyto>
      <Location>Shanghai</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>002c5e0f-f56</externalid>
      <Title>Member of Technical Staff, Software Co-Design AI HPC Systems</Title>
      <Description><![CDATA[<p>Our team&#39;s mission is to architect, co-design, and productionize next-generation AI systems at datacenter scale. We operate at the intersection of models, systems software, networking, storage, and AI hardware, optimizing end-to-end performance, efficiency, reliability, and cost.</p>
<p>We pursue this mission through deep hardware–software co-design, combining rigorous systems thinking with hands-on engineering. The team invests heavily in understanding real production workloads large-scale training, inference, and emerging multimodal models and translating those insights into concrete improvements across the stack: from kernels, runtimes, and distributed systems, all the way down to silicon-level trade-offs and datacenter-scale architectures.</p>
<p>This role sits at the boundary between exploration and production. You will work closely with internal infrastructure, hardware, compiler, and product teams, as well as external partners across the hardware and systems ecosystem. Our operating model emphasizes rapid ideation and prototyping, followed by disciplined execution to drive high-leverage ideas into production systems that operate at massive scale.</p>
<p>In addition to delivering real-world impact on large-scale AI platforms, the team actively contributes to the broader research and engineering community. Our work aligns closely with leading communities in ML systems, distributed systems, computer architecture, and high-performance computing, and we regularly publish, prototype, and open-source impactful technologies where appropriate.</p>
<p>About the Team</p>
<p>We build foundational AI infrastructure that enables large-scale training and inference across diverse workloads and rapidly evolving hardware generations. Our work directly shapes how AI systems are designed, deployed, and scaled today and into the future. Engineers on this team operate with end-to-end ownership, deep technical rigor, and a strong bias toward real-world impact.</p>
<p>Microsoft Superintelligence Team</p>
<p>Microsoft Superintelligence team’s mission is to empower every person and every organization on the planet to achieve more. As employees we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.</p>
<p>This role is part of Microsoft AI’s Superintelligence Team. The MAIST is a startup-like team inside Microsoft AI, created to push the boundaries of AI toward Humanist Superintelligence,ultra-capable systems that remain controllable, safety-aligned, and anchored to human values. Our mission is to create AI that amplifies human potential while ensuring humanity remains firmly in control. We aim to deliver breakthroughs that benefit society,advancing science, education, and global well-being.</p>
<p>We’re also fortunate to partner with incredible product teams giving our models the chance to reach billions of users and create immense positive impact. If you’re a brilliant, highly-ambitious and low ego individual, you’ll fit right in,come and join us as we work on our next generation of models!</p>
<p>Responsibilities</p>
<p>Lead the co-design of AI systems across hardware and software boundaries, spanning accelerators, interconnects, memory systems, storage, runtimes, and distributed training/inference frameworks.</p>
<p>Drive architectural decisions by analyzing real workloads, identifying bottlenecks across compute, communication, and data movement, and translating findings into actionable system and hardware requirements.</p>
<p>Co-design and optimize parallelism strategies, execution models, and distributed algorithms to improve scalability, utilization, reliability, and cost efficiency of large-scale AI systems.</p>
<p>Develop and evaluate what-if performance models to project system behavior under future workloads, model architectures, and hardware generations, providing early guidance to hardware and platform roadmaps.</p>
<p>Partner with compiler, kernel, and runtime teams to unlock the full performance of current and next-generation accelerators, including custom kernels, scheduling strategies, and memory optimizations.</p>
<p>Influence and guide AI hardware design at system and silicon levels, including accelerator microarchitecture, interconnect topology, memory hierarchy, and system integration trade-offs.</p>
<p>Lead cross-functional efforts to prototype, validate, and productionize high-impact co-design ideas, working across infrastructure, hardware, and product teams.</p>
<p>Mentor senior engineers and researchers, set technical direction, and raise the overall bar for systems rigor, performance engineering, and co-design thinking across the organization.</p>
<p>Qualifications</p>
<p>Minimum Qualifications</p>
<p>Bachelor’s degree in Computer Science, Computer Engineering, Electrical Engineering, or a related technical field, or equivalent practical experience.</p>
<p>10+ years of experience (or equivalent depth) working across systems software, hardware architecture, or AI infrastructure , with demonstrated impact at scale.</p>
<p>Strong background in one or more of the following areas: AI accelerator or GPU architectures Distributed systems and large-scale AI training/inference High-performance computing (HPC) and collective communications ML systems, runtimes, or compilers Performance modeling, benchmarking, and systems analysis Hardware–software co-design for AI workloads Proficiency in systems-level programming (e.g., C/C++, CUDA, Python) and performance-critical software development.</p>
<p>Proven ability to work across organizational boundaries and influence technical decisions involving multiple stakeholders.</p>
<p>Preferred Qualifications</p>
<p>Experience designing or operating large-scale AI clusters for training or inference.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>AI accelerator or GPU architectures, Distributed systems and large-scale AI training/inference, High-performance computing (HPC) and collective communications, ML systems, runtimes, or compilers, Performance modeling, benchmarking, and systems analysis, Hardware–software co-design for AI workloads, Proficiency in systems-level programming (e.g., C/C++, CUDA, Python)</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a subsidiary of Microsoft Corporation, a multinational technology company that develops, manufactures, licenses, and supports a wide range of software products, services, and devices.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-software-co-design-ai-hpc-systems-mai-superintelligence-team-5/</Applyto>
      <Location>Zürich</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>51b57192-d10</externalid>
      <Title>Member of Technical Staff, Capacity &amp; Efficiency Infrastructure - MAI Superintelligence Team</Title>
      <Description><![CDATA[<p>Microsoft AI is looking for a Member of Technical Staff – Capacity &amp; Efficiency Infrastructure to help us improve manage, and improve the efficiency of, our compute fleet. We&#39;re seeking someone who brings an abundance of positive energy, empathy, and kindness to the team every day, in addition to being highly effective. The ideal candidate enjoys building world-class consumer experiences and products in a fast-paced environment. You will actively contribute to the development of AI models powering our innovative products. Expect to wear multiple hats and work across engineering, research, and everything in between.</p>
<p>Your contributions will span model architecture, data curation, training and inference infrastructure, evaluation protocols, alignment and reinforcement learning from human feedback (RLHF), and many other exciting topics at the cutting edge of AI. Microsoft AI is building the training infrastructure that powers frontier-scale models and advances research toward humanist superintelligence. As a Member of Technical Staff – Capacity &amp; Efficiency, you will contribute to a fast-moving codebase that enables training at an unprecedented scale. This role will require building software and mathematical models for measuring the effectiveness of our capacity usage and then developing tools and techniques to help us improve. This will require you to partner with ML researchers to scale up the latest research recipes, implement new forms of distributed training parallelism, and ensure the reliability and performance of thousands of GPUs across our supercomputing fleet. Profiling, benchmarking, debugging, and fine-grained optimization are core to this role, demanding both engineering rigor and creativity.</p>
<p>Microsoft Superintelligence Team:</p>
<p>The MAIST is a startup-like team inside Microsoft AI, created to push the boundaries of AI toward Humanist Superintelligence,ultra-capable systems that remain controllable, safety-aligned, and anchored to human values. Our mission is to create AI that amplifies human potential while ensuring humanity remains firmly in control. We aim to deliver breakthroughs that benefit society,advancing science, education, and global well-being. We’re also fortunate to partner with incredible product teams giving our models the chance to reach billions of users and create immense positive impact.</p>
<p>Responsibilities:</p>
<ul>
<li>Design, implement, test, and optimize distributed training infrastructure in Python and C++ for large-scale GPU clusters.</li>
<li>Build and evolve telemetry systems to provide visibility into infrastructure &amp; ML model performance, utilization, and cost related metrics.</li>
<li>Profile, benchmark, and debug performance bottlenecks across compute, memory, networking, and storage subsystems.</li>
<li>Drive architectural improvements across various ML services which deliver measurable efficiency improvements.</li>
<li>Build and evolve tools to automatically provide insights and recommendations to improve fleet-wide efficiency.</li>
<li>Optimize collective communication libraries (e.g., NCCL) for emerging NVLink and InfiniBand topologies.</li>
<li>Partner with ML researchers and infrastructure engineers to understand their plans and future needs and develop plans to balance growth with efficiency.</li>
<li>Collaborate with hardware teams to optimize for next-generation accelerators (NVIDIA, MAIA, and beyond).</li>
</ul>
<p>Qualifications:</p>
<ul>
<li>Bachelor’s Degree in Computer Science, or related technical discipline AND 6+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR equivalent experience.</li>
<li>Deep understanding of the fundamentals of GPU architectures and DL/LLM architectures.</li>
<li>Deep experience in profiling and analyzing performance in large-scale distributed computing systems.</li>
<li>Experience with low-level GPU programming (CUDA, Triton, NCCL) and frameworks such as PyTorch or JAX.</li>
<li>Experience in leading technical projects and supporting architectural decisions with data.</li>
<li>Experience building infrastructure for large-scale machine learning or generative AI workloads.</li>
<li>Experience in networking (InfiniBand, NVLink), storage systems, or distributed training parallelisms.</li>
<li>Track record of contributing to high-performance computing or large-scale AI infrastructure projects.</li>
</ul>
<p>Software Engineering IC4 – The typical base pay range for this role across the U.S. is USD $119,800 – $234,700 per year.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>C, C++, Python, GPU architectures, DL/LLM architectures, low-level GPU programming, PyTorch, JAX, networking, storage systems, distributed training parallelisms</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a technology company that develops and markets software products and services. It is one of the largest and most successful companies in the world.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-capacity-efficiency-infrastructure-mai-superintelligence-team-2/</Applyto>
      <Location>Redmond</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>540d480d-7d6</externalid>
      <Title>Member of Technical Staff, Capacity &amp; Efficiency Infrastructure - MAI Superintelligence Team</Title>
      <Description><![CDATA[<p>Microsoft AI is looking for a Member of Technical Staff – Capacity &amp; Efficiency Infrastructure to help us improve manage, and improve the efficiency of, our compute fleet. We&#39;re seeking someone who brings an abundance of positive energy, empathy, and kindness to the team every day, in addition to being highly effective. The ideal candidate enjoys building world-class consumer experiences and products in a fast-paced environment. You will actively contribute to the development of AI models powering our innovative products. Expect to wear multiple hats and work across engineering, research, and everything in between. Your contributions will span model architecture, data curation, training and inference infrastructure, evaluation protocols, alignment and reinforcement learning from human feedback (RLHF), and many other exciting topics at the cutting edge of AI. Microsoft AI is building the training infrastructure that powers frontier-scale models and advances research toward humanist superintelligence. As a Member of Technical Staff – Capacity &amp; Efficiency, you will contribute to a fast-moving codebase that enables training at an unprecedented scale. This role will require building software and mathematical models for measuring the effectiveness of our capacity usage and then developing tools and techniques to help us improve. This will require you to partner with ML researchers to scale up the latest research recipes, implement new forms of distributed training parallelism, and ensure the reliability and performance of thousands of GPUs across our supercomputing fleet. Profiling, benchmarking, debugging, and fine-grained optimization are core to this role, demanding both engineering rigor and creativity.</p>
<p>Responsibilities: Design, implement, test, and optimize distributed training infrastructure in Python and C++ for large-scale GPU clusters. Build and evolve telemetry systems to provide visibility into infrastructure &amp; ML model performance, utilization, and cost related metrics. Profile, benchmark, and debug performance bottlenecks across compute, memory, networking, and storage subsystems. Drive architectural improvements across various ML services which deliver measurable efficiency improvements. Build and evolve tools to automatically provide insights and recommendations to improve fleet-wide efficiency. Optimize collective communication libraries (e.g., NCCL) for emerging NVLink and InfiniBand topologies. Partner with ML researchers and infrastructure engineers to understand their plans and future needs and develop plans to balance growth with efficiency. Collaborate with hardware teams to optimize for next-generation accelerators (NVIDIA, MAIA, and beyond). Embody our Culture and Values.</p>
<p>Qualifications: Required Qualifications: Bachelor’s Degree in Computer Science, or related technical discipline AND 6+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR equivalent experience. Preferred Qualifications: Bachelor’s Degree in Computer Science or related technical field AND 10+ years technical engineering experience with coding in languages including, but not limited to, C++ or Python OR Master’s Degree in Computer Science or related technical field AND 8+ years technical engineering experience with coding in languages including, but not limited to, C++ or Python OR equivalent experience. Deep understanding of the fundamentals of GPU architectures and DL/LLM architectures. Deep experience in profiling and analyzing performance in large-scale distributed computing systems. Deep experience in profiling and analyzing performance in ML models especially GenAI models. Experience with low-level GPU programming (CUDA, Triton, NCCL) and frameworks such as PyTorch or JAX. Experience in leading technical projects and supporting architectural decisions with data. Experience building infrastructure for large-scale machine learning or generative AI workloads. Experience in networking (InfiniBand, NVLink), storage systems, or distributed training parallelisms. Track record of contributing to high-performance computing or large-scale AI infrastructure projects.</p>
<p>Software Engineering IC4 – The typical base pay range for this role across the U.S. is USD $119,800 – $234,700 per year. There is a different range applicable to specific work locations, within the San Francisco Bay area and New York City.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$119,800 – $234,700 per year</Salaryrange>
      <Skills>C, C++, C#, Java, JavaScript, Python, GPU architectures, DL/LLM architectures, low-level GPU programming, PyTorch, JAX, networking, storage systems, distributed training parallelisms</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a technology company that develops and markets software products and services. It is one of the largest and most influential technology companies in the world.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-capacity-efficiency-infrastructure-mai-superintelligence-team/</Applyto>
      <Location>Mountain View</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>a51375e8-30e</externalid>
      <Title>Member of Technical Staff, Software Co-Design AI HPC Systems</Title>
      <Description><![CDATA[<p>Our team&#39;s mission is to architect, co-design, and productionize next-generation AI systems at datacenter scale. We operate at the intersection of models, systems software, networking, storage, and AI hardware, optimizing end-to-end performance, efficiency, reliability, and cost. Our work spans today&#39;s frontier AI workloads and directly shapes the next generation of accelerators, system architectures, and large-scale AI platforms. We pursue this mission through deep hardware–software co-design, combining rigorous systems thinking with hands-on engineering. The team invests heavily in understanding real production workloads large-scale training, inference, and emerging multimodal models and translating those insights into concrete improvements across the stack: from kernels, runtimes, and distributed systems, all the way down to silicon-level trade-offs and datacenter-scale architectures. This role sits at the boundary between exploration and production. You will work closely with internal infrastructure, hardware, compiler, and product teams, as well as external partners across the hardware and systems ecosystem. Our operating model emphasizes rapid ideation and prototyping, followed by disciplined execution to drive high-leverage ideas into production systems that operate at massive scale. In addition to delivering real-world impact on large-scale AI platforms, the team actively contributes to the broader research and engineering community. Our work aligns closely with leading communities in ML systems, distributed systems, computer architecture, and high-performance computing, and we regularly publish, prototype, and open-source impactful technologies where appropriate.</p>
<p>About the Team</p>
<p>We build foundational AI infrastructure that enables large-scale training and inference across diverse workloads and rapidly evolving hardware generations. Our work directly shapes how AI systems are designed, deployed, and scaled today and into the future. Engineers on this team operate with end-to-end ownership, deep technical rigor, and a strong bias toward real-world impact.</p>
<p>Microsoft Superintelligence Team</p>
<p>Microsoft Superintelligence team’s mission is to empower every person and every organization on the planet to achieve more. As employees we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.</p>
<p>This role is part of Microsoft AI’s Superintelligence Team. The MAIST is a startup-like team inside Microsoft AI, created to push the boundaries of AI toward Humanist Superintelligence—ultra-capable systems that remain controllable, safety-aligned, and anchored to human values. Our mission is to create AI that amplifies human potential while ensuring humanity remains firmly in control. We aim to deliver breakthroughs that benefit society—advancing science, education, and global well-being. We’re also fortunate to partner with incredible product teams giving our models the chance to reach billions of users and create immense positive impact. If you’re a brilliant, highly-ambitious and low ego individual, you’ll fit right in—come and join us as we work on our next generation of models!</p>
<p>Responsibilities</p>
<p>Lead the co-design of AI systems across hardware and software boundaries, spanning accelerators, interconnects, memory systems, storage, runtimes, and distributed training/inference frameworks. Drive architectural decisions by analyzing real workloads, identifying bottlenecks across compute, communication, and data movement, and translating findings into actionable system and hardware requirements. Co-design and optimize parallelism strategies, execution models, and distributed algorithms to improve scalability, utilization, reliability, and cost efficiency of large-scale AI systems. Develop and evaluate what-if performance models to project system behavior under future workloads, model architectures, and hardware generations, providing early guidance to hardware and platform roadmaps. Partner with compiler, kernel, and runtime teams to unlock the full performance of current and next-generation accelerators, including custom kernels, scheduling strategies, and memory optimizations. Influence and guide AI hardware design at system and silicon levels, including accelerator microarchitecture, interconnect topology, memory hierarchy, and system integration trade-offs. Lead cross-functional efforts to prototype, validate, and productionize high-impact co-design ideas, working across infrastructure, hardware, and product teams. Mentor senior engineers and researchers, set technical direction, and raise the overall bar for systems rigor, performance engineering, and co-design thinking across the organization.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>AI accelerator or GPU architectures, Distributed systems and large-scale AI training/inference, High-performance computing (HPC) and collective communications, ML systems, runtimes, or compilers, Performance modeling, benchmarking, and systems analysis, Hardware–software co-design for AI workloads, Proficiency in systems-level programming (e.g., C/C++, CUDA, Python) and performance-critical software development, Experience designing or operating large-scale AI clusters for training or inference, Deep familiarity with LLMs, multimodal models, or recommendation systems, and their systems-level implications, Experience with accelerator interconnects and communication stacks (e.g., NCCL, MPI, RDMA, high-speed Ethernet or InfiniBand), Background in performance modeling and capacity planning for future hardware generations, Prior experience contributing to or leading hardware roadmaps, silicon bring-up, or platform architecture reviews, Publications, patents, or open-source contributions in systems, architecture, or ML systems</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a technology company that develops and markets software products and services. It is one of the largest and most successful technology companies in the world.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-software-co-design-ai-hpc-systems-mai-superintelligence-team-3/</Applyto>
      <Location>London</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>cd1a0d16-311</externalid>
      <Title>Member of Technical Staff, Software Co-Design AI HPC Systems</Title>
      <Description><![CDATA[<p>Our team&#39;s mission is to architect, co-design, and productionize next-generation AI systems at datacenter scale. We operate at the intersection of models, systems software, networking, storage, and AI hardware, optimizing end-to-end performance, efficiency, reliability, and cost.</p>
<p>We pursue this mission through deep hardware–software co-design, combining rigorous systems thinking with hands-on engineering. The team invests heavily in understanding real production workloads large-scale training, inference, and emerging multimodal models and translating those insights into concrete improvements across the stack: from kernels, runtimes, and distributed systems, all the way down to silicon-level trade-offs and datacenter-scale architectures.</p>
<p>This role sits at the boundary between exploration and production. You will work closely with internal infrastructure, hardware, compiler, and product teams, as well as external partners across the hardware and systems ecosystem. Our operating model emphasizes rapid ideation and prototyping, followed by disciplined execution to drive high-leverage ideas into production systems that operate at massive scale.</p>
<p>In addition to delivering real-world impact on large-scale AI platforms, the team actively contributes to the broader research and engineering community. Our work aligns closely with leading communities in ML systems, distributed systems, computer architecture, and high-performance computing, and we regularly publish, prototype, and open-source impactful technologies where appropriate.</p>
<p>Microsoft Superintelligence Team
Microsoft Superintelligence team’s mission is to empower every person and every organization on the planet to achieve more. As employees we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.</p>
<p>This role is part of Microsoft AI’s Superintelligence Team. The MAIST is a startup-like team inside Microsoft AI, created to push the boundaries of AI toward Humanist Superintelligence—ultra-capable systems that remain controllable, safety-aligned, and anchored to human values. Our mission is to create AI that amplifies human potential while ensuring humanity remains firmly in control. We aim to deliver breakthroughs that benefit society—advancing science, education, and global well-being. We’re also fortunate to partner with incredible product teams giving our models the chance to reach billions of users and create immense positive impact.</p>
<p>Responsibilities
Lead the co-design of AI systems across hardware and software boundaries, spanning accelerators, interconnects, memory systems, storage, runtimes, and distributed training/inference frameworks.</p>
<p>Drive architectural decisions by analyzing real workloads, identifying bottlenecks across compute, communication, and data movement, and translating findings into actionable system and hardware requirements.</p>
<p>Co-design and optimize parallelism strategies, execution models, and distributed algorithms to improve scalability, utilization, reliability, and cost efficiency of large-scale AI systems.</p>
<p>Develop and evaluate what-if performance models to project system behavior under future workloads, model architectures, and hardware generations, providing early guidance to hardware and platform roadmaps.</p>
<p>Partner with compiler, kernel, and runtime teams to unlock the full performance of current and next-generation accelerators, including custom kernels, scheduling strategies, and memory optimizations.</p>
<p>Influence and guide AI hardware design at system and silicon levels, including accelerator microarchitecture, interconnect topology, memory hierarchy, and system integration trade-offs.</p>
<p>Lead cross-functional efforts to prototype, validate, and productionize high-impact co-design ideas, working across infrastructure, hardware, and product teams.</p>
<p>Mentor senior engineers and researchers, set technical direction, and raise the overall bar for systems rigor, performance engineering, and co-design thinking across the organization.</p>
<p>Qualifications
Bachelor’s Degree in Computer Science or related technical field AND 6+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR equivalent experience.</p>
<p>Additional or Preferred Qualifications
Master’s Degree in Computer Science or related technical field AND 8+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR Bachelor’s Degree in Computer Science or related technical field AND 12+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python OR equivalent experience.</p>
<p>Strong background in one or more of the following areas: AI accelerator or GPU architectures Distributed systems and large-scale AI training/inference High-performance computing (HPC) and collective communications ML systems, runtimes, or compilers Performance modeling, benchmarking, and systems analysis Hardware–software co-design for AI workloads Proficiency in systems-level programming (e.g., C/C++, CUDA, Python) and performance-critical software development.</p>
<p>Proven ability to work across organizational boundaries and influence technical decisions involving multiple stakeholders. Experience designing or operating large-scale AI clusters for training or inference. Deep familiarity with LLMs, multimodal models, or recommendation systems, and their systems-level implications. Experience with accelerator interconnects and communication stacks (e.g., NCCL, MPI, RDMA, high-speed Ethernet or InfiniBand). Background in performance modeling and capacity planning for future hardware generations. Prior experience contributing to or leading hardware roadmaps, silicon bring-up, or platform architecture reviews. Publications, patents, or open-source contributions in systems, architecture, or ML systems are a plus.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$139,900 – $274,800 per year</Salaryrange>
      <Skills>C, C++, C#, Java, JavaScript, Python, AI accelerator or GPU architectures, Distributed systems and large-scale AI training/inference, High-performance computing (HPC) and collective communications, ML systems, runtimes, or compilers, Performance modeling, benchmarking, and systems analysis, Hardware–software co-design for AI workloads, Proficiency in systems-level programming (e.g., C/C++, CUDA, Python) and performance-critical software development, LLMs, multimodal models, or recommendation systems, and their systems-level implications, Accelerator interconnects and communication stacks (e.g., NCCL, MPI, RDMA, high-speed Ethernet or InfiniBand), Performance modeling and capacity planning for future hardware generations, Contributing to or leading hardware roadmaps, silicon bring-up, or platform architecture reviews, Publications, patents, or open-source contributions in systems, architecture, or ML systems</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Microsoft AI</Employername>
      <Employerlogo>https://logos.yubhub.co/microsoft.ai.png</Employerlogo>
      <Employerdescription>Microsoft AI is a technology company that develops and markets software products and services. It is one of the largest and most successful technology companies in the world.</Employerdescription>
      <Employerwebsite>https://microsoft.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://microsoft.ai/job/member-of-technical-staff-software-co-design-ai-hpc-systems-mai-superintelligence-team-2/</Applyto>
      <Location>Redmond</Location>
      <Country></Country>
      <Postedate>2026-03-08</Postedate>
    </job>
    <job>
      <externalid>4054dca1-a4f</externalid>
      <Title>AI Inference Engineer</Title>
      <Description><![CDATA[<p>We are looking for an AI Inference engineer to join our growing team. Our current stack is Python, Rust, C++, PyTorch, Triton, CUDA, Kubernetes. You will have the opportunity to work on large-scale deployment of machine learning models for real-time inference.</p>
<p><strong>What you&#39;ll do</strong></p>
<p>Develop APIs for AI inference that will be used by both internal and external customers.</p>
<ul>
<li>Develop APIs for AI inference that will be used by both internal and external customers</li>
<li>Benchmark and address bottlenecks throughout our inference stack</li>
<li>Improve the reliability and observability of our systems and respond to system outages</li>
<li>Explore novel research and implement LLM inference optimizations</li>
</ul>
<p><strong>What you need</strong></p>
<ul>
<li>Experience with ML systems and deep learning frameworks (e.g. PyTorch, TensorFlow, ONNX)</li>
<li>Familiarity with common LLM architectures and inference optimization techniques (e.g. continuous batching, quantization, etc.)</li>
<li>Understanding of GPU architectures or experience with GPU kernel programming using CUDA</li>
</ul>
<p><strong>Why this matters</strong></p>
<p>As an AI Inference engineer, you will play a critical role in the development and deployment of our machine learning models. Your work will have a direct impact on the performance and reliability of our systems, and will help us to continue to innovate and improve our products.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>Final offer amounts are determined by multiple factors, including, experience and expertise.</Salaryrange>
      <Skills>ML systems, deep learning frameworks, GPU architectures, LLM architectures, inference optimization techniques</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Perplexity</Employername>
      <Employerlogo>https://logos.yubhub.co/perplexity.com.png</Employerlogo>
      <Employerdescription>Perplexity is a company that is looking for an AI Inference engineer to join their growing team. They are a technology company that is working on large-scale deployment of machine learning models for real-time inference.</Employerdescription>
      <Employerwebsite>https://jobs.ashbyhq.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/perplexity/e4777627-ff8f-4257-8612-3a016bb58592</Applyto>
      <Location>London</Location>
      <Country></Country>
      <Postedate>2026-03-04</Postedate>
    </job>
    <job>
      <externalid>e37be4c0-4be</externalid>
      <Title>AI Inference Engineer</Title>
      <Description><![CDATA[<p>Perplexity is looking for an AI Inference Engineer to join their team. The successful candidate will be responsible for developing APIs for AI inference, benchmarking and addressing bottlenecks throughout the inference stack, improving the reliability and observability of systems, and exploring novel research and implementing LLM inference optimisations.</p>
<p><strong>What you&#39;ll do</strong></p>
<p>As an AI Inference Engineer at Perplexity, you will have the opportunity to work on large-scale deployment of machine learning models for real-time inference. You will be responsible for developing APIs for AI inference that will be used by both internal and external customers.</p>
<ul>
<li>Develop APIs for AI inference that will be used by both internal and external customers</li>
<li>Benchmark and address bottlenecks throughout our inference stack</li>
<li>Improve the reliability and observability of our systems and respond to system outages</li>
<li>Explore novel research and implement LLM inference optimisations</li>
</ul>
<p><strong>What you need</strong></p>
<p>To be successful in this role, you will need to have experience with ML systems and deep learning frameworks (e.g. PyTorch, TensorFlow, ONNX), familiarity with common LLM architectures and inference optimisation techniques (e.g. continuous batching, quantisation, etc.), and understanding of GPU architectures or experience with GPU kernel programming using CUDA.</p>
<ul>
<li>Experience with ML systems and deep learning frameworks (e.g. PyTorch, TensorFlow, ONNX)</li>
<li>Familiarity with common LLM architectures and inference optimisation techniques (e.g. continuous batching, quantisation, etc.)</li>
<li>Understanding of GPU architectures or experience with GPU kernel programming using CUDA</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$220K – $405K</Salaryrange>
      <Skills>ML systems, deep learning frameworks, LLM architectures, inference optimisation techniques, GPU architectures, GPU kernel programming, continuous batching, quantisation, PyTorch, TensorFlow, ONNX</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Perplexity</Employername>
      <Employerlogo>https://logos.yubhub.co/perplexity.ai.png</Employerlogo>
      <Employerdescription>Perplexity is a cutting-edge technology company that specialises in artificial intelligence and machine learning. They are looking for talented individuals to join their team and contribute to the development of their AI products.</Employerdescription>
      <Employerwebsite>https://www.perplexity.ai/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.ashbyhq.com/perplexity/8a976851-9bef-4b07-8d36-567fa9540aef</Applyto>
      <Location>San Francisco, New York City, Palo Alto</Location>
      <Country></Country>
      <Postedate>2026-03-04</Postedate>
    </job>
  </jobs>
</source>