{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/coherent"},"x-facet":{"type":"skill","slug":"coherent","display":"Coherent","count":5},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ac45e205-e7d"},"title":"Engineering Manager, Inference Routing and Performance","description":"<p><strong>About the role\\nEvery request that hits Claude , from claude.ai, the API, our cloud partners, or internal research , passes through a routing decision. Not a generic load balancer round-robin, but a decision that accounts for what&#39;s already cached where, which accelerator the request runs best on, and what else is in flight across the fleet.\\n\\nGet it right and you extract meaningfully more throughput from the same hardware. Get it wrong and you burn capacity, miss latency SLOs, or shed load that shouldn&#39;t have been shed.\\n\\nThe Inference Routing team owns this layer. We build the cluster-level routing and coordination plane for Anthropic&#39;s inference fleet , the system that sits between the API surface and the inference engines themselves, making fleet-wide efficiency decisions in real time.\\n\\nAs Anthropic moves from &quot;many independent inference replicas&quot; toward &quot;a single warehouse-scale computer running a coordinated program,&quot; Dystro is the coordination layer. This is a deeply technical team.\\n\\nThe engineers here design custom load-balancing algorithms, build quantitative models of system performance, debug latency spikes that cross kernel, network, and framework boundaries, and reason carefully about cache placement across thousands of accelerators.\\n\\nThey work shoulder-to-shoulder with teams that write kernels and ML framework internals.\\n\\nThe EM for this team doesn&#39;t need to write kernels , but they do need the systems depth to make architectural calls, evaluate deeply technical candidates, and spot when a proposed optimization will have second-order effects on the fleet.\\n\\nYou&#39;ll inherit a strong team of distributed-systems engineers, and you&#39;ll be accountable for two things that pull in different directions: shipping system-level performance improvements that measurably increase fleet throughput and efficiency, and running the team operationally so that deploys are safe, incidents are rare, and the teams who depend on Dystro can plan around you with confidence.\\n\\nThe job is holding both.\\n\\n## Representative work:\\nThings the Inference Routing EM actually spends time on:\\n- Deciding whether a proposed routing algorithm change is worth the deploy risk, given the modeled throughput gain and the blast radius if it regresses\\n- Sequencing a quarter where KV-cache offload, a new coordination protocol, and two model launches all compete for the same engineers\\n- Working through a persistent tail-latency regression with the team , walking down from fleet-level metrics to per-replica behavior to a root cause in the networking stack\\n- Building the case (with numbers) to peer teams for why a cross-team protocol change unlocks the next efficiency win\\n- Running the post-incident review after a cache-eviction bug caused a capacity event, and turning it into process changes that stick\\n- Interviewing a candidate who has built schedulers at supercomputing scale, and deciding whether they&#39;d be additive to a team that already goes deep\\n\\n## What you&#39;ll do:\\nDrive system-level performance\\n- Own the technical roadmap for cluster-level inference efficiency , routing decisions, cache placement and eviction, cross-replica coordination, and the protocols that keep routing and inference engines in sync\\n- Partner with the inference engine, kernels, and performance teams to identify fleet-level throughput and latency wins, then turn those into shipped improvements with measurable results\\n- Build the team&#39;s habit of quantitative performance modeling: claim a win only when you can measure it, and know before you ship what the expected effect is\\n\\nDeliver reliably and operate cleanly\\n- Set technical strategy for how routing evolves across heterogeneous hardware (GPUs, TPUs, Trainium) and across all our serving surfaces\\n- Run the team&#39;s operational backbone , on-call rotation, incident response, postmortem review, deploy safety , so the team can ship aggressively without the system becoming fragile\\n- Create clarity at a seam: Inference Routing sits between the API surface, the inference engines, and the cloud deployment teams. You&#39;ll make sure commitments are realistic, dependencies are understood, and nobody is surprised\\n\\nBuild and grow the team\\n- Develop and retain a strong existing team, and hire against the bar described above: people who can go to the OS and framework level when the problem demands it, and who care about production reliability\\n- Coach engineers through a roadmap where priorities shift with model launches, new hardware, and scaling demands. We pair a lot here , you&#39;ll help make that collaboration pattern productive\\n- Pick up slack when it matters. This is a small team in a critical path; sometimes the EM is the one unblocking a stuck deploy or synthesizing a design debate\\n\\n## You may be a good fit if you:\\n- Have 5+ years of engineering management experience, ideally with at least part of that leading teams on critical-path production infrastructure at scale\\n- Have a deep systems background , load balancing, scheduling, cache-coherent distributed state, high-performance networking, or similar. You need enough depth to make architectural calls about routing and efficiency, and to evaluate candidates who go to the kernel and framework level\\n- Have shipped performance improvements in large-scale systems and can explain, with numbers, what the impact was\\n- Have run production infrastructure with real operational stakes: on-call, incident response, capacity events, deploy discipline\\n- Are results-oriented with a bias toward impact, and comfortable working in a space where throughput, latency, stability, and feature velocity all pull in different directions\\n- Build strong relationships across team boundaries , this is a seam role, and much of the job is making sure other teams can rely on yours\\n- Are curious about machine learning systems. You don&#39;t need an ML research background, but you should want to learn how transformer inference actually works and how that shapes the systems problems\\n\\nStrong candidates may also have:\\n- Experience with LLM inference serving , KV caching, continuous batching, request scheduling, prefill/decode disaggregation\\n- Background in cluster schedulers, load balancers, service meshes, or coordination planes at scale\\n- Familiarity with heterogeneous accelerator fleets (GPU/TPU/Trainium) and how hardware differences affect workload placement\\n- Experience with GPU/accelerator programming, ML framework internals, or OS-level performance debugging , enough to follow and evaluate the technical work, not necessarily to do it daily\\n- Led teams at supercomputing or hyperscaler infrastructure scale\\n- Led teams through rapid-growth periods where hiring and onboarding competed with roadmap delivery\\n\\nThe annual compensation range for this role is listed below. For sales roles, the range provided is the role’s On Target Earnings (&quot;OTE&quot;) range, meaning that the range includes both the sales commissions/sales bonuses target and annual base salary for the role.\\nAnnual Salary: $405,000-$485,000 USD</strong></p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ac45e205-e7d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5155391008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["engineering management","distributed systems","load balancing","scheduling","cache-coherent distributed state","high-performance networking","machine learning systems"],"x-skills-preferred":["LLM inference serving","cluster schedulers","load balancers","service meshes","coordination planes","heterogeneous accelerator fleets","GPU/TPU/Trainium","GPU/accelerator programming","ML framework internals","OS-level performance debugging"],"datePosted":"2026-04-18T15:56:48.587Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"engineering management, distributed systems, load balancing, scheduling, cache-coherent distributed state, high-performance networking, machine learning systems, LLM inference serving, cluster schedulers, load balancers, service meshes, coordination planes, heterogeneous accelerator fleets, GPU/TPU/Trainium, GPU/accelerator programming, ML framework internals, OS-level performance debugging","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9cd0420a-99d"},"title":"Network Engineer, Capacity and Efficiency","description":"<p><strong>About the Role</strong></p>\n<p>We&#39;re looking for a network engineer who thinks in metrics first. You will use deep networking knowledge and rigorous measurement to figure out where and how bandwidth, latency, and dollars are being used, find optimization opportunities and land them.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build the network observability stack. Design and deploy telemetry pipelines , sFlow/IPFIX, gNMI streaming, eBPF host probes , that turn packet counters into per-flow, per-tenant, per-workload cost and utilization data. Own the SLIs for backbone and DCN fabric health.</li>\n<li>Hunt for efficiency. Analyze inter-region traffic patterns, identify hot links and stranded capacity, and quantify the dollar impact. Build the models that tell us whether we should buy more capacity, or move the workload.</li>\n<li>Own QoS and traffic engineering. Design and operate traffic classification, marking, and shaping across the backbone. Make sure bulk checkpoint transfers don’t starve latency-sensitive inference, and that we’re not paying premium cross-region rates for traffic that could take the cheap path.</li>\n<li>Drive cost attribution. Tie network spend , egress, interconnect ports, transit, optical leases , back to the teams and workloads that generate it. Make network cost a first-class input to capacity planning and workload placement decisions.</li>\n<li>Influence decisions you don&#39;t own. A large fraction of this role is convincing other teams to act on what your data shows: making the case to research that a traffic pattern needs to change, to finance that an interconnect tranche is worth buying, to Systems Networking that a QoS policy needs rewriting.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Have 5+ years operating large-scale production networks , data center fabrics (spine-leaf, Clos), backbone/WAN, or hyperscaler-adjacent environments.</li>\n<li>Are genuinely fluent across the stack: BGP (including policy and communities), ECMP, VXLAN/EVPN or equivalent overlays, QoS (DSCP, queuing, shaping), and L1/optical basics (DWDM, coherent, LAGs).</li>\n<li>Know at least one major CSP’s networking model deeply , AWS (VPC, TGW, Direct Connect, Gateway Load Balancer) or GCP (Shared VPC, Interconnect, Cloud Router, Network Connectivity Center) , and understand how their overlays interact with physical underlays.</li>\n<li>Have built or operated network telemetry at scale: streaming telemetry (gNMI/OpenConfig), flow export (sFlow, IPFIX, NetFlow), or eBPF-based host-side instrumentation. You can reason about sampling, cardinality, and storage tradeoffs.</li>\n<li>Comfortable writing Python or Go to build tooling, telemetry pipelines, infrastructure-as-code, config management for network devices and automation, that you’ll ship to production.</li>\n<li>Think quantitatively by default. You reach for a notebook or a Grafana query before you reach for an opinion, and you can turn messy counter data into a defensible cost model.</li>\n<li>Communicate crisply. You can explain to a finance partner why a 10% egress reduction matters, and to a network engineer why a specific ECMP imbalance is costing real money.</li>\n</ul>\n<p><strong>Nice to Have</strong></p>\n<ul>\n<li>SRE experience for large-scale network infrastructure , designing for reliability, defining SLOs/SLIs for network services, capacity planning with error budgets, and incident response for network-impacting outages at scale.</li>\n<li>Background on a cloud provider&#39;s networking team or a cloud networking product team , building or operating the interconnect, backbone, or SDN control plane from the provider side, not just consuming it as a customer.</li>\n<li>Familiarity with AI/ML infrastructure traffic patterns like collective communication (all-reduce, all-gather), checkpoint/weight transfer, inference serving, and how these stress networks differ than traditional workloads in terms of burst behavior, flow synchronization, and bandwidth symmetry.</li>\n<li>Experience with HPC fabrics like InfiniBand, RoCE v2, lossless Ethernet, or custom high-radix topologies and an understanding of how job placement, congestion management, and adaptive routing interact at scale.</li>\n<li>Background in traffic engineering for large backbones and the operational judgment to know when TE is worth the complexity.</li>\n<li>Hands-on time with multi-cloud connectivity: cross-cloud peering, private interconnect products, and the billing models that come with them.</li>\n<li>Experience building cost/chargeback systems for shared infrastructure, or FinOps exposure in a large cloud environment.</li>\n</ul>\n<p><strong>Representative Projects</strong></p>\n<ul>\n<li>Build a per-flow cost attribution pipeline that traces every byte of cross-region egress back to the team and workload that generated it</li>\n<li>Design QoS policy for the private backbone that prevents bulk checkpoint transfers from starving inference traffic</li>\n<li>Model whether it&#39;s cheaper to buy an additional 1.6Tb interconnect tranche or to re-route traffic through existing capacity</li>\n<li>Instrument DCN fabric utilization with streaming telemetry and build the Grafana dashboards that become the team&#39;s source of truth for network observability</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9cd0420a-99d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://anthropic.com","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5177143008","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["network engineering","network observability","telemetry pipelines","sFlow/IPFIX","gNMI streaming","eBPF host probes","BGP","ECMP","VXLAN/EVPN","QoS","DSCP","queuing","shaping","L1/optical basics","DWDM","coherent","LAGs","AWS","GCP","cloud networking","infrastructure-as-code","config management","automation","Python","Go","quantitative analysis","cost modeling","communication"],"x-skills-preferred":["SRE","cloud provider's networking team","cloud networking product team","AI/ML infrastructure traffic patterns","HPC fabrics","traffic engineering","multi-cloud connectivity","cost/chargeback systems","FinOps"],"datePosted":"2026-04-18T15:42:29.482Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"network engineering, network observability, telemetry pipelines, sFlow/IPFIX, gNMI streaming, eBPF host probes, BGP, ECMP, VXLAN/EVPN, QoS, DSCP, queuing, shaping, L1/optical basics, DWDM, coherent, LAGs, AWS, GCP, cloud networking, infrastructure-as-code, config management, automation, Python, Go, quantitative analysis, cost modeling, communication, SRE, cloud provider's networking team, cloud networking product team, AI/ML infrastructure traffic patterns, HPC fabrics, traffic engineering, multi-cloud connectivity, cost/chargeback systems, FinOps"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_63af8568-789"},"title":"Engineering Manager, Inference Routing and Performance","description":"<p><strong>About the role\\nEvery request that hits Claude , from claude.ai, the API, our cloud partners, or internal research , passes through a routing decision. Not a generic load balancer round-robin, but a decision that accounts for what&#39;s already cached where, which accelerator the request runs best on, and what else is in flight across the fleet.\\n\\nGet it right and you extract meaningfully more throughput from the same hardware. Get it wrong and you burn capacity, miss latency SLOs, or shed load that shouldn&#39;t have been shed.\\n\\nThe Inference Routing team owns this layer. We build the cluster-level routing and coordination plane for Anthropic&#39;s inference fleet , the system that sits between the API surface and the inference engines themselves, making fleet-wide efficiency decisions in real time.\\n\\nAs Anthropic moves from &quot;many independent inference replicas&quot; toward &quot;a single warehouse-scale computer running a coordinated program,&quot; Dystro is the coordination layer. This is a deeply technical team.\\n\\nThe engineers here design custom load-balancing algorithms, build quantitative models of system performance, debug latency spikes that cross kernel, network, and framework boundaries, and reason carefully about cache placement across thousands of accelerators.\\n\\nThey work shoulder-to-shoulder with teams that write kernels and ML framework internals.\\n\\nThe EM for this team doesn&#39;t need to write kernels , but they do need the systems depth to make architectural calls, evaluate deeply technical candidates, and spot when a proposed optimization will have second-order effects on the fleet.\\n\\nYou&#39;ll inherit a strong team of distributed-systems engineers, and you&#39;ll be accountable for two things that pull in different directions: shipping system-level performance improvements that measurably increase fleet throughput and efficiency, and running the team operationally so that deploys are safe, incidents are rare, and the teams who depend on Dystro can plan around you with confidence.\\n\\nThe job is holding both.\\n\\n## Representative work:\\nThings the Inference Routing EM actually spends time on:\\n- Deciding whether a proposed routing algorithm change is worth the deploy risk, given the modeled throughput gain and the blast radius if it regresses\\n- Sequencing a quarter where KV-cache offload, a new coordination protocol, and two model launches all compete for the same engineers\\n- Working through a persistent tail-latency regression with the team , walking down from fleet-level metrics to per-replica behavior to a root cause in the networking stack\\n- Building the case (with numbers) to peer teams for why a cross-team protocol change unlocks the next efficiency win\\n- Running the post-incident review after a cache-eviction bug caused a capacity event, and turning it into process changes that stick\\n- Interviewing a candidate who has built schedulers at supercomputing scale, and deciding whether they&#39;d be additive to a team that already goes deep\\n\\n## What you&#39;ll do:\\nDrive system-level performance\\n- Own the technical roadmap for cluster-level inference efficiency , routing decisions, cache placement and eviction, cross-replica coordination, and the protocols that keep routing and inference engines in sync\\n- Partner with the inference engine, kernels, and performance teams to identify fleet-level throughput and latency wins, then turn those into shipped improvements with measurable results\\n- Build the team&#39;s habit of quantitative performance modeling: claim a win only when you can measure it, and know before you ship what the expected effect is\\n\\nDeliver reliably and operate cleanly\\n- Set technical strategy for how routing evolves across heterogeneous hardware (GPUs, TPUs, Trainium) and across all our serving surfaces\\n- Run the team&#39;s operational backbone , on-call rotation, incident response, postmortem review, deploy safety , so the team can ship aggressively without the system becoming fragile\\n- Create clarity at a seam: Inference Routing sits between the API surface, the inference engines, and the cloud deployment teams. You&#39;ll make sure commitments are realistic, dependencies are understood, and nobody is surprised\\n\\nBuild and grow the team\\n- Develop and retain a strong existing team, and hire against the bar described above: people who can go to the OS and framework level when the problem demands it, and who care about production reliability\\n- Coach engineers through a roadmap where priorities shift with model launches, new hardware, and scaling demands. We pair a lot here , you&#39;ll help make that collaboration pattern productive\\n- Pick up slack when it matters. This is a small team in a critical path; sometimes the EM is the one unblocking a stuck deploy or synthesizing a design debate\\n\\n## You may be a good fit if you:\\n- Have 5+ years of engineering management experience, ideally with at least part of that leading teams on critical-path production infrastructure at scale\\n- Have a deep systems background , load balancing, scheduling, cache-coherent distributed state, high-performance networking, or similar. You need enough depth to make architectural calls about routing and efficiency, and to evaluate candidates who go to the kernel and framework level\\n- Have shipped performance improvements in large-scale systems and can explain, with numbers, what the impact was\\n- Have run production infrastructure with real operational stakes: on-call, incident response, capacity events, deploy discipline\\n- Are results-oriented with a bias toward impact, and comfortable working in a space where throughput, latency, stability, and feature velocity all pull in different directions\\n- Build strong relationships across team boundaries , this is a seam role, and much of the job is making sure other teams can rely on yours\\n- Are curious about machine learning systems. You don&#39;t need an ML research background, but you should want to learn how transformer inference actually works and how that shapes the systems problems\\n\\nStrong candidates may also have:\\n- Experience with LLM inference serving , KV caching, continuous batching, request scheduling, prefill/decode disaggregation\\n- Background in cluster schedulers, load balancers, service meshes, or coordination planes at scale\\n- Familiarity with heterogeneous accelerator fleets (GPU/TPU/Trainium) and how hardware differences affect workload placement\\n- Experience with GPU/accelerator programming, ML framework internals, or OS-level performance debugging , enough to follow and evaluate the technical work, not necessarily to do it daily\\n- Led teams at supercomputing or hyperscaler infrastructure scale\\n- Led teams through rapid-growth periods where hiring and onboarding competed with roadmap delivery\\n\\nThe annual compensation range for this role is listed below. For sales roles, the range provided is the role’s On Target Earnings (&quot;OTE&quot;) range, meaning that the range includes both the sales commissions/sales bonuses target and annual base salary for the role.\\nAnnual Salary: $405,000-$485,000 USD</strong></p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_63af8568-789","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5155391008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["engineering management","deep systems background","load balancing","scheduling","cache-coherent distributed state","high-performance networking"],"x-skills-preferred":["LLM inference serving","cluster schedulers","load balancers","service meshes","coordination planes","heterogeneous accelerator fleets","GPU/TPU/Trainium","GPU/accelerator programming","ML framework internals","OS-level performance debugging"],"datePosted":"2026-04-18T15:37:38.038Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"engineering management, deep systems background, load balancing, scheduling, cache-coherent distributed state, high-performance networking, LLM inference serving, cluster schedulers, load balancers, service meshes, coordination planes, heterogeneous accelerator fleets, GPU/TPU/Trainium, GPU/accelerator programming, ML framework internals, OS-level performance debugging","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_94e4ded2-2fd"},"title":"Quantum Engineer","description":"<p>As a Quantum Engineer at Rigetti Computing, you will contribute to the development of our gate-based, superconducting quantum computers. Your work will involve characterizing experimental devices, assessing their performance, identifying areas for improvement, and providing feedback to internal development teams.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li><p>Providing technical focus and leadership on a particular R&amp;D topic, such as gate calibration, qubit relaxation and/or dephasing, piloting new research ideas in small processors, integration of new ideas into large processors, or implementation of quantum error correction.</p>\n</li>\n<li><p>Collaborating across diverse teams of expert engineers to turn experimental concepts into robust technologies that work at large QPU scales.</p>\n</li>\n<li><p>Staying current with cutting-edge research, synthesizing results from the literature, and implementing state-of-the-art calibration, error diagnostic, and system optimization techniques to enhance device performance.</p>\n</li>\n</ul>\n<p>Required qualifications include:</p>\n<ul>\n<li><p>PhD in Physics or similar applied scientific field, or Masters&#39; in Physics with 2+ years relevant industry experience.</p>\n</li>\n<li><p>Demonstrated abilities in experimental problem solving and experimental research.</p>\n</li>\n<li><p>Expertise in the practical implementation of coherent quantum effects and control.</p>\n</li>\n<li><p>Proficiency using Python in a scientific context.</p>\n</li>\n<li><p>Ability to thrive in a strongly collaborative environment, with a team-first mindset.</p>\n</li>\n</ul>\n<p>Nice to have qualifications include:</p>\n<ul>\n<li><p>Experience calibrating and characterizing single- and two-qubit quantum gates in superconducting quantum devices.</p>\n</li>\n<li><p>Experience utilizing quantum simulation frameworks, such as QuTip.</p>\n</li>\n<li><p>Proven capability to execute and effectively communicate outcomes from innovative research and development initiatives.</p>\n</li>\n<li><p>Software development experience in a collaborative industry setting.</p>\n</li>\n<li><p>Previous technical role in an industrial or startup setting.</p>\n</li>\n</ul>\n<p>Additional information:</p>\n<ul>\n<li><p>As engineering leaders, we value diversity and are committed to building a culture of inclusion to attract and engage innovative thinkers.</p>\n</li>\n<li><p>Applications from women, minorities, and other under-represented groups are encouraged.</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_94e4ded2-2fd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Rigetti Computing","sameAs":"https://www.rigetti.com","logo":"https://logos.yubhub.co/rigetti.com.png"},"x-apply-url":"https://jobs.lever.co/rigetti/0f66c1c3-12be-4e60-a7e1-25e494ee2841","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Experimental problem solving","Experimental research","Coherent quantum effects and control","Quantum simulation frameworks"],"x-skills-preferred":["Calibrating and characterizing single- and two-qubit quantum gates","Utilizing quantum simulation frameworks","Software development","Previous technical role in an industrial or startup setting"],"datePosted":"2026-04-17T12:54:13.460Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Experimental problem solving, Experimental research, Coherent quantum effects and control, Quantum simulation frameworks, Calibrating and characterizing single- and two-qubit quantum gates, Utilizing quantum simulation frameworks, Software development, Previous technical role in an industrial or startup setting"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b4d3cb52-7c4"},"title":"Senior ASIC Verification Engineer, Coherent High Speed Interconnect","description":"<p>We are now looking for a Senior ASIC Verification Engineer for our Coherent High Speed Interconnect team. For two decades, NVIDIA has pioneered visual computing, the art and science of computer graphics. With our invention of the GPU - the engine of modern visual computing - the field has grown to encompass video games, movie production, product design, medical diagnosis, and scientific research.</p>\n<p>Today, we stand at the beginning of the next era, the AI computing era, ignited by a new computing model, GPU deep learning. This new model - where deep neural networks are trained to recognize patterns from meaningful amounts of data - has shown to be deeply effective at solving the most sophisticated problems in everyday life.</p>\n<p>As a Senior ASIC Verification Engineer at NVIDIA, you will verify the design and implementation of our innovative high speed coherent interconnects for our mobile SoCs and GPUs. This position offers the opportunity to have real impact in a multifaceted, technology-focused company impacting product lines ranging from consumer graphics to self-driving cars and the growing field of artificial intelligence.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>In this position, you will be responsible for verification of high-speed coherent interconnect design, architecture and golden models.</li>\n<li>You will be responsible for micro-architecture using sophisticated verification methodologies.</li>\n<li>As a member of our verification team, you&#39;ll understand the design &amp; implementation, define the verification scope, develop the verification infrastructure (Testbenches, BFMs, Checkers, Monitors), complete test/coverage plans, and verify the correctness of the design. This role will collaborate with architects, designers, emulation, and silicon verification teams to accomplish your tasks.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>Bachelors or Master’s Degree (or equivalent experience)</li>\n<li>3+ years of relevant verification experience</li>\n<li>Experience in architecting test bench environments for unit level verification</li>\n<li>Background in verification using random stimulus along with functional coverage and assertion-based verification methodologies</li>\n<li>Prior Design or Verification experience of Coherent high-speed interconnects</li>\n<li>Knowledge of industry standard interconnect protocols like PCIE, CXL, CHI will be useful</li>\n<li>Strong background developing TB&#39;s from scratch using SV and UVM methodology is desired</li>\n<li>C++ programming language experience, scripting ability and an expertise in System Verilog</li>\n<li>Exposure to design and verification tools (VCS or equivalent simulation tools, debug tools like Debussy, GDB)</li>\n<li>Strong debugging and analytical skills</li>\n<li>Experienced communication and interpersonal skills are required. A history of mentoring junior engineers and interns a huge plus.</li>\n</ul>\n<p>NVIDIA is widely considered to be one of the technology world’s most desirable employers! We have some of the most forward-thinking and dedicated people in the world working for us. If you&#39;re creative and autonomous, we want to hear from you.</p>\n<p>You will also be eligible for equity and benefits.</p>\n<p>Applications for this job will be accepted at least until March 13, 2026.</p>\n<p>This posting is for an existing vacancy.</p>\n<p>NVIDIA uses AI tools in its recruiting processes.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b4d3cb52-7c4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"NVIDIA","sameAs":"https://nvidia.wd5.myworkdayjobs.com","logo":"https://logos.yubhub.co/nvidia.com.png"},"x-apply-url":"https://nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAExternalCareerSite/job/US-CA-Santa-Clara/Senior-ASIC-Verification-Engineer--Coherent-High-Speed-Interconnect_JR2010025","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Verification of high-speed coherent interconnect design, architecture and golden models","Micro-architecture using sophisticated verification methodologies","Testbenches, BFMs, Checkers, Monitors","System Verilog","C++ programming language","Design and verification tools (VCS or equivalent simulation tools, debug tools like Debussy, GDB)"],"x-skills-preferred":["Random stimulus along with functional coverage and assertion-based verification methodologies","Prior Design or Verification experience of Coherent high-speed interconnects","Knowledge of industry standard interconnect protocols like PCIE, CXL, CHI"],"datePosted":"2026-03-09T20:46:52.056Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"US, CA, Santa ClaraUS, MA, WestfordUS, TX, AustinUS, OR, Hillsboro"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Verification of high-speed coherent interconnect design, architecture and golden models, Micro-architecture using sophisticated verification methodologies, Testbenches, BFMs, Checkers, Monitors, System Verilog, C++ programming language, Design and verification tools (VCS or equivalent simulation tools, debug tools like Debussy, GDB), Random stimulus along with functional coverage and assertion-based verification methodologies, Prior Design or Verification experience of Coherent high-speed interconnects, Knowledge of industry standard interconnect protocols like PCIE, CXL, CHI"}]}