{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/request-routing"},"x-facet":{"type":"skill","slug":"request-routing","display":"Request Routing","count":10},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_37765ea5-447"},"title":"Staff Software Engineer, Voices","description":"<p>You will work on the core systems powering Synthesia&#39;s script preview and voice generation experience, ensuring users can reliably generate high-quality voiceovers across a wide range of languages, providers, and use cases.</p>\n<p>You will build and operate backend services that orchestrate multiple text-to-speech (TTS) providers, alongside Synthesia&#39;s in-house models, delivering a seamless and consistent experience to end users despite underlying system complexity.</p>\n<p>You will be responsible for designing and evolving systems that handle provider reliability, request routing, and output consistency, ensuring users can generate and regenerate voice content with predictable, high-quality results.</p>\n<p>You will contribute to user-facing product problems from a backend perspective, working closely with frontend engineers to ensure APIs and workflows integrate cleanly into the product experience.</p>\n<p>You will own projects that span multiple systems and domains, such as:</p>\n<ul>\n<li>Building robustness layers (retries, throttling, failover) to handle unreliable third-party providers</li>\n</ul>\n<ul>\n<li>Designing persistence systems to ensure consistent voice outputs across generations</li>\n</ul>\n<ul>\n<li>Improving how voice data is stored, retrieved, and reused</li>\n</ul>\n<p>You will also work on newer initiatives around voice discovery and recommendations, helping users find the best voices by building evaluation systems, improving recommendation logic, and adding observability to monitor quality.</p>\n<p>You will evaluate your work through system performance and user experience metrics, using observability to debug and continuously improve reliability.</p>\n<p>You will collaborate closely with product, frontend, and R&amp;D teams, ensuring backend systems support both current product needs and future innovation.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_37765ea5-447","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Synthesia","sameAs":"https://www.synthesia.io","logo":"https://logos.yubhub.co/synthesia.io.png"},"x-apply-url":"https://jobs.ashbyhq.com/synthesia/dfe4614e-040d-4447-885b-8f0827428f6e","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"Full time","x-salary-range":null,"x-skills-required":["Python","FastAPI","text-to-speech (TTS) providers","backend services","designing and evolving systems","provider reliability","request routing","output consistency","APIs","workflows","systems","domains","robustness layers","persistence systems","voice data","evaluation systems","recommendation logic","observability"],"x-skills-preferred":["observability tools","workflow systems","evaluation/recommendation systems"],"datePosted":"2026-04-24T13:18:05.862Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Europe"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, FastAPI, text-to-speech (TTS) providers, backend services, designing and evolving systems, provider reliability, request routing, output consistency, APIs, workflows, systems, domains, robustness layers, persistence systems, voice data, evaluation systems, recommendation logic, observability, observability tools, workflow systems, evaluation/recommendation systems"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a3e876d0-ca3"},"title":"Technical Integration Specialist","description":"<p>As a Technical Integration Specialist, you will improve the ecommerce experience for hundreds of millions of users across the world by providing first level technical support to ecommerce companies (our customers) as they integrate their websites and mobile applications with Constructor&#39;s platform.</p>\n<p>This includes onboarding support, API and library education, troubleshooting, index configuration, communication, and process improvement.</p>\n<p>Key responsibilities include collaborating with customers to understand their technical integration needs, assisting them through the integration process, providing expert guidance on integrating with Constructor&#39;s APIs and utilizing open source libraries, helping resolve technical challenges and integration issues alongside Customer Success teams, working with customers to ensure their product catalogs are properly configured, and acting as a liaison between customers and other technical resources.</p>\n<p>We&#39;re looking for someone with proven experience in customer support or technical support roles, knowledge of REST-based web services, APIs, and integration best practices, excellent communication and interpersonal skills, a detail-oriented mentality with a commitment to delivering exceptional customer support, strong problem-solving skills, and a proactive approach to issue resolution.</p>\n<p>Preferred qualifications include familiarity with structured data formats (JSON, CSV) and basic search/index concepts, familiarity with debugging tools such as browser developer tools, API clients (e.g., Postman), or logs, familiarity with A/B testing, understanding of server side and client side contexts, network request routing / HTTP, and familiarity working with open source libraries, JavaScript web application development, and integration patterns.</p>\n<p>Benefits include working with smart and empathetic people who will help you grow and make a meaningful impact, regular team offsite events to connect and collaborate, fully remote team - choose where you live, unlimited vacation time - we strongly encourage all of our employees take at least 3 weeks per year, work from home stipend, Apple laptops provided for new employees, training and development budget for every employee, refreshed each year, maternity &amp; paternity leave for qualified employees, and a base salary range between $80k and $110k USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a3e876d0-ca3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://constructor.com","logo":"https://logos.yubhub.co/constructor.com.png"},"x-apply-url":"https://apply.workable.com/j/D34BF95979","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$80k-$110k USD","x-skills-required":["REST-based web services","APIs","integration best practices","structured data formats (JSON, CSV)","basic search/index concepts","debugging tools (browser developer tools, API clients, logs)","A/B testing","server side and client side contexts","network request routing / HTTP","open source libraries","JavaScript web application development","integration patterns"],"x-skills-preferred":["familiarity with debugging tools","familiarity with A/B testing","familiarity working with open source libraries"],"datePosted":"2026-04-24T13:11:03.209Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"REST-based web services, APIs, integration best practices, structured data formats (JSON, CSV), basic search/index concepts, debugging tools (browser developer tools, API clients, logs), A/B testing, server side and client side contexts, network request routing / HTTP, open source libraries, JavaScript web application development, integration patterns, familiarity with debugging tools, familiarity with A/B testing, familiarity working with open source libraries","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":110000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_72d8bd0c-244"},"title":"Technical Integration Specialist","description":"<p>As a Technical Integration Specialist, you will improve the e-commerce experience for hundreds of millions of users across the world by providing first-level technical support to e-commerce companies as they integrate their websites and mobile applications with Constructor&#39;s platform.</p>\n<p>This includes onboarding support, API and library education, troubleshooting, index configuration, communication, and process improvement.</p>\n<p>You will be working with customers to understand their technical integration needs and assist them through the integration process.</p>\n<p>You will provide expert guidance on integrating with Constructor&#39;s APIs and utilizing open-source libraries.</p>\n<p>You will help resolve technical challenges and integration issues alongside Customer Success teams.</p>\n<p>You will work with customers to ensure their product catalogs are properly configured, including uploading catalog data, configuring facets, searchabilities, sort orders, and retrieving results.</p>\n<p>You will act as a liaison between customers and other technical resources, responding promptly to inquiries with clear explanations.</p>\n<p>You will document customer interactions, common issues, and solutions, improving integration documentation and customer onboarding.</p>\n<p>Requirements:</p>\n<ul>\n<li>Proven experience in customer support or technical support roles.</li>\n</ul>\n<ul>\n<li>Knowledge of REST-based web services, APIs, and integration best practices.</li>\n</ul>\n<ul>\n<li>Excellent communication and interpersonal skills, with the ability to explain technical concepts to non-technical audiences.</li>\n</ul>\n<ul>\n<li>Detail-oriented mentality with a commitment to delivering exceptional customer support.</li>\n</ul>\n<ul>\n<li>Strong problem-solving skills and a proactive approach to issue resolution.</li>\n</ul>\n<p>Preferred qualifications:</p>\n<ul>\n<li>Familiarity with structured data formats (JSON, CSV) and basic search/index concepts.</li>\n</ul>\n<ul>\n<li>Familiarity with debugging tools such as browser developer tools, API clients (e.g., Postman), or logs.</li>\n</ul>\n<ul>\n<li>Familiarity with A/B testing, understanding of server-side and client-side contexts, network request routing / HTTP.</li>\n</ul>\n<ul>\n<li>Familiarity working with open-source libraries, JavaScript web application development, and integration patterns.</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Work with smart and empathetic people who will help you grow and make a meaningful impact.</li>\n</ul>\n<ul>\n<li>Regular team offsite events to connect and collaborate.</li>\n</ul>\n<ul>\n<li>Fully remote team - choose where you live.</li>\n</ul>\n<ul>\n<li>Unlimited vacation time - we strongly encourage all of our employees take at least 3 weeks per year.</li>\n</ul>\n<ul>\n<li>Work from home stipend!</li>\n</ul>\n<ul>\n<li>Apple laptops provided for new employees.</li>\n</ul>\n<ul>\n<li>Training and development budget for every employee, refreshed each year.</li>\n</ul>\n<ul>\n<li>Base salary: $80k–$110k USD, depending on knowledge, skills, experience, and interview results.</li>\n</ul>\n<ul>\n<li>Stock options - offered in addition to the base salary.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_72d8bd0c-244","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://constructor.io","logo":"https://logos.yubhub.co/constructor.io.png"},"x-apply-url":"https://apply.workable.com/j/FFDF6D632C","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$80k–$110k USD","x-skills-required":["REST-based web services","APIs","integration best practices","structured data formats","basic search/index concepts"],"x-skills-preferred":["debugging tools","A/B testing","server-side and client-side contexts","network request routing / HTTP","open-source libraries","JavaScript web application development","integration patterns"],"datePosted":"2026-04-24T13:10:53.748Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Brisbane"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"REST-based web services, APIs, integration best practices, structured data formats, basic search/index concepts, debugging tools, A/B testing, server-side and client-side contexts, network request routing / HTTP, open-source libraries, JavaScript web application development, integration patterns","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":110000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_61e346b2-915"},"title":"Sr. Software Engineer, Inference","description":"<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>The team has a dual mandate: maximizing compute efficiency to serve our explosive customer growth, while enabling breakthrough research by giving our scientists the high-performance inference infrastructure they need to develop next-generation models. We tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.</p>\n<p>Strong candidates may also have experience with:</p>\n<ul>\n<li>High-performance, large-scale distributed systems</li>\n<li>Implementing and deploying machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP)</li>\n<li>Python or Rust</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Have significant software engineering experience, particularly with distributed systems</li>\n<li>Are results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p>Representative projects across the org:</p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p>Deadline to apply: None. Applications will be reviewed on a rolling basis.</p>\n<p>The annual compensation range for this role is £225,000-£325,000 GBP.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_61e346b2-915","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5152348008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"£225,000-£325,000 GBP","x-skills-required":["High-performance, large-scale distributed systems","Implementing and deploying machine learning systems at scale","Load balancing, request routing, or traffic management systems","LLM inference optimization, batching, and caching strategies","Kubernetes and cloud infrastructure (AWS, GCP)","Python or Rust"],"x-skills-preferred":[],"datePosted":"2026-04-18T16:00:17.377Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"High-performance, large-scale distributed systems, Implementing and deploying machine learning systems at scale, Load balancing, request routing, or traffic management systems, LLM inference optimization, batching, and caching strategies, Kubernetes and cloud infrastructure (AWS, GCP), Python or Rust","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":225000,"maxValue":325000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7d4c3fc5-2ed"},"title":"Senior Software Engineer, Inference","description":"<p>About the role:</p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>The team has a dual mandate: maximizing compute efficiency to serve our explosive customer growth, while enabling breakthrough research by giving our scientists the high-performance inference infrastructure they need to develop next-generation models. We tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.</p>\n<p>Strong candidates may also have experience with:</p>\n<ul>\n<li>High-performance, large-scale distributed systems</li>\n<li>Implementing and deploying machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP)</li>\n<li>Python or Rust</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Have significant software engineering experience, particularly with distributed systems</li>\n<li>Are results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p>Representative projects across the org:</p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p>Annual compensation range for this role is €235,000-€295,000 EUR.</p>\n<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</p>\n<p>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</p>\n<p>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</p>\n<p>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links,visit anthropic.com/careers directly for confirmed position openings.</p>\n<p>How we&#39;re different:</p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>\n<p>Come work with us!</p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7d4c3fc5-2ed","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4641822008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"€235,000-€295,000 EUR","x-skills-required":["High-performance, large-scale distributed systems","Implementing and deploying machine learning systems at scale","Load balancing, request routing, or traffic management systems","LLM inference optimization, batching, and caching strategies","Kubernetes and cloud infrastructure (AWS, GCP)","Python or Rust"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:09.302Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, IE"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"High-performance, large-scale distributed systems, Implementing and deploying machine learning systems at scale, Load balancing, request routing, or traffic management systems, LLM inference optimization, batching, and caching strategies, Kubernetes and cloud infrastructure (AWS, GCP), Python or Rust","baseSalary":{"@type":"MonetaryAmount","currency":"EUR","value":{"@type":"QuantitativeValue","minValue":235000,"maxValue":295000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_32c0c69a-037"},"title":"Staff Software Engineer, Inference","description":"<p><strong>About the role:</strong></p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>As a Staff Software Engineer on our Inference team, you will work end to end, identifying and addressing key infrastructure blockers to serve Claude to millions of users while enabling breakthrough AI research. Strong candidates should have familiarity with performance optimization, distributed systems, large-scale service orchestration, and intelligent request routing. Familiarity with LLM inference optimization, batching strategies, and multi-accelerator deployments is highly encouraged but not strictly necessary.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Work end to end on identifying and addressing key infrastructure blockers to serve Claude to millions of users while enabling breakthrough AI research</li>\n<li>Collaborate with the team to design and implement solutions to complex problems</li>\n<li>Develop and maintain large-scale distributed systems</li>\n<li>Implement and deploy machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP)</li>\n<li>Python or Rust</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>Significant software engineering experience, particularly with distributed systems</li>\n<li>Results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p><strong>Benefits:</strong></p>\n<ul>\n<li>Competitive compensation and benefits</li>\n<li>Optional equity donation matching</li>\n<li>Generous vacation and parental leave</li>\n<li>Flexible working hours</li>\n<li>Lovely office space in which to collaborate with colleagues</li>\n</ul>\n<p><strong>Application Instructions:</strong></p>\n<p>If you&#39;re interested in this role, please submit your application through our website. We look forward to hearing from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_32c0c69a-037","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5150472008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"€295.000-€355.000 EUR","x-skills-required":["performance optimization","distributed systems","large-scale service orchestration","intelligent request routing","LLM inference optimization","batching strategies","multi-accelerator deployments","Kubernetes","cloud infrastructure","Python","Rust"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:56:14.384Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, IE"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"performance optimization, distributed systems, large-scale service orchestration, intelligent request routing, LLM inference optimization, batching strategies, multi-accelerator deployments, Kubernetes, cloud infrastructure, Python, Rust"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e394b0fa-2ba"},"title":"Staff Software Engineer, Inference","description":"<p><strong>About the role</strong></p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>As a Staff Software Engineer on our Inference team, you will work end to end, identifying and addressing key infrastructure blockers to serve Claude to millions of users while enabling breakthrough AI research. Strong candidates should have familiarity with performance optimization, distributed systems, large-scale service orchestration, and intelligent request routing. Familiarity with LLM inference optimization, batching strategies, and multi-accelerator deployments is highly encouraged but not strictly necessary.</p>\n<p><strong>Strong candidates may also have experience with</strong></p>\n<ul>\n<li>High-performance, large-scale distributed systems</li>\n<li>Implementing and deploying machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP)</li>\n<li>Python or Rust</li>\n</ul>\n<p><strong>You may be a good fit if you</strong></p>\n<ul>\n<li>Have significant software engineering experience, particularly with distributed systems</li>\n<li>Are results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p><strong>Representative projects across the org</strong></p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p><strong>Deadline to apply</strong></p>\n<p>None. Applications will be reviewed on a rolling basis.</p>\n<p><strong>Annual compensation range</strong></p>\n<p>The annual compensation range for this role is £325,000-£390,000 GBP.</p>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>\n<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>\n<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>Why work with us?</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e394b0fa-2ba","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5097742008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"£325,000-£390,000 GBP","x-skills-required":["performance optimization","distributed systems","large-scale service orchestration","intelligent request routing","LLM inference optimization","batching strategies","multi-accelerator deployments","Kubernetes","cloud infrastructure","Python","Rust"],"x-skills-preferred":["high-performance distributed systems","machine learning systems","load balancing","request routing","traffic management","caching strategies"],"datePosted":"2026-04-18T15:50:52.588Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"performance optimization, distributed systems, large-scale service orchestration, intelligent request routing, LLM inference optimization, batching strategies, multi-accelerator deployments, Kubernetes, cloud infrastructure, Python, Rust, high-performance distributed systems, machine learning systems, load balancing, request routing, traffic management, caching strategies","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":325000,"maxValue":390000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e5a3deb2-908"},"title":"Senior Software Engineer, Inference","description":"<p>Job Title: Senior Software Engineer, Inference</p>\n<p>About the Role:</p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>The team has a dual mandate: maximizing compute efficiency to serve our explosive customer growth, while enabling breakthrough research by giving our scientists the high-performance inference infrastructure they need to develop next-generation models. We tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Significant software engineering experience, particularly with distributed systems</li>\n<li>Results-oriented, with a bias towards flexibility and impact</li>\n<li>Ability to pick up slack, even if it goes outside your job description</li>\n<li>Willingness to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Competitive compensation and benefits</li>\n<li>Optional equity donation matching</li>\n<li>Generous vacation and parental leave</li>\n<li>Flexible working hours</li>\n<li>Lovely office space in which to collaborate with colleagues</li>\n</ul>\n<p>Note: The salary range for this role is €235,000-€295,000 EUR per year.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e5a3deb2-908","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4641822008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"€235,000-€295,000 EUR per year","x-skills-required":["High-performance, large-scale distributed systems","Implementing and deploying machine learning systems at scale","Load balancing, request routing, or traffic management systems","LLM inference optimization, batching, and caching strategies","Kubernetes and cloud infrastructure (AWS, GCP)","Python or Rust"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:39.086Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, IE"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"High-performance, large-scale distributed systems, Implementing and deploying machine learning systems at scale, Load balancing, request routing, or traffic management systems, LLM inference optimization, batching, and caching strategies, Kubernetes and cloud infrastructure (AWS, GCP), Python or Rust","baseSalary":{"@type":"MonetaryAmount","currency":"EUR","value":{"@type":"QuantitativeValue","minValue":235000,"maxValue":295000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f95fe525-8fd"},"title":"Staff Software Engineer, Inference","description":"<p><strong>About the role</strong></p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators. The team has a dual mandate: maximizing compute efficiency to serve our explosive customer growth, while enabling breakthrough research by giving our scientists the high-performance inference infrastructure they need to develop next-generation models. We tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.</p>\n<p><strong>As a Staff Software Engineer on our Inference team, you will work end to end, identifying and addressing key infrastructure blockers to serve Claude to millions of users while enabling breakthrough AI research. Strong candidates should have familiarity with performance optimization, distributed systems, large-scale service orchestration, and intelligent request routing. Familiarity with LLM inference optimization, batching strategies, and multi-accelerator deployments is highly encouraged but not strictly necessary.</strong></p>\n<p><strong>Strong candidates may also have experience with</strong></p>\n<ul>\n<li>High-performance, large-scale distributed systems</li>\n<li>Implementing and deploying machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP)</li>\n<li>Python or Rust</li>\n</ul>\n<p><strong>You may be a good fit if you</strong></p>\n<ul>\n<li>Have significant software engineering experience, particularly with distributed systems</li>\n<li>Are results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p><strong>Representative projects across the org</strong></p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p><strong>Deadline to apply: None. Applications will be reviewed on a rolling basis.</strong></p>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Education requirements: We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</li>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</strong></p>\n<p><strong>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</strong></p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f95fe525-8fd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5097742008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"£325,000 - £390,000GBP","x-skills-required":["performance optimization","distributed systems","large-scale service orchestration","intelligent request routing","LLM inference optimization","batching strategies","multi-accelerator deployments","Kubernetes","cloud infrastructure","Python","Rust"],"x-skills-preferred":["high-performance, large-scale distributed systems","implementing and deploying machine learning systems at scale","load balancing, request routing, or traffic management systems","caching strategies"],"datePosted":"2026-03-08T13:49:42.673Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"performance optimization, distributed systems, large-scale service orchestration, intelligent request routing, LLM inference optimization, batching strategies, multi-accelerator deployments, Kubernetes, cloud infrastructure, Python, Rust, high-performance, large-scale distributed systems, implementing and deploying machine learning systems at scale, load balancing, request routing, or traffic management systems, caching strategies","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":325000,"maxValue":390000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ca53b3f7-f72"},"title":"Staff / Senior Software Engineer, Inference","description":"<p><strong>About the role</strong></p>\n<p>Our Inference team is responsible for building and maintaining the critical systems that serve Claude to millions of users worldwide. We bring Claude to life by serving our models via the industry&#39;s largest compute-agnostic inference deployments. We are responsible for the entire stack from intelligent request routing to fleet-wide orchestration across diverse AI accelerators.</p>\n<p>The team has a dual mandate: <strong>maximizing compute efficiency</strong> to serve our explosive customer growth, while <strong>enabling breakthrough research</strong> by giving our scientists the high-performance inference infrastructure they need to develop next-generation models. We tackle complex, distributed systems challenges across multiple accelerator families and emerging AI hardware running in multiple cloud platforms.</p>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have significant software engineering experience, particularly with distributed systems</li>\n<li>Are results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Enjoy pair programming (we love to pair!)</li>\n<li>Want to learn more about machine learning systems and infrastructure</li>\n<li>Thrive in environments where technical excellence directly drives both business results and research breakthroughs</li>\n<li>Care about the societal impacts of your work</li>\n</ul>\n<p><strong>Strong candidates may also have experience with:</strong></p>\n<ul>\n<li>High-performance, large-scale distributed systems</li>\n<li>Implementing and deploying machine learning systems at scale</li>\n<li>Load balancing, request routing, or traffic management systems</li>\n<li>LLM inference optimization, batching, and caching strategies</li>\n<li>Kubernetes and cloud infrastructure (AWS, GCP, Azure)</li>\n<li>Python or Rust</li>\n</ul>\n<p><strong>Representative projects:</strong></p>\n<ul>\n<li>Designing intelligent routing algorithms that optimize request distribution across thousands of accelerators</li>\n<li>Autoscaling our compute fleet to dynamically match supply with demand across production, research, and experimental workloads</li>\n<li>Building production-grade deployment pipelines for releasing new models to millions of users</li>\n<li>Integrating new AI accelerator platforms to maintain our hardware-agnostic competitive advantage</li>\n<li>Contributing to new inference features (e.g., structured sampling, prompt caching)</li>\n<li>Supporting inference for new model architectures</li>\n<li>Analyzing observability data to tune performance based on real-world production workloads</li>\n<li>Managing multi-region deployments and geographic routing for global customers</li>\n</ul>\n<p><strong>Deadline to apply:</strong></p>\n<p>None. Applications will be reviewed on a rolling basis.</p>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</strong></p>\n<p><strong>Your safety matters to us. To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</strong></p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p>The easiest way to understand our research directions is to read our recent research. This research co</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ca53b3f7-f72","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4951696008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$300,000 - $485,000 USD","x-skills-required":["distributed systems","machine learning systems","load balancing","request routing","traffic management","LLM inference optimization","Kubernetes","cloud infrastructure","Python","Rust"],"x-skills-preferred":["high-performance distributed systems","implementing and deploying machine learning systems at scale","structured sampling","prompt caching"],"datePosted":"2026-03-08T13:49:03.736Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"distributed systems, machine learning systems, load balancing, request routing, traffic management, LLM inference optimization, Kubernetes, cloud infrastructure, Python, Rust, high-performance distributed systems, implementing and deploying machine learning systems at scale, structured sampling, prompt caching","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":300000,"maxValue":485000,"unitText":"YEAR"}}}]}