{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/analytics-infrastructure"},"x-facet":{"type":"skill","slug":"analytics-infrastructure","display":"Analytics Infrastructure","count":4},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1aad838f-387"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability.</p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li>Data Governance &amp; Access Control: Design and implement robust access control systems ensuring only authorized users can access sensitive data.</li>\n<li>Financial Data Infrastructure: Build and maintain data pipelines and warehouses powering business-critical reporting.</li>\n<li>Cloud Storage &amp; Reliability: Architect disaster recovery, backup, and replication systems for petabyte-scale data.</li>\n<li>Data Platform &amp; Tooling: Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark.</li>\n</ul>\n<p>You&#39;ll work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p>To be successful in this role, you&#39;ll need:</p>\n<ul>\n<li>10+ years of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems.</li>\n<li>3+ years of experience leading large scale, complex projects or teams as an engineer or tech lead.</li>\n<li>Deep experience with at least one of:</li>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar.</li>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS).</li>\n<li>Can navigate complex technical tradeoffs between performance, cost, security, and maintainability.</li>\n<li>Have excellent collaboration skills - you work well with both technical and non-technical stakeholders.</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure.</li>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures.</li>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale.</li>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks.</li>\n<li>Experience working in fintech, financial services, or highly regulated environments.</li>\n<li>Security engineering background with focus on data protection and access controls.</li>\n</ul>\n<p>Technologies We Use:</p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran.</li>\n<li>Storage: GCS, S3.</li>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS.</li>\n<li>Languages: Python, Go, SQL.</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1aad838f-387","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures"],"x-skills-preferred":["data warehousing","ETL/ELT pipelines","analytics infrastructure","data reliability","availability","cost efficiency","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"datePosted":"2026-04-18T15:52:47.297Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, data reliability, availability, cost efficiency, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a738803a-64f"},"title":"Head of Enterprise Marketing Strategy & Analytics","description":"<p><strong>About the Role</strong></p>\n<p>This foundational leadership role will build and lead the Enterprise Marketing Strategy &amp; Analytics function, serving as the operating system for a rapidly scaling marketing organisation. The primary mandate is to define and measure success across all marketing programmes,from demand generation (field events, ABM, EBCs, partner co-marketing) to pipeline contribution,creating a clear line from investment to pipeline to revenue.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Define and own the Enterprise Marketing measurement framework, targets, and reporting, covering the full funnel from top-of-funnel demand through pipeline influence and closed-won attribution.</li>\n<li>Build and maintain core analytics infrastructure (data models, attribution logic, dashboards) in partnership with Revenue Operations and Data Science, ensuring marketing and sales alignment on key metrics.</li>\n<li>Serve as one of the primary operating partner to Finance, HR, and Recruiting, leading budget tracking, headcount planning, and vendor management.</li>\n<li>Partner with marketing leadership and the central Marketing Ops &amp; Strategy team on annual and quarterly planning, resource allocation, and performance reviews.</li>\n<li>Establish the operating cadence for Enterprise Marketing (QBRs, pipeline reviews, program retros), coordinating with the central Marketing Ops &amp; Strategy team on organisation-wide rhythms, and drive the preparation needed to make these forums decision-useful.</li>\n<li>Lead the identification of high-leverage workflows to automate, partnering with the central GTM AI team on implementation and measuring productivity gains.</li>\n<li>Build and manage the Marketing Operations, Demand Analytics, and MarTech team, setting a high bar for analytical rigor and business partnership.</li>\n<li>Drive cross-functional alignment on shared definitions, tooling, and a single source of truth for marketing performance across the broader Marketing organisation and with Revenue Operations.</li>\n<li>Conduct strategic analyses to inform key organisational decisions, such as resource deployment, coverage ratios, and campaign capacity planning.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>10+ years in marketing operations, analytics, revenue operations, or strategy roles, including at least 3 years leading a team.</li>\n<li>Experience building or significantly scaling a marketing ops/analytics function at a high-growth B2B technology company undergoing significant organisational expansion.</li>\n<li>Deep fluency in the enterprise demand funnel, including lead scoring, MQL/SQL definitions, pipeline attribution, and campaign influence models.</li>\n<li>Hands-on expertise with the modern GTM data stack (CRM, Marketing Automation, BI tools).</li>\n<li>Proven track record of strategic partnership with Finance and Revenue Operations, including experience building budget models and sitting in planning cycles.</li>\n<li>Expertise in running the core operational rhythm of a marketing organisation: QBRs, headcount tracking, budget pacing, and vendor renewals.</li>\n<li>Strong written and verbal communication, capable of translating complex datasets into clear business narratives.</li>\n<li>Genuine curiosity about AI and a willingness to be an early, hands-on adopter of automation tools in your team’s workflows.</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. Visa sponsorship: We do sponsor visas!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a738803a-64f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5169101008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$400,000-$400,000 USD","x-skills-required":["marketing operations","analytics","revenue operations","strategy","demand generation","field events","ABM","EBCs","partner co-marketing","pipeline contribution","marketing programmes","investment","pipeline","revenue","measurement framework","targets","reporting","funnel","top-of-funnel demand","pipeline influence","closed-won attribution","core analytics infrastructure","data models","attribution logic","dashboards","Data Science","marketing","sales","alignment","key metrics","budget tracking","headcount planning","vendor management","marketing leadership","central Marketing Ops & Strategy team","annual planning","quarterly planning","resource allocation","performance reviews","operating cadence","QBRs","pipeline reviews","program retros","organisation-wide rhythms","decision-useful","high-leverage workflows","automation","GTM AI team","implementation","productivity gains","Demand Analytics","MarTech team","analytical rigor","business partnership","cross-functional alignment","shared definitions","tooling","single source of truth","marketing performance","strategic analyses","resource deployment","coverage ratios","campaign capacity planning","lead scoring","MQL/SQL definitions","pipeline attribution","campaign influence models","modern GTM data stack","CRM","Marketing Automation","BI tools","strategic partnership","Finance","budget models","planning cycles","core operational rhythm","headcount tracking","budget pacing","vendor renewals","written communication","verbal communication","complex datasets","business narratives","AI","automation tools"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:21.649Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Marketing","industry":"Technology","skills":"marketing operations, analytics, revenue operations, strategy, demand generation, field events, ABM, EBCs, partner co-marketing, pipeline contribution, marketing programmes, investment, pipeline, revenue, measurement framework, targets, reporting, funnel, top-of-funnel demand, pipeline influence, closed-won attribution, core analytics infrastructure, data models, attribution logic, dashboards, Data Science, marketing, sales, alignment, key metrics, budget tracking, headcount planning, vendor management, marketing leadership, central Marketing Ops & Strategy team, annual planning, quarterly planning, resource allocation, performance reviews, operating cadence, QBRs, pipeline reviews, program retros, organisation-wide rhythms, decision-useful, high-leverage workflows, automation, GTM AI team, implementation, productivity gains, Demand Analytics, MarTech team, analytical rigor, business partnership, cross-functional alignment, shared definitions, tooling, single source of truth, marketing performance, strategic analyses, resource deployment, coverage ratios, campaign capacity planning, lead scoring, MQL/SQL definitions, pipeline attribution, campaign influence models, modern GTM data stack, CRM, Marketing Automation, BI tools, strategic partnership, Finance, budget models, planning cycles, core operational rhythm, headcount tracking, budget pacing, vendor renewals, written communication, verbal communication, complex datasets, business narratives, AI, automation tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":400000,"maxValue":400000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7599c887-a7e"},"title":"Design Engineer, AI Capability Development (Education Labs)","description":"<p><strong>About Anthropic</strong></p>\n<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>\n<p><strong>About the role</strong></p>\n<p>We believe skill with AI is fundamental to human agency. Education Labs builds the paradigms that help people become genuinely more capable—not just more engaged.</p>\n<p>This is a new kind of role: part researcher, part product builder, part interaction designer. You&#39;ll be the second technical builder on a small team studying how AI transforms human capability—and shipping features based on what we discover. You&#39;ll have significant creative license to define what &#39;good&#39; looks like, exploring new interaction patterns rather than optimizing existing ones.</p>\n<p>We&#39;re skeptical of tutorials, onboarding flows, and engagement metrics. We care about experiences that make users progressively more capable, curious, and empowered over time. This means integrating skill development into product design, using Claude itself as a capability-building partner, and measuring success by how users actually grow.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Ship features that help users develop real skill with AI—measuring success by capability growth, not time-on-site</li>\n</ul>\n<ul>\n<li>Architect end-to-end prototypes (front-end and back-end) that test new interaction paradigms, with particular attention to the front-of-the-frontend: motion, polish, and interaction feel</li>\n</ul>\n<ul>\n<li>Define technical direction for the team—establish patterns others can follow</li>\n</ul>\n<ul>\n<li>Build relationships across Product, Design, and Research to influence how skill development principles shape Anthropic&#39;s broader product strategy</li>\n</ul>\n<ul>\n<li>Shape team strategy and roadmap—identify the highest-leverage opportunities and build conviction across stakeholders</li>\n</ul>\n<ul>\n<li>Translate research insights about skill development and human-AI collaboration into shipped product through close collaboration with researchers</li>\n</ul>\n<ul>\n<li>Document and share your work through clear writing, prototypes, and presentations that influence thinking across the organization</li>\n</ul>\n<p><strong>You may be a good fit if you have:</strong></p>\n<p><strong>Strong full-stack engineering with design sensibility</strong></p>\n<ul>\n<li>6+ years building and shipping web products, with deep expertise across the stack</li>\n</ul>\n<ul>\n<li>Strong front-end craft: TypeScript/JavaScript, React, CSS—with an eye for interaction design, motion, and visual detail</li>\n</ul>\n<ul>\n<li>Solid back-end and data pipeline experience: Python, API design, analytics infrastructure</li>\n</ul>\n<ul>\n<li>A portfolio showcasing innovative interaction designs and high-quality implementations</li>\n</ul>\n<ul>\n<li>Track record of independently driving features from prototype to production</li>\n</ul>\n<p><strong>Deep conviction about human capability</strong></p>\n<ul>\n<li>Strong perspective on how technology should enhance human capabilities rather than diminish them</li>\n</ul>\n<ul>\n<li>Experience or genuine passion for skill development, HCI, developer tools, or products that help people become more capable</li>\n</ul>\n<ul>\n<li>Skepticism of purely engagement-driven metrics; interest in measuring capability outcomes</li>\n</ul>\n<p><strong>Research mindset with product execution</strong></p>\n<ul>\n<li>Comfort with ambiguity and exploring undefined problem spaces</li>\n</ul>\n<ul>\n<li>Ability to rapidly prototype, test with users, and iterate toward production</li>\n</ul>\n<ul>\n<li>Strong instincts for product design and user experience, even without formal design training</li>\n</ul>\n<p><strong>Strategic leadership and coalition building</strong></p>\n<ul>\n<li>Experience setting vision, shaping team strategy, and building conviction across cross-functional stakeholders</li>\n</ul>\n<ul>\n<li>Ability to build productive relationships with Product, Design, Research, and Engineering teams—especially when your team isn&#39;t the owner</li>\n</ul>\n<ul>\n<li>Strong sense of prioritization—knowing what to build now, what to defer, and what to cut</li>\n</ul>\n<ul>\n<li>Track record of influencing roadmaps and decisions beyond your immediate team</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Experience in developer tools, creative tools, learning platforms, or other products where user skill development and mastery matter more than time-on-site</li>\n</ul>\n<ul>\n<li>Background in learning sciences, cognitive science, HCI, skill acquisition research, or educational psychology (formal or self-directed)</li>\n</ul>\n<ul>\n<li>Experience with experimentation frameworks, A/B testing, or analytics that measure capability development in production</li>\n</ul>\n<ul>\n<li>Previous experience in research labs, frontier tech companies, or startups with high autonomy and ambiguity</li>\n</ul>\n<ul>\n<li>Published writing, talks, or open-source work on skill development, human-AI interaction, or product philosophy</li>\n</ul>\n<ul>\n<li>Experience building AI-native product experiences or working with LLMs in production contexts</li>\n</ul>\n<p><strong>Strong candidates may also have</strong></p>\n<ul>\n<li>Background in learning sciences, cognitive science, HCI, or educational psychology</li>\n</ul>\n<ul>\n<li>Experience in developer tools, creative tools, or learning platforms where mastery matters more than engagement</li>\n</ul>\n<ul>\n<li>Published writing, talks, or open-source work on skill development or human-AI interaction</li>\n</ul>\n<ul>\n<li>Experience building AI-native product experiences or working with LLMs in production</li>\n</ul>\n<p><strong>What this role is not</strong></p>\n<p>This is a hands-on technical role building product features, embedded within a research team. You&#39;ll provide technical guidance and help set direction, but this role doesn&#39;t involve people management off the bat. If you&#39;re looking to immediately transition into engineering management or lead a large team, this likely isn&#39;t the right fit.</p>\n<p>The annual compensation range for this role is not specified.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7599c887-a7e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5097186008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["full-stack engineering","design sensibility","TypeScript/JavaScript","React","CSS","Python","API design","analytics infrastructure"],"x-skills-preferred":["developer tools","creative tools","learning platforms","skill development","HCI","developer tools","experimentation frameworks","A/B testing","analytics"],"datePosted":"2026-03-08T13:57:54.483Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"full-stack engineering, design sensibility, TypeScript/JavaScript, React, CSS, Python, API design, analytics infrastructure, developer tools, creative tools, learning platforms, skill development, HCI, developer tools, experimentation frameworks, A/B testing, analytics"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1ace7478-7a2"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p><strong>About the role</strong></p>\n<p>Data Infrastructure designs, operates, and scales secure, privacy-respecting systems that power data-driven decisions across Anthropic. Our mission is to provide data processing, storage, and access that are trusted, fast, and easy to use.</p>\n<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability. This role offers the opportunity to work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p><strong>Responsibilities:</strong></p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li><strong>Data Governance &amp; Access Control:</strong> Design and implement robust access control systems ensuring only authorized users can access sensitive data. Build infrastructure for permission management, audit logging, and compliance requirements. Work on IAM policies, ACLs, and security controls that scale across thousands of users and systems.</li>\n</ul>\n<ul>\n<li><strong>Financial Data Infrastructure:</strong> Build and maintain data pipelines and warehouses powering business-critical reporting. Ensure data integrity, accuracy, and availability for complex financial systems, including third party revenue ingestion pipelines; manage the external relationships as needed to drive upstream dependencies. Own the reliability of systems processing revenue, usage, and business metrics.</li>\n</ul>\n<ul>\n<li><strong>Cloud Storage &amp; Reliability:</strong> Architect disaster recovery, backup, and replication systems for petabyte-scale data. Ensure high availability and durability of data stored in cloud object storage (GCS, S3). Build systems that protect against data loss and enable rapid recovery.</li>\n</ul>\n<ul>\n<li><strong>Data Platform &amp; Tooling:</strong> Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark. Optimize query performance, manage costs, and enable self-service analytics across the organization.</li>\n</ul>\n<p><strong>You might be a good fit if you:</strong></p>\n<ul>\n<li>Have 10+ years (not including internships or co-ops) of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems</li>\n</ul>\n<ul>\n<li>Have 3+ years (not including internships or co-ops) of experience leading large scale, complex projects or teams as an engineer or tech lead</li>\n</ul>\n<ul>\n<li>Can set technical direction for a team, not just execute within it</li>\n</ul>\n<ul>\n<li>Have deep experience with at least one of:</li>\n</ul>\n<ul>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar</li>\n</ul>\n<ul>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS)</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure</li>\n</ul>\n<ul>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures</li>\n</ul>\n<ul>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale</li>\n</ul>\n<ul>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks</li>\n</ul>\n<ul>\n<li>Experience working in fintech, financial services, or highly regulated environments</li>\n</ul>\n<ul>\n<li>Security engineering background with focus on data protection and access controls</li>\n</ul>\n<p><strong>Technologies We Use:</strong></p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran</li>\n</ul>\n<ul>\n<li>Storage: GCS, S3</li>\n</ul>\n<ul>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS</li>\n</ul>\n<ul>\n<li>Languages: Python, Go, SQL</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</p>\n<p><strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1ace7478-7a2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000 - $485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures","data warehousing","ETL/ELT pipelines","analytics infrastructure","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"x-skills-preferred":["data governance","access control","cloud storage","reliability","data platform","tooling","self-service analytics","data processing infrastructure","query performance","cost management"],"datePosted":"2026-03-08T13:52:03.469Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls, data governance, access control, cloud storage, reliability, data platform, tooling, self-service analytics, data processing infrastructure, query performance, cost management","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}}]}