{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/pulumi"},"x-facet":{"type":"skill","slug":"pulumi","display":"Pulumi","count":11},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1aad838f-387"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability.</p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li>Data Governance &amp; Access Control: Design and implement robust access control systems ensuring only authorized users can access sensitive data.</li>\n<li>Financial Data Infrastructure: Build and maintain data pipelines and warehouses powering business-critical reporting.</li>\n<li>Cloud Storage &amp; Reliability: Architect disaster recovery, backup, and replication systems for petabyte-scale data.</li>\n<li>Data Platform &amp; Tooling: Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark.</li>\n</ul>\n<p>You&#39;ll work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p>To be successful in this role, you&#39;ll need:</p>\n<ul>\n<li>10+ years of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems.</li>\n<li>3+ years of experience leading large scale, complex projects or teams as an engineer or tech lead.</li>\n<li>Deep experience with at least one of:</li>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar.</li>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS).</li>\n<li>Can navigate complex technical tradeoffs between performance, cost, security, and maintainability.</li>\n<li>Have excellent collaboration skills - you work well with both technical and non-technical stakeholders.</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure.</li>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures.</li>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale.</li>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks.</li>\n<li>Experience working in fintech, financial services, or highly regulated environments.</li>\n<li>Security engineering background with focus on data protection and access controls.</li>\n</ul>\n<p>Technologies We Use:</p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran.</li>\n<li>Storage: GCS, S3.</li>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS.</li>\n<li>Languages: Python, Go, SQL.</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1aad838f-387","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures"],"x-skills-preferred":["data warehousing","ETL/ELT pipelines","analytics infrastructure","data reliability","availability","cost efficiency","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"datePosted":"2026-04-18T15:52:47.297Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, data reliability, availability, cost efficiency, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_fa9a54d7-549"},"title":"Senior Site Reliability Engineer, Data Infrastructure","description":"<p>As a Senior Site Reliability Engineer, you will own the reliability and performance of our Kubernetes-based data platform. You will design and operate highly available, multi-region systems, ensuring our services meet strict uptime and latency targets.</p>\n<p>Day-to-day, you’ll work on scaling infrastructure, improving deployment pipelines, and hardening our security posture. You’ll play a key role in evolving our DevSecOps practices while partnering closely with engineering teams to ensure services are built for reliability from day one.</p>\n<p>We operate with production-grade discipline, supporting mission-critical services with stringent uptime requirements and a focus on automation, observability, and resilience.</p>\n<p>The Platform &amp; Infrastructure Engineering team in the Data Infrastructure organization is responsible for the reliability, scalability, and security of the company’s data platform. The team builds and operates the foundational systems that power data ingestion, transformation, analytics, and internal AI workloads at scale.</p>\n<p>About the role:</p>\n<ul>\n<li>5+ years of experience in Site Reliability Engineering, Platform Engineering, or Infrastructure Engineering roles</li>\n<li>Deep expertise in Kubernetes and containerized software services, including cluster design, operations, and troubleshooting in production environments</li>\n<li>Strong experience building and operating CI/CD systems, including tools such as Argo CD and GitHub Actions</li>\n<li>Proven experience owning production systems with high availability requirements (≥99.99% uptime), including incident response, SLI/SLO/SLA definition, error budgets, and postmortems</li>\n<li>Hands-on experience designing and operating geo-replicated, multi-region, active-active systems, including traffic routing, failover strategies, and data consistency tradeoffs</li>\n<li>Strong experience building and owning observability components, including metrics, logging, and tracing (e.g., Prometheus, Grafana, OpenTelemetry).</li>\n<li>Experience with infrastructure as code (e.g., Helm, Terraform, Pulumi) and automated environment provisioning</li>\n<li>Strong understanding of system performance tuning, capacity planning, and resource optimization in distributed systems</li>\n<li>Experience implementing and operating security best practices in cloud-native environments (e.g., secrets management, network policies, vulnerability scanning)</li>\n</ul>\n<p>Preferred:</p>\n<ul>\n<li>Experience operating data platforms or data-intensive workloads (e.g., Spark, Airflow, Kafka, Flink)</li>\n<li>Familiarity with service mesh technologies (e.g., Istio, Linkerd)</li>\n<li>Experience working in regulated environments with compliance frameworks such as GDPR, SOC 2, HIPAA, or SOX</li>\n<li>Background in building internal developer platforms or self-service infrastructure</li>\n</ul>\n<p>Wondering if you’re a good fit?</p>\n<p>We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren’t a 100% skill or experience match.</p>\n<p>Here are a few qualities we’ve found compatible with our team. If some of this describes you, we’d love to talk.</p>\n<ul>\n<li>You love building highly reliable systems that operate at scale</li>\n<li>You’re curious about how to continuously improve system resilience, security, and operations</li>\n<li>You’re an expert in diagnosing and solving complex distributed systems problems</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning.</p>\n<p>Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems.</p>\n<p>As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too.</p>\n<p>Come join us!</p>\n<p>The base salary range for this role is $165,000 to $242,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation.</p>\n<p>In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.</p>\n<p>In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n<li>100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets.</p>\n<p>New hires will be invited to attend onboarding at one of our hubs within their first month.</p>\n<p>Teams also gather quarterly to support collaboration.</p>\n<p>California Consumer Privacy Act - California applicants only</p>\n<p>CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace.</p>\n<p>All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information.</p>\n<p>As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship.</p>\n<p>If reasonable accommodation is needed, please contact: careers@coreweave.com.</p>\n<p>Export Control Compliance</p>\n<p>This position requires access to export controlled information.</p>\n<p>To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without restrictions, or (C) otherwise exempt from the export regulations.</p>\n<p>If you are not a U.S. person, you will be required to provide documentation of your eligibility to access the export controlled information before being considered for this position.</p>\n<p>Please note that CoreWeave is subject to the requirements of the U.S. Department of Commerce&#39;s Export Administration Regulations (EAR) and the U.S. Department of State&#39;s International Traffic in Arms Regulations (ITAR).</p>\n<p>By applying for this position, you acknowledge that you have read and understood the export control requirements and that you will comply with them.</p>\n<p>If you have any questions or concerns regarding the export control requirements, please contact: careers@coreweave.com.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_fa9a54d7-549","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4671535006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["Kubernetes","containerized software services","cluster design","operations","troubleshooting","CI/CD systems","Argo CD","GitHub Actions","production systems","high availability","incident response","SLI/SLO/SLA definition","error budgets","postmortems","geo-replicated","multi-region","active-active systems","traffic routing","failover strategies","data consistency tradeoffs","observability components","metrics","logging","tracing","Prometheus","Grafana","OpenTelemetry","infrastructure as code","Helm","Terraform","Pulumi","automated environment provisioning","system performance tuning","capacity planning","resource optimization","distributed systems","security best practices","cloud-native environments","secrets management","network policies","vulnerability scanning"],"x-skills-preferred":["Spark","Airflow","Kafka","Flink","service mesh technologies","Istio","Linkerd","regulated environments","compliance frameworks","GDPR","SOC 2","HIPAA","SOX","internal developer platforms","self-service infrastructure"],"datePosted":"2026-04-18T15:51:59.035Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, containerized software services, cluster design, operations, troubleshooting, CI/CD systems, Argo CD, GitHub Actions, production systems, high availability, incident response, SLI/SLO/SLA definition, error budgets, postmortems, geo-replicated, multi-region, active-active systems, traffic routing, failover strategies, data consistency tradeoffs, observability components, metrics, logging, tracing, Prometheus, Grafana, OpenTelemetry, infrastructure as code, Helm, Terraform, Pulumi, automated environment provisioning, system performance tuning, capacity planning, resource optimization, distributed systems, security best practices, cloud-native environments, secrets management, network policies, vulnerability scanning, Spark, Airflow, Kafka, Flink, service mesh technologies, Istio, Linkerd, regulated environments, compliance frameworks, GDPR, SOC 2, HIPAA, SOX, internal developer platforms, self-service infrastructure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c7fe95f3-dcf"},"title":"Site Reliability Engineer (SRE)","description":"<p>You will work on the team responsible for the backend services that power our products such as grok.com and the API. We focus on writing and maintaining highly scalable and reliable services that can efficiently process tens of thousands of queries per second. The services are hosted on a number of Kubernetes clusters (on-prem &amp; cloud).</p>\n<p>Our team is small, highly motivated, and focused on engineering excellence. We operate with a flat organisational structure. All employees are expected to be hands-on and to contribute directly to the company&#39;s mission. Leadership is given to those who show initiative and consistently deliver excellence.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Work on the team that is responsible for the backend services that power our products such as grok.com and the API.</li>\n<li>Write and maintain highly scalable and reliable services that can efficiently process tens of thousands of queries per second.</li>\n<li>Ensure the services are hosted on a number of Kubernetes clusters (on-prem &amp; cloud).</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>Expert knowledge of Kubernetes.</li>\n<li>Expert knowledge of continuous deployment systems such as Buildkite and ArgoCD.</li>\n<li>Expert knowledge of monitoring technologies such as Prometheus, Grafana, and PagerDuty.</li>\n<li>Expert knowledge of infrastructure as code technologies such as Pulumi or Terraform.</li>\n<li>Familiarity with a systems programming language like Rust, C++ or Go.</li>\n<li>Experience with traffic management and HTTP proxies such as nginx and envoy.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c7fe95f3-dcf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://xai.com","logo":"https://logos.yubhub.co/xai.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/4681662007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kubernetes","Buildkite","ArgoCD","Prometheus","Grafana","PagerDuty","Pulumi","Terraform","Rust","C++","Go","nginx","envoy"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:48:59.475Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, Buildkite, ArgoCD, Prometheus, Grafana, PagerDuty, Pulumi, Terraform, Rust, C++, Go, nginx, envoy"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_04884ef5-f9e"},"title":"Software Engineer, Compute (8+ YOE)","description":"<p>We&#39;re looking for an experienced software engineer to help lead the next phase of platform maturity in how we run Kubernetes at Airtable. As a member of the Compute Platform team, you&#39;ll be responsible for building and evolving the infrastructure that powers Airtable&#39;s services at scale.</p>\n<p>Your primary focus will be on designing, implementing, and scaling core Kubernetes platform capabilities used across ~70 clusters, spread across multiple environments. You&#39;ll also lead foundational modernization efforts, such as migrating to a new CNI plugin to overhaul IP security rule management across clusters and regions.</p>\n<p>In addition to your technical expertise, you&#39;ll collaborate closely with product and security teams to power a rapidly growing enterprise business. You&#39;ll spend roughly 70% of your time in hands-on engineering and 30% in design reviews, mentorship, and cross-team collaboration.</p>\n<p>To succeed in this role, you&#39;ll need 8+ years of software engineering experience, with deep expertise building and operating a Kubernetes-based internal service platform. You&#39;ll also need a strong understanding of Kubernetes internals, including controllers/operators, CRDs, networking, and cluster architecture.</p>\n<p>If you&#39;re excited about building internal platforms, shaping infrastructure strategy, and partnering closely with product and security teams, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_04884ef5-f9e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airtable","sameAs":"https://airtable.com/","logo":"https://logos.yubhub.co/airtable.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airtable/jobs/8442397002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kubernetes","Typescript","Golang","Cloud Native Infrastructure","CI/CD","Infrastructure as Code","Terraform","CloudFormation","OpenTofu","Pulumi"],"x-skills-preferred":["AWS infrastructure","EKS","Spinnaker","ArgoCD","Flux","Jenkins"],"datePosted":"2026-04-18T15:39:46.997Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY; Remote - US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, Typescript, Golang, Cloud Native Infrastructure, CI/CD, Infrastructure as Code, Terraform, CloudFormation, OpenTofu, Pulumi, AWS infrastructure, EKS, Spinnaker, ArgoCD, Flux, Jenkins"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_22fe5cb2-ba9"},"title":"Engineering Manager, Datastores","description":"<p>At Webflow, we&#39;re building the world&#39;s leading AI-native Digital Experience Platform, and we&#39;re doing it as a remote-first company built on trust, transparency, and a whole lot of creativity.</p>\n<p>This work takes grit, because we move fast, without ever sacrificing craft or quality. Our mission is to bring development superpowers to everyone. From entrepreneurs launching their first idea to global enterprises scaling their digital presence, we empower teams to design, launch, and optimize for the web without barriers.</p>\n<p>We believe the future of the web, and work, is more open, more creative, and more equitable. And we’re here to build it together.</p>\n<p>We&#39;re looking for an Engineering Manager, Datastores to lead the team responsible for the reliability, scalability, and evolution of Webflow’s core production databases , primarily MongoDB and PostgreSQL. This team operates at the heart of our application and hosting stack, enabling product teams to ship confidently while maintaining high standards of performance, durability, security, and data residency.</p>\n<p>Webflow’s product and hosting platform operates at a significant scale. The Datastores team sits at a critical boundary between application velocity and system durability. This is a high-leverage leadership role at the core of Webflow’s infrastructure strategy.</p>\n<p><strong>About the role:</strong></p>\n<ul>\n<li>Lead and grow a team of Database engineers responsible for MongoDB and PostgreSQL in production.</li>\n<li>Own the operational excellence of our database layer, including availability, durability, performance, cost efficiency, and data residency.</li>\n<li>Drive roadmap and strategy for multi-region architecture, backup and disaster recovery, indexing and schema governance, capacity planning, and infrastructure automation (Pulumi/Terraform).</li>\n<li>Partner with Product Engineering to guide new access patterns, review high-impact launches for database risk, and establish guardrails that enable velocity without compromising reliability.</li>\n<li>Improve reliability through proactive failure-mode detection, clear SLOs, actionable alerting, and high-quality incident response and retrospectives.</li>\n<li>Build self-service tooling and paved roads for migrations, connection management, indexing, and query best practices.</li>\n<li>Mentor and grow senior and staff engineers while contributing to broader infrastructure strategy across AWS, Kubernetes, and stateful systems architecture.</li>\n</ul>\n<p><strong>About you:</strong></p>\n<ul>\n<li>BS / BA college degree or relevant experience</li>\n<li>Business-level fluency to read, write and speak in English</li>\n<li>2+ years of experience leading high-performing engineering teams.</li>\n<li>6+ years of hands-on experience operating and scaling production databases (MongoDB and/or PostgreSQL preferred).</li>\n<li>Experience running business-critical, high-throughput systems with strong availability and durability requirements.</li>\n</ul>\n<p>You’ll thrive in this role if you:</p>\n<ul>\n<li>Bring deep expertise in operating and scaling production databases (e.g., replication, failover, indexing, query planning, migrations) and have led teams supporting stateful, multi-region systems with strict uptime requirements.</li>\n<li>Balance strong architectural judgment with pragmatism , evolving our datastore strategy while enabling product teams to ship quickly and safely.</li>\n<li>Think in terms of SLOs, capacity models, and long-term architectural trade-offs, with hands-on experience in infrastructure as code (Pulumi/Terraform), Kubernetes, and AWS.</li>\n<li>Bring strong systems-level thinking to performance and reliability, identifying root causes across application, database, and infrastructure layers and building preventative solutions.</li>\n<li>Lead calmly through high-severity incidents, drive blameless postmortems and systemic improvements, and build strong cross-functional relationships grounded in craftsmanship and continuous improvement.</li>\n<li>Stay curious and open to growth-Demonstrate a proactive embrace of AI, actively building and applying fluency in emerging technologies to elevate how we work, drive faster outcomes, and expand collective impact.</li>\n</ul>\n<p><strong>Our Core Behaviors:</strong></p>\n<ul>\n<li>Build lasting customer trust.</li>\n<li>Win together.</li>\n<li>Reinvent ourselves.</li>\n<li>Deliver with speed, quality, and craft.</li>\n</ul>\n<p><strong>Benefits:</strong></p>\n<ul>\n<li>Ownership in what you help build.</li>\n<li>Health coverage that actually covers you.</li>\n<li>Support for every stage of family life.</li>\n<li>Time off that’s actually off.</li>\n<li>Wellness for the whole you.</li>\n<li>Invest in your future.</li>\n<li>Monthly stipends that flex with your life.</li>\n<li>Bonus for building together.</li>\n</ul>\n<p><strong>Be you, with us:</strong></p>\n<p>At Webflow, equality is a core tenet of our culture. We are an Equal Opportunity (EEO)/Veterans/Disabled Employer and are committed to building an inclusive global team that represents a variety of backgrounds, perspectives, beliefs, and experiences.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_22fe5cb2-ba9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Webflow","sameAs":"https://webflow.com/","logo":"https://logos.yubhub.co/webflow.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/webflow/jobs/7648674","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["database engineering","MongoDB","PostgreSQL","infrastructure automation","Pulumi/Terraform","Kubernetes","AWS","leadership","team management","operational excellence","availability","durability","performance","cost efficiency","data residency","multi-region architecture","backup and disaster recovery","indexing and schema governance","capacity planning","self-service tooling","paved roads","migrations","connection management","query best practices"],"x-skills-preferred":[],"datePosted":"2026-03-31T18:17:49.720Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Argentina Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"database engineering, MongoDB, PostgreSQL, infrastructure automation, Pulumi/Terraform, Kubernetes, AWS, leadership, team management, operational excellence, availability, durability, performance, cost efficiency, data residency, multi-region architecture, backup and disaster recovery, indexing and schema governance, capacity planning, self-service tooling, paved roads, migrations, connection management, query best practices"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_34566519-beb"},"title":"Software Engineer III","description":"<p>Electronic Arts creates next-level entertainment experiences that inspire players and fans around the world. Here, everyone is part of the story. Part of a community that connects across the globe. A place where creativity thrives, new perspectives are invited, and ideas matter. A team where everyone makes play happen.</p>\n<p>We are looking for a Senior Software Engineer to lead our efforts in building and scaling infrastructure service for game development. This is a high-impact role focused on designing, implementing and managing scalable, reliable infrastructure solutions that power GPS tools and services used by game production teams across the company.</p>\n<p>As part of Game Production Solutions (GPS), you&#39;ll have a direct impact on empowering game developers and improving how games are built and played around the world. You&#39;ll work with talented, creative, and driven individuals who are passionate about games and technology.</p>\n<p>Key responsibilities include:\nArchitect Orchestration Tools: Assist designing and implementing a unified service for large-scale virtualization, managing provisioning, scaling and monitoring across hybrid environments (Azure/AWS/On-prem)\nAPI Development and Launch: Help drive the production launch of a new VM creation API, ensuring high availability through rigorous load testing and integration validation.\nInfrastructure as Code: Build and maintain modular IaC patterns to automate the lifecycle of compute resources at scale\nObservability and Reliability: Establish robust monitoring, logging and alerting frameworks (SLIs/SLOs) to provide deep visibility into API health and infrastructure performance\nCross-functional Leadership: Drive defect resolution and performance by collaborating with IT, Security and other partner teams.\nRelease Management: Manage phased rollouts, including lighthouse customer pilots, production deployment validation and go-live execution.\nDocumentation: Author high-quality technical specs, production runbooks and troubleshooting guides for our engineering team.</p>\n<p>Technical skills required include:\nProgramming Languages: scripting and programming languages such as Powershell, GoLang.\nInfrastructure as Code: infrastructure-as-code, configuration-as-code automation tools, such as Packer, Terraform, Pulumi, Ansible, Chef, etc.\nInfrastructure background: Extensive experience managing large-scale compute environments on-premise (vSphere, OpenShift, etc.) and in the public cloud (Azure, etc.)\nVersion Control &amp; CI/CD: Deep understanding of Git-based workflows (GitHub/GitLab) and CI/CD pipeline construction.\nContainerization: Kubernetes, Docker.\nBonus: Experience with Prometheus, Grafana, ELK, CloudBolt, SQL.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_34566519-beb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Electronic Arts","sameAs":"https://jobs.ea.com","logo":"https://logos.yubhub.co/jobs.ea.com.png"},"x-apply-url":"https://jobs.ea.com/en_US/careers/JobDetail/Software-Engineer/212286","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"temporary","x-salary-range":"$119,600 - $167,300 CAD","x-skills-required":["Powershell","GoLang","Packer","Terraform","Pulumi","Ansible","Chef","vSphere","OpenShift","Azure","Git","CI/CD","Kubernetes","Docker"],"x-skills-preferred":["Prometheus","Grafana","ELK","CloudBolt","SQL"],"datePosted":"2026-03-10T12:21:08.911Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver"}},"employmentType":"TEMPORARY","occupationalCategory":"Engineering","industry":"Technology","skills":"Powershell, GoLang, Packer, Terraform, Pulumi, Ansible, Chef, vSphere, OpenShift, Azure, Git, CI/CD, Kubernetes, Docker, Prometheus, Grafana, ELK, CloudBolt, SQL","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":119600,"maxValue":167300,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_24c98a4b-7b9"},"title":"Platform Engineer: Platform Solutions","description":"<p><strong>About the Role</strong></p>\n<p>The Platform Solutions team is dedicated to building and maintaining foundational systems that empower all product teams by providing highly reliable and scalable platform services. We focus on infrastructure-level solutions, enabling seamless functionality while abstracting away complexities for developers across the organization.</p>\n<p>As a Platform Engineer in the Platform Solutions team, you will design, develop, and optimize custom platform services to ensure our systems are resilient, secure, and performant. You&#39;ll collaborate with engineers across the organization to deliver cutting-edge infrastructure solutions that align with business needs and support the growth of our platform.</p>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Proficiency in at least one programming language such as Go (preferred), Java, Python, C# or similar</li>\n<li>Hands-on experience with AWS services (e.g., ECS, DynamoDB, S3, Lambda, SQS)</li>\n<li>Familiarity with cloud-native system design best practices</li>\n<li>Experience with infrastructure-as-code tools like CloudFormation (preferred), Terraform or Pulumi</li>\n<li>Understanding of CI/CD pipelines and tools such as GitHub Actions, or CircleCI</li>\n<li>Experience with containerization and orchestration tools (e.g., Docker, ECS)</li>\n<li>Knowledge of networking concepts, security, and distributed systems</li>\n<li>Understanding of web concepts and security, including OAuth, SAML, CORS, Cookies, and HTTP security</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Unlimited vacation time - we strongly encourage all of our employees take at least 3 weeks per year</li>\n<li>Fully remote team - choose where you live</li>\n<li>Work from home stipend! We want you to have the resources you need to set up your home office</li>\n<li>Apple laptops provided for new employees</li>\n<li>Training and development budget for every employee, refreshed each year</li>\n<li>Maternity &amp; Paternity leave for qualified employees</li>\n<li>Work with smart people who will help you grow and make a meaningful impact</li>\n<li>This position has a base salary range between $80k and $120k USD.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_24c98a4b-7b9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/16EC82BD53","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$80k - $120k USD","x-skills-required":["Go","Java","Python","C#","AWS services","CloudFormation","Terraform","Pulumi","CI/CD pipelines","GitHub Actions","CircleCI","containerization","orchestration","networking concepts","security","distributed systems","web concepts","OAuth","SAML","CORS","Cookies","HTTP security"],"x-skills-preferred":["infrastructure-as-code tools","cloud-native system design best practices"],"datePosted":"2026-03-09T10:56:46.410Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Portugal"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Java, Python, C#, AWS services, CloudFormation, Terraform, Pulumi, CI/CD pipelines, GitHub Actions, CircleCI, containerization, orchestration, networking concepts, security, distributed systems, web concepts, OAuth, SAML, CORS, Cookies, HTTP security, infrastructure-as-code tools, cloud-native system design best practices","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":120000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e9d432ac-fb7"},"title":"Implementation Engineer","description":"<p>We are seeking an Implementation Engineer to join our team in Pune. As an Implementation Engineer, you will be responsible for developing Infrastructure-as-Code using Terraform, CDK, or Pulumi, and developing CI/CD pipelines for Cloud Deployments. You will also be responsible for developing custom automation using Lambda Functions or Azure functions. Your tasks will include managing code repositories, performing peer code reviews, and maintaining code hygiene. You will also be responsible for mentoring and training junior engineers in your pod. You will work side-by-side with customers to design, diagram, and document complex integrations, and then build and deploy these designs. You will also establish partnerships and strategic relationships with contacts at our biggest brands. Time management is critical, and you should be able to manage multiple tasks and projects simultaneously. You will also be responsible for analysing and auditing existing Helpshift implementations to make improvements. You will become an expert at using Helpshift&#39;s administrative tools, which include a suite of AI products, bots, and other mission-critical support functions. You will work collaboratively with Account Managers, Customer Success Managers, and Sales teams to ensure customers&#39; overall success with the product. You will continually optimise the overall development process with improvements to documentation, trainings, and other customer-facing content.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Develop Infrastructure-as-Code using Terraform, CDK, or Pulumi</li>\n<li>Develop CI/CD pipelines for Cloud Deployments</li>\n<li>Develop custom automation using Lambda Functions or Azure functions</li>\n<li>Manage code repositories and perform peer code reviews</li>\n<li>Maintain code hygiene and write test cases for solutions</li>\n<li>Mentor and train junior engineers in your pod</li>\n<li>Work side-by-side with customers to design, diagram, and document complex integrations</li>\n<li>Establish partnerships and strategic relationships with contacts at our biggest brands</li>\n<li>Analyse and audit existing Helpshift implementations to make improvements</li>\n<li>Become an expert at using Helpshift&#39;s administrative tools</li>\n<li>Work collaboratively with Account Managers, Customer Success Managers, and Sales teams</li>\n<li>Continually optimise the overall development process</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>3 years of SaaS experience in a specialization such as consulting services, technical pre-sales, or solution engineering</li>\n<li>Proven experience translating ambiguous customer requirements into actionable technical solutions</li>\n<li>Proficiency with Python, Go, C#, Node.js, or Powershell</li>\n<li>Proficiency with deploying cloud solutions on AWS or Azure</li>\n<li>Familiarity with technical SaaS concepts such as SDKs, APIs, and cloud computing</li>\n<li>Exceptional organisational skills and a project management mindset</li>\n<li>Understanding of Object-Oriented Programming concepts</li>\n<li>Excellent communication skills and ability to lead meetings with customer executives and analysts</li>\n<li>Proficiency in G-Suite and ability to perform data analysis tasks</li>\n<li>Curiosity about complex systems and natural problem-solving skills</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Hybrid setup</li>\n<li>Worker&#39;s insurance</li>\n<li>Paid Time Offs</li>\n<li>Other employee benefits to be discussed by our Talent Acquisition team in Pune</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e9d432ac-fb7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Helpshift","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/450F76EA64","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Infrastructure-as-Code","Terraform","CDK","Pulumi","CI/CD pipelines","Cloud Deployments","custom automation","Lambda Functions","Azure functions","code repositories","peer code reviews","code hygiene","test cases","mentoring","training","junior engineers","customer success","account management","sales"],"x-skills-preferred":["Python","Go","C#","Node.js","Powershell","AWS","Azure","SDKs","APIs","cloud computing","G-Suite","data analysis"],"datePosted":"2026-03-09T10:56:00.023Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Pune"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Infrastructure-as-Code, Terraform, CDK, Pulumi, CI/CD pipelines, Cloud Deployments, custom automation, Lambda Functions, Azure functions, code repositories, peer code reviews, code hygiene, test cases, mentoring, training, junior engineers, customer success, account management, sales, Python, Go, C#, Node.js, Powershell, AWS, Azure, SDKs, APIs, cloud computing, G-Suite, data analysis"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_eebf21c4-d1f"},"title":"Staff Site Reliability Engineer","description":"<p>Join our Site Reliability Engineering (SRE) team and help ensure the reliability, scalability, and performance of Replit&#39;s infrastructure that serves millions of developers worldwide.</p>\n<p>As a Staff Site Reliability Engineer, you will bridge the gap between development and operations, implementing automation and establishing best practices that enable our platform to scale efficiently while maintaining high availability.</p>\n<p>We are seeking Staff SREs who are passionate about building and maintaining resilient systems at scale. Your mission will be to proactively find and analyze reliability problems across our stack, then design and implement software and systems to create step-function improvements.</p>\n<p>You will design robust observability solutions, lead incident response, automate operational tasks, and continuously improve our infrastructure&#39;s reliability, all while mentoring and educating the broader engineering team to make reliability a core value at Replit.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Architect and Implement Observability: Design, build, and lead the implementation of comprehensive monitoring, logging, and tracing solutions. Create dashboards and metrics that provide real-time visibility into system health and performance, enabling proactive issue detection.</li>\n</ul>\n<ul>\n<li>Define and Drive Reliability Standards: Work with product and engineering teams to define, implement, and track Service Level Objectives (SLOs) and Service Level Indicators (SLIs). Build systems to monitor and report on these metrics, holding teams accountable and ensuring we maintain high reliability standards while balancing innovation speed.</li>\n</ul>\n<ul>\n<li>Lead Incident Management and Response: Act as a senior leader during high-impact incidents, guiding the team to rapid resolution. Conduct thorough, blameless post-mortems and drive the implementation of preventative measures. Develop and refine runbooks and build automation to reduce Mean Time To Recovery (MTTR).</li>\n</ul>\n<ul>\n<li>Drive Automation and Infrastructure as Code: Architect, build, and improve automation to eliminate toil and operational work. Design and maintain CI/CD pipelines and infrastructure automation using tools like Terraform or Pulumi. Create self-healing systems that can automatically respond to common failure scenarios.</li>\n</ul>\n<ul>\n<li>Optimize Performance on Kubernetes: Collaborate with core infrastructure and product teams to performance-tune and optimize our large-scale cloud deployments, with a deep focus on Kubernetes, Docker, and GCP. Identify and resolve performance bottlenecks, implement capacity planning strategies, and reduce latency across global regions.</li>\n</ul>\n<ul>\n<li>Debug and Harden Distributed Systems: Dive deep into debugging extremely difficult technical problems across the stack. Use your findings to design and implement long-term fixes that make our systems and products more robust, operable, and easier to diagnose.</li>\n</ul>\n<ul>\n<li>Provide Staff-Level Guidance: Review feature and system designs from across the company, acting as a key owner for the reliability, scalability, security, and operational integrity of those designs.</li>\n</ul>\n<ul>\n<li>Educate and Mentor: Educate, mentor, and hold accountable the broader engineering team to improve the reliability of our systems, making reliability a core value of the Replit engineering culture.</li>\n</ul>\n<ul>\n<li>Build and Integrate: Write high-quality, well-tested code in Python or Go to meet the needs of your customers, whether it&#39;s building new internal tools or integrating with third-party vendors.</li>\n</ul>\n<p><strong>Required Skills and Experience</strong></p>\n<ul>\n<li>8-10 years of experience in Site Reliability Engineering or similar roles (e.g., DevOps, Systems Engineering, Infrastructure Engineering).</li>\n</ul>\n<ul>\n<li>Strong programming skills in languages like Python or Go. You write high-quality, well-tested code.</li>\n</ul>\n<ul>\n<li>Deep understanding of distributed systems. You’ve designed, built, scaled, and maintained production services and know how to compose a service-oriented architecture.</li>\n</ul>\n<ul>\n<li>Deep experience with container orchestration platforms, specifically Kubernetes, and cloud-native technologies.</li>\n</ul>\n<ul>\n<li>Proven track record of designing, implementing, and maintaining sophisticated monitoring and observability solutions (e.g., metrics, logging, tracing).</li>\n</ul>\n<ul>\n<li>Strong incident management skills with extensive experience leading incident response for complex systems and demonstrated critical thinking under pressure.</li>\n</ul>\n<ul>\n<li>Experience with infrastructure as code (e.g., Terraform, Pulumi) and configuration management tools.</li>\n</ul>\n<ul>\n<li>Excellent written and verbal communication skills, with an ability to explain complex technical concepts clearly and simply and a bias toward open, transparent cultural practices.</li>\n</ul>\n<ul>\n<li>Strong interpersonal skills, with experience working with and mentoring engineers from junior to principal levels.</li>\n</ul>\n<ul>\n<li>A willingness to dive into understanding, debugging, and improving any layer of the stack.</li>\n</ul>\n<ul>\n<li>You&#39;re passionate about making software creation accessible and empowering the next generation of builders.</li>\n</ul>\n<p><strong>Bonus Points</strong></p>\n<ul>\n<li>Deep experience with Google Cloud Platform (GCP) services and tools.</li>\n</ul>\n<ul>\n<li>Expert-level knowledge of modern observability platforms (e.g., Prometheus, Grafana, Datadog, OpenTelemetry).</li>\n</ul>\n<ul>\n<li>Experience designing and building reliable systems capable of handling high throughput and low latency.</li>\n</ul>\n<ul>\n<li>Significant experience with Go and Terraform.</li>\n</ul>\n<ul>\n<li>Familiarity with working in rapid-growth, startup environments.</li>\n</ul>\n<ul>\n<li>Experience writing company-facing blog posts and training materials.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_eebf21c4-d1f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Replit","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/replit.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/replit/d50ad15b-82d4-452f-b4ea-2a7f5e796170","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"Full time","x-salary-range":"$220K - $325K","x-skills-required":["Site Reliability Engineering","DevOps","Systems Engineering","Infrastructure Engineering","Python","Go","Distributed Systems","Container Orchestration","Kubernetes","Cloud-Native Technologies","Monitoring and Observability","Incident Management","Infrastructure as Code","Terraform","Pulumi","Configuration Management"],"x-skills-preferred":["Google Cloud Platform","Prometheus","Grafana","Datadog","OpenTelemetry","Go","Terraform"],"datePosted":"2026-03-08T22:20:23.639Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote (United States)"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Site Reliability Engineering, DevOps, Systems Engineering, Infrastructure Engineering, Python, Go, Distributed Systems, Container Orchestration, Kubernetes, Cloud-Native Technologies, Monitoring and Observability, Incident Management, Infrastructure as Code, Terraform, Pulumi, Configuration Management, Google Cloud Platform, Prometheus, Grafana, Datadog, OpenTelemetry, Go, Terraform","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220000,"maxValue":325000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1ace7478-7a2"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p><strong>About the role</strong></p>\n<p>Data Infrastructure designs, operates, and scales secure, privacy-respecting systems that power data-driven decisions across Anthropic. Our mission is to provide data processing, storage, and access that are trusted, fast, and easy to use.</p>\n<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability. This role offers the opportunity to work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p><strong>Responsibilities:</strong></p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li><strong>Data Governance &amp; Access Control:</strong> Design and implement robust access control systems ensuring only authorized users can access sensitive data. Build infrastructure for permission management, audit logging, and compliance requirements. Work on IAM policies, ACLs, and security controls that scale across thousands of users and systems.</li>\n</ul>\n<ul>\n<li><strong>Financial Data Infrastructure:</strong> Build and maintain data pipelines and warehouses powering business-critical reporting. Ensure data integrity, accuracy, and availability for complex financial systems, including third party revenue ingestion pipelines; manage the external relationships as needed to drive upstream dependencies. Own the reliability of systems processing revenue, usage, and business metrics.</li>\n</ul>\n<ul>\n<li><strong>Cloud Storage &amp; Reliability:</strong> Architect disaster recovery, backup, and replication systems for petabyte-scale data. Ensure high availability and durability of data stored in cloud object storage (GCS, S3). Build systems that protect against data loss and enable rapid recovery.</li>\n</ul>\n<ul>\n<li><strong>Data Platform &amp; Tooling:</strong> Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark. Optimize query performance, manage costs, and enable self-service analytics across the organization.</li>\n</ul>\n<p><strong>You might be a good fit if you:</strong></p>\n<ul>\n<li>Have 10+ years (not including internships or co-ops) of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems</li>\n</ul>\n<ul>\n<li>Have 3+ years (not including internships or co-ops) of experience leading large scale, complex projects or teams as an engineer or tech lead</li>\n</ul>\n<ul>\n<li>Can set technical direction for a team, not just execute within it</li>\n</ul>\n<ul>\n<li>Have deep experience with at least one of:</li>\n</ul>\n<ul>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar</li>\n</ul>\n<ul>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS)</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure</li>\n</ul>\n<ul>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures</li>\n</ul>\n<ul>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale</li>\n</ul>\n<ul>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks</li>\n</ul>\n<ul>\n<li>Experience working in fintech, financial services, or highly regulated environments</li>\n</ul>\n<ul>\n<li>Security engineering background with focus on data protection and access controls</li>\n</ul>\n<p><strong>Technologies We Use:</strong></p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran</li>\n</ul>\n<ul>\n<li>Storage: GCS, S3</li>\n</ul>\n<ul>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS</li>\n</ul>\n<ul>\n<li>Languages: Python, Go, SQL</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</p>\n<p><strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1ace7478-7a2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000 - $485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures","data warehousing","ETL/ELT pipelines","analytics infrastructure","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"x-skills-preferred":["data governance","access control","cloud storage","reliability","data platform","tooling","self-service analytics","data processing infrastructure","query performance","cost management"],"datePosted":"2026-03-08T13:52:03.469Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls, data governance, access control, cloud storage, reliability, data platform, tooling, self-service analytics, data processing infrastructure, query performance, cost management","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e9e336c5-ad3"},"title":"Software Engineer, Privacy Infrastructure","description":"<p><strong>Software Engineer, Privacy Infrastructure</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Location Type</strong></p>\n<p>Hybrid</p>\n<p><strong>Department</strong></p>\n<p>Security</p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$230K – $325K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>OpenAI’s Privacy Engineering team sits at the intersection of Security, Privacy, Legal, and Core Infrastructure. Our mission is to build data infrastructure and systems to support our privacy, legal, and security teams—securely, quickly, and at scale. Our guiding principles include: defensibility by default, enabling researchers, preparing for future transformative technologies, and building a robust security culture.</p>\n<p><strong>About the Role</strong></p>\n<p>We’re looking for a Software Engineer who can design and operate technical systems that support legal compliance workflows, including secure data processing and document review. You’ll partner daily with Legal, Security, IT, and partner engineering teams to turn legal processes into concrete technical workflows. This role is ideal for an engineer who loves large-scale data problems and understands the rigor required when the results may be scrutinized.</p>\n<p>This position is located in San Francisco. Relocation assistance is available.</p>\n<p><strong>In this role, you will:</strong></p>\n<ul>\n<li>Design and operate data storage pipelines that can operate at scale.</li>\n</ul>\n<ul>\n<li>Build search &amp; discovery services (e.g., Spark/Databricks, index layers, metadata catalogs) based on the needs of partner teams.</li>\n</ul>\n<ul>\n<li>Automate secure data transfers—encrypting, checksumming, and auditing exports to reviewers.</li>\n</ul>\n<ul>\n<li>Stand up locked-down compute environments that balance usability with security controls.</li>\n</ul>\n<ul>\n<li>Instrument monitoring and KPIs that maintain accountability of data holds and productions.</li>\n</ul>\n<ul>\n<li>Collaborate cross-functionally to codify SOPs, threat models, and chain-of-custody documentation that withstand scrutiny.</li>\n</ul>\n<p><strong>You might thrive in this role if you:</strong></p>\n<ul>\n<li>Have hands-on experience building or operating large-scale data-lake or backup systems (Azure, AWS, GCP).</li>\n</ul>\n<ul>\n<li>Know your way around Terraform or Pulumi, CI/CD, and can turn ad-hoc legal requests into repeatable pipelines.</li>\n</ul>\n<ul>\n<li>Comfortable working with discovery workflows (legal holds, enterprise document collections, secure review) or eager to build expertise quickly.</li>\n</ul>\n<ul>\n<li>Able to communicate technical concepts — from storage governance to block-ID APIs — clearly to teams such as Legal, Engineering, and others.</li>\n</ul>\n<ul>\n<li>Have shipped secure solutions that balance speed, cost, and evidentiary defensibility—and can articulate the trade-offs.</li>\n</ul>\n<ul>\n<li>Communicate crisply, document rigorously, and enjoy working across disciplines under tight deadlines.</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e9e336c5-ad3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/07153f7c-7e8b-4283-a879-cb07a224e083","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$230K – $325K • Offers Equity","x-skills-required":["Terraform","Pulumi","CI/CD","Spark/Databricks","index layers","metadata catalogs","Azure","AWS","GCP","large-scale data-lake or backup systems","secure data transfers","compute environments","monitoring and KPIs","SOPs","threat models","chain-of-custody documentation"],"x-skills-preferred":["hands-on experience building or operating large-scale data-lake or backup systems","comfortable working with discovery workflows","able to communicate technical concepts clearly to teams such as Legal, Engineering, and others","have shipped secure solutions that balance speed, cost, and evidentiary defensibility"],"datePosted":"2026-03-06T18:29:40.108Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Terraform, Pulumi, CI/CD, Spark/Databricks, index layers, metadata catalogs, Azure, AWS, GCP, large-scale data-lake or backup systems, secure data transfers, compute environments, monitoring and KPIs, SOPs, threat models, chain-of-custody documentation, hands-on experience building or operating large-scale data-lake or backup systems, comfortable working with discovery workflows, able to communicate technical concepts clearly to teams such as Legal, Engineering, and others, have shipped secure solutions that balance speed, cost, and evidentiary defensibility","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":230000,"maxValue":325000,"unitText":"YEAR"}}}]}