{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/major-cloud-provider"},"x-facet":{"type":"skill","slug":"major-cloud-provider","display":"Major Cloud Provider","count":3},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8bd53be2-6cf"},"title":"Senior Site Reliability Data Engineer","description":"<p>For over 31,000 growing businesses and HR teams seeking a comprehensive, all-in-one HR suite, Workable emerges as the premier solution. We uniquely combine the world’s most widely adopted Applicant Tracking System (Workable Recruiting) with a full-spectrum employee management system (Workable HR).</p>\n<p>At Workable, we empower companies to focus on what truly matters: hiring the right people and fostering their growth. While we take HR seriously, we maintain a lighthearted and collaborative culture. At Workable, you’ll find smart people who have fun, learn, innovate, and help others do the same.</p>\n<p>We respect everyone, we hire the best, and make sure every experience is special.</p>\n<p>As a Senior Site Reliability Data Engineer based in Athens, Greece, you will play a critical role in ensuring the reliability, scalability, and performance of Workable&#39;s data and cloud infrastructure. This is a high-impact position where your expertise will directly influence the operational excellence and growth of our data platform.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Design, build, and maintain core data engineering infrastructure including ETL/ELT pipelines, Apache Spark workloads, and data warehouse systems.</li>\n</ul>\n<ul>\n<li>Ensure availability, scalability, and performance of data infrastructure and pipelines with deep operational ownership.</li>\n</ul>\n<ul>\n<li>Design, implement, and maintain scalable reliability tooling and automation to streamline deployment, monitoring, and incident response across distributed services.</li>\n</ul>\n<ul>\n<li>Operate and optimize Kubernetes-based cloud infrastructure to ensure high availability, performance, and cost-efficiency.</li>\n</ul>\n<ul>\n<li>Partner cross-functionally with developers and analysts to design, release, and troubleshoot production systems; provide data engineering expertise.</li>\n</ul>\n<ul>\n<li>Lead cross-functional projects with development teams to improve system reliability, automate capacity planning, and enforce SRE best practices.</li>\n</ul>\n<ul>\n<li>Develop and maintain centralized observability, including logging, metrics, tracing, and alerting pipelines; continuously improve incident detection and response workflows.</li>\n</ul>\n<ul>\n<li>Own observability for data pipelines (freshness, completeness, latency, error rates) and ensure SLOs are met.</li>\n</ul>\n<ul>\n<li>Plan platform growth and manage capacity for the data platform and related infrastructure.</li>\n</ul>\n<ul>\n<li>Operate, deploy, and monitor data platform components and broader cloud services from development through production.</li>\n</ul>\n<ul>\n<li>Develop tools and automation to simplify data operations and make deployments more robust and self-service.</li>\n</ul>\n<ul>\n<li>Collaborate with peer SREs to roll out production changes and mitigate data/infrastructure incidents.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8bd53be2-6cf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Workable"},"x-apply-url":"https://apply.workable.com/j/22CEAF6027","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Apache Spark","ETL/ELT pipelines","cloud data warehouses","major cloud provider","infrastructure automation tools","centralized logging","monitoring","observability frameworks"],"x-skills-preferred":["production experience with Kubernetes","streaming systems","data quality","data observability tooling","relational and NoSQL databases","proficiency in programming languages","deep knowledge of Linux systems"],"datePosted":"2026-04-24T14:13:36.167Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Athens"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Apache Spark, ETL/ELT pipelines, cloud data warehouses, major cloud provider, infrastructure automation tools, centralized logging, monitoring, observability frameworks, production experience with Kubernetes, streaming systems, data quality, data observability tooling, relational and NoSQL databases, proficiency in programming languages, deep knowledge of Linux systems"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ed4bd662-c67"},"title":"Senior Solutions Architect, Commercial - San Francisco","description":"<p>We are looking for a Senior Solutions Architect to support our Commercial Sales team in a consumption-based business where customer success drives revenue growth. You&#39;ll work across the full sales cycle, from initial technical evaluations with new prospects through helping existing customers expand their use of Temporal in production.</p>\n<p>The nature of our business means you&#39;ll spend significant time helping customers who&#39;ve already adopted Temporal unlock more value by expanding into additional use cases, teams, and workloads. This is a high-velocity, technically deep role.</p>\n<p>You&#39;ll partner with developers, architects, and engineering leaders at fast-moving companies to help them understand how Temporal fits into their existing architecture and prove out value through hands-on technical work.</p>\n<p>You&#39;ll be working in a consumption model where usage grows over time, which means building strong technical relationships and staying engaged with accounts as they scale.</p>\n<p>As an early member of a growing team, you should be comfortable with ambiguity, frequent context switching, and creating leverage through reusable assets that help the broader team move faster.</p>\n<p>Must reside in San Francisco, CA</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ed4bd662-c67","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Temporal","sameAs":"https://temporal.io/","logo":"https://logos.yubhub.co/temporal.io.png"},"x-apply-url":"https://job-boards.greenhouse.io/temporaltechnologies/jobs/5037692007","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$200,000 - $250,000 OTE","x-skills-required":["Strong development background with hands-on coding experience in at least one modern language (Go, Java, TypeScript, or Python)","Deep understanding of distributed systems (reliability, observability, and fault tolerance)","Proven experience in a pre-sales, customer-facing engineering, or solutions architecture role working with technical buyers","Exceptional time management and prioritization skills with the ability to thrive in high-volume environments","Enthusiasm for AI/ML technologies and eagerness to learn about emerging use cases in agentic workflows and LLM orchestration"],"x-skills-preferred":["Experience with workflow engines, event-driven architectures, or orchestration technologies (Temporal, Cadence, or similar)","Background articulating the value of commercial SaaS offerings that compete with open source alternatives (Redis, Kafka, Databricks, etc.)","Contributions to developer tooling, open source projects, or technical content","Strong cross-functional collaboration skills with the ability to serve as a technical bridge between customers and internal teams","Certifications with any of the major cloud providers (AWS, GCP, or Azure) or foundational AI model providers (OpenAI, Anthropic, or Google)"],"datePosted":"2026-04-18T15:56:33.427Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States - Remote Opportunity"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Strong development background with hands-on coding experience in at least one modern language (Go, Java, TypeScript, or Python), Deep understanding of distributed systems (reliability, observability, and fault tolerance), Proven experience in a pre-sales, customer-facing engineering, or solutions architecture role working with technical buyers, Exceptional time management and prioritization skills with the ability to thrive in high-volume environments, Enthusiasm for AI/ML technologies and eagerness to learn about emerging use cases in agentic workflows and LLM orchestration, Experience with workflow engines, event-driven architectures, or orchestration technologies (Temporal, Cadence, or similar), Background articulating the value of commercial SaaS offerings that compete with open source alternatives (Redis, Kafka, Databricks, etc.), Contributions to developer tooling, open source projects, or technical content, Strong cross-functional collaboration skills with the ability to serve as a technical bridge between customers and internal teams, Certifications with any of the major cloud providers (AWS, GCP, or Azure) or foundational AI model providers (OpenAI, Anthropic, or Google)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":200000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_58d220e6-02a"},"title":"Senior Site Reliability Engineer, Tenant Services: Geo","description":"<p>Job Title: Senior Site Reliability Engineer, Tenant Services: Geo</p>\n<p>We are looking for a skilled Senior Site Reliability Engineer to join our Tenant Services, Geo team. As a Senior Site Reliability Engineer, you will be responsible for ensuring the smooth operation of our user-facing services and production systems.</p>\n<p>About Us</p>\n<p>GitLab is the intelligent orchestration platform for DevSecOps. It enables organisations to increase developer productivity, improve operational efficiency, reduce security and compliance risk, and accelerate digital transformation.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Execute Dedicated Geo migrations and cutovers end-to-end, including planning, pre-cutover validation, execution, and post-cutover verification and cleanup.</li>\n<li>Join the team&#39;s shift and weekend coverage rotation for Dedicated cutovers across EMEA and US hours, and participate in the SaaS Site Reliability Engineering (SRE) on-call rotation to respond to incidents that impact GitLab.com availability.</li>\n<li>Operate and improve the Geo operational surface for Dedicated, including:</li>\n<li>Environment preparation and data hygiene checks prior to migrations.</li>\n<li>Execution of replication, validation, and cutover procedures.</li>\n<li>Handling Geo-related escalations from Support and internal partners.</li>\n<li>Design, build, and maintain automation, tooling, and runbooks that make migrations, cutovers, and Geo escalations as &#39;boring&#39; and repeatable as possible.</li>\n<li>Run our infrastructure with tools such as Ansible, Chef, Terraform, GitLab CI/CD, and Kubernetes; contribute improvements back to GitLab&#39;s product and infrastructure where appropriate.</li>\n<li>Build and maintain monitoring, alerting, and dashboards that:</li>\n<li>Detect symptoms early, not just outages.</li>\n<li>Track migration and cutover success rates, duration, rollback frequency, and related SLOs.</li>\n<li>Collaborate closely with:</li>\n<li>The core Geo team on improving Geo features and operability.</li>\n<li>Dedicated migrations and Support on migration planning, customer communications, and escalation handling.</li>\n<li>Other Infrastructure teams on capacity planning, disaster recovery, and reliability improvements.</li>\n<li>Contribute to readiness reviews, incident reviews, and root cause analyses, turning learnings into changes in automation, process, or product.</li>\n<li>Document every action, including runbooks, architecture decisions, and post-incident reviews, so your findings turn into repeatable practices and automation.</li>\n<li>Proactively identify and reduce toil by automating repetitive operational work and simplifying migration workflows.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>Experience operating highly-available distributed systems at scale, ideally in a SaaS environment with customer-facing SLAs.</li>\n<li>Hands-on experience with at least one major cloud provider (e.g., Google Cloud Platform or Amazon Web Services), including networking, storage, and managed services.</li>\n<li>Experience with Kubernetes and its ecosystem (e.g., Helm), including deploying and troubleshooting workloads.</li>\n<li>Experience with infrastructure as code and configuration management tools such as Terraform, Ansible, or Chef.</li>\n<li>Strong programming skills in at least one general-purpose language (preferably Go or Ruby) and proficiency with scripting (e.g., Shell, Python).</li>\n<li>Experience with observability systems (e.g., Prometheus, Grafana, logging stacks) and using metrics and logs to troubleshoot performance and reliability issues.</li>\n<li>Practical exposure to data replication, backup/restore, or migration scenarios (e.g., database replication, storage replication, or Geo-like technologies) where data integrity and downtime risk must be carefully managed.</li>\n<li>Comfort participating in an on-call rotation, investigating incidents across the stack, and driving follow-through on corrective actions.</li>\n<li>Ability to engage directly with enterprise customers during migrations and incidents, including on live calls and through clear written updates.</li>\n<li>Ability to clearly define problems, propose options, and think beyond immediate fixes to improve systems and processes over time.</li>\n<li>Ability to be a &#39;manager of one&#39;: self-directed, organized, and able to drive work to completion in a remote, asynchronous environment.</li>\n<li>Strong written and verbal communication skills, with a bias toward clear, asynchronous documentation and collaboration.</li>\n<li>Alignment with our company values and a commitment to working in accordance with those values.</li>\n</ul>\n<p>Nice to Have</p>\n<ul>\n<li>Experience working with disaster recovery technologies.</li>\n<li>Experience with managed/hosted environments similar to GitLab Dedicated, including regulated or compliance-sensitive customers (e.g., SOC2, ISO).</li>\n<li>Prior work on large-scale data migrations or cutovers where customer data integrity, performance, and downtime risk had to be carefully balanced.</li>\n<li>Hands-on experience designing and operating database replication, backup/restore, and cutover workflows (for example, PostgreSQL or cloud-managed equivalents such as AWS RDS), including planning and executing low-risk migrations for large datasets.</li>\n<li>Experience with multi-tenant architectures, sharding, or routing strategies in high-traffic SaaS platforms.</li>\n<li>Familiarity with GitLab (self-managed or SaaS), and/or contributions to open source projects.</li>\n</ul>\n<p>Benefits</p>\n<ul>\n<li>Benefits to support your health, finances, and well-being</li>\n<li>Flexible Paid Time Off</li>\n<li>Team Member Resource Groups</li>\n<li>Equity Compensation &amp; Employee Stock Purchase Plan</li>\n<li>Growth and Development Fund</li>\n<li>Parental leave</li>\n<li>Home office support</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_58d220e6-02a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"GitLab","sameAs":"https://about.gitlab.com/","logo":"https://logos.yubhub.co/about.gitlab.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gitlab/jobs/8490453002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Experience operating highly-available distributed systems at scale","Hands-on experience with at least one major cloud provider","Experience with Kubernetes and its ecosystem","Experience with infrastructure as code and configuration management tools","Strong programming skills in at least one general-purpose language"],"x-skills-preferred":["Experience working with disaster recovery technologies","Experience with managed/hosted environments similar to GitLab Dedicated","Prior work on large-scale data migrations or cutovers","Hands-on experience designing and operating database replication, backup/restore, and cutover workflows","Experience with multi-tenant architectures, sharding, or routing strategies in high-traffic SaaS platforms"],"datePosted":"2026-04-18T15:51:05.184Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, India"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Experience operating highly-available distributed systems at scale, Hands-on experience with at least one major cloud provider, Experience with Kubernetes and its ecosystem, Experience with infrastructure as code and configuration management tools, Strong programming skills in at least one general-purpose language, Experience working with disaster recovery technologies, Experience with managed/hosted environments similar to GitLab Dedicated, Prior work on large-scale data migrations or cutovers, Hands-on experience designing and operating database replication, backup/restore, and cutover workflows, Experience with multi-tenant architectures, sharding, or routing strategies in high-traffic SaaS platforms"}]}