{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/queue"},"x-facet":{"type":"skill","slug":"queue","display":"Queue","count":35},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_90cf3dbc-2b3"},"title":"FBS Network Engineer","description":"<p>FBS – Farmer Business Services is part of Farmers operations with the purpose of building a global approach to identifying, recruiting, hiring, and retaining top talent. We believe that the foundation of every successful business lies in having the right people with the right skills. As a Network Engineer, you would support for any/all project activities within network and telecom, including gathering customer requirements and providing pricing support for needed requirements. This includes the supporting the design, development, documentation, implementation and operation of the network and telecom systems, including prototypes into the production environment.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Ensuring the business logic and call flows remain intact throughout the migration.</li>\n<li>Managing the carrier porting, SIP trunking, and infrastructure configuration.</li>\n<li>Support the collection and management of customer requirements (RFS), including attending customer pre-project discussions to represent network and telecom.</li>\n<li>Support the process of developing ROMs (project estimates), defend the ROM and solicit formal ROM approvals</li>\n<li>Support the direct implementation of small network and/or telecom projects, including the development and management of project plans</li>\n<li>Assist in the on-boarding of external resources to support larger projects or to support a large number of projects. In addition, provide oversight to the 3rd party engineering resources to assist with the implementation of the project(s).</li>\n<li>Provide monthly reporting data for all project activities. This role will act as a mentor to lower roles, and support a continuous process improvement activities.</li>\n<li>Working with their manager and/or director, this role will help provide individual and team goals to the management to be used in the annual performance management tool, Additional responsibilities include assisting in the training of new hires.</li>\n<li>Support the design, development, documentation, implementation and operation of the network and telecom systems, including prototypes into the production environment Project management</li>\n<li>Understands basics of project management, timelines and critical path Budget awareness</li>\n<li>Understanding of project budgets including understanding of basic finance treatments and timings Dependability and Time Management</li>\n<li>Completes assigned tasks on time and predictably.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_90cf3dbc-2b3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://www.capgemini.com/","logo":"https://logos.yubhub.co/capgemini.com.png"},"x-apply-url":"https://jobs.workable.com/view/pRd2TCTkwU2dEo2ipYoB7J/remote-fbs-network-engineer-in-brazil-at-capgemini","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Telecom system experience","Advanced Scripting: Expert-level knowledge of Cisco Unified Contact Center Enterprise (UCCE) scripting and call studio.","Call Flow Logic: Ability to reconfigure and validate complex call flows, IVR menus, and queueing logic to align with new Bandwidth entry points.","Platform Knowledge: Strong grasp of Finesse, CUIC (reporting), and CVP (Voice Portal).","Post-Migration Support: Experience providing Tier 3 support for contact center agents and managing 'Day 2' optimization based on business feedback.","Carrier & Porting Expertise: Proven experience managing large-scale Toll-Free and DID migrations (preferable to Bandwidth).","SIP & Gateway Management: Deep knowledge of Cisco VCube (CUBE) setup, dial-peer configuration, and SIP normalization.","Infrastructure: Proficiency in Cisco Unified Communications Manager (CUCM) and Cisco Unified Border Element."],"x-skills-preferred":[],"datePosted":"2026-04-24T14:18:29.546Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Brazil"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Telecom system experience, Advanced Scripting: Expert-level knowledge of Cisco Unified Contact Center Enterprise (UCCE) scripting and call studio., Call Flow Logic: Ability to reconfigure and validate complex call flows, IVR menus, and queueing logic to align with new Bandwidth entry points., Platform Knowledge: Strong grasp of Finesse, CUIC (reporting), and CVP (Voice Portal)., Post-Migration Support: Experience providing Tier 3 support for contact center agents and managing 'Day 2' optimization based on business feedback., Carrier & Porting Expertise: Proven experience managing large-scale Toll-Free and DID migrations (preferable to Bandwidth)., SIP & Gateway Management: Deep knowledge of Cisco VCube (CUBE) setup, dial-peer configuration, and SIP normalization., Infrastructure: Proficiency in Cisco Unified Communications Manager (CUCM) and Cisco Unified Border Element."},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_85f879d7-d2d"},"title":"MS Dynamics CRM Consultant","description":"<p>We are seeking an experienced MS Dynamics CRM (Dynamics 365 CE) Consultant with strong hands-on expertise in Contact Center, Omnichannel or Customer Service, and/or Unified Routing. The ideal candidate will drive solution design, configuration, integration, and end-to-end testing of customer service and interaction management features within Microsoft Dynamics 365.</p>\n<p>Key Responsibilities:</p>\n<p>Configure and customize Dynamics 365 Customer Engagement modules. Implement and support Customer Service, Case Management, Queues, SLAs, Routing Rules, and Workflows. Build UI customizations, business rules, forms, and entities based on business requirements.</p>\n<p>Configure Omnichannel capabilities, including Live chat, Voice channel, SMS &amp; WhatsApp (if applicable), Social channels, Agent experience profiles, Workstreams and conversation flows. Set up agent scripts, macros, smart assist, and session templates.</p>\n<p>Configure Unified Routing for automatic work item assignment. Set up routing rules, classification models, queues, and capacity profiles. Enable and optimize work distribution models across channels.</p>\n<p>Work with telephony integration platforms such as Azure Communication Services, Webchat/Voice channels, CTI tool integrations. Support the implementation of intelligent call routing and agent desktop features.</p>\n<p>Gather requirements and translate business needs into CRM functional solutions. Prepare solution documentation, configuration guides, and user documentation. Participate in design reviews, impact analysis, and architectural discussions.</p>\n<p>Create and execute unit tests, integration tests, and UAT support. Assist in deployment activities across multiple environments. Ensure solution stability, performance, and adherence to best practices.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_85f879d7-d2d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://www.capgemini.com/us-en/about-us/who-we-are/","logo":"https://logos.yubhub.co/capgemini.com.png"},"x-apply-url":"https://jobs.workable.com/view/mmune3oe8rZNKH3FD9BMuP/hybrid-ms-dynamics-crm-consultant-in-kolkata-at-capgemini","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MS Dynamics CRM","Dynamics 365 CE","Contact Center","Omnichannel","Unified Routing","Customer Service","Case Management","Queues","SLAs","Routing Rules","Workflows","UI customizations","Business rules","Forms","Entities","Agent experience profiles","Workstreams","Conversation flows","Agent scripts","Macros","Smart assist","Session templates","Telephony integration platforms","Azure Communication Services","Webchat/Voice channels","CTI tool integrations","Intelligent call routing","Agent desktop features","Solution design","Configuration","Integration","End-to-end testing","Unit tests","Integration tests","UAT support","Deployment activities","Solution stability","Performance","Best practices"],"x-skills-preferred":[],"datePosted":"2026-04-24T14:17:06.624Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Kolkata"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Technology","skills":"MS Dynamics CRM, Dynamics 365 CE, Contact Center, Omnichannel, Unified Routing, Customer Service, Case Management, Queues, SLAs, Routing Rules, Workflows, UI customizations, Business rules, Forms, Entities, Agent experience profiles, Workstreams, Conversation flows, Agent scripts, Macros, Smart assist, Session templates, Telephony integration platforms, Azure Communication Services, Webchat/Voice channels, CTI tool integrations, Intelligent call routing, Agent desktop features, Solution design, Configuration, Integration, End-to-end testing, Unit tests, Integration tests, UAT support, Deployment activities, Solution stability, Performance, Best practices"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dfc341df-398"},"title":"MS Dynamics CRM Consultant","description":"<p>We are seeking an experienced MS Dynamics CRM (Dynamics 365 CE) Consultant with strong hands-on expertise in Contact Center, Omnichannel OR Customer Service, and/or Unified Routing. The ideal candidate will drive solution design, configuration, integration, and end-to-end testing of customer service and interaction management features within Microsoft Dynamics 365.</p>\n<p><strong>Key Responsibilities</strong></p>\n<p>Configure and customize Dynamics 365 Customer Engagement modules. Implement and support Customer Service, Case Management, Queues, SLAs, Routing Rules, and Workflows. Build UI customizations, business rules, forms, and entities based on business requirements.</p>\n<p>Configure Omnichannel capabilities, including: Live chat Voice channel SMS &amp; WhatsApp (if applicable) Social channels Agent experience profiles Workstreams and conversation flows Set up agent scripts, macros, smart assist, and session templates.</p>\n<p>Configure Unified Routing for automatic work item assignment. Set up routing rules, classification models, queues, and capacity profiles. Enable and optimize work distribution models across channels.</p>\n<p>Work with telephony integration platforms such as: Azure Communication Services Webchat/Voice channels CTI tool integrations Support the implementation of intelligent call routing and agent desktop features.</p>\n<p>Gather requirements and translate business needs into CRM functional solutions. Prepare solution documentation, configuration guides, and user documentation. Participate in design reviews, impact analysis, and architectural discussions.</p>\n<p>Create and execute unit tests, integration tests, and UAT support. Assist in deployment activities across multiple environments. Ensure solution stability, performance, and adherence to best practices.</p>\n<p><strong>Requirements</strong></p>\n<p>6–12 years of experience in MS Dynamics CRM / Dynamics 365 CE. Strong experience in Omnichannel for Customer Service, Contact Center, or Unified Routing Plugins and Custom Workflow Activities (C#) – optional but beneficial.</p>\n<p><strong>Benefits</strong></p>\n<p>Competitive compensation and benefits package:</p>\n<ol>\n<li>Competitive salary and performance-based bonuses</li>\n<li>Comprehensive benefits package</li>\n<li>Career development and training opportunities</li>\n<li>Flexible work arrangements (remote and/or office-based)</li>\n<li>Dynamic and inclusive work culture within a globally renowned group</li>\n<li>Private Health Insurance</li>\n<li>Pension Plan</li>\n<li>Paid Time Off</li>\n<li>Training &amp; Development</li>\n</ol>\n<p>Note: Benefits differ based on employee level.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dfc341df-398","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://www.capgemini.com/us-en/about-us/who-we-are/","logo":"https://logos.yubhub.co/capgemini.com.png"},"x-apply-url":"https://jobs.workable.com/view/ojUycfQkVEMQcai7DsxQHp/hybrid-ms-dynamics-crm-consultant-in-hyderabad-at-capgemini","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MS Dynamics CRM","Dynamics 365 CE","Contact Center","Omnichannel","Unified Routing","Customer Service","Case Management","Queues","SLAs","Routing Rules","Workflows","UI customizations","Business rules","Forms","Entities","Telephony integration platforms","Azure Communication Services","Webchat/Voice channels","CTI tool integrations","Intelligent call routing","Agent desktop features"],"x-skills-preferred":[],"datePosted":"2026-04-24T14:16:08.731Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hyderabad"}},"employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Technology","skills":"MS Dynamics CRM, Dynamics 365 CE, Contact Center, Omnichannel, Unified Routing, Customer Service, Case Management, Queues, SLAs, Routing Rules, Workflows, UI customizations, Business rules, Forms, Entities, Telephony integration platforms, Azure Communication Services, Webchat/Voice channels, CTI tool integrations, Intelligent call routing, Agent desktop features"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9a2a83a9-dc8"},"title":"MS Dynamics CRM Consultant","description":"<p>We are seeking an experienced MS Dynamics CRM (Dynamics 365 CE) Consultant with strong hands-on expertise in Contact Center, Omnichannel OR Customer Service, and/or Unified Routing. The ideal candidate will drive solution design, configuration, integration, and end-to-end testing of customer service and interaction management features within Microsoft Dynamics 365.</p>\n<p>Key Responsibilities:</p>\n<p>Configure and customize Dynamics 365 Customer Engagement modules. Implement and support Customer Service, Case Management, Queues, SLAs, Routing Rules, and Workflows. Build UI customizations, business rules, forms, and entities based on business requirements.</p>\n<p>Configure Omnichannel capabilities, including: Live chat Voice channel SMS &amp; WhatsApp (if applicable) Social channels Agent experience profiles Workstreams and conversation flows Set up agent scripts, macros, smart assist, and session templates.</p>\n<p>Configure Unified Routing for automatic work item assignment. Set up routing rules, classification models, queues, and capacity profiles. Enable and optimize work distribution models across channels.</p>\n<p>Work with telephony integration platforms such as: Azure Communication Services Webchat/Voice channels CTI tool integrations Support the implementation of intelligent call routing and agent desktop features.</p>\n<p>Gather requirements and translate business needs into CRM functional solutions. Prepare solution documentation, configuration guides, and user documentation. Participate in design reviews, impact analysis, and architectural discussions.</p>\n<p>Create and execute unit tests, integration tests, and UAT support. Assist in deployment activities across multiple environments. Ensure solution stability, performance, and adherence to best practices.</p>\n<p>Required Experience: 6–12 years of experience in MS Dynamics CRM / Dynamics 365 CE. Strong experience in Omnichannel for Customer Service, Contact Center, or Unified Routing</p>\n<p>Good to have: Hands-on experience with Power Platform (Power Automate, Power Apps). Knowledge of Chatbots / Virtual Agents (Power Virtual Agents) Exposure to Azure components: Functions, ACS, Event Hub, etc</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9a2a83a9-dc8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://www.capgemini.com/us-en/about-us/who-we-are/","logo":"https://logos.yubhub.co/capgemini.com.png"},"x-apply-url":"https://jobs.workable.com/view/3k6qdMzZ65fJK3oGLnZEfo/hybrid-ms-dynamics-crm-consultant-in-bengaluru-at-capgemini","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MS Dynamics CRM","Dynamics 365 CE","Contact Center","Omnichannel","Unified Routing","Customer Service","Case Management","Queues","SLAs","Routing Rules","Workflows","UI customizations","Business rules","Forms","Entities","Telephony integration","Azure Communication Services","Webchat","Voice channels","CTI tool integrations","Intelligent call routing","Agent desktop features","Solution design","Configuration","Integration","Testing","Deployment","Solution stability","Performance","Best practices"],"x-skills-preferred":[],"datePosted":"2026-04-24T14:14:42.601Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"MS Dynamics CRM, Dynamics 365 CE, Contact Center, Omnichannel, Unified Routing, Customer Service, Case Management, Queues, SLAs, Routing Rules, Workflows, UI customizations, Business rules, Forms, Entities, Telephony integration, Azure Communication Services, Webchat, Voice channels, CTI tool integrations, Intelligent call routing, Agent desktop features, Solution design, Configuration, Integration, Testing, Deployment, Solution stability, Performance, Best practices"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b8361772-263"},"title":"Software Engineer","description":"<p>In this role, you will join the Ford Pro Intelligence (FPI) Telematics team as a Software Engineer. The team creates back-end services and APIs that help customers understand, manage, and control their fleets of vehicles via web, mobile, and API applications.</p>\n<p>Your primary responsibilities will include:</p>\n<ul>\n<li>Participating in and/or leading the development of requirements, features, user stories, use cases, and test cases.</li>\n<li>Authoring process, technical design, and support documents.</li>\n<li>Collaborating with the broader FPI Telematics team on solution designs, development, and deployment.</li>\n<li>Participating and/or leading incident, problem, change, and service request-related activities, including root cause analysis (RCA).</li>\n</ul>\n<p>You will work on delivering products that include Spring/Cloud services that support processing and storing telematics information while providing a secure set of APIs accessible to customers.</p>\n<p>As a Software Engineer, you will have the opportunity to work on a wide range of projects and contribute to the growth and success of the FPI Telematics team.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b8361772-263","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Ford Motor Company","sameAs":"https://www.ford.com/","logo":"https://logos.yubhub.co/ford.com.png"},"x-apply-url":"https://efds.fa.em5.oraclecloud.com/hcmUI/CandidateExperience/en/sites/CX_1/job/62744","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"A range of salary grades 6-8.","x-skills-required":["Java","Springboot","Kotlin","Node.js","GCP","AWS","Azure","serverless functions","databases","messaging queues","caching systems","relational databases","SQL like PostgreSQL"],"x-skills-preferred":[],"datePosted":"2026-04-24T12:21:05.191Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"Java, Springboot, Kotlin, Node.js, GCP, AWS, Azure, serverless functions, databases, messaging queues, caching systems, relational databases, SQL like PostgreSQL"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_68a62835-66b"},"title":"Senior DevOps Engineer","description":"<p>We are seeking a highly skilled and self-motivated Senior Embedded DevOps Engineer to support our engineering teams. This role will focus on driving changes and ensuring adherence to company-established standards for data infrastructure and CI/CD pipelines.</p>\n<p>The ideal candidate will have strong experience working with AWS and/or GCP, cloud-based data streaming and processing services, containerized application deployments, infrastructure automation, and Site Reliability Engineering (SRE) best practices for performance and cost optimization.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Drive initiatives to implement and enforce best practices for data streaming, processing, analytics and monitoring infrastructure.</li>\n<li>Deploy and manage services on Kubernetes-based platforms such as Amazon EKS and Google Kubernetes Engine (GKE).</li>\n<li>Provision and manage cloud infrastructure using Terraform, ensuring best practices in security, scalability, and cost-efficiency.</li>\n<li>Maintain and optimize CI/CD pipelines using Jenkins, ArgoCD, and GitHub Enterprise Actions to support automated deployments and testing.</li>\n<li>Work with cloud-native data services such as AWS Kinesis, AWS Glue, Google Dataflow, and Google Pub/Sub, BigQuery, BigTable</li>\n<li>Familiarity with workflow orchestration services such as Apache Airflow and Google Cloud Composer.</li>\n<li>Develop and maintain automation scripts and tooling using Python to support DevOps processes.</li>\n<li>Monitor system performance, troubleshoot issues, and implement proactive solutions to enhance reliability and efficiency.</li>\n<li>Implement SRE practices to improve service reliability, scalability, and cost-effectiveness.</li>\n<li>Analyze and optimize cloud costs, identifying areas for improvement and implementing cost-saving strategies.</li>\n<li>Ensure compliance with security policies and best practices in cloud environments.</li>\n<li>Drive adoption of company standards and influence data teams to align with best DevOps and SRE practices.</li>\n<li>Collaborate with cross-functional teams to improve development workflows and infrastructure.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>7+ years of experience in a DevOps, Site Reliability Engineering, or Cloud Infrastructure role.</li>\n<li>Strong experience with AWS and GCP data services, including Kinesis, Glue, Pub/Sub, and Dataflow.</li>\n<li>Proficiency in deploying and managing workloads on Kubernetes (EKS/GKE) in production environments.</li>\n<li>Hands-on experience with Infrastructure-as-Code (IaC) using Terraform.</li>\n<li>Expertise in CI/CD pipeline management using Jenkins, ArgoCD, and GitHub Enterprise Actions.</li>\n<li>Programming skills in Python for automation and scripting.</li>\n<li>Experience with observability and monitoring tools (e.g., Prometheus, Grafana, Datadog, or CloudWatch).</li>\n<li>Strong understanding of SRE principles, including performance monitoring, incident response, and reliability engineering.</li>\n<li>Experience with cost optimization strategies for cloud infrastructure.</li>\n<li>Self-motivated and driven, with a strong ability to influence and drive changes across multiple teams.</li>\n<li>Ability to work collaboratively in an agile environment and support multiple teams.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with data lake architectures and big data processing frameworks (e.g., Apache Spark, Flink, Snowflake, BigQuery).</li>\n<li>Familiarity with event-driven architectures and message queues (e.g., Kafka, RabbitMQ).</li>\n<li>Experience with workflow orchestration tools such as Apache Airflow and Google Cloud Composer.</li>\n<li>Knowledge of service mesh technologies like Istio.</li>\n<li>Experience with GitOps workflows and Kubernetes-native tooling.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_68a62835-66b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8496473002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["AWS","GCP","Kubernetes","Terraform","Jenkins","ArgoCD","GitHub Enterprise Actions","Python","Apache Airflow","Google Cloud Composer","CloudWatch","Prometheus","Grafana","Datadog"],"x-skills-preferred":["Data lake architectures","Big data processing frameworks","Event-driven architectures","Message queues","Workflow orchestration tools","Service mesh technologies","GitOps workflows","Kubernetes-native tooling"],"datePosted":"2026-04-24T12:19:32.227Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Toronto, Ontario, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"AWS, GCP, Kubernetes, Terraform, Jenkins, ArgoCD, GitHub Enterprise Actions, Python, Apache Airflow, Google Cloud Composer, CloudWatch, Prometheus, Grafana, Datadog, Data lake architectures, Big data processing frameworks, Event-driven architectures, Message queues, Workflow orchestration tools, Service mesh technologies, GitOps workflows, Kubernetes-native tooling"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dfe3062f-992"},"title":"Staff Backend Software Engineer — Privileged Access Management (PAM)","description":"<p>Secure Every Identity, from AI to Human</p>\n<p>Identity is the key to unlocking the potential of AI. Okta secures AI by building the trusted, neutral infrastructure that enables organisations to safely embrace this new era.</p>\n<p>We are looking for builders and owners who operate with speed and urgency and execute with excellence. This is an opportunity to do career-defining work. We&#39;re all in on this mission. If you are too, let&#39;s talk.</p>\n<p>Backend Software Engineer , Okta Privileged Access Management (PAM)</p>\n<p>Ever wonder how large organisations make sure the right people can access their most critical systems? That&#39;s the problem we solve. Our team builds the infrastructure that controls who can reach sensitive servers, databases, and cloud resources, grants access only when it&#39;s needed. It is the security layer between people (and non-human-interfaces) and the systems they need to do their jobs.</p>\n<p>We&#39;re looking for a Backend Software Engineer who wants to work on hard problems: distributed systems and building software where getting it right really matters. You&#39;ll ship code that protects real infrastructure for real organisations. You’ll build foundations that multiple feature teams depend on. When you make something faster, more reliable, or easier to use, it multiplies across the entire product.</p>\n<p>This is a role for someone who likes thinking about how systems fit together. You&#39;ll need strong opinions about what makes a good abstraction, and the flexibility to evolve those abstractions as the product grows.</p>\n<p>The Okta Privileged Access Management (PAM) is an identity-centric approach to a common and critical privileged access use case. Our elegant Zero Trust architecture is purpose-built for the modern cloud and helps customers solve challenging security and operations pain points at scale.</p>\n<p>We are looking for a Backend Software Engineer to join our fast-growing team with a focus on scalability, reliability, and enhancing the building blocks of the product. In this role you will:</p>\n<ul>\n<li>Be deeply involved in evolving the core architecture of PAM.</li>\n</ul>\n<ul>\n<li>Work in our product development teams to build scalable, composable components of our platform.</li>\n</ul>\n<ul>\n<li>Be responsible for designing and implementing scalable architecture patterns.</li>\n</ul>\n<ul>\n<li>Design and build APIs with OpenAPI Specification that customers rely on for access to production infrastructure.</li>\n</ul>\n<ul>\n<li>Work on backend systems written in Go</li>\n</ul>\n<ul>\n<li>Participate in the rotational on-call activities with SRE and product development teams.</li>\n</ul>\n<p>You might be a good fit if you:</p>\n<ul>\n<li>Are an experienced software engineer with a background in Golang (other languages are also acceptable)</li>\n</ul>\n<ul>\n<li>Experienced working with relational databases like PostgreSQL or similar RDBMS technologies.</li>\n</ul>\n<ul>\n<li>Have the ability to design database models and backend APIs.</li>\n</ul>\n<ul>\n<li>Have experience working with cloud services like Caching, Queues, NoSQL Databases etc.</li>\n</ul>\n<ul>\n<li>Experienced working with any cloud provider such as AWS, GCP or Azure.</li>\n</ul>\n<ul>\n<li>Thrive in a collaborative environment built on end-to-end ownership.</li>\n</ul>\n<ul>\n<li>Love thinking about distributed systems, and the reliability, availability, and performance implications of the decisions made in their design.</li>\n</ul>\n<ul>\n<li>Enjoy deep-diving into production metrics, and familiarity with monitoring tools like Splunk, DataDog etc.</li>\n</ul>\n<ul>\n<li>Think in terms of systems, services, and APIs.</li>\n</ul>\n<p>Required education and experience:</p>\n<ul>\n<li>8+ years working as a software engineer.</li>\n</ul>\n<ul>\n<li>Experience working with production systems.</li>\n</ul>\n<ul>\n<li>Bachelors in CS, or equivalent</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dfe3062f-992","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7826456","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$160,000-$200,000 CAD","x-skills-required":["Golang","PostgreSQL","database models","backend APIs","cloud services","caching","queues","NoSQL Databases","AWS","GCP","Azure"],"x-skills-preferred":[],"datePosted":"2026-04-24T12:18:58.053Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Toronto, Ontario, Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Golang, PostgreSQL, database models, backend APIs, cloud services, caching, queues, NoSQL Databases, AWS, GCP, Azure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":200000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2dd530a1-6a3"},"title":"Staff Backend Engineer","description":"<p>Secure Every Identity, from AI to Human Identity is the key to unlocking the potential of AI. Okta secures AI by building the trusted, neutral infrastructure that enables organisations to safely embrace this new era.</p>\n<p>We are looking for builders and owners who operate with speed and urgency and execute with excellence. This is an opportunity to do career-defining work. We&#39;re all in on this mission. If you are too, let&#39;s talk.</p>\n<p>The PAM Team</p>\n<p>Ever wonder how large organisations make sure the right people can access their most critical systems? That&#39;s the problem the Okta Privileged Access Management (PAM) team solves. Our solution controls who can reach sensitive servers, databases and cloud resources and grants access only when it&#39;s needed. It is the security layer between people (and non-human-identities) and the systems they need to do their jobs.</p>\n<p>The Staff Backend Engineer Opportunity</p>\n<p>We&#39;re looking for a Backend Software Engineer who wants to work on hard problems: distributed systems and building software where getting it right really matters. You&#39;ll ship code that protects real infrastructure for real organisations. You’ll build foundations that multiple feature teams depend on. When you make something faster, more reliable, or easier to use, it multiplies across the entire product.</p>\n<p>This is a role for someone who likes thinking about how systems fit together. You&#39;ll need strong opinions about what makes a good abstraction, and the flexibility to evolve those abstractions as the product grows.</p>\n<p>What you’ll be doing</p>\n<p>Be deeply involved in evolving the core architecture of PAM.</p>\n<p>Work in our product development teams to build scalable, composable components of our platform.</p>\n<p>Be responsible for designing and implementing scalable architecture patterns.</p>\n<p>Design and build APIs with OpenAPI Specification that customers rely on for access to production infrastructure.</p>\n<p>Work on backend systems written in Go</p>\n<p>Participate in the rotational on-call activities with SRE and product development teams</p>\n<p>What you’ll bring to the role</p>\n<p>8+ years of experience as a SWE</p>\n<p>Are an experienced software engineer with a background in Golang (other languages are also acceptable)</p>\n<p>Experienced working with relational databases like PostgreSQL or similar RDBMS technologies.</p>\n<p>Have the ability to design database models and backend APIs.</p>\n<p>Have experience working with cloud services like Caching, Queues, NoSQL Databases etc.</p>\n<p>Experienced working with any cloud provider such as AWS, GCP or Azure.</p>\n<p>Thrive in a collaborative environment built on end-to-end ownership.</p>\n<p>Love thinking about distributed systems, and the reliability, availability, and performance implications of the decisions made in their design.</p>\n<p>Enjoy deep-diving into production metrics, and familiarity with monitoring tools like Splunk, DataDog etc.</p>\n<p>Think in terms of systems, services, and APIs</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2dd530a1-6a3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7819478","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$160,000-$220,000 CAD","x-skills-required":["Golang","PostgreSQL","Cloud services","Caching","Queues","NoSQL Databases","AWS","GCP","Azure"],"x-skills-preferred":[],"datePosted":"2026-04-24T12:17:47.932Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Toronto, Ontario, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Golang, PostgreSQL, Cloud services, Caching, Queues, NoSQL Databases, AWS, GCP, Azure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6071d39f-afb"},"title":"Staff Software Engineer, Continuous Integration","description":"<p>We are seeking a talented and experienced Staff Software Engineer to join our Continuous Integration (CI) team within the Developer Productivity organization. The CI team is responsible for the infrastructure that enables hundreds of engineers to ship code safely and efficiently to production.</p>\n<p>The CI team manages Anthropic&#39;s continuous integration system, which includes:</p>\n<ul>\n<li>CI infrastructure for automated testing and code quality assurance across our monorepo, designed to scale with rapid growth</li>\n<li>Test infrastructure that runs on Kubernetes clusters across multiple cloud providers, handling intelligent test selection, execution, and reporting for a large and complex test suite</li>\n<li>Merge queue management and complex branching strategies that ensure code quality at scale</li>\n<li>CI tooling and automation to improve developer productivity and reduce operational overhead</li>\n</ul>\n<p>As a Staff Software Engineer, you will design and build highly reliable, scalable CI infrastructure that supports thousands of daily builds across multiple cloud providers. You will also develop intelligent test selection systems that reduce CI time while maintaining code quality.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build highly reliable, scalable CI infrastructure that supports thousands of daily builds across multiple cloud providers</li>\n<li>Develop intelligent test selection systems that reduce CI time while maintaining code quality</li>\n<li>Build and improve incident response automation, including cluster load shedding, automatic recovery, and observability tooling</li>\n<li>Improve test infrastructure reliability through flake detection, quarantine systems, and test state management</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>10+ years of relevant industry experience building and operating large-scale CI/CD systems</li>\n<li>Deep experience with CI orchestration tools (Buildkite, Jenkins, GitHub Actions, or similar)</li>\n<li>Excellent communication skills and enjoy supporting internal partners</li>\n<li>Care deeply about reliability and building systems that &#39;never fail the same way twice&#39;</li>\n</ul>\n<p>Preferred qualifications include experience with merge queues and branch management at scale, test infrastructure, including intelligent test selection and flake management, and GitHub API and automation experience.</p>\n<p>We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6071d39f-afb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5073998008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"£325,000-£390,000 GBP","x-skills-required":["CI orchestration tools","Kubernetes","Cloud providers","Test infrastructure","Merge queue management","Branching strategies","CI tooling","Automation"],"x-skills-preferred":["Merge queues","Branch management","Intelligent test selection","Flake management","GitHub API"],"datePosted":"2026-04-24T12:14:53.102Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"CI orchestration tools, Kubernetes, Cloud providers, Test infrastructure, Merge queue management, Branching strategies, CI tooling, Automation, Merge queues, Branch management, Intelligent test selection, Flake management, GitHub API","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":325000,"maxValue":390000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c73333c8-f80"},"title":"Software Engineer, Safeguards Foundations (Internal Tooling)","description":"<p>We are seeking a software engineer to join our Safeguards Foundations team, which builds the platforms, infrastructure, and internal tools that support the development of beneficial AI systems. As a software engineer on this team, you will design, build, and maintain internal review and enforcement tooling used by Safeguards analysts, including case queues, content review surfaces, decision/audit logging, and account-actioning workflows. You will work closely with Trust &amp; Safety operations, policy, and detection-engineering teams to turn messy operational workflows into well-designed, durable software.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and maintain internal review and enforcement tooling used by Safeguards analysts</li>\n<li>Understand user workflows and establish tooling for well processes that may be distributed across a number of tools and UIs</li>\n<li>Develop the &#39;base layer&#39; of reusable APIs, data storage, and backend services that let new review workflows be stood up quickly and safely</li>\n<li>Partner with operations and policy teams to understand reviewer pain points, then translate them into clear product improvements that reduce handling time and decision error</li>\n<li>Integrate tooling with upstream detection systems and downstream enforcement infrastructure so that flagged behaviour flows cleanly from signal → human review → action</li>\n<li>Build in the guardrails that sensitive internal tools require: granular permissions, audit trails, data-access controls, and reviewer wellbeing features (e.g. content blurring, exposure limits)</li>\n<li>Instrument the tools you ship , surfacing metrics on queue health, reviewer throughput, and decision quality so the team can see what&#39;s working</li>\n<li>Contribute to the Foundations team&#39;s shared platform and on-call responsibilities</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>4+ years of experience as a software engineer, with meaningful time spent building internal tools, operations platforms, or back-office products</li>\n<li>Comfortable using agentic coding tools (e.g. Claude Code) as a core part of your workflow, and can direct them to ship well-tested, production-quality software at a high cadence without lowering the bar</li>\n<li>Take a product-minded approach to internal users: you work with the people using your tools, watch where they struggle, and fix it</li>\n<li>Results-oriented, with a bias towards flexibility and impact</li>\n<li>Pick up slack, even if it goes outside your job description</li>\n<li>Communicate clearly with non-engineering stakeholders and can explain technical trade-offs to operations and policy partners</li>\n<li>Care about the societal impacts of your work and want to apply your engineering skills directly to AI safety</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience building tooling in a trust &amp; safety, content moderation, fraud, integrity, or risk-operations setting</li>\n<li>Experience designing case-management or workflow systems (queues, SLAs, escalation paths, audit logs)</li>\n<li>Experience working with sensitive data and understanding the privacy, access-control, and reviewer-wellbeing considerations that come with it</li>\n<li>Experience with GCP/AWS, Postgres/BigQuery, and CI/CD in a production environment</li>\n<li>Experience using LLMs as a building block inside operational tools (e.g. assisted triage, summarisation, or classification in the review loop)</li>\n</ul>\n<p>Representative projects:</p>\n<ul>\n<li>Rebuilding the analyst review queue so cases are routed by severity and skill, with full decision history and one-click escalation</li>\n<li>Shipping a unified account-investigation view that pulls signals from multiple detection systems into a single, permissioned surface</li>\n<li>Adding content-obfuscation and exposure-tracking features to protect reviewers working with harmful material</li>\n<li>Building an internal labelling tool that feeds high-quality ground truth back to the detection and research teams</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c73333c8-f80","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5191433008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"£255,000-£325,000 GBP","x-skills-required":["agentic coding tools","APIs","backend services","case queues","content review surfaces","decision/audit logging","account-actioning workflows","CI/CD","GCP/AWS","Postgres/BigQuery","LLMs","sensitive data","privacy","access-control","reviewer-wellbeing"],"x-skills-preferred":[],"datePosted":"2026-04-24T12:14:27.217Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"agentic coding tools, APIs, backend services, case queues, content review surfaces, decision/audit logging, account-actioning workflows, CI/CD, GCP/AWS, Postgres/BigQuery, LLMs, sensitive data, privacy, access-control, reviewer-wellbeing","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":255000,"maxValue":325000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_12f710b8-5d6"},"title":"Systems Ops Administrator","description":"<p>At Twilio, we&#39;re shaping the future of communications, all from the comfort of our homes. We deliver innovative solutions to hundreds of thousands of businesses and empower millions of developers worldwide to craft personalized customer experiences.</p>\n<p>This role is the backbone of our Global Sales organisations, providing a blend of technical Salesforce administration and high-level data stewardship. You will ensure that Salesforce CPQ, OLM, and multiple Salesforce instances run with high process integrity while simultaneously managing the data quality and firmographic insights that drive our Sales organisation forward.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Salesforce Administration: Manage user setup, roles, profiles, permissions, and security settings. Serve as the escalation point for advanced user requests and troubleshoot complex and multi-instance issues. Ensure Compliance with security and access governance policies. Monitor system usage and adoption to recommend improvements.</li>\n</ul>\n<ul>\n<li>System Maintenance: Maintain page layouts, workflows, validation rules, and approval processes. Configure and maintain Salesforce support processes.</li>\n</ul>\n<ul>\n<li>Data Stewardship: Perform data audits and investigations across systems to enhance firmographic information. Fix data issues following standard operating procedures and data policies.</li>\n</ul>\n<ul>\n<li>Operational Excellence: Monitor and Manage Salesforce case queues to meet SLAs. Oversee the synchronisation of Salesforce data with ERP/Billing systems to ensure seamless order-to-cash cycles.</li>\n</ul>\n<ul>\n<li>Strategy &amp; Documentation: Provide documentation for implementation changes to data quality standards. Document data requirements, strategy, and rules (standardisation, cleansing, and validation) for both legacy and new environments.</li>\n</ul>\n<ul>\n<li>Collaboration: Partner with Sales Ops, Finance, IT, and Business Stakeholders to identify and drive automation advances that improve operational efficiency.</li>\n</ul>\n<ul>\n<li>Testing &amp; Training: Support UAT and deployments while developing training materials to empower our end users.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Salesforce Expertise: 3+ years of experience as a Salesforce Administrator (Sales Cloud) with an active Salesforce Administrator certification (ADM-201).</li>\n</ul>\n<ul>\n<li>Data Analysis: Demonstrable experience executing data quality analysis within complex, federated system architectures.</li>\n</ul>\n<ul>\n<li>Process Mindset: Experience in Sales Operations support and a basic understanding of data providers like ZoomInfo.</li>\n</ul>\n<ul>\n<li>Technical Logic: Ability to evaluate data risks, troubleshoot integrated system issues, and provide solutions in a timely manner.</li>\n</ul>\n<p>Desired:</p>\n<ul>\n<li>Advanced Certifications: Salesforce Certified Advanced Admin.</li>\n</ul>\n<ul>\n<li>Technical Stack: Experience with SQL, Heroku, or Python (specifically for data pipelining).</li>\n</ul>\n<ul>\n<li>Analytics: Familiarity with Tableau and Tableau CRM (CRM Analytics).</li>\n</ul>\n<ul>\n<li>Automation: Knowledge of Apex, Flow, SOQL, and advanced Salesforce automation.</li>\n</ul>\n<ul>\n<li>Soft Skills: Excellent written and verbal communication skills; ability to build relationships across all levels of the organisation.</li>\n</ul>\n<ul>\n<li>Linguistic Diversity: Proficiency in multiple languages applicable to the NAMER region is a plus.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_12f710b8-5d6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Twilio","sameAs":"https://www.twilio.com/","logo":"https://logos.yubhub.co/twilio.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/twilio/jobs/7837170","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Salesforce","Sales Cloud","Salesforce Administrator","Data Analysis","Data Quality","Firmographic Information","Salesforce CPQ","OLM","Salesforce Instances","User Setup","Roles","Profiles","Permissions","Security Settings","System Maintenance","Page Layouts","Workflows","Validation Rules","Approval Processes","Data Stewardship","Data Audits","Investigations","Standard Operating Procedures","Data Policies","Operational Excellence","SLAs","Case Queues","ERP/Billing Systems","Order-to-Cash Cycles","Strategy & Documentation","Implementation Changes","Data Quality Standards","Data Requirements","Strategy","Rules","Collaboration","Sales Ops","Finance","IT","Business Stakeholders","Automation Advances","Operational Efficiency","Testing & Training","UAT","Deployments","Training Materials","End Users"],"x-skills-preferred":["SQL","Heroku","Python","Tableau","Tableau CRM","Apex","Flow","SOQL","Advanced Salesforce Automation"],"datePosted":"2026-04-24T12:14:01.841Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - India"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Sales","industry":"Technology","skills":"Salesforce, Sales Cloud, Salesforce Administrator, Data Analysis, Data Quality, Firmographic Information, Salesforce CPQ, OLM, Salesforce Instances, User Setup, Roles, Profiles, Permissions, Security Settings, System Maintenance, Page Layouts, Workflows, Validation Rules, Approval Processes, Data Stewardship, Data Audits, Investigations, Standard Operating Procedures, Data Policies, Operational Excellence, SLAs, Case Queues, ERP/Billing Systems, Order-to-Cash Cycles, Strategy & Documentation, Implementation Changes, Data Quality Standards, Data Requirements, Strategy, Rules, Collaboration, Sales Ops, Finance, IT, Business Stakeholders, Automation Advances, Operational Efficiency, Testing & Training, UAT, Deployments, Training Materials, End Users, SQL, Heroku, Python, Tableau, Tableau CRM, Apex, Flow, SOQL, Advanced Salesforce Automation"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ceaa6a31-fb0"},"title":"Software Engineer, Workers Observability","description":"<p>At Cloudflare, we&#39;re on a mission to help build a better Internet. We&#39;re not looking for people who wait for a polished roadmap; we&#39;re looking for the builders who see the cracks in the Internet that everyone else has simply learned to live with.</p>\n<p>We value candidates who have the instinct to spot a &quot;normalized&quot; problem and the AI-native curiosity to create a solution using the latest tools. Our culture is built on iteration, leveraging AI to ship faster today to make it better tomorrow, while ensuring that every improvement, no matter how small, is shared across the team to lift everyone up.</p>\n<p>If you&#39;re the type of person who values curiosity over bureaucracy, and that AI is a partner in solving tough problems to keep the Internet moving forward, you&#39;ll fit right in.</p>\n<p>The Workers Observability (WOBS) Team provides a growing suite of tools to monitor, troubleshoot and optimize applications on the Cloudflare Developer platform. These tools increasingly work across Cloudflare products to provide complete visibility into user applications.</p>\n<p>As a Software Engineer on the WOBS Team, you will own code across the development lifecycle from design to release including QA, incident management, and performance optimization. You will be responsible for designing and shipping core software features across the observability stack - including the frontend customer experience, integrations with other Cloudflare primitives, and our internal data stores and ingestion pipelines.</p>\n<p>You will work closely with cross-functional teams to identify and prioritize features, and collaborate with engineers, designers, and product managers to deliver high-quality software that meets the performance requirements of Cloudflare&#39;s systems.</p>\n<p>Success in this role will depend on becoming an expert not just on the observability platform, but on the Cloudflare Developer Platform as a whole. You will need to have strong proficiency in at least one programming language, preferably a typed language (e.g. Go, TypeScript, Rust) as well as SQL, and experience building high-volume distributed systems using cloud primitives (AWS, GCP, CF), including queues, key-value stores, relational databases, blob storage, and serverless compute.</p>\n<p>You will also need to have experience troubleshooting distributed data systems including working with logs and data stores. A working knowledge of lower-level languages like Rust or C++ and building frontend applications in Typescript and React is a plus.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ceaa6a31-fb0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7505637","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","TypeScript","Rust","SQL","112","queue","key-value store","relational database","blob storage","serverless compute"],"x-skills-preferred":["C++","React"],"datePosted":"2026-04-24T12:12:24.027Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, TypeScript, Rust, SQL, 112, queue, key-value store, relational database, blob storage, serverless compute, C++, React"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bf1554e6-c64"},"title":"Software Engineer, AI Agents","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>Why This Role Matters</p>\n<p>At Cloudflare, we&#39;re building industrial-scale AI agents that support customers directly. This isn&#39;t research theater. Your code will power real customer interactions from day one, at global scale. Cloudflare already has the parts. You will assemble Workers, Durable Objects, KV, R2, D1, Vectorize, Workers AI, AI Gateway, and the Agent SDK into real agents customers use every day.</p>\n<p>Role Intent</p>\n<p>Ship production agents on the Cloudflare stack. Build, deploy, learn, repeat. Your code is the front door for Cloudflare customers.</p>\n<p>What You Will Do</p>\n<ul>\n<li>Build agents on Workers with Durable Objects for state and short term memory</li>\n<li>Wire tools with the Agent SDK, MCP, and function calling</li>\n<li>Use Vectorize, KV, R2, and D1 for semantic memory, cache, files, and config</li>\n<li>Run models through Workers AI and AI Gateway; integrate third parties when needed</li>\n<li>Create evals, guardrails, and audits. Measure, tune, re-ship fast</li>\n<li>Build agents that summarize, propose fixes, and escalate cleanly to humans</li>\n<li>Expose agent health and metrics in transparent dashboards. No mystery boxes</li>\n<li>Integrate with queues and webhooks; publish events on Queues or Pub/Sub</li>\n<li>Cut cost per case and time to first response. Prove it with data.</li>\n<li>Take end to end ownership including on call for what you ship (with team support)</li>\n<li>Design and maintain robust observability for distributed AI workflows, implementing structured logging and end-to-end tracing across async service boundaries to ensure visibility into agent reasoning and execution.</li>\n<li>Architect security boundaries for agent-led operations; implementing secure credential handling, multi-layer approval gates, and fine-grained trust scoping for mutative actions.</li>\n</ul>\n<p>Must Have</p>\n<ul>\n<li>Demonstrated success shipping production systems. Repos and releases that show real work.</li>\n<li>Strong in TypeScript or Rust on Workers. HTTP, queues, async, performance</li>\n<li>Fluency with Durable Objects, KV or R2, and either D1 or Postgres</li>\n<li>Hands on with model tooling. Prompt I/O, tool calling, evals, safety checks</li>\n<li>Observability mindset. Logs, traces, metrics, redlines</li>\n<li>Experience with a2a/multi-agent frameworks</li>\n<li>Experience developing LLM evaluation frameworks; automated scoring systems, CI-integrated quality gates.</li>\n</ul>\n<p>Nice to Have</p>\n<ul>\n<li>Workers AI, AI Gateway, and Vectorize in production</li>\n<li>Salesforce or Service Cloud experience. Webhooks and case APIs</li>\n<li>Security depth. Prompt injection protection, secrets detection, PII handling</li>\n<li>OSS agent frameworks. Know what to borrow and what to throw away.</li>\n</ul>\n<p>How We Build</p>\n<p>Align fast on what matters.</p>\n<p>Divide and conquer. Own your piece.</p>\n<p>Ship. Watch customers use it.</p>\n<p>Learn and repeat.</p>\n<p>Why Join Cloudflare in India?</p>\n<p>Impact at global scale: Your code will serve Cloudflare&#39;s customers across every region. Tens of millions of Internet properties depend on us.</p>\n<p>Work on the edge: Few companies give engineers the chance to build AI directly into an edge platform that runs in 300+ cities worldwide.</p>\n<p>Career growth: As one of the early engineers in our India based AI team, you&#39;ll have visibility, leadership opportunities, and a direct hand in shaping Cloudflare&#39;s AI roadmap.</p>\n<p>Culture of ownership: We believe in autonomy, accountability, and trust. Engineers here own outcomes, not just tickets.</p>\n<p>Learn and grow fast: Collaborate with peers across Support, Product, Security, and AI Platform teams. We encourage knowledge sharing, mentorship, and continuous learning.</p>\n<p>Interview Signal</p>\n<p>Expect to demonstrate your ability to:</p>\n<ul>\n<li>Build a mini agent on Workers using the Agent SDK</li>\n<li>Store session memory in Durable Objects</li>\n<li>Add semantic recall with Vectorize</li>\n<li>Ship behind a KV flag with traces and observability</li>\n<li>Push to production fast and take ownership</li>\n</ul>\n<p>Team Mission</p>\n<p>The Agent Tech team owns the end to end stack for customer facing agents on Cloudflare. Everything runs at the edge.</p>\n<p>Core Stack: Workers, Durable Objects, KV, R2, D1, Queues, Pub/Sub, Vectorize, Workers AI, AI Gateway, Pages, Zero Trust.</p>\n<p>Principles: Ship fast. Measure truth. Simplify relentlessly. Own outcomes.</p>\n<p>Fraud Alert</p>\n<p>Do not fall victim to recruitment fraud. Cloudflare never charges application fees or requires candidates to purchase third-party certifications or training as a condition of employment. All official communication comes strictly from @cloudflare.com email addresses.</p>\n<p>What Makes Cloudflare Special?</p>\n<p>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul.</p>\n<p>Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunities to all employees and applicants for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, veteran status, genetic information, or any other characteristic protected by law.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bf1554e6-c64","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7831810","x-work-arrangement":null,"x-experience-level":null,"x-job-type":null,"x-salary-range":null,"x-skills-required":["TypeScript","Rust","Workers","HTTP","queues","async","performance","Durable Objects","KV","R2","Postgres","model tooling","prompt I/O","tool calling","evals","safety checks","observability mindset","logs","traces","metrics","redlines","a2a/multi-agent frameworks","LLM evaluation frameworks","automated scoring systems","CI-integrated quality gates"],"x-skills-preferred":[],"datePosted":"2026-04-24T12:12:03.262Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"In-Office"}},"occupationalCategory":"Engineering","industry":"Technology","skills":"TypeScript, Rust, Workers, HTTP, queues, async, performance, Durable Objects, KV, R2, Postgres, model tooling, prompt I/O, tool calling, evals, safety checks, observability mindset, logs, traces, metrics, redlines, a2a/multi-agent frameworks, LLM evaluation frameworks, automated scoring systems, CI-integrated quality gates"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1811cb81-878"},"title":"SAP EWM Senior Technical Consultant","description":"<p>We are seeking an experienced SAP EWM Senior Technical Consultant with strong expertise in SAP Extended Warehouse Management (EWM) to design, develop, and support advanced warehouse management solutions.</p>\n<p>The ideal candidate will possess deep technical knowledge, hands-on development experience, and the ability to work closely with functional consultants and business stakeholders to deliver high-quality SAP EWM implementations and enhancements.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Design, develop, and implement technical solutions in SAP EWM.</li>\n<li>Lead technical architecture and development activities for EWM projects and support engagements.</li>\n<li>Develop and enhance, RF Framework developments, PPF enhancements, BADI implementations and user exits</li>\n<li>Enhancements and custom reports</li>\n<li>Interfaces (IDocs, BAPIs, Proxy, APIs)</li>\n<li>Development of forms using SmartForms / Adobe Forms.</li>\n<li>Build and support integration of EWM with PP and TM. Indepth understanding of PMR and ASR functionality for integration.</li>\n<li>Build the integration of eWM with other modules of SAP i.e. SD, MM etc..</li>\n<li>Work with middleware technologies (PI/PO, CPI).</li>\n<li>Perform performance optimization and troubleshooting.</li>\n<li>Support cutover activities and hypercare.</li>\n<li>Provide technical leadership and mentor junior consultants.</li>\n<li>Prepare technical documentation (FS/TS, design documents, test scripts).</li>\n<li>Participate in client workshops and requirement gathering sessions in close alignment with functional consultant.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>10+ years of SAP technical experience.</li>\n<li>5+ years of hands-on experience in SAP EWM (Embedded and/or Decentralized).</li>\n<li>Experience in at least 2 full-cycle SAP EWM implementations.</li>\n<li>Strong ABAP and ABAP OO development skills.</li>\n<li>Experience with:</li>\n</ul>\n<p>EWM RF development   Enhancements and BADI implementations   PPF framework   CIF integration   EWM queues and monitoring</p>\n<ul>\n<li>Experience in S/4HANA EWM implementation projects.</li>\n<li>Strong debugging and performance tuning skills.</li>\n<li>Experience with OData services and Fiori is a plus.</li>\n<li>Knowledge of warehouse processes (inbound, outbound, internal movements, physical inventory).</li>\n<li>Strong communication and stakeholder management skills.</li>\n</ul>\n<p>At MHP, you grow continuously through your tasks in an innovative and supportive environment. This makes us the perfect sparring partner for your career , both when it comes to professional input and career networking. Among other things, we offer:</p>\n<ul>\n<li>Appreciation. We support and value our colleagues for who they are and celebrate our successes together.</li>\n<li>We always welcome creativity and fresh ideas.</li>\n<li>With us, you have the opportunity to grow – in your tasks, your expertise, and your responsibilities.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1811cb81-878","directApply":true,"hiringOrganization":{"@type":"Organization","name":"MHP","sameAs":"https://www.mhp.com/","logo":"https://logos.yubhub.co/mhp.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=20072","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SAP EWM","ABAP","ABAP OO","RF Framework","PPF framework","CIF integration","EWM queues and monitoring","OData services","Fiori"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:28:04.152Z","employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Consulting","skills":"SAP EWM, ABAP, ABAP OO, RF Framework, PPF framework, CIF integration, EWM queues and monitoring, OData services, Fiori"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_32af9ee8-9f1"},"title":"SAP EWM Senior Technical Consultant","description":"<p>We are seeking an experienced SAP EWM Senior Technical Consultant with strong expertise in SAP Extended Warehouse Management (EWM) to design, develop, and support advanced warehouse management solutions.</p>\n<p>The ideal candidate will possess deep technical knowledge, hands-on development experience, and the ability to work closely with functional consultants and business stakeholders to deliver high-quality SAP EWM implementations and enhancements.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Design, develop, and implement technical solutions in SAP EWM.</li>\n<li>Develop and enhance, RF Framework developments, PPF enhancements, BADI implementations and user exits</li>\n<li>Enhancements and custom reports</li>\n<li>Interfaces (IDocs, BAPIs, Proxy, APIs)</li>\n<li>Development of forms using SmartForms / Adobe Forms.</li>\n<li>Build and support integration of EWM with PP and TM. Indepth understanding of PMR and ASR functionality for integration.</li>\n<li>Build the integration of eWM with other modules of SAP i.e. SD, MM etc..</li>\n<li>Work with middleware technologies (PI/PO, CPI).</li>\n<li>Perform performance optimization and troubleshooting.</li>\n<li>Support cutover activities and hypercare.</li>\n<li>Provide technical leadership and mentor junior consultants.</li>\n<li>Prepare technical documentation (FS/TS, design documents, test scripts).</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>8+ years of SAP technical experience.</li>\n<li>5+ years of hands-on experience in SAP EWM (Embedded and/or Decentralized).</li>\n<li>Experience in at least 2 full-cycle SAP EWM implementations.</li>\n<li>Strong ABAP and ABAP OO development skills.</li>\n<li>Experience with:</li>\n<li>EWM RF development</li>\n<li>Enhancements and BADI implementations</li>\n<li>PPF framework</li>\n<li>CIF integration</li>\n<li>EWM queues and monitoring</li>\n<li>Experience in S/4HANA EWM implementation projects.</li>\n<li>Strong debugging and performance tuning skills.</li>\n<li>Experience with OData services and Fiori is a plus.</li>\n<li>Knowledge of warehouse processes (inbound, outbound, internal movements, physical inventory).</li>\n<li>Strong communication and stakeholder management skills.</li>\n</ul>\n<p>At MHP, you will continuously grow with your projects and objectives in an innovative and supportive environment. That makes us the perfect sparring partner for your career, fueling your growth as an expert in your field while expanding your business network.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_32af9ee8-9f1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"MHP","sameAs":"http://www.mhp.com/","logo":"https://logos.yubhub.co/mhp.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=19969","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SAP EWM","ABAP","ABAP OO","RF Framework","PPF framework","CIF integration","EWM queues and monitoring","OData services","Fiori"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:26:20.212Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Consulting","skills":"SAP EWM, ABAP, ABAP OO, RF Framework, PPF framework, CIF integration, EWM queues and monitoring, OData services, Fiori"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_27168e63-72b"},"title":"Systems Engineer, SSL/TLS Team","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>Cloudflare was named to Entrepreneur Magazine&#39;s Top Company Cultures list and ranked among the World&#39;s Most Innovative Companies by Fast Company.</p>\n<p>Available Locations: Austin</p>\n<p>Role Summary</p>\n<p>Back in 2014, Cloudflare launched Universal SSL to provide all of our customers with the security of encrypted traffic between visitors and Cloudflare&#39;s network. On the day we flipped the switch, we doubled the amount of sites using SSL/TLS on the internet. Since then, we&#39;ve been working on providing even more ways to encrypt traffic easier and more effectively.</p>\n<p>To better serve our customers, we&#39;re constantly improving our systems for issuing, managing, and deploying TLS certificates. By joining Cloudflare&#39;s SSL/TLS Team, you can help ensure that every site owner has the ability to provide their users with a secure connection. Your code will help ensure millions of websites have access to encryption. As such, the results of your code will be felt by millions of site owners and billions of visitors every day.</p>\n<p>We&#39;re continuously providing technology to site owners that was previously only available to the largest organizations. Your contributions will help make the web a faster, safer, better, and more heavily encrypted place.</p>\n<p>Cloudflare is looking for a Systems Engineer to help us build services with RESTful APIs and their own datastores using Go and other languages. You will also write and maintain functional specifications for your service, so others know how it works. In this role you will integrate multiple services via message queue systems and evaluate 3rd party and open source tools. You will design and implement alerts, write procedures for ops to handle such alerts, or serve as an escalation point for unresolved operational issues. You will also need to read and understand code from a legacy codebase to migrate logic to a modern services architecture and review the code of fellow engineers via pull requests.</p>\n<p>We strive to build reliable, fault-tolerant systems that can operate at Cloudflare&#39;s scale.</p>\n<p>Role Requirements (Must-Have Skills)</p>\n<ul>\n<li>3+ years of professional experience building backend systems at a technology company</li>\n</ul>\n<ul>\n<li>Experience writing production-quality code in Go (or other languages with a willingness to learn Go)</li>\n</ul>\n<ul>\n<li>Proven experience building and operating distributed systems</li>\n</ul>\n<ul>\n<li>Proficiency with modern Unix/Linux development and runtime environments</li>\n</ul>\n<ul>\n<li>Strong debugging and troubleshooting skills with attention to detail</li>\n</ul>\n<ul>\n<li>Track record of delivering well-tested, reliable code for production environments</li>\n</ul>\n<p>Desirable Skills, Knowledge, and Experience</p>\n<ul>\n<li>Experience designing and building RESTful APIs and full-stack applications</li>\n</ul>\n<ul>\n<li>Familiarity with PKI/TLS concepts and certificate management</li>\n</ul>\n<ul>\n<li>Experience writing technical documentation and specifications</li>\n</ul>\n<ul>\n<li>Working knowledge of SQL and relational databases such as PostgreSQL</li>\n</ul>\n<ul>\n<li>Understanding of distributed systems constraints: performance, caching, data storage/retrieval</li>\n</ul>\n<ul>\n<li>Experience with message queues (Kafka, RabbitMQ, etc)</li>\n</ul>\n<p>Benefits</p>\n<p>Cloudflare offers a complete package of benefits and programs to support you and your family. Our benefits programs can help you pay health care expenses, support caregiving, build capital for the future and make life a little easier and fun!</p>\n<p>The below is a description of our benefits for employees in the United States, and benefits may vary for employees based outside the U.S.</p>\n<p>Health &amp; Welfare Benefits</p>\n<ul>\n<li>Medical/Rx Insurance</li>\n</ul>\n<ul>\n<li>Dental Insurance</li>\n</ul>\n<ul>\n<li>Vision Insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Accounts</li>\n</ul>\n<ul>\n<li>Commuter Spending Accounts</li>\n</ul>\n<ul>\n<li>Fertility &amp; Family Forming Benefits</li>\n</ul>\n<ul>\n<li>On-demand mental health support and Employee Assistance Program</li>\n</ul>\n<ul>\n<li>Global Travel Medical Insurance</li>\n</ul>\n<p>Financial Benefits</p>\n<ul>\n<li>Short and Long Term Disability Insurance</li>\n</ul>\n<ul>\n<li>Life &amp; Accident Insurance</li>\n</ul>\n<ul>\n<li>401(k) Retirement Savings Plan</li>\n</ul>\n<ul>\n<li>Employee Stock Participation Plan</li>\n</ul>\n<p>Time Off</p>\n<ul>\n<li>Flexible paid time off covering vacation and sick leave</li>\n</ul>\n<ul>\n<li>Leave programs, including parental, pregnancy health, medical, and bereavement leave</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We&#39;re not just a highly ambitious, large-scale technology company. We&#39;re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare&#39;s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law. We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals with disabilities.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_27168e63-72b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7542754","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Unix/Linux development and runtime environments","Distributed systems","Debugging and troubleshooting","SQL and relational databases"],"x-skills-preferred":["RESTful APIs","PKI/TLS concepts and certificate management","Message queues","Technical documentation and specifications"],"datePosted":"2026-04-18T15:51:38.124Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Unix/Linux development and runtime environments, Distributed systems, Debugging and troubleshooting, SQL and relational databases, RESTful APIs, PKI/TLS concepts and certificate management, Message queues, Technical documentation and specifications"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_50df2bb4-aba"},"title":"Product Operations Lead, Tier 1","description":"<p>We are looking for a Product Operations Lead, Tier 1 to pilot our Help Desk operations and join our rapidly growing team in Costa Mesa, CA. In this position, you will play a key role in supporting our customers and their fielded assets by driving issues spanning multiple software and hardware products to resolution in real time.</p>\n<p>Using your intimate familiarity with these products and related issues, you must be able to coach teammates, interface directly with customers, work independently, and resolve complex issues using your discretion and sound judgment. The challenges you will encounter are innumerable and unpredictable, but you will be expected to take ownership of delivering the technical solution.</p>\n<p>You should have an aptitude for debugging and an appetite for real-time response, rapid resolution, and root-causing complex issues across multiple platforms and organizations.</p>\n<p>WHAT YOU&#39;LL DO:</p>\n<ul>\n<li>Supervise and mentor Product Operations Associates, ensuring effective performance, adherence to protocols, and professional development</li>\n<li>Manage shift schedules and staffing levels to maintain 24/7 operational readiness</li>\n<li>Develop Tier 1 scoped processes and runbooks</li>\n<li>Create and implement process and product improvements based on customer feedback and available metrics/analytics, utilizing internal monitoring dashboards and alerting software</li>\n<li>Interface directly with Anduril&#39;s customers and hardware/software products to provide world-class account level and technical support on complex issues</li>\n<li>Manage ticket flow, triage, and response through Salesforce&#39;s Service Desk management system</li>\n<li>Identify and solve persistent problems and reoccurring failures</li>\n<li>Author and issue product-specific knowledge articles and SOPs for end-user triaging</li>\n</ul>\n<p>REQUIRED QUALIFICATIONS:</p>\n<ul>\n<li>Bachelor&#39;s degree or equivalent experience in a related field</li>\n<li>Minimum of 2-3 years of experience in a Operations/Support Center or similar operational environment</li>\n<li>Prior experience supporting complex hardware systems, networks, and related technologies (military assets are a plus)</li>\n<li>Excellent communication skills, both written and verbal</li>\n<li>Ability to work independently and as part of a team</li>\n<li>Extremely detail oriented with excellent customer service</li>\n<li>Strong aptitude for problem solving in unstructured situations at the interface of hardware, software, and networking. Ability to drive challenging and vague technical problems to clarity and resolution</li>\n<li>Must be able to obtain and hold a U.S. Secret security clearance; Holding clearances or Background Investigations in the past a plus</li>\n</ul>\n<p>PREFERRED QUALIFICATIONS:</p>\n<ul>\n<li>Experience leading help desk operations and guiding team members</li>\n<li>Experience with ticketing systems and managing a ticket queue (Service Desk, Service Now, Jira, etc.)</li>\n<li>Experience with on-call support operations and working in limited risk tolerance production environments</li>\n<li>Familiarity with US military, hardware, software interfaces, and product operations are all pluses</li>\n<li>Government agency experience preferred</li>\n<li>Willingness to work non-standard hours and weekends</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_50df2bb4-aba","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5041072007","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$77,000-$102,000 USD","x-skills-required":["Product Operations","Help Desk Operations","Ticketing Systems","Customer Service","Problem Solving","Communication","Teamwork","Security Clearance"],"x-skills-preferred":["Leadership","Ticket Queue Management","On-call Support","US Military Hardware","Software Interfaces"],"datePosted":"2026-04-18T15:48:39.850Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Product Operations, Help Desk Operations, Ticketing Systems, Customer Service, Problem Solving, Communication, Teamwork, Security Clearance, Leadership, Ticket Queue Management, On-call Support, US Military Hardware, Software Interfaces","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":77000,"maxValue":102000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4f71a295-9c1"},"title":"Staff Software Engineer","description":"<p>We are looking for an experienced Staff Software Engineer to work on our next-generation Imports Platform team. The Imports Platform team is leading a strategic initiative to modernize Okta&#39;s identity lifecycle management capabilities by architecting and migrating from a legacy monolithic system to a highly scalable, distributed microservices platform.</p>\n<p>As a Staff Software Engineer on the Imports Platform team, you will be a technical leader who independently owns projects end-to-end, from ideation and architectural design through implementation, deployment, and operational excellence. You will drive technical strategy, make critical architectural decisions, and influence both your immediate team and cross-team initiatives.</p>\n<p>You will work on complex distributed systems challenges including massive-scale batch processing, real-time synchronization, and user matching algorithms that serve thousands of enterprise customers. This role requires strong technical leadership, strategic thinking, and the ability to balance short-term delivery with long-term platform health.</p>\n<p>You will mentor senior and junior engineers, partner with Product Management on feature strategy, and help shape the future of Okta&#39;s Imports platform. This is a hybrid position requiring a blend of remote and in-office collaboration.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Provide technical leadership on major development projects, including architectural design and implementation strategy</li>\n<li>Independently own and deliver projects end-to-end within the team, including technical prioritization and tradeoffs</li>\n<li>Generate design ideas and solutions for ambiguous problems, taking complete ownership from conception through production impact</li>\n<li>Design and architect core, high-performance, scalable software components with full ownership of all production aspects (scalability, reliability, monitoring, alerting, resource efficiency, testing, documentation)</li>\n<li>Lead technical design discussions and guide the team in making architectural decisions</li>\n<li>Drive the migration strategy from monolithic to microservices architecture, including planning, scoping, and execution</li>\n<li>Interface extensively with cross-functional teams (Architects, QA, Product, Technical Support, Documentation, UX, and SRE) to deliver comprehensive import and sync solutions</li>\n<li>Analyze and refine requirements with Product Management, partnering on product features and helping define the &#39;how&#39;</li>\n<li>Conduct code reviews with focus on systems design, reliability, performance, scalability, security, and maintainability</li>\n<li>Share knowledge widely, coordinate across teams, and manage risk and dependencies for projects</li>\n<li>Work with QA and SRE teams to define comprehensive testing strategies and operational excellence practices</li>\n<li>Independently troubleshoot complex production incidents spanning the home team, perform root cause analysis, and drive operational improvement projects</li>\n<li>Use data and metrics to drive technical decisions and validate the impact of architectural changes</li>\n<li>Mentor and provide technical guidance to senior and junior engineers on the team</li>\n<li>Help resolve difficult customer issues and work closely with Field teams and CSMs to identify patterns and drive product improvements</li>\n<li>Participate in group strategy discussions and help break down strategic initiatives into actionable technical milestones</li>\n<li>Proactively identify and advocate for improvements in team velocity, engineering practices, and operational processes</li>\n<li>Drive improvements in observability, monitoring, and production support capabilities</li>\n</ul>\n<p>Required Knowledge, Skills, and Abilities:</p>\n<ul>\n<li>7+ years of software development experience building highly-reliable, mission-critical software at scale</li>\n<li>Deep expertise with object-oriented languages, particularly Java, with proven ability to architect large-scale systems</li>\n<li>Expert-level knowledge of Spring Boot framework, Maven, and modern Java development practices</li>\n<li>Deep understanding of infrastructure-level technologies: distributed systems, caching strategies, stream processing, resilient architectures</li>\n<li>Solid experience with data stores including relational databases (MySQL), caching layers (Redis), and cloud storage (S3)</li>\n<li>Experience with one or more Directory services: Active Directory, LDAP, Office 365, Azure AD</li>\n<li>Strong experience with RESTful APIs, gRPC, and microservices architecture patterns</li>\n<li>Proven track record of working with systems at massive scale, including batch processing and real-time sync capabilities</li>\n<li>Experience with cloud platforms (AWS, GCP) including services like SQS, S3, and multi-region architectures</li>\n<li>Strong understanding of distributed job processing, message queues, and event-driven architectures</li>\n<li>Demonstrated ability to lead technical projects independently and influence cross-team initiatives</li>\n<li>Excellent communication skills with ability to share information widely and coordinate across teams</li>\n<li>Strong mentorship capabilities with experience guiding senior and junior engineers</li>\n<li>Customer-focused mindset with experience working with Field teams to resolve complex issues</li>\n<li>Strategic thinking ability to participate in and contribute to platform strategy</li>\n<li>Experience with operational excellence including incident management, root cause analysis, and driving systemic improvements</li>\n</ul>\n<p>Nice to Haves:</p>\n<ul>\n<li>Experience with Protocol Buffers (Protos) and building event-driven systems</li>\n<li>Experience with server-side technologies including advanced caching, asynchronous processing, multi-threading, and concurrency patterns</li>\n<li>Experience in Test-Driven Development (TDD) and automated testing strategies</li>\n<li>Deep knowledge of Identity and Access Management protocols and technologies: OAuth, OpenID Connect, SAML, SCIM, LDAP</li>\n<li>Experience with Microsoft Azure management APIs, Microsoft Graph API, Office 365, or ADFS</li>\n<li>Experience automating and deploying large-scale production services in AWS, GCP, or similar cloud platforms</li>\n<li>Experience with feature flag frameworks and gradual rollout strategies for large-scale migrations</li>\n<li>Understanding of user matching, correlation algorithms, and identity resolution at scale</li>\n<li>Experience with observability platforms, creating comprehensive monitoring and alerting strategies</li>\n<li>Experience migrating monolithic applications to microservices architecture</li>\n<li>Knowledge of data modeling for graph databases and relationship management</li>\n<li>Experience with incremental sync, delta detection, and change data capture patterns</li>\n<li>Background in building resilient systems with retry logic, circuit breakers, and failure handling</li>\n<li>Experience with performance optimization and capacity planning for high-throughput systems</li>\n</ul>\n<p>Education and Training:</p>\n<p>B.S. Computer Science or related field</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4f71a295-9c1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7725948","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Spring Boot","Maven","distributed systems","caching strategies","stream processing","resilient architectures","relational databases","caching layers","cloud storage","Directory services","RESTful APIs","gRPC","microservices architecture patterns","batch processing","real-time sync capabilities","cloud platforms","distributed job processing","message queues","event-driven architectures"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:19.715Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Spring Boot, Maven, distributed systems, caching strategies, stream processing, resilient architectures, relational databases, caching layers, cloud storage, Directory services, RESTful APIs, gRPC, microservices architecture patterns, batch processing, real-time sync capabilities, cloud platforms, distributed job processing, message queues, event-driven architectures"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_396fe53d-121"},"title":"Consulting Architect - Observability","description":"<p>As a Consulting Architect – Observability, you will play a pivotal role in helping our customers realise the value of Elastic’s Solutions. Acting as a trusted technical advisor, you will work with enterprises to design, deliver, and scale architectures that improve application performance, infrastructure visibility, and end-user experience.</p>\n<p>You&#39;ll collaborate with Elastic’s Professional Services, Engineering, Product, and Sales teams to accelerate adoption of the Elastic Observability platform, ensuring customers maximise the value of their data while achieving business outcomes. This is a highly impactful role, with opportunities to guide strategy, lead complex implementations, and mentor both customers and teammates.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Translating business and technical requirements into scalable, outcome-driven solutions built on the Elastic Stack.</li>\n<li>Leading end-to-end delivery of customer engagements , from discovery and design through implementation, enablement, and optimisation.</li>\n<li>Partnering with customers to architect, deploy, and operationalise Elastic solutions that drive measurable value and adoption.</li>\n<li>Providing technical oversight, guidance, and enablement to customers and teammates throughout project lifecycles.</li>\n<li>Collaborating cross-functionally with Sales, Product, Engineering, and Support to ensure successful outcomes and continuous improvement.</li>\n<li>Capturing and sharing best practices, lessons learned, and solution patterns across the Elastic Services community.</li>\n<li>Contributing to internal enablement, mentoring, and a culture of continuous learning and collaboration</li>\n</ul>\n<p>Required skills include:</p>\n<ul>\n<li>5+ years as a consultant, architect, or engineer with expertise in observability, monitoring, or related domains.</li>\n<li>Expertise in the Telecommunications domain, especially with Mobile networks and devices.</li>\n<li>Strong experience with time-series data ingestion and processing, including pipelines with Elastic Agents, Beats, and Logstash.</li>\n<li>Knowledge of messaging queues (Kafka, Redis) and ingestion optimisation strategies.</li>\n<li>Understanding of observability concepts like distributed tracing, metrics pipelines, log aggregation, anomaly detection, SLOs/SLIs.</li>\n<li>Experience with one or more: Kubernetes, cloud platforms (AWS, Azure, GCP), or infrastructure as code.</li>\n<li>Familiarity with Elastic Common Schema (ECS), data parsing, and normalisation.</li>\n<li>Proven experience deploying Elastic Observability (APM, UEM, logs, metrics, infra, network monitoring) or similar solutions at enterprise scale.</li>\n<li>Hands-on expertise in distributed systems and large-scale infrastructure.</li>\n<li>Ability to design and build dashboards, visualisations, and alerting thresholds that drive actionable insights.</li>\n<li>Experience with Kubernetes, Linux, Java, databases, Docker, AWS/Azure/GCP, VMs, Lucene.</li>\n<li>Strong communication and presentation skills, with experience engaging directly with customers.</li>\n<li>Bachelor’s, Master’s, or PhD in Computer Science, Engineering, or related field, or equivalent experience.</li>\n<li>Comfortable working in highly distributed teams, both remote and on-site when needed.</li>\n<li>May require significant travel to customer sites to support engagements and solution implementations; candidates should be comfortable with varying levels of travel based on business needs.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_396fe53d-121","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Elastic","sameAs":"https://www.elastic.co/","logo":"https://logos.yubhub.co/elastic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/elastic/jobs/7440232","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["observability","monitoring","Elastic Stack","time-series data ingestion","Elastic Agents","Beats","Logstash","messaging queues","Kafka","Redis","distributed tracing","metrics pipelines","log aggregation","anomaly detection","SLOs/SLIs","Kubernetes","cloud platforms","infrastructure as code","Elastic Common Schema","data parsing","normalisation","databases","Docker","VMs","Lucene"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:40:26.428Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Tokyo, Japan"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"observability, monitoring, Elastic Stack, time-series data ingestion, Elastic Agents, Beats, Logstash, messaging queues, Kafka, Redis, distributed tracing, metrics pipelines, log aggregation, anomaly detection, SLOs/SLIs, Kubernetes, cloud platforms, infrastructure as code, Elastic Common Schema, data parsing, normalisation, databases, Docker, VMs, Lucene"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3a7e27c3-92a"},"title":"Pipeline Engineer (Graphics/3D)","description":"<p>We are building a production-grade web application for 3D Gaussian Splat scene generation, editing, and publishing. We&#39;re looking for a Pipeline Engineer to help integrate cutting-edge research features and make them reliable, debuggable, and delightful to use.</p>\n<p>This is a high-ownership, fullstack-but-backend-heavy role that sits between R&amp;D and frontend. You will work end-to-end across graphics/ML algorithms, backend services, and frontend UI , turning proof-of-concepts into shipped features that users can rely on. The ideal candidate enjoys making complex, messy systems work smoothly in production and improving them continuously based on both internal testing and external user feedback.</p>\n<p><strong>Key Responsibilities</strong></p>\n<ul>\n<li>Bridge research and product by working closely with both graphics/computer vision researchers and frontend engineers to ship usable features.</li>\n<li>Turn standalone Python scripts into clean, production-ready systems with clear inputs, outputs, validations, and failure modes.</li>\n<li>Develop backend services, APIs, and tooling that expose complex 3D workflows in a reliable and scalable way.</li>\n<li>Assist in integrations across the 3D ecosystem, including asset import/export and format conversion with common DCC tools.</li>\n</ul>\n<p><strong>Ideal Candidate Profile</strong></p>\n<ul>\n<li>You have a strong pipeline mindset, with experience turning scripts into production systems with clear inputs, outputs, validations, and failure modes.</li>\n<li>You enjoy building tools and infrastructure that enable others, and you take pride in making complex systems understandable and usable.</li>\n<li>You have fluency in the 3D ecosystem, including familiarity with 3D algorithms, DCC tools and common 3D file formats, sufficient to design integrations and debug workflow issues.</li>\n</ul>\n<p><strong>Minimum Qualifications</strong></p>\n<ul>\n<li>Strong proficiency in Python, including packaging, typing, tooling, debugging, and performance profiling.</li>\n<li>Strong literacy in core 3D graphics and computer vision concepts, such as transforms, cameras, coordinate systems, rendering, and visual artifact debugging.</li>\n<li>Demonstrated experience taking prototypes to production, including refactoring, testing, CI/CD, versioned artifacts, and reproducibility.</li>\n<li>Solid backend fundamentals, including HTTP APIs, FastAPI (or similar frameworks), async/concurrency basics, cloud deployment, and service reliability.</li>\n</ul>\n<p><strong>Strongly Preferred / Nice-to-Haves</strong></p>\n<ul>\n<li>Experience with photogrammetry, 3D reconstruction, or Gaussian splat rendering pipelines.</li>\n<li>Hands-on experience with DCC tools such as Blender, Maya, Houdini, Unreal, or Unity.</li>\n<li>Familiarity with the GPU stack (CUDA, PyTorch), batch/queue systems, and containerization (Docker, Kubernetes).</li>\n<li>Frontend adjacency, with comfort collaborating on React-based parameter plumbing and UX for technical controls.</li>\n<li>Experience with production pipelines at VFX, animation, or gaming studios.</li>\n<li>A production support mindset, including willingness to iterate on documentation, tutorials, and error messages to improve usability and reduce misuse.</li>\n</ul>\n<p><strong>Example Projects You Might Work On</strong></p>\n<ul>\n<li>Packaging ML and 3D Python pipelines into GPU-backed FastAPI services with request validation, reproducible outputs, and well-defined request/response schemas.</li>\n<li>Designing parameter schemas and defaults that map cleanly from frontend controls to backend APIs and internal pipeline configurations.</li>\n<li>Integrating import/export workflows with popular DCC tools (e.g., Blender, Maya, Houdini, Unity, Unreal, USD), identifying workflow friction, and producing lightweight documentation, tutorials, and example code/scripts to help users succeed.</li>\n</ul>\n<p><strong>Who You Are</strong></p>\n<ul>\n<li>Fearless Innovator: We need people who thrive on challenges and aren&#39;t afraid to tackle the impossible.</li>\n<li>Resilient Builder: Impacting Large World Models isn&#39;t a sprint; it&#39;s a marathon with hurdles. We&#39;re looking for builders who can weather the storms of groundbreaking research and come out stronger.</li>\n<li>Mission-Driven Mindset: Everything we do is in service of creating the best spatially intelligent AI systems, and using them to empower people.</li>\n<li>Collaborative Spirit: We&#39;re building something bigger than any one person. We need team players who can harness the power of collective intelligence.</li>\n</ul>\n<p>We&#39;re hiring the brightest minds from around the globe to bring diverse perspectives to our cutting-edge work. If you&#39;re ready to work on technology that will reshape how machines perceive and interact with the world - then World Labs is your launchpad.</p>\n<p>Join us, and let&#39;s make history together.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3a7e27c3-92a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"World Labs","sameAs":"https://www.worldlabs.ai/","logo":"https://logos.yubhub.co/worldlabs.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/worldlabs/jobs/4093035009","x-work-arrangement":"remote","x-experience-level":null,"x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","3D graphics","computer vision","FastAPI","HTTP APIs","async/concurrency basics","cloud deployment","service reliability"],"x-skills-preferred":["photogrammetry","3D reconstruction","Gaussian splat rendering","Blender","Maya","Houdini","Unreal","Unity","GPU stack","batch/queue systems","containerization","React-based parameter plumbing","UX for technical controls","production pipelines","VFX","animation","gaming studios"],"datePosted":"2026-04-17T13:09:23.876Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, 3D graphics, computer vision, FastAPI, HTTP APIs, async/concurrency basics, cloud deployment, service reliability, photogrammetry, 3D reconstruction, Gaussian splat rendering, Blender, Maya, Houdini, Unreal, Unity, GPU stack, batch/queue systems, containerization, React-based parameter plumbing, UX for technical controls, production pipelines, VFX, animation, gaming studios"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9d27e558-af6"},"title":"Senior Site Reliability Engineer","description":"<p><strong>Role</strong></p>\n<p>We are building a global operating network that finally enables supply-chain companies to collaborate within one platform. Our workflow engine empowers non-technical industry experts to model their complex manufacturing and operational processes. Our forms engine enables unprecedented data exchange between companies. And our upcoming AI engine can generate entire new processes and summarize the complex goings-on across thousands of workflows, identifying inefficiencies and driving optimization as companies react to a constantly-shifting global landscape.</p>\n<p>As an SRE you will have the opportunity to shape our developer platform, work directly with customers, and architect solutions that balance the rigorous security and reliability requirements of global enterprises with the speed and flexibility of a rapidly growing series A organization.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Contribute to SRE-owned portions of application codebases related to infrastructure clients, SaaS clients, observability, and reliability patterns.</li>\n<li>Contribute to the developer platform interfaces to enable a growing number of engineers, microservices, and environments (helm charts, CI platform, and deploy processes).</li>\n<li>Advocate for new tools and processes that will help Regrello grow.</li>\n<li>Take part in on-call rotations.</li>\n<li>Collaborate with cross-functional teams, including Development, QA, Product Management, to ensure successful releases.</li>\n</ul>\n<p><strong>Stack</strong></p>\n<ul>\n<li>GCP: GKE, CloudRun, Memorystore, CloudSQL, BigQuery</li>\n<li>Kubernetes: helm, helmfile</li>\n<li>Automation: Terraform, shell</li>\n<li>Queue: Temporal, Machinery, Celery</li>\n<li>Launchdarkly</li>\n<li>Otel / Prometheus / Grafana / Splunk</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Bachelor’s degree in Computer Science or a related field.</li>\n<li>4-8 years of experience in site reliability, software engineering, or a related role.</li>\n<li>Strong understanding of software development lifecycle (SDLC) and Agile methodologies.</li>\n<li>Experience with CI/CD tools such as Github Actions, GitLab CI, or CircleCI.</li>\n<li>Proficiency in scripting languages for automation tasks.</li>\n<li>Fluency with cloud platforms (AWS, Azure, GCP), kubernetes, feature flags, and modern backend technologies (experience with Go is strongly preferred, with the ability to quickly learn new technologies as needed).</li>\n<li>A builder’s spirit (you have a track record of building projects for fun, staying updated with open-source developments, etc.)</li>\n<li>Excellent problem-solving and communications skills, and attention to detail, with the ability to work effectively in a remote team environment.</li>\n</ul>\n<p><strong>Culture and Compensation</strong></p>\n<p>We are a customer-obsessed, product-driven company that is building a flexible, hybrid/remote culture to enable the brightest minds in the industry. We are particularly interested in candidates based in our hubs of Seattle, San Francisco, and New York, but we will consider candidates who live anywhere in the US, Canada, or Mexico. We have industry-leading compensation packages, including equity and health benefits. We are willing to sponsor US work authorization if needed.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9d27e558-af6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Regrello","sameAs":"https://regrello.com","logo":"https://logos.yubhub.co/regrello.com.png"},"x-apply-url":"https://jobs.lever.co/regrello/e4222908-c38b-4c4c-9067-9f66d94c0be2","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$150,000-200,000 per year","x-skills-required":["Bachelor’s degree in Computer Science or a related field","4-8 years of experience in site reliability, software engineering, or a related role","Strong understanding of software development lifecycle (SDLC) and Agile methodologies","Experience with CI/CD tools such as Github Actions, GitLab CI, or CircleCI","Proficiency in scripting languages for automation tasks","Fluency with cloud platforms (AWS, Azure, GCP), kubernetes, feature flags, and modern backend technologies (experience with Go is strongly preferred, with the ability to quickly learn new technologies as needed)","A builder’s spirit (you have a track record of building projects for fun, staying updated with open-source developments, etc.)","Excellent problem-solving and communications skills, and attention to detail, with the ability to work effectively in a remote team environment"],"x-skills-preferred":["GCP: GKE, CloudRun, Memorystore, CloudSQL, BigQuery","Kubernetes: helm, helmfile","Automation: Terraform, shell","Queue: Temporal, Machinery, Celery","Launchdarkly","Otel / Prometheus / Grafana / Splunk"],"datePosted":"2026-04-17T12:54:41.965Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Bachelor’s degree in Computer Science or a related field, 4-8 years of experience in site reliability, software engineering, or a related role, Strong understanding of software development lifecycle (SDLC) and Agile methodologies, Experience with CI/CD tools such as Github Actions, GitLab CI, or CircleCI, Proficiency in scripting languages for automation tasks, Fluency with cloud platforms (AWS, Azure, GCP), kubernetes, feature flags, and modern backend technologies (experience with Go is strongly preferred, with the ability to quickly learn new technologies as needed), A builder’s spirit (you have a track record of building projects for fun, staying updated with open-source developments, etc.), Excellent problem-solving and communications skills, and attention to detail, with the ability to work effectively in a remote team environment, GCP: GKE, CloudRun, Memorystore, CloudSQL, BigQuery, Kubernetes: helm, helmfile, Automation: Terraform, shell, Queue: Temporal, Machinery, Celery, Launchdarkly, Otel / Prometheus / Grafana / Splunk","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":150000,"maxValue":200000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dd7fb909-289"},"title":"Web Crawling Engineer","description":"<p>About Mistral AI</p>\n<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>\n<p>We are looking for a skilled and motivated Web Crawling Engineer to join our dynamic engineering team. The ideal candidate should have a solid background in distributed web crawling, scraping and data extraction, with experience using advanced tools and technologies to collect and process large-scale data from diverse web sources at large scale.</p>\n<p>Responsibilities</p>\n<p>As a Web crawling engineer, you will be responsible for:</p>\n<ul>\n<li>Developing and maintaining web crawlers using Go to extract data from target websites.</li>\n<li>Utilizing headless browsing techniques, such as Chrome DevTools, to automate and optimize data collection processes.</li>\n<li>Collaborating with cross-functional teams to identify, scrape, and integrate data from APIs and web pages to support business objectives.</li>\n<li>Creating and implementing efficient parsing patterns using tokenizers, regular expressions, XPaths, and CSS selectors to ensure accurate data extraction.</li>\n<li>Designing and managing distributed job queues using technologies such as Redis, Aerospike and Kubernetes to handle large-scale distributed crawling and processing tasks.</li>\n<li>Developing strategies to monitor and ensure data quality, accuracy, and integrity throughout the crawling and indexing process.</li>\n<li>Continuously improving and optimizing existing web crawling infrastructure to maximize efficiency and adapt to new challenges.</li>\n</ul>\n<p>About You</p>\n<p>Core programming and web technologies</p>\n<ul>\n<li>Proficiency in Go (Golang)/Rust/Zig for building scalable and efficient web crawlers.</li>\n<li>Deep understanding of TCP, UDP, TLS and HTTP/1.1,2,3 protocols and web communication.</li>\n<li>Knowledge of HTML, CSS, and JavaScript for parsing and navigating web content.</li>\n<li>Familiarity with cloud platforms (AWS, GCP), orchestration (Kubernetes, Nomad), and containerization (Docker) for deployment.</li>\n</ul>\n<p>Data Structures &amp; Algorithms</p>\n<ul>\n<li>Mastery of queues, stacks, hash maps, and other data structures for efficient data handling.</li>\n<li>Ability to design and optimize algorithms for large-scale web crawling.</li>\n</ul>\n<p>Web Scraping &amp; Data Acquisition</p>\n<ul>\n<li>Hands-on experience with networking and web scraping libraries.</li>\n<li>Understanding of how search engines work and best practices for web crawling optimization.</li>\n</ul>\n<p>Databases &amp; Data Storage</p>\n<ul>\n<li>Experience with SQL and/or NoSQL databases (knowing Aerospike is a bonus) for storing and managing crawled data.</li>\n<li>Familiarity with data warehousing and scalable storage solutions.</li>\n</ul>\n<p>Distributed Systems &amp; Big Data</p>\n<ul>\n<li>Knowledge of distributed systems (e.g., Hadoop, Spark) for processing large datasets.</li>\n</ul>\n<p>Bonus Skills (Nice-to-Have)</p>\n<ul>\n<li>Experience with web archiving projects &amp; tooling, open-source archiving is a big plus!</li>\n<li>Experience applying Machine Learning to improve crawling efficiency or accuracy.</li>\n<li>Experience with low-level networking programming and/or userspace TCP/IP stacks.</li>\n</ul>\n<p>Hiring Process</p>\n<p>Here is what you should expect:</p>\n<ul>\n<li>Introduction call - 35 min</li>\n<li>Hiring Manager Interview - 30 min</li>\n<li>Live-coding Interview - 45 min</li>\n<li>System Design Interview - 45 min</li>\n<li>Deep dive interview (optional) - 60min</li>\n<li>Culture-fit discussion - 30 min</li>\n<li>Reference checks</li>\n</ul>\n<p>Additional Information</p>\n<p>Location &amp; Remote</p>\n<p>This role is primarily based in one of our European offices , Paris, France and London, UK. We will prioritize candidates who either reside there or are open to relocating. We strongly believe in the value of in-person collaboration to foster strong relationships and seamless communication within our team. In certain specific situations, we will also consider remote candidates based in one of the countries listed in this job posting , currently France, UK, Germany, Belgium, Netherlands, Spain and Italy. In any case, we ask all new hires to visit our Paris HQ office:</p>\n<ul>\n<li>for the first week of their onboarding (accommodation and travelling covered)</li>\n<li>then at least 2 days per month</li>\n</ul>\n<p>What we offer</p>\n<p>💰 Competitive salary and equity</p>\n<p>🧑‍⚕️ Health insurance</p>\n<p>🚴 Transportation allowance</p>\n<p>🥎 Sport allowance</p>\n<p>🥕 Meal vouchers</p>\n<p>💰 Private pension plan</p>\n<p>🍼 Parental : Generous parental leave policy</p>\n<p>🌎 Visa sponsorship</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dd7fb909-289","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Mistral AI","sameAs":"https://mistral.ai","logo":"https://logos.yubhub.co/mistral.ai.png"},"x-apply-url":"https://jobs.lever.co/mistral/c96bf665-7d73-406b-8d8f-ddf8df5d160f","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Rust","Zig","TCP","UDP","TLS","HTTP/1.1","HTTP/2","HTTP/3","HTML","CSS","JavaScript","cloud platforms","orchestration","containerization","queues","stacks","hash maps","SQL","NoSQL databases","data warehousing","scalable storage solutions","distributed systems","Hadoop","Spark"],"x-skills-preferred":["web archiving projects","Machine Learning","low-level networking programming","userspace TCP/IP stacks"],"datePosted":"2026-04-17T12:48:06.790Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Paris"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Rust, Zig, TCP, UDP, TLS, HTTP/1.1, HTTP/2, HTTP/3, HTML, CSS, JavaScript, cloud platforms, orchestration, containerization, queues, stacks, hash maps, SQL, NoSQL databases, data warehousing, scalable storage solutions, distributed systems, Hadoop, Spark, web archiving projects, Machine Learning, low-level networking programming, userspace TCP/IP stacks"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_386ee13c-ffd"},"title":"Principal Backend Engineer","description":"<p>We&#39;re looking for a Senior Backend Engineer to join our LATAM engineering team. You will design and build the backend systems that power Jeeves&#39;s financial platform , working across payments, cards, spend management, and compliance infrastructure that serves businesses across the Americas and beyond.</p>\n<p>This is a backend engineering role at its core, we&#39;re looking for a strong backend engineer who knows how to work effectively with AI tools, understands where AI can accelerate development and product capabilities, and is comfortable integrating AI-powered features into production backend systems.</p>\n<p>Given the global nature of our business and the collaborative nature of our team, fluency in English is required for daily work with engineering, product, and business teams across multiple regions. Fluency in Spanish or Portuguese is equally required , our LATAM teams, customers, and operational partners work in both languages, and you will too.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Backend Engineering - Design, build, and maintain scalable, reliable backend services that process financial transactions and serve Jeeves customers across 20+ countries.</li>\n<li>Write clean, testable, production-quality code in Go, Python, or Node.js/TypeScript; participate actively in design and code reviews.</li>\n<li>Build and consume RESTful and GraphQL APIs; design inter-service communication using gRPC, message queues, and event-driven architectures.</li>\n<li>Design and optimize relational and non-relational database schemas (PostgreSQL, MongoDB, Redis) for correctness, performance, and scale.</li>\n<li>Own backend features end-to-end , from scoping and technical design through deployment, monitoring, and iteration.</li>\n<li>Implement security best practices: authentication, authorization, input validation, and data protection across distributed services.</li>\n</ul>\n<p><strong>AI-Assisted Feature Development</strong></p>\n<ul>\n<li>Integrate LLM API calls (e.g., OpenAI, Anthropic) into backend services as product features , such as spend categorization, document parsing, or natural language workflows , ensuring those integrations are reliable, observable, and cost-efficient.</li>\n<li>Build backend pipelines that consume AI-generated outputs safely: validate structured outputs, handle fallback scenarios, and design graceful degradation when AI services are unavailable or return low-confidence results.</li>\n<li>Collaborate with AI and data science teams to integrate model outputs into backend APIs , bridging experimental AI work and production systems.</li>\n<li>Use AI coding tools (GitHub Copilot, Claude, Cursor, etc.) fluently as part of your everyday development workflow.</li>\n</ul>\n<p><strong>Reliability &amp; Operations</strong></p>\n<ul>\n<li>Instrument services with structured logging, distributed tracing, and metrics for full operational visibility.</li>\n<li>Participate in on-call rotation; respond to production incidents and contribute to post-incident reviews.</li>\n<li>Contribute to CI/CD pipeline improvements, testing infrastructure, and deployment practices.</li>\n</ul>\n<p><strong>Cross-Regional Collaboration</strong></p>\n<ul>\n<li>Work closely with engineering, product, compliance, and data teams across multiple time zones and regions , communicating in both English and Spanish or Portuguese as the situation requires.</li>\n<li>Contribute to a globally distributed engineering culture through thorough documentation, async design reviews, and thoughtful pull request feedback.</li>\n<li>Bring your regional perspective to product and engineering conversations , our LATAM customers have specific needs, and engineers who understand those markets make our product better.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>5+ years of professional backend engineering experience building and operating production systems.</li>\n<li>Fluent in English , professional fluency required for daily work with global teams in written and spoken contexts.</li>\n<li>Fluent in Spanish or Portuguese , required for collaboration with LATAM teammates, customers, and operational partners.</li>\n<li>Strong proficiency in at least one backend language: Go, Python, or Node.js/TypeScript.</li>\n<li>Experience designing and building RESTful APIs, microservices, and event-driven backend systems.</li>\n<li>Solid understanding of relational databases (PostgreSQL preferred): schema design, query optimization, and data modeling.</li>\n<li>Experience with cloud infrastructure (AWS, GCP, or Azure), containerization (Docker, Kubernetes), and CI/CD pipelines.</li>\n<li>Demonstrated ability to integrate third-party APIs reliably in production , including error handling, retry logic, and observability.</li>\n<li>Experience working on globally distributed teams across time zones and regions.</li>\n<li>Comfortable using AI tools as part of everyday engineering work , integrating LLM API outputs into backend services and using AI coding assistants fluently.</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Experience in fintech, financial services, payments, or a regulated industry , familiarity with ledger systems, payment rails, or financial compliance (KYC/AML, PCI-DSS) is a strong plus.</li>\n<li>Prior experience at a startup or high-growth scale-up, comfortable building in ambiguity without heavy process support.</li>\n<li>Experience with multi-currency systems or cross-border payment processing.</li>\n<li>Familiarity with message queue systems (Kafka, RabbitMQ) and event-driven architecture.</li>\n<li>Global work experience , prior roles at companies operating across multiple countries and regulatory environments.</li>\n<li>Fluency in both Spanish and Portuguese.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_386ee13c-ffd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Jeeves","sameAs":"https://www.jeeves.com/","logo":"https://logos.yubhub.co/jeeves.com.png"},"x-apply-url":"https://jobs.lever.co/tryjeeves/6cfaf109-e538-45cd-bd0f-ed0bc360fc7f","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Python","Node.js/TypeScript","RESTful APIs","microservices","event-driven backend systems","relational databases","cloud infrastructure","containerization","CI/CD pipelines","third-party APIs","AI tools","LLM API outputs","backend services","financial transactions","cross-border payments","compliance infrastructure"],"x-skills-preferred":["fintech","financial services","payments","regulated industry","ledger systems","payment rails","financial compliance","multi-currency systems","cross-border payment processing","message queue systems","event-driven architecture","global work experience"],"datePosted":"2026-04-17T12:39:56.666Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Brazil"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Go, Python, Node.js/TypeScript, RESTful APIs, microservices, event-driven backend systems, relational databases, cloud infrastructure, containerization, CI/CD pipelines, third-party APIs, AI tools, LLM API outputs, backend services, financial transactions, cross-border payments, compliance infrastructure, fintech, financial services, payments, regulated industry, ledger systems, payment rails, financial compliance, multi-currency systems, cross-border payment processing, message queue systems, event-driven architecture, global work experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1946e60d-b41"},"title":"Principal Backend Engineer","description":"<p>Jeeves is looking for a Senior Backend Engineer to join our LATAM engineering team. You will design and build the backend systems that power Jeeves&#39;s financial platform , working across payments, cards, spend management, and compliance infrastructure that serves businesses across the Americas and beyond.</p>\n<p>This is a backend engineering role at its core, we&#39;re looking for a strong backend engineer who knows how to work effectively with AI tools, understands where AI can accelerate development and product capabilities, and is comfortable integrating AI-powered features into production backend systems.</p>\n<p>Given the global nature of our business and the collaborative nature of our team, fluency in English is required for daily work with engineering, product, and business teams across multiple regions. Fluency in Spanish or Portuguese is equally required , our LATAM teams, customers, and operational partners work in both languages, and you will too.</p>\n<p><strong>Backend Engineering</strong></p>\n<ul>\n<li>Design, build, and maintain scalable, reliable backend services that process financial transactions and serve Jeeves customers across 20+ countries.</li>\n<li>Write clean, testable, production-quality code in Go, Python, or Node.js/TypeScript; participate actively in design and code reviews.</li>\n<li>Build and consume RESTful and GraphQL APIs; design inter-service communication using gRPC, message queues, and event-driven architectures.</li>\n<li>Design and optimize relational and non-relational database schemas (PostgreSQL, MongoDB, Redis) for correctness, performance, and scale.</li>\n<li>Own backend features end-to-end , from scoping and technical design through deployment, monitoring, and iteration.</li>\n<li>Implement security best practices: authentication, authorization, input validation, and data protection across distributed services.</li>\n</ul>\n<p><strong>AI-Assisted Feature Development</strong></p>\n<ul>\n<li>Integrate LLM API calls (e.g., OpenAI, Anthropic) into backend services as product features , such as spend categorization, document parsing, or natural language workflows , ensuring those integrations are reliable, observable, and cost-efficient.</li>\n<li>Build backend pipelines that consume AI-generated outputs safely: validate structured outputs, handle fallback scenarios, and design graceful degradation when AI services are unavailable or return low-confidence results.</li>\n<li>Collaborate with AI and data science teams to integrate model outputs into backend APIs , bridging experimental AI work and production systems.</li>\n<li>Use AI coding tools (GitHub Copilot, Claude, Cursor, etc.) fluently as part of your everyday development workflow.</li>\n</ul>\n<p><strong>Reliability &amp; Operations</strong></p>\n<ul>\n<li>Instrument services with structured logging, distributed tracing, and metrics for full operational visibility.</li>\n<li>Participate in on-call rotation; respond to production incidents and contribute to post-incident reviews.</li>\n<li>Contribute to CI/CD pipeline improvements, testing infrastructure, and deployment practices.</li>\n</ul>\n<p><strong>Cross-Regional Collaboration</strong></p>\n<ul>\n<li>Work closely with engineering, product, compliance, and data teams across multiple time zones and regions , communicating in both English and Spanish or Portuguese as the situation requires.</li>\n<li>Contribute to a globally distributed engineering culture through thorough documentation, async design reviews, and thoughtful pull request feedback.</li>\n<li>Bring your regional perspective to product and engineering conversations , our LATAM customers have specific needs, and engineers who understand those markets make our product better.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1946e60d-b41","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Jeeves","sameAs":"https://www.jeeves.com/","logo":"https://logos.yubhub.co/jeeves.com.png"},"x-apply-url":"https://jobs.lever.co/tryjeeves/3bc7c001-f114-414d-a65a-63519eec59e6","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Python","Node.js/TypeScript","RESTful APIs","microservices","event-driven backend systems","relational databases","cloud infrastructure","containerization","CI/CD pipelines","security best practices","authentication","authorization","input validation","data protection"],"x-skills-preferred":["fintech","financial services","payments","regulated industry","ledger systems","payment rails","financial compliance","KYC/AML","PCI-DSS","multi-currency systems","cross-border payment processing","message queue systems","event-driven architecture"],"datePosted":"2026-04-17T12:39:32.789Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Argentina"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Go, Python, Node.js/TypeScript, RESTful APIs, microservices, event-driven backend systems, relational databases, cloud infrastructure, containerization, CI/CD pipelines, security best practices, authentication, authorization, input validation, data protection, fintech, financial services, payments, regulated industry, ledger systems, payment rails, financial compliance, KYC/AML, PCI-DSS, multi-currency systems, cross-border payment processing, message queue systems, event-driven architecture"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_306ceadd-888"},"title":"Principal Backend Engineer","description":"<p>We&#39;re looking for a Senior Backend Engineer to join our LATAM engineering team. You will design and build the backend systems that power Jeeves&#39;s financial platform , working across payments, cards, spend management, and compliance infrastructure that serves businesses across the Americas and beyond.</p>\n<p>This is a backend engineering role at its core, we&#39;re looking for a strong backend engineer who knows how to work effectively with AI tools, understands where AI can accelerate development and product capabilities, and is comfortable integrating AI-powered features into production backend systems.</p>\n<p>Given the global nature of our business and the collaborative nature of our team, fluency in English is required for daily work with engineering, product, and business teams across multiple regions. Fluency in Spanish or Portuguese is equally required , our LATAM teams, customers, and operational partners work in both languages, and you will too.</p>\n<p>As a Senior Backend Engineer, you will:</p>\n<ul>\n<li><p>Design, build, and maintain scalable, reliable backend services that process financial transactions and serve Jeeves customers across 20+ countries.</p>\n</li>\n<li><p>Write clean, testable, production-quality code in Go, Python, or Node.js/TypeScript; participate actively in design and code reviews.</p>\n</li>\n<li><p>Build and consume RESTful and GraphQL APIs; design inter-service communication using gRPC, message queues, and event-driven architectures.</p>\n</li>\n<li><p>Design and optimize relational and non-relational database schemas (PostgreSQL, MongoDB, Redis) for correctness, performance, and scale.</p>\n</li>\n<li><p>Own backend features end-to-end , from scoping and technical design through deployment, monitoring, and iteration.</p>\n</li>\n<li><p>Implement security best practices: authentication, authorization, input validation, and data protection across distributed services.</p>\n</li>\n<li><p>Integrate LLM API calls (e.g., OpenAI, Anthropic) into backend services as product features , such as spend categorization, document parsing, or natural language workflows , ensuring those integrations are reliable, observable, and cost-efficient.</p>\n</li>\n<li><p>Build backend pipelines that consume AI-generated outputs safely: validate structured outputs, handle fallback scenarios, and design graceful degradation when AI services are unavailable or return low-confidence results.</p>\n</li>\n<li><p>Collaborate with AI and data science teams to integrate model outputs into backend APIs , bridging experimental AI work and production systems.</p>\n</li>\n<li><p>Use AI coding tools (GitHub Copilot, Claude, Cursor, etc.) fluently as part of your everyday development workflow.</p>\n</li>\n<li><p>Instrument services with structured logging, distributed tracing, and metrics for full operational visibility.</p>\n</li>\n<li><p>Participate in on-call rotation; respond to production incidents and contribute to post-incident reviews.</p>\n</li>\n<li><p>Contribute to CI/CD pipeline improvements, testing infrastructure, and deployment practices.</p>\n</li>\n<li><p>Work closely with engineering, product, compliance, and data teams across multiple time zones and regions , communicating in both English and Spanish or Portuguese as the situation requires.</p>\n</li>\n<li><p>Contribute to a globally distributed engineering culture through thorough documentation, async design reviews, and thoughtful pull request feedback.</p>\n</li>\n<li><p>Bring your regional perspective to product and engineering conversations , our LATAM customers have specific needs, and engineers who understand those markets make our product better.</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_306ceadd-888","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Jeeves","sameAs":"https://www.jeeves.com/","logo":"https://logos.yubhub.co/jeeves.com.png"},"x-apply-url":"https://jobs.lever.co/tryjeeves/e73e6e7c-a787-4d3d-bf2d-bdf8ce56d2c5","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Python","Node.js/TypeScript","RESTful APIs","GraphQL APIs","gRPC","message queues","event-driven architectures","relational databases","non-relational databases","security best practices","authentication","authorization","input validation","data protection"],"x-skills-preferred":["fintech","financial services","payments","regulated industry","ledger systems","payment rails","financial compliance","KYC/AML","PCI-DSS","multi-currency systems","cross-border payment processing","message queue systems","event-driven architecture","global work experience"],"datePosted":"2026-04-17T12:38:13.894Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Colombia"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Go, Python, Node.js/TypeScript, RESTful APIs, GraphQL APIs, gRPC, message queues, event-driven architectures, relational databases, non-relational databases, security best practices, authentication, authorization, input validation, data protection, fintech, financial services, payments, regulated industry, ledger systems, payment rails, financial compliance, KYC/AML, PCI-DSS, multi-currency systems, cross-border payment processing, message queue systems, event-driven architecture, global work experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_71befc94-b2f"},"title":"Principal Backend Engineer","description":"<p>We&#39;re looking for a Senior Backend Engineer to join our LATAM engineering team. You will design and build the backend systems that power Jeeves&#39;s financial platform , working across payments, cards, spend management, and compliance infrastructure that serves businesses across the Americas and beyond.</p>\n<p>This is a backend engineering role at its core, we&#39;re looking for a strong backend engineer who knows how to work effectively with AI tools, understands where AI can accelerate development and product capabilities, and is comfortable integrating AI-powered features into production backend systems.</p>\n<p>Given the global nature of our business and the collaborative nature of our team, fluency in English is required for daily work with engineering, product, and business teams across multiple regions. Fluency in Spanish or Portuguese is equally required , our LATAM teams, customers, and operational partners work in both languages, and you will too.</p>\n<p><strong>Backend Engineering</strong></p>\n<ul>\n<li>Design, build, and maintain scalable, reliable backend services that process financial transactions and serve Jeeves customers across 20+ countries.</li>\n<li>Write clean, testable, production-quality code in Go, Python, or Node.js/TypeScript; participate actively in design and code reviews.</li>\n<li>Build and consume RESTful and GraphQL APIs; design inter-service communication using gRPC, message queues, and event-driven architectures.</li>\n<li>Design and optimize relational and non-relational database schemas (PostgreSQL, MongoDB, Redis) for correctness, performance, and scale.</li>\n<li>Own backend features end-to-end , from scoping and technical design through deployment, monitoring, and iteration.</li>\n<li>Implement security best practices: authentication, authorization, input validation, and data protection across distributed services.</li>\n</ul>\n<p><strong>AI-Assisted Feature Development</strong></p>\n<ul>\n<li>Integrate LLM API calls (e.g., OpenAI, Anthropic) into backend services as product features , such as spend categorization, document parsing, or natural language workflows , ensuring those integrations are reliable, observable, and cost-efficient.</li>\n<li>Build backend pipelines that consume AI-generated outputs safely: validate structured outputs, handle fallback scenarios, and design graceful degradation when AI services are unavailable or return low-confidence results.</li>\n<li>Collaborate with AI and data science teams to integrate model outputs into backend APIs , bridging experimental AI work and production systems.</li>\n<li>Use AI coding tools (GitHub Copilot, Claude, Cursor, etc.) fluently as part of your everyday development workflow.</li>\n</ul>\n<p><strong>Reliability &amp; Operations</strong></p>\n<ul>\n<li>Instrument services with structured logging, distributed tracing, and metrics for full operational visibility.</li>\n<li>Participate in on-call rotation; respond to production incidents and contribute to post-incident reviews.</li>\n<li>Contribute to CI/CD pipeline improvements, testing infrastructure, and deployment practices.</li>\n</ul>\n<p><strong>Cross-Regional Collaboration</strong></p>\n<ul>\n<li>Work closely with engineering, product, compliance, and data teams across multiple time zones and regions , communicating in both English and Spanish or Portuguese as the situation requires.</li>\n<li>Contribute to a globally distributed engineering culture through thorough documentation, async design reviews, and thoughtful pull request feedback.</li>\n<li>Bring your regional perspective to product and engineering conversations , our LATAM customers have specific needs, and engineers who understand those markets make our product better.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_71befc94-b2f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Jeeves","sameAs":"https://www.jeeves.com/","logo":"https://logos.yubhub.co/jeeves.com.png"},"x-apply-url":"https://jobs.lever.co/tryjeeves/c1a4c8bd-0be1-48d9-bdaa-4f37a1e21fc4","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Python","Node.js/TypeScript","RESTful APIs","GraphQL APIs","gRPC","message queues","event-driven architectures","relational databases","non-relational databases","security best practices","authentication","authorization","input validation","data protection"],"x-skills-preferred":["fintech","financial services","payments","regulated industry","ledger systems","payment rails","financial compliance","KYC/AML","PCI-DSS","multi-currency systems","cross-border payment processing","message queue systems","event-driven architecture"],"datePosted":"2026-04-17T12:38:03.434Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mexico"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Go, Python, Node.js/TypeScript, RESTful APIs, GraphQL APIs, gRPC, message queues, event-driven architectures, relational databases, non-relational databases, security best practices, authentication, authorization, input validation, data protection, fintech, financial services, payments, regulated industry, ledger systems, payment rails, financial compliance, KYC/AML, PCI-DSS, multi-currency systems, cross-border payment processing, message queue systems, event-driven architecture"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ac012549-95e"},"title":"Fraud Operations Manager","description":"<p>We are hiring a Fraud Operations Manager to lead our Fraud Operations Centre in Bangalore. As a people leader, you will manage team leads and senior associates, setting clear expectations, coaching performance, and building a high-judgement culture rooted in strong decisioning and customer empathy.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Lead the Bangalore Fraud Operations Centre, managing team leads and senior associates</li>\n<li>Own core fraud investigations and complex investigations, including escalations and time-sensitive incident response in partnership with global Fraud/Risk stakeholders</li>\n<li>Establish and run the operating system for the org: goal setting, team cadences, SOPs, change management, queue health, SLAs, QA programs, calibration, and readiness for new products/markets</li>\n<li>Define, develop, and represent key operational and risk metrics for the Bangalore centre (through dashboards and recurring readouts), translating data into actions and priorities</li>\n<li>Drive process development and continuous improvement: root cause analysis, defect reduction, policy adherence, and consistency across reviewers and teams</li>\n<li>Partner with other Risk teams to evolve fraud tooling and workflows (manual and automated actioning), including safe rollout and measurement of changes</li>\n<li>Support outsourcing relevant workflows: identify transition candidates, design/standardize processes for hand off, and support training as needed</li>\n<li>Own staffing models, capacity planning, scheduling, hiring plans, onboarding, and training strategy to meet business needs while maintaining a high bar for quality and user experience</li>\n<li>Maintain close “floor” engagement: regularly dive into cases, review decisions, support frontline judgment, and ensure feedback loops translate into measurable improvements</li>\n<li>Partner with recruiting to define role profiles and interview loops, drive bar-raising hiring decisions, and build a diverse team</li>\n<li>Own career development and succession planning, team health and engagement</li>\n<li>Build a great culture, a destination workplace, fostering our values, serving as a beacon of Stripe’s user-centric philosophy and culture of transparency, empathy, inclusion, and empowerment</li>\n</ul>\n<p>You are an operations leader who can scale fraud investigation and review programs while staying close to the details. You are fluent in running high-performing teams through clear metrics and operating cadence, and you can partner cross-functionally to evolve tooling and processes. You bring sound judgment, calm execution under pressure, and a bias toward data-backed solutions and measurable improvements in both fraud outcomes and user experience.</p>\n<p>Minimum requirements:</p>\n<ul>\n<li>Experience leading cross-functional initiatives and driving change through ambiguity, including new workflows for new products, markets, or emerging fraud patterns</li>\n<li>Experience leading fraud, risk, trust &amp; safety, or investigations operations in a high-scale environment, including managing managers/team leads</li>\n<li>Ability to partner effectively with internal globally distributed stakeholders and comfortable leading internal teams and vendors</li>\n<li>Proven ability to define and operationalize metrics (e.g., accuracy/quality, SLAs, throughput, loss outcomes, appeals/rework, user friction) and use data to drive priorities</li>\n<li>Comfort working across detection systems and manual/automated actioning systems; ability to translate operational needs into product/engineering requirements</li>\n<li>Strong operational background including new process launches, service delivery and strong at building operating models</li>\n<li>Excellent written and verbal communication skills, including presenting in operational reviews and representing the team in high-pressure escalations in English</li>\n<li>Skill and credibility doing the core work of a user-facing team with a high bar for quality, and a willingness to lead by example</li>\n<li>At least 7+ years of direct people management experience; this includes handling performance management and proactively coach/mentor individuals</li>\n<li>Experience with payments, risk, compliance, or financial industry operations</li>\n</ul>\n<p>Preferred qualifications:</p>\n<ul>\n<li>Demonstrated ability to run an operations centre: capacity planning, scheduling, queue management, performance management, and quality programs</li>\n<li>Strong fraud domain depth across investigation workflows, risk vectors, and decision frameworks; able to independently review cases and guide frontline judgment</li>\n<li>Experience scaling or transforming fraud review operations through process redesign, tooling improvements, and/or vendor partnerships</li>\n<li>Experience with payments risk (CNP/card testing, ATO, merchant fraud, transaction fraud) and familiarity with how upstream detection and downstream enforcement interact</li>\n<li>Strong analytics proficiency (building/owning dashboards; querying data directly) and comfort partnering deeply with data science/engineering</li>\n<li>Experience operating in a global, follow-the-sun model and partnering across time zones</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ac012549-95e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com/","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7608396","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["fraud investigation","risk management","operations leadership","team management","process development","continuous improvement","data analysis","fraud tooling","workflow management","capacity planning","scheduling","queue management","performance management","quality programs","operating models","payments risk","compliance","financial industry operations"],"x-skills-preferred":["fraud domain depth","investigation workflows","risk vectors","decision frameworks","process redesign","tooling improvements","vendor partnerships","CNP/card testing","ATO","merchant fraud","transaction fraud","analytics proficiency","data science","engineering"],"datePosted":"2026-03-31T18:15:48.331Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Technology","skills":"fraud investigation, risk management, operations leadership, team management, process development, continuous improvement, data analysis, fraud tooling, workflow management, capacity planning, scheduling, queue management, performance management, quality programs, operating models, payments risk, compliance, financial industry operations, fraud domain depth, investigation workflows, risk vectors, decision frameworks, process redesign, tooling improvements, vendor partnerships, CNP/card testing, ATO, merchant fraud, transaction fraud, analytics proficiency, data science, engineering"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3437e4dc-7d6"},"title":"后端工程师 (Kotlin) - 深圳, 中国 (Senior)","description":"<p><strong>Job Overview</strong></p>\n<p>We are seeking a senior backend engineer to join our team in Shenzhen, China. As a senior backend engineer, you will be responsible for designing, developing, and optimizing our core backend systems, including payment gateways, settlement systems, mobile POS integrations, and financial service APIs.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, develop, and optimize high-performance backend services and APIs (REST/gRPC) using Kotlin or Java</li>\n<li>Participate in backend architecture design to ensure system high availability and scalability</li>\n<li>Troubleshoot and optimize system performance issues to ensure system stability</li>\n<li>Collaborate closely with frontend, product, and testing teams to drive project delivery</li>\n<li>Write high-quality, maintainable code and participate in code reviews</li>\n<li>Stay up-to-date with the latest trends in backend and cloud technologies</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>5+ years of experience in backend development (Kotlin or Java)</li>\n<li>Familiarity with microservices architecture, cloud deployment (AWS/GCP), and CI/CD pipelines (GitHub Actions)</li>\n<li>Proficiency in SQL/NoSQL databases (PostgreSQL, MySQL, Redis)</li>\n<li>Familiarity with Docker, Kubernetes, and message queue systems (Kafka)</li>\n<li>Familiarity with API design (REST/gRPC) and version control (GitHub)</li>\n<li>Familiarity with agile development processes and good team collaboration skills</li>\n<li>Good English communication skills (Mandarin or Cantonese is a plus)</li>\n<li>Experience in financial technology, payment, or settlement systems is a plus</li>\n</ul>\n<p><strong>Preferred Skills</strong></p>\n<ul>\n<li>High concurrency and large-scale system development experience</li>\n<li>Understanding of DevOps, CI/CD pipelines, and ability to drive automation deployment and operations</li>\n<li>Experience in distributed systems or data architecture design</li>\n<li>Contributions to open-source communities or personal technical blogs</li>\n</ul>\n<p><strong>Why Join Kody?</strong></p>\n<ul>\n<li>Global technology company with offices in Singapore, London, and Hong Kong</li>\n<li>Flexible work arrangements in Shenzhen and Hong Kong</li>\n<li>Technology-driven culture with engineers having core influence on product decisions</li>\n<li>Challenging projects, including large-scale backend architecture design and optimization</li>\n<li>Competitive salary and benefits to reward your technical contributions</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3437e4dc-7d6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Kody","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/nt225i4DCrp5uzosmH4FXU/%E5%90%8E%E7%AB%AF%E5%B7%A5%E7%A8%8B%E5%B8%88-(kotlin)---%E6%B7%B1%E5%9C%B3%2C-%E4%B8%AD%E5%9B%BD-(senior)-in-shenzhen-at-kody","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kotlin","Java","microservices architecture","cloud deployment","CI/CD pipelines","SQL/NoSQL databases","Docker","Kubernetes","message queue systems","API design","version control","agile development processes"],"x-skills-preferred":["high concurrency and large-scale system development experience","DevOps","CI/CD pipelines","distributed systems or data architecture design","open-source communities","personal technical blogs"],"datePosted":"2026-03-09T17:01:45.087Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"深圳, 中国"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Kotlin, Java, microservices architecture, cloud deployment, CI/CD pipelines, SQL/NoSQL databases, Docker, Kubernetes, message queue systems, API design, version control, agile development processes, high concurrency and large-scale system development experience, DevOps, CI/CD pipelines, distributed systems or data architecture design, open-source communities, personal technical blogs"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_555d0422-965"},"title":"后端工程师 (Kotlin) - 深圳, 中国","description":"<p><strong><strong>Job Overview</strong></strong></p>\n<p>We are seeking a senior Kotlin backend engineer to join our team in Shenzhen, China. As a backend engineer, you will be responsible for designing, developing, and optimizing our backend services to support high-traffic and high-performance applications.</p>\n<p><strong><strong>Responsibilities</strong></strong></p>\n<ul>\n<li><strong>Backend Development</strong>: Design, develop, and optimize backend APIs and services using Kotlin (or Java).</li>\n<li><strong>Architecture Design</strong>: Participate in system architecture design to ensure code quality, scalability, and high performance.</li>\n<li><strong>Database Management</strong>: Optimize SQL/NoSQL databases to improve data access efficiency.</li>\n<li><strong>API Design and Development</strong>: Build RESTful or gRPC APIs to integrate with frontend and mobile teams.</li>\n<li><strong>Microservices and Cloud Deployment</strong>: Develop scalable microservices architecture and deploy on Docker, Kubernetes, AWS/GCP.</li>\n<li><strong>Performance Optimization</strong>: Identify and resolve system performance bottlenecks to improve application efficiency.</li>\n<li><strong>Testing and Code Quality</strong>: Write unit tests, integration tests to ensure code stability.</li>\n<li><strong>Cross-Team Collaboration</strong>: Work closely with frontend engineers, product managers, QA teams to drive product delivery.</li>\n</ul>\n<p><strong><strong>Requirements</strong></strong></p>\n<ul>\n<li><strong>3+ years of experience in Kotlin or Java backend development</strong>, familiar with Spring Boot, Ktor, or similar frameworks.</li>\n<li>Familiar with PostgreSQL, MySQL, MongoDB, or Redis for database management and optimization.</li>\n<li>Familiar with RESTful API, GraphQL, or gRPC design and development.</li>\n<li>Familiar with microservices architecture, Docker, Kubernetes, and cloud deployment (AWS, GCP, or Azure).</li>\n<li>Familiar with multi-threading, asynchronous programming, and message queues (Kafka, RabbitMQ) for performance optimization.</li>\n<li><strong>Good English communication skills</strong> to collaborate with international teams.</li>\n<li>Currently residing in Shenzhen or surrounding areas, with regular office work and occasional trips to Hong Kong.</li>\n</ul>\n<p><strong><strong>Preferred Skills</strong></strong></p>\n<ul>\n<li>Experience with high-concurrency and large-scale system development.</li>\n<li>Understanding of DevOps, CI/CD pipelines to drive automation deployment and operations.</li>\n<li>Experience with distributed systems or data architecture design.</li>\n<li>Contributions to open-source projects or personal technical blogs.</li>\n</ul>\n<p><strong><strong>Why Join Kody?</strong></strong></p>\n<ul>\n<li><strong>Global fintech company</strong> with offices in Singapore, London, and Hong Kong.</li>\n<li><strong>Remote work + Hong Kong office</strong> with flexible work arrangements.</li>\n<li><strong>Technology-driven culture</strong> with engineers having core influence in product decisions.</li>\n<li><strong>Challenging projects</strong> with large-scale backend architecture design and optimization.</li>\n<li><strong>Competitive salary and benefits</strong> to reward your technical contributions.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_555d0422-965","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Kody","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/j11pr7wKpF3kZfgRFDHhpD/%E5%90%8E%E7%AB%AF%E5%B7%A5%E7%A8%8B%E5%B8%88-(kotlin)---%E6%B7%B1%E5%9C%B3%2C-%E4%B8%AD%E5%9B%BD-in-shenzhen-at-kody","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kotlin","Java","Spring Boot","Ktor","PostgreSQL","MySQL","MongoDB","Redis","RESTful API","GraphQL","gRPC","Docker","Kubernetes","AWS","GCP","Azure","multi-threading","asynchronous programming","message queues"],"x-skills-preferred":["high-concurrency","large-scale system development","DevOps","CI/CD pipelines","distributed systems","data architecture design","open-source projects","personal technical blogs"],"datePosted":"2026-03-09T17:00:04.481Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"深圳, 中国"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Kotlin, Java, Spring Boot, Ktor, PostgreSQL, MySQL, MongoDB, Redis, RESTful API, GraphQL, gRPC, Docker, Kubernetes, AWS, GCP, Azure, multi-threading, asynchronous programming, message queues, high-concurrency, large-scale system development, DevOps, CI/CD pipelines, distributed systems, data architecture design, open-source projects, personal technical blogs"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b06e2fe0-1bd"},"title":"后端工程师 (Kotlin)","description":"<p><strong><strong>Job Description</strong></strong></p>\n<p>We are seeking a senior Kotlin backend engineer to join our team. As a backend engineer, you will be responsible for designing, developing, and optimizing backend services to support high-traffic and high-performance applications.</p>\n<p><strong><strong>Responsibilities</strong></strong></p>\n<ul>\n<li><strong>Backend Development</strong>: Design, develop, and optimize backend APIs and services using Kotlin (or Java).</li>\n<li><strong>Architecture Design</strong>: Participate in system architecture design to ensure code quality, scalability, and high performance.</li>\n<li><strong>Database Management</strong>: Optimize SQL/NoSQL databases to improve data access efficiency.</li>\n<li><strong>API Design and Development</strong>: Develop RESTful or gRPC APIs to integrate with frontend and mobile teams.</li>\n<li><strong>Microservices and Cloud Deployment</strong>: Develop scalable microservices architecture and deploy using Docker, Kubernetes, AWS/GCP.</li>\n<li><strong>Performance Optimization</strong>: Identify and resolve system performance bottlenecks to improve application efficiency.</li>\n<li><strong>Testing and Code Quality</strong>: Write unit tests, integration tests to ensure code stability.</li>\n<li><strong>Cross-Team Collaboration</strong>: Work closely with frontend engineers, product managers, QA teams to drive product delivery.</li>\n</ul>\n<p><strong><strong>Requirements</strong></strong></p>\n<ul>\n<li><strong>3+ years of experience in Kotlin or Java backend development</strong>, familiar with Spring Boot, Ktor, or similar frameworks.</li>\n<li>Familiar with PostgreSQL, MySQL, MongoDB, or Redis for database management and optimization.</li>\n<li>Familiar with RESTful API, GraphQL, or gRPC design and development.</li>\n<li>Familiar with microservices architecture, Docker, Kubernetes, and cloud deployment (AWS, GCP, or Azure).</li>\n<li>Familiar with multi-threading, asynchronous programming, and message queues (Kafka, RabbitMQ) for performance optimization.</li>\n<li><strong>Good English communication skills</strong> to collaborate with international teams.</li>\n<li>Currently residing in Shenzhen or surrounding areas, with regular office work and occasional trips to Hong Kong.</li>\n</ul>\n<p><strong><strong>Preferred Skills</strong></strong></p>\n<ul>\n<li>Experience with high-concurrency and large-scale system development.</li>\n<li>Understanding of DevOps, CI/CD pipelines to drive automation deployment and operations.</li>\n<li>Experience with distributed systems or data architecture design.</li>\n<li>Contributions to open-source projects or personal technical blogs.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b06e2fe0-1bd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Kody","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/gt1SZS2RXmB6Dmj8oD99Uj/%E5%90%8E%E7%AB%AF%E5%B7%A5%E7%A8%8B%E5%B8%88-(kotlin)---%E6%B7%B1%E5%9C%B3%2C-%E4%B8%AD%E5%9B%BD-in-shenzhen-at-kody","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kotlin","Java","Spring Boot","Ktor","PostgreSQL","MySQL","MongoDB","Redis","RESTful API","GraphQL","gRPC","Docker","Kubernetes","AWS","GCP","Azure","Multi-threading","Asynchronous programming","Message queues"],"x-skills-preferred":["High-concurrency system development","DevOps","CI/CD pipelines","Distributed systems","Data architecture design"],"datePosted":"2026-03-09T16:59:50.050Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"深圳, 中国"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kotlin, Java, Spring Boot, Ktor, PostgreSQL, MySQL, MongoDB, Redis, RESTful API, GraphQL, gRPC, Docker, Kubernetes, AWS, GCP, Azure, Multi-threading, Asynchronous programming, Message queues, High-concurrency system development, DevOps, CI/CD pipelines, Distributed systems, Data architecture design"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_055a769a-c68"},"title":"后端工程师 (Kotlin) - 深圳, 中国 (Senior)","description":"<p><strong>Job Overview</strong></p>\n<p>As a Senior Backend Engineer at Kody, you will be responsible for designing, developing, and optimizing our core backend systems, including payment gateways, settlement systems, mobile POS integrations, and financial service APIs.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, develop, and optimize high-performance backend services and APIs (REST/gRPC) using Kotlin or Java</li>\n<li>Participate in backend architecture design to ensure system high availability and scalability</li>\n<li>Troubleshoot and optimize system performance issues to ensure system stability</li>\n<li>Collaborate closely with frontend, product, and testing teams to drive project delivery</li>\n<li>Write high-quality, maintainable code and participate in code reviews</li>\n<li>Stay up-to-date with the latest developments in backend and cloud technologies</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>5+ years of experience in backend development (Kotlin or Java)</li>\n<li>Familiarity with microservices architecture, cloud deployment (AWS/GCP), and CI/CD pipelines (GitHub Actions)</li>\n<li>Familiarity with SQL/NoSQL databases (PostgreSQL, MySQL, Redis)</li>\n<li>Familiarity with Docker, Kubernetes, and message queue systems (Kafka)</li>\n<li>Familiarity with API design (REST/gRPC) and version control (GitHub)</li>\n<li>Familiarity with agile development processes and good team collaboration skills</li>\n<li>Good English communication skills (Mandarin or Cantonese is a plus)</li>\n<li>Experience in financial technology, payment, or settlement systems is a plus</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Experience in high-concurrency, large-scale system development</li>\n<li>Understanding of DevOps, CI/CD pipelines, and ability to drive automation deployment and operations</li>\n<li>Experience in distributed systems or data architecture design</li>\n<li>Contributions to open-source communities or personal technical blogs</li>\n</ul>\n<p><strong>Why Join Kody?</strong></p>\n<ul>\n<li>Global technology company with offices in Singapore, London, and Hong Kong</li>\n<li>Flexible work arrangements in Shenzhen and Hong Kong</li>\n<li>Technology-driven culture with engineers having core influence on product decisions</li>\n<li>Challenging projects, including large-scale backend architecture design and optimization</li>\n<li>Competitive salary and benefits to reward your technical contributions</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_055a769a-c68","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Kody","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/2mZgGN3RMgykf9oavxvD49/%E5%90%8E%E7%AB%AF%E5%B7%A5%E7%A8%8B%E5%B8%88-(kotlin)---%E6%B7%B1%E5%9C%B3%2C-%E4%B8%AD%E5%9B%BD-(senior)-in-shenzhen-at-kody","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kotlin","Java","microservices architecture","cloud deployment","CI/CD pipelines","SQL/NoSQL databases","Docker","Kubernetes","message queue systems","API design","version control","agile development processes"],"x-skills-preferred":["high-concurrency","large-scale system development","DevOps","CI/CD pipelines","distributed systems","data architecture design","open-source communities","personal technical blogs"],"datePosted":"2026-03-09T16:58:56.332Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"深圳, 中国"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Kotlin, Java, microservices architecture, cloud deployment, CI/CD pipelines, SQL/NoSQL databases, Docker, Kubernetes, message queue systems, API design, version control, agile development processes, high-concurrency, large-scale system development, DevOps, CI/CD pipelines, distributed systems, data architecture design, open-source communities, personal technical blogs"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7b60b203-14a"},"title":"Integrations Platform Administrator","description":"<p><strong>Why Valvoline Global Operations?</strong></p>\n<p>At Valvoline Global Operations, we&#39;re proud to be The Original Motor Oil, but we&#39;ve never rested on being first. Founded in 1866, we introduced the world&#39;s first branded motor oil, staking our claim as a pioneer in the automotive and industrial solutions industry. Today, as an affiliate of Aramco, one of the world&#39;s largest integrated energy and chemicals companies, we are driven by innovation and committed to creating sustainable solutions for a better future.</p>\n<p>With a global presence, we develop future-ready products and provide best-in-class services for our partners around the world. For us, originality isn&#39;t just about where we began; it&#39;s about where we&#39;re headed and how we&#39;ll lead the way. We are originality in motion.</p>\n<p>Our corporate values—Care, Integrity, Passion, Unity, and Excellence—are at the heart of everything we do. These values define how we operate, how we treat one another, and how we engage with our partners, customers, and the communities we serve. At Valvoline Global, we are united in our commitment to:</p>\n<ul>\n<li>Treating everyone with care.</li>\n<li>Acting with unwavering integrity.</li>\n<li>Striving for excellence in all endeavors.</li>\n<li>Delivering on our commitments with passion.</li>\n<li>Collaborating as one unified team.</li>\n</ul>\n<p><strong>Job Purpose</strong></p>\n<p>The Senior Platform Engineer – MuleSoft Platform &amp; Ops Automation is responsible for the stability, security, scalability, and automation of the MuleSoft Anypoint Platform across cloud and hybrid environments. This role owns the operational health of enterprise integration services, ensuring reliable API and integration execution while reducing manual effort through automation. The role combines deep platform administration expertise with DevOps and operational leadership, acting as the escalation point for complex incidents and platform design decisions. Working closely with architecture, application engineering, DevOps, and security teams, this role ensures MuleSoft remains a resilient, compliant, and high-performing enterprise integration platform.</p>\n<p><strong>How You Make an Impact (Job Accountabilities)</strong></p>\n<ul>\n<li>Own the operational stability, performance, and availability of the MuleSoft Anypoint Platform across production and non-production environments</li>\n<li>Administer and optimize MuleSoft Runtime Manager, API Manager, Access Management, AnypointMQ, RPA, Intelligent Document Processing, Partner Manager, and Runtime Fabric or CloudHub deployments</li>\n<li>Administer and optimize use of Design Center, Exchange, and Anypoint Experience Hub for presentation and consumption of APIs to internal and external audiences.</li>\n<li>Monitor, report, and forecast on MuleSoft platform consumption to ensure accuracy in budget forecasting and environment expansion based on platform licensing.</li>\n<li>Administer and optimize delivery of software used on developer workstations such as Anypoint Studio, Anypoint Code Builder, Postman, and Github Desktop</li>\n</ul>\n<ul>\n<li>Design, implement, and maintain CI/CD pipelines to automate MuleSoft deployments, environment promotion, and rollback strategies</li>\n</ul>\n<ul>\n<li>Lead production incident response, root cause analysis, and preventative remediation for integration and API failures</li>\n<li>Implement observability, monitoring, alerting, logging, and SLA frameworks to ensure proactive detection and resolution of platform issues</li>\n<li>Enforce platform standards for security, configuration, and operational readiness in alignment with enterprise architecture and compliance requirements</li>\n<li>Support API gateway configuration, TLS certificates, policy enforcement, and secure integration patterns</li>\n<li>Plan and execute MuleSoft platform upgrades, patches, and capacity scaling initiatives</li>\n<li>Collaborate with infrastructure, security, network, application, DevOps, and architecture teams to deliver reliable, secure, and scalable integration solutions</li>\n<li>Engage with MuleSoft Account Managers, committees, and/or user groups to stay ahead of platform changes, new capabilities, and emerging technologies</li>\n<li>Manage platform documentation in tools such as LeanIX and ServiceNow CMDB</li>\n<li>Architect, implement, and support Disaster Recovery plans for the MuleSoft application environments</li>\n<li>Implement security best practices and solutions for users and applications consuming MuleSoft capabilities</li>\n</ul>\n<p><strong>What You Bring to the Role (Qualifications)</strong></p>\n<ul>\n<li>Seven (7) or more years of experience supporting enterprise-class applications, middleware, or integration platforms</li>\n<li>Three (3) or more years of hands-on administration and engineering of the MuleSoft Anypoint Platform</li>\n<li>Strong experience operating MuleSoft in production environments including CloudHub, Runtime Fabric, or hybrid deployments</li>\n<li>Hands-on experience with Code, Container, and Artifact Repository tools such as Bitbucket, Github, Maven, JFrog, or similar</li>\n<li>Experience with CI/CD automation tools such as Jenkins, Azure DevOps, GitHub Actions, Sonarqube, or similar</li>\n<li>Hands-on experience with API-led connectivity, REST and SOAP web services, RAML, and OpenAPI</li>\n<li>Hands-on experience supporting EDI and B2B connectivity and message exchanges with 3rd parties</li>\n<li>Hands-on experience with Event Driven Architecture, Message Queueing, and Pub/Sub integration patterns</li>\n<li>Strong troubleshooting skills using MuleSoft debuggers, flow analyzers, and performance tuning techniques</li>\n<li>Experience with Linux systems, scripting (Python, Bash, etc), and automation for platform operations</li>\n<li>Ability to communicate technical issues and remediation plans to engineering teams and leadership</li>\n</ul>\n<p><strong>Competencies Desired</strong></p>\n<ul>\n<li>Hands on admin Mulesoft platform experience – must have</li>\n<li>Hands on admin Webmethods experience – nice to have (legacy architecture)</li>\n<li>Platform Stability and Reliability</li>\n<li>Automation and DevOps Enablement</li>\n<li>Production Incident Leadership</li>\n<li>Security and Compliance Alignment</li>\n<li>Cross-Team Collaboration</li>\n<li>Technical Problem Solving</li>\n<li>Operational Documentation Discipline</li>\n</ul>\n<p><strong>Certifications Desired</strong></p>\n<ul>\n<li>MuleSoft Certified Integration Architect</li>\n<li>MuleSoft Certified Platform Integration Architect</li>\n</ul>\n<p><strong>Working Conditions / Physical Requirements / Travel Requirements</strong></p>\n<ul>\n<li>Primarily office or remote work with standard computer and virtual collaboration tools</li>\n<li>May participate in 24x7 on-call rotation to support production systems</li>\n<li>Travel requirements: Some may be required for team building and major project rollouts. Travel 5%.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7b60b203-14a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Valvoline Global Operations","sameAs":"https://jobs.valvolineglobal.com","logo":"https://logos.yubhub.co/jobs.valvolineglobal.com.png"},"x-apply-url":"https://jobs.valvolineglobal.com/job/Integrations-Platform-Administrator-OH/1366647100/","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["MuleSoft Anypoint Platform","MuleSoft Runtime Manager","API Manager","Access Management","AnypointMQ","RPA","Intelligent Document Processing","Partner Manager","Runtime Fabric","CloudHub","Design Center","Exchange","Anypoint Experience Hub","Postman","Github Desktop","CI/CD pipelines","Linux systems","scripting","automation","API-led connectivity","REST and SOAP web services","RAML","OpenAPI","EDI and B2B connectivity","Event Driven Architecture","Message Queueing","Pub/Sub integration patterns"],"x-skills-preferred":[],"datePosted":"2026-03-09T11:05:17.158Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"OH, US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"MuleSoft Anypoint Platform, MuleSoft Runtime Manager, API Manager, Access Management, AnypointMQ, RPA, Intelligent Document Processing, Partner Manager, Runtime Fabric, CloudHub, Design Center, Exchange, Anypoint Experience Hub, Postman, Github Desktop, CI/CD pipelines, Linux systems, scripting, automation, API-led connectivity, REST and SOAP web services, RAML, OpenAPI, EDI and B2B connectivity, Event Driven Architecture, Message Queueing, Pub/Sub integration patterns"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_11a9548c-a4f"},"title":"Staff+ Software Engineer, Developer Productivity","description":"<p><strong>About Anthropic</strong></p>\n<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>\n<p><strong>About the Role</strong></p>\n<p>Anthropic&#39;s Infrastructure organisation is foundational to our mission of developing AI systems that are reliable, interpretable, and steerable. The systems we build determine how quickly we can train new models, how reliably we can run safety experiments, and how effectively we can scale Claude to millions of users — demonstrating that safe, reliable infrastructure and frontier capabilities can go hand in hand.</p>\n<p>Developer Productivity owns the end-to-end experience of how engineers and researchers at Anthropic develop, build, test, and ship code at scale — from the source control and language ecosystems that underpin our monorepo, to the build and CI infrastructure that keeps thousands of daily builds running reliably across multiple cloud providers, to the developer acceleration tooling that deeply integrates Claude into engineering workflows.</p>\n<p>_Team Matching: Team matching is determined after the interview process based on interview performance, interests, and business priorities. Please note we may also consider you for different Infrastructure teams._</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Own the technical strategy and roadmap for your area, translating team-level goals into concrete execution plans</li>\n<li>Define infrastructure architecture, ensuring the hardest problems get solved — whether by you directly or by working through others</li>\n<li>Design and build scalable, reliable distributed infrastructure and shared libraries that support high-volume workloads across all engineering teams</li>\n<li>Own and evolve build environments, package management, and dependency systems to enable fast, reproducible builds</li>\n<li>Define and implement language ecosystem standards, tooling, and frameworks that drive developer productivity across research and production workloads</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have 10+ years (not including internships or co-ops) of experience in a Software Engineer role, building and operating large-scale developer infrastructure</li>\n<li>Have 3+ years (not including internships or co-ops) of experience leading large scale, complex projects or teams as an engineer or tech lead</li>\n<li>Have deep experience with build systems, CI/CD pipelines, and/or developer tooling in a large monorepo environment</li>\n<li>Have strong proficiency in Python, Rust and/or Go</li>\n<li>Are obsessed with developer productivity and reducing friction in the software development lifecycle</li>\n<li>Have experience with container orchestration and infrastructure at scale</li>\n<li>Have excellent communication skills and enjoy supporting internal partners to improve their development experience</li>\n<li>Are excited about designing foundational systems and are comfortable working independently on ambiguous, high-impact technical challenges</li>\n</ul>\n<p><strong>Strong candidates may have:</strong></p>\n<ul>\n<li>Experience with CI orchestration tools (Buildkite, Jenkins, GitHub Actions, or similar) and merge queue management at scale</li>\n<li>Experience building or operating remote build execution systems (Bazel Remote Execution API, BuildBarn, BuildBuddy, or similar)</li>\n<li>Experience with Nix/NixOS/Docker and managing large image / package sets at scale</li>\n<li>Experience building CLI tools, developer-facing services, and GitHub API and automation workflows</li>\n</ul>\n<p>_Deadline to apply: None. Applications will be reviewed on a rolling basis._</p>\n<p>The annual compensation range for this role is listed below.</p>\n<p>For sales roles, the range provided is the role’s On Target Earnings (&quot;OTE&quot;) range, meaning that the range includes both the sales commissions/sales bonuses target and annual base salary for the role.</p>\n<p>Annual Salary:</p>\n<p>$405,000 - $485,000USD</p>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</p>\n<p><strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p><strong>Your safety matters to us.</strong> To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact work in AI safety and development happens at the intersection of technical expertise and societal responsibility. We&#39;re committed to building a team that reflects a wide range of backgrounds, perspectives, and experiences. We believe that diversity in all its forms drives better decision-making, more innovative solutions, and greater impact.</p>\n<p>We&#39;re an equal opportunities employer and welcome applications from all qualified candidates.</p>\n<p>If you&#39;re excited about this role and want to learn more, please don&#39;t hesitate to reach out to us. We look forward to hearing from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_11a9548c-a4f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5110511008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000 - $485,000USD","x-skills-required":["Python","Rust","Go","Build systems","CI/CD pipelines","Developer tooling","Container orchestration","Infrastructure at scale"],"x-skills-preferred":["CI orchestration tools","Merge queue management","Remote build execution systems","Nix/NixOS/Docker","Large image/package sets","CLI tools","Developer-facing services","GitHub API and automation workflows"],"datePosted":"2026-03-08T13:53:03.879Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Rust, Go, Build systems, CI/CD pipelines, Developer tooling, Container orchestration, Infrastructure at scale, CI orchestration tools, Merge queue management, Remote build execution systems, Nix/NixOS/Docker, Large image/package sets, CLI tools, Developer-facing services, GitHub API and automation workflows","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ea50b076-736"},"title":"Infrastructure Engineer","description":"<p>Our mission is to automate coding. The first step in our journey is to build the best tool for professional programmers, using a combination of inventive research, design, and engineering. We&#39;re looking for talented infrastructure engineers to join our team and help us build the platform that supports our functionality.</p>\n<p>As an infrastructure engineer at Cursor, you will be responsible for building highly-available distributed systems, dealing with the woes of scaling to millions of programmers, and working with various databases, proxies, caches, task queues, and orchestration systems. You will also be involved in shipping infra for safely computing import graphs and shadow lints.</p>\n<p>Some sample projects you may work on include creating a retrieval system that processes 10,000,000,000+ files, staring at esoteric flame graphs to performance engineer our reranking library, and working with many databases, proxies, caches, task queues, and orchestration systems.</p>\n<p>We&#39;re a team that enjoys spirited debate, crazy ideas, and shipping code. We&#39;re looking for people who are truth-seeking, passionate, and creative, and who are comfortable working in a flat organisational structure.</p>\n<p>Our offices are located in North Beach, San Francisco and Manhattan, New York, and are replete with well-stocked libraries. We&#39;re an in-person team, and we enjoy working together to build something amazing.</p>\n<p>If you&#39;re passionate about building highly-available distributed systems, and you&#39;re looking for a challenging and rewarding role, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ea50b076-736","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cursor","sameAs":"https://cursor.com","logo":"https://logos.yubhub.co/cursor.com.png"},"x-apply-url":"https://cursor.com/careers/software-engineer-infrastructure","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"Competitive salary and benefits package","x-skills-required":["Distributed systems","Database management","Proxy management","Cache management","Task queue management","Orchestration systems","Performance engineering","Reranking library"],"x-skills-preferred":["Flame graph analysis","Import graph computation","Shadow linting"],"datePosted":"2026-03-08T00:18:53.519Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, New York"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Distributed systems, Database management, Proxy management, Cache management, Task queue management, Orchestration systems, Performance engineering, Reranking library, Flame graph analysis, Import graph computation, Shadow linting"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1f8b41f2-3cf"},"title":"Senior Integration Developer","description":"<p>The Valvoline integration development team is responsible for establishing and maintaining communication between computer applications, both internal to Valvoline and with external business partners. The integration developer senior will be responsible for leading the development of integration solutions using Valvoline’s integration toolsets, webMethods, SAP Integration Suite and Mulesoft AnyPoint.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>The integration developer senior will design, develop and maintain integration services for Valvoline’s business applications and external partner’s integrations. Lead reviews of junior developers integration designs and coding. Ensure developers are using common integration patterns where applicable. Enforce the practice of creating and maintaining documentation on the integrations within the tools specified for the team.</p>\n<ul>\n<li>Design, develop and maintain integration services for Valvoline’s business applications and external partner’s integrations.</li>\n<li>Lead reviews of junior developers integration designs and coding.</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>BA/BS degree in Computer Science or related field of study.</li>\n<li>5-10 years of experience working with following: Software AG’s webMethods, Mulesoft Anypoint or SAP Integration Suite, Integration with ERP systems, Document exchange protocols such as AS2, SFTP, https, ftp, Publishing and subscribing messages to queues and topics in a messaging provider, Service Oriented Architecture and Web Services, ANSI X12 and EDIFACT EDI standards, XML based standards such as CIDX, xCBL or OAIGs, Cloud Integration, API development, API Management.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1f8b41f2-3cf","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Valvoline","sameAs":"https://jobs.valvolineglobal.com","logo":"https://logos.yubhub.co/jobs.valvolineglobal.com.png"},"x-apply-url":"https://jobs.valvolineglobal.com/job/Senior-Integration-Developer/1353398400/","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Software AG’s webMethods","Mulesoft Anypoint","SAP Integration Suite","Integration with ERP systems","Document exchange protocols such as AS2, SFTP, https, ftp","Publishing and subscribing messages to queues and topics in a messaging provider","Service Oriented Architecture and Web Services","ANSI X12 and EDIFACT EDI standards","XML based standards such as CIDX, xCBL or OAIGs","Cloud Integration","API development","API Management"],"x-skills-preferred":["Relational Databases and SQL","Problem-solving & organizational skills","The ability to communicate with both business (non-technical) and technical persons"],"datePosted":"2026-01-08T18:07:47.029Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"Software AG’s webMethods, Mulesoft Anypoint, SAP Integration Suite, Integration with ERP systems, Document exchange protocols such as AS2, SFTP, https, ftp, Publishing and subscribing messages to queues and topics in a messaging provider, Service Oriented Architecture and Web Services, ANSI X12 and EDIFACT EDI standards, XML based standards such as CIDX, xCBL or OAIGs, Cloud Integration, API development, API Management, Relational Databases and SQL, Problem-solving & organizational skills, The ability to communicate with both business (non-technical) and technical persons"}]}