{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/pipelines"},"x-facet":{"type":"skill","slug":"pipelines","display":"Pipelines","count":100},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c05eef2e-517"},"title":"Lead Data Scientist and AI Engineer","description":"<p>If you want to work on innovative AI solutions and lead a team, this might be the right opportunity for you. As a Lead Data Scientist and AI Engineer, you will be responsible for leading a team of data scientists, AI engineers, and computer vision specialists, as well as overseeing the successful implementation of data-driven and intelligent projects.</p>\n<p>Your tasks will include:</p>\n<ul>\n<li>Leading a team of data scientists, AI engineers, and computer vision specialists in a technical and disciplinary manner</li>\n<li>Managing AI, computer vision, and physical AI projects and being responsible for their successful implementation - from use case to productive deployment</li>\n<li>Steering the operational project deployment of your team and ensuring high utilization and delivery quality</li>\n<li>Supporting architectural decisions and ensuring compliance with technical standards (AI engineering, MLOps, vision pipelines)</li>\n<li>Actively participating in presales, developing demos, and supporting technical solution designs</li>\n<li>Developing your team technically and disciplinarily and building targeted competences in the areas of ML and physical AI</li>\n</ul>\n<p>To be successful in this role, you will need to have a strong background in data science, AI engineering, or machine learning, as well as leadership experience. You should also have a passion for implementing AI projects and a good understanding of technical solution designs.</p>\n<p>In addition to your technical expertise, you should have excellent communication and leadership skills, as well as the ability to work in a fast-paced environment.</p>\n<p>We offer a competitive salary and a range of benefits, including flexible working hours, a generous holiday allowance, and opportunities for professional development.</p>\n<p>If you are interested in this opportunity, please submit your application, including your resume and cover letter, through our online portal.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c05eef2e-517","directApply":true,"hiringOrganization":{"@type":"Organization","name":"MHP","sameAs":"https://www.mhp.com","logo":"https://logos.yubhub.co/mhp.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=20142","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"Competitive salary","x-skills-required":["data science","AI engineering","machine learning","leadership","communication","team management"],"x-skills-preferred":["computer vision","MLOps","vision pipelines","presales","demos","technical solution design"],"datePosted":"2026-04-22T17:26:27.506Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data science, AI engineering, machine learning, leadership, communication, team management, computer vision, MLOps, vision pipelines, presales, demos, technical solution design"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_272750a8-710"},"title":"Consultant","description":"<p>As a Consultant at MHP, you will operate infrastructure in AWS using Terraform, create technical concepts for new features and enhancements within a Scrum Team, develop and maintain scalable Java Spring Boot microservices, and work with AWS and Kubernetes.</p>\n<p>You will have expertise in backend programming using Java and Spring Boot, experience with AWS, including services like S3, EC2, and Lambda, and experience with Terraform for creating and managing AWS infrastructure.</p>\n<p>You will also have experience with tools such as IntelliJ and REST tools (Postman or similar), proficiency in working with Kubernetes for microservices, advanced-level AWS certification, experience with Apache Kafka event streaming, experience working with MongoDB database, and experience working with GitLab CI/CD pipelines.</p>\n<p>You will start by arrangement, work full-time (40h) with 27 vacation days, and have an unlimited employment contract. You will need a valid work permit and be fluent in written and spoken English.</p>\n<p>At MHP, you will continuously grow with your projects and objectives in an innovative and supportive environment. You will be part of a strong team spirit, where every win, big or small, belongs to all of us. You will welcome curiosity, creativity, and unconventional thinking patterns, and recognize the importance of healthy, tight-knit communities and sustainable environmental changes.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_272750a8-710","directApply":true,"hiringOrganization":{"@type":"Organization","name":"MHP","sameAs":"http://www.mhp.com/","logo":"https://logos.yubhub.co/mhp.com.png"},"x-apply-url":"https://jobs.porsche.com/index.php?ac=jobad&id=18226","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Spring Boot","AWS","Terraform","Kubernetes","IntelliJ","REST tools","Apache Kafka","MongoDB","GitLab CI/CD pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-22T17:25:42.569Z","employmentType":"FULL_TIME","occupationalCategory":"IT","industry":"Consulting","skills":"Java, Spring Boot, AWS, Terraform, Kubernetes, IntelliJ, REST tools, Apache Kafka, MongoDB, GitLab CI/CD pipelines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b33cbd91-bc9"},"title":"Systematic Production Support Engineer","description":"<p>We are seeking an experienced Systematic Production Support Engineer to help us scale our systematic operations and support engineering capabilities. This role directly supports portfolio management teams across Millennium, with operational excellence at the core. Our efforts are focused on delivering the highest quality returns to our investors – providing a world-class and reliable trading and technology platform is essential to this mission.</p>\n<p>As a Systematic Production Support Engineer, you will be responsible for building, developing, and maintaining a reliable, scalable, and integrated platform for trading strategy monitoring, reporting, and operations. You will work closely with portfolio managers and other internal customers to reduce operational risk through the implementation of monitoring, reporting, and trade workflow solutions, as well as automated systems and processes focused on trading and operations.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Building, developing, and maintaining a reliable, scalable, and integrated platform for trading strategy monitoring, reporting, and operations</li>\n<li>Working with portfolio managers and other internal customers to reduce operational risk through the implementation of monitoring, reporting, and trade workflow solutions</li>\n<li>Implementing automated systems and processes focused on trading and operations</li>\n<li>Streamlining development and deployment processes</li>\n</ul>\n<p>Technical qualifications include:</p>\n<ul>\n<li>5+ years of development experience in Python</li>\n<li>Experience working in a Linux/Unix environment</li>\n<li>Experience working with PostgreSQL or other relational databases</li>\n</ul>\n<p>Preferred skills and experience include:</p>\n<ul>\n<li>Understanding of NLP, supervised/non-supervised learning, and Generative AI models</li>\n<li>Experience operating and monitoring low-latency trading environments</li>\n<li>Familiarity with quantitative finance and electronic trading concepts</li>\n<li>Familiarity with financial data</li>\n<li>Broad understanding of equities, futures, FX, or other financial instruments</li>\n<li>Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#</li>\n<li>Experience with Apache/Confluent Kafka</li>\n<li>Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline)</li>\n<li>Experience with containerization and orchestration technologies</li>\n<li>Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure</li>\n<li>Contributions to open-source projects</li>\n</ul>\n<p>This is a unique opportunity to drive significant value creation for one of the world&#39;s leading investment managers.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b33cbd91-bc9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Unknown","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954716155","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Linux/Unix","PostgreSQL","NLP","supervised/non-supervised learning","Generative AI models","low-latency trading environments","quantitative finance","electronic trading concepts","financial data","equities","futures","FX","distributed systems","backend development","C/C++","Java","Scala","Go","C#","Apache/Confluent Kafka","SDLC pipelines","containerization","orchestration technologies","AWS","GCP","Azure"],"x-skills-preferred":["Understanding of NLP, supervised/non-supervised learning, and Generative AI models","Experience operating and monitoring low-latency trading environments","Familiarity with quantitative finance and electronic trading concepts","Familiarity with financial data","Broad understanding of equities, futures, FX, or other financial instruments","Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#","Experience with Apache/Confluent Kafka","Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline)","Experience with containerization and orchestration technologies","Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure","Contributions to open-source projects"],"datePosted":"2026-04-18T22:14:36.583Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Linux/Unix, PostgreSQL, NLP, supervised/non-supervised learning, Generative AI models, low-latency trading environments, quantitative finance, electronic trading concepts, financial data, equities, futures, FX, distributed systems, backend development, C/C++, Java, Scala, Go, C#, Apache/Confluent Kafka, SDLC pipelines, containerization, orchestration technologies, AWS, GCP, Azure, Understanding of NLP, supervised/non-supervised learning, and Generative AI models, Experience operating and monitoring low-latency trading environments, Familiarity with quantitative finance and electronic trading concepts, Familiarity with financial data, Broad understanding of equities, futures, FX, or other financial instruments, Experience designing and developing distributed systems with a focus on backend development in C/C++, Java, Scala, Go, or C#, Experience with Apache/Confluent Kafka, Experience automating SDLC pipelines (e.g., Jenkins, TeamCity, or AWS CodePipeline), Experience with containerization and orchestration technologies, Experience building and deploying systems that utilize services provided by AWS, GCP, or Azure, Contributions to open-source projects"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f77c41bb-0ad"},"title":"Application Security Engineer","description":"<p>We are seeking an experienced Application Security Engineer to join our team. As a subject matter expert, you will have direct experience in a wide range of security technologies, tools, and methodologies. The role is suited for an experienced Application Security engineer with proven understanding in enterprise security and AI security and will focus on building toolsets and processes to drive adoption of secure practices across the enterprise.</p>\n<p>The team fosters a collaborative environment and is building a best-in-class program to partner with the business to protect the Firm’s information and computer systems. Millennium is a complex and robust technical environment and securing the Firm from external and internal threats is a top priority.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Define and implement security guardrails for Generative AI, LLMs, and Agentic frameworks, ensuring safe enterprise adoption.</li>\n<li>Conduct specialized threat modeling, red teaming, and risk assessments for AI/ML models (e.g., testing for prompt injection, model theft, and data poisoning).</li>\n<li>Lead risk management activities, including application risk assessments, design reviews, and mitigation strategies for IT projects.</li>\n<li>Engage throughout the SDLC to identify vulnerabilities, conduct code reviews/penetration testing, and enforce secure coding standards.</li>\n<li>Evangelize AppSec and AI security best practices through developer education, training materials, and outreach.</li>\n<li>Design robust security architectures and integrate automated security testing (SAST/DAST/SCA) into CI/CD pipelines.</li>\n<li>Partner with Technology, Trading, Legal, and Compliance to create policies and communicate technical risks to non-technical stakeholders.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>Bachelor&#39;s degree or higher in Computer Science, Computer Engineering, IT Security or related field.</li>\n<li>5+ years’ experience working as an Application Security Engineer, Software Engineer, or similar role.</li>\n<li>Deep understanding of AI-specific risks (OWASP Top 10 for LLMs) and experience securing applications utilizing LLMs.</li>\n<li>Experience working with AI models, Agentic frameworks and security risks associated with AI.</li>\n<li>Experience in working with global teams, collaborating on code and presentations.</li>\n<li>Demonstrated work experience in hybrid on-premise and Public Cloud environments (AWS/GCP/Azure)</li>\n<li>Strong understanding of security architectures, secure configuration principles/coding practices, cryptography fundamentals and encryption protocols.</li>\n<li>Experience with common SCM &amp; CI/CD technologies like GitHub, Jenkins, Artifactory, etc. and integrating Security Scanning and Vulnerability Management into the CI/CD Pipelines</li>\n<li>Familiarity with static and dynamic security analysis tools, and SCA/SBOM solutions.</li>\n<li>Hands on experience with Secrets Management &amp; Password Vault technologies such as Delinea Secret Server and/or Hashicorp Vault, etc.</li>\n<li>Strong experience in secure programming in languages such as Python, Java, C++, C#, or similar.</li>\n<li>Familiarity with Infrastructure as Code tools (CloudFormation, Terraform, Ansible, etc.)</li>\n<li>Familiarity with web application security testing tools and methodologies.</li>\n<li>Knowledge of various security frameworks and standards such as ISO 27001, NIST, OWASP, etc.</li>\n<li>Knowledge of Linux, OS internals and containers is a plus.</li>\n<li>Certifications like CISSP, CISM, CompTIA Security+, or CEH are advantageous.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f77c41bb-0ad","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT Infrastructure","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955629927","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["AI-specific risks","Generative AI","LLMs","Agentic frameworks","Security guardrails","Threat modeling","Red teaming","Risk assessments","Application risk assessments","Design reviews","Mitigation strategies","Secure coding standards","Automated security testing","CI/CD pipelines","Security architectures","Secure configuration principles","Cryptography fundamentals","Encryption protocols","SCM & CI/CD technologies","Security scanning","Vulnerability management","Static and dynamic security analysis tools","SCA/SBOM solutions","Secrets management","Password vault technologies","Secure programming","Infrastructure as Code tools","Web application security testing tools","Methodologies","Security frameworks","Standards","Linux","OS internals","Containers"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:14:17.280Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, Ireland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"AI-specific risks, Generative AI, LLMs, Agentic frameworks, Security guardrails, Threat modeling, Red teaming, Risk assessments, Application risk assessments, Design reviews, Mitigation strategies, Secure coding standards, Automated security testing, CI/CD pipelines, Security architectures, Secure configuration principles, Cryptography fundamentals, Encryption protocols, SCM & CI/CD technologies, Security scanning, Vulnerability management, Static and dynamic security analysis tools, SCA/SBOM solutions, Secrets management, Password vault technologies, Secure programming, Infrastructure as Code tools, Web application security testing tools, Methodologies, Security frameworks, Standards, Linux, OS internals, Containers"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_af8ed06d-a9a"},"title":"Forward Deployed Software Engineer - Equities Technology","description":"<p>We are seeking a hands-on, business-facing engineer to join our team. In this role, you will partner directly with some of the most sophisticated quantitative researchers, developers, and portfolio managers in the industry.</p>\n<p>Our team is a specialized group of engineers operating at the intersection of technology and quantitative finance. We function as an internal centre of excellence, providing expert-level solutions, architecture, and hands-on development in AI, Cloud (AWS/GCP), DevOps, and high-performance computing.</p>\n<p>As a forward deployed software engineer, you will be responsible for translating complex research requirements into robust, scalable, and secure technical architectures across on-prem, hybrid, and cloud environments. You will write high-quality, production-ready code across the full stack, including Python libraries, infrastructure-as-code (Terraform), CI/CD pipelines, automation scripts, and ML/AI proof-of-concepts.</p>\n<p>You will also develop and maintain our suite of managed products, reusable patterns, and best practice guides to provide self-service options and accelerate onboarding for new and existing teams. Additionally, you will act as the primary technical point of contact for embedded engagements, owning projects from discovery and planning through to implementation, knowledge transfer, and support.</p>\n<p>To succeed in this role, you will need to have a deep understanding of computer science principles, including data structures, algorithms, and system design. You will also need to have experience working with cloud providers, such as AWS or GCP, and be familiar with infrastructure-as-code concepts. Excellent verbal and written communication skills are also essential, as you will need to build strong relationships with stakeholders and articulate complex ideas to diverse audiences.</p>\n<p>Innovative thinking and a passion for AI/ML and its practical applications are highly desirable. Experience designing systems and architectures from ambiguous business needs, as well as experience with scheduling or asynchronous workflow frameworks/services, is also preferred.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_af8ed06d-a9a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755953439247","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Cloud computing (AWS/GCP)","DevOps","Infrastructure-as-code (Terraform)","CI/CD pipelines","Automation scripts","ML/AI proof-of-concepts","Data structures","Algorithms","System design"],"x-skills-preferred":["Experience in the financial services or fintech space","Experience building applications on top of LLMs using frameworks like LangChain or LlamaIndex","Experience with MLOps tooling and concepts","Cloud certifications (AWS or GCP)"],"datePosted":"2026-04-18T22:14:13.794Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Miami, Florida, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Cloud computing (AWS/GCP), DevOps, Infrastructure-as-code (Terraform), CI/CD pipelines, Automation scripts, ML/AI proof-of-concepts, Data structures, Algorithms, System design, Experience in the financial services or fintech space, Experience building applications on top of LLMs using frameworks like LangChain or LlamaIndex, Experience with MLOps tooling and concepts, Cloud certifications (AWS or GCP)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6690b2fa-cab"},"title":"(Senior) Team Lead Data Analytics (all genders)","description":"<p>At Holidu, data isn&#39;t just a support function, it&#39;s how we make decisions. The Analytics team builds the products and foundations that keep the whole organisation sharp, from day-to-day operations to long-term strategy.</p>\n<p>This role is on-site in Munich, with two office days per week.</p>\n<p>As a Senior Team Lead Data Analytics, you will lead one of Holidu&#39;s core analytics teams, a function at the intersection of data, strategy, and real business impact. The team has four direct reports and entails collaborating cross-functionally with data engineers and data scientists.</p>\n<p>Engage with senior leadership on strategic projects, providing insights that influence product strategy, internal operations, and revenue growth.</p>\n<p>You and your team will support a range of stakeholders across the company (e.g. Customer Support, Host Experience, Sales and Account Management).</p>\n<p>As a member of the BI leadership team, you will help shape the department strategy and the future of AI-powered data products.</p>\n<p>Understand problems and identify opportunities across a diverse range of stakeholder use cases, translating them into analytical requirements and communicating complex findings clearly to both technical and commercial audiences.</p>\n<p>Lead from the front: this role carries meaningful individual contributor responsibility. You&#39;ll be expected to do real analytical work, diving deep into the data, building solutions, and setting the bar for quality in your team.</p>\n<p>Shape the future of analytics at Holidu by recruiting top talent, setting clear goals, and developing your team personally and professionally.</p>\n<p>The ideal candidate will have 5+ years of data analytics experience, people management experience, a collaborative mindset, a mission-driven mentality, excellent analytical and technical skills, and a genuine commitment to AI enablement.</p>\n<p>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</p>\n<p>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</p>\n<p>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</p>\n<p>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</p>\n<p>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</p>\n<p>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6690b2fa-cab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2598226","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Database: AWS Stack (Redshift, Athena, Glue, S3)","Data Pipelines: Airflow, dbt","Data Visualisation: Looker","Data Analytics: SQL, Python","Collaboration: Git, Jira, Confluence, Slack"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:28.264Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Technology","industry":"Travel Technology","skills":"Database: AWS Stack (Redshift, Athena, Glue, S3), Data Pipelines: Airflow, dbt, Data Visualisation: Looker, Data Analytics: SQL, Python, Collaboration: Git, Jira, Confluence, Slack"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c2995faa-123"},"title":"Software Engineer – Equity Derivatives Pricing & Risk System","description":"<p>We are seeking a highly skilled Java Developer with a strong background in Equity Derivatives to join our team in London.</p>\n<p>In this role, you will play a pivotal part in building and enhancing Equity Volatility Risk and P&amp;L system that supports our Equity Volatility Managers.</p>\n<p>This is an exciting opportunity to work in a fast-paced hedge fund environment, where your contributions will directly impact trading performance and risk management capabilities.</p>\n<p>The ideal candidate will bring a combination of technical expertise and business domain knowledge for developing robust, scalable systems.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Design, develop, and implement a robust risk system for Equity Volatility trading strategies.</li>\n<li>Build and maintain scalable, high-performance server-side application using Java and Spring Boot frameworks.</li>\n<li>Build and integrate exotic pricing models to handle pricing and lifecycle of the product.</li>\n<li>Provide level-3 support, troubleshooting, and performance tuning for production systems.</li>\n<li>Proactively address system bottlenecks and implement solutions to ensure the platform remains robust.</li>\n<li>Conduct code reviews and implement automated testing to ensure the reliability and quality of the system.</li>\n<li>Write clean, maintainable, and testable code, adhering to best practices in software engineering.</li>\n</ul>\n<p>Qualifications/Skills Required:</p>\n<ul>\n<li>Proficiency in Java development with experience in building scalable, high-performance systems.</li>\n<li>Strong knowledge of Spring Boot and its ecosystem for developing microservices.</li>\n<li>Experience with Python for scripting and automation.</li>\n<li>Experience in distributed caching technologies (e.g. Ignite, or similar).</li>\n<li>Familiarity with containerization technologies (e.g. Podman, Kubernetes) and cloud computing platforms (e.g. AWS).</li>\n<li>Solid understanding of software development best practices, including version control (e.g. Git), CI/CD pipelines, and automated testing frameworks.</li>\n<li>Previous experience working with Equity Derivatives in a sell-side or buy-side firm.</li>\n<li>Strong understanding of equity derivative products such as options and futures.</li>\n<li>Some understanding of structured products in terms of pricing, lifecycle, and risk characteristics.</li>\n<li>Strong problem-solving skills and the ability to work effectively in a fast-paced, high-pressure environment.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c2995faa-123","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955392398","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Spring Boot","Python","Distributed caching technologies","Containerization technologies","Cloud computing platforms","Version control","CI/CD pipelines","Automated testing frameworks","Equity Derivatives"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:24.304Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Spring Boot, Python, Distributed caching technologies, Containerization technologies, Cloud computing platforms, Version control, CI/CD pipelines, Automated testing frameworks, Equity Derivatives"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1a20521b-6ce"},"title":"Senior Execution Quantitative Analyst - Fixed Income","description":"<p>We are seeking a Senior Execution Quantitative Analyst to lead the expansion of our central execution capabilities into fixed income markets, covering corporate credit (IG/HY), Treasuries (cash and futures), and interest rate swaps.</p>\n<p>This is a hands-on role requiring deep fixed income market structure knowledge combined with strong quantitative and software development skills. The successful candidate will be expected to assess the firm&#39;s existing data and workflow landscape, identify and size near-term P&amp;L opportunities, and lead the build-out of execution and analysis infrastructure.</p>\n<p><strong>Principal Responsibilities</strong></p>\n<ul>\n<li>Assess the firm&#39;s existing fixed income data assets (dealer axes, evaluated pricing, TRACE prints, swap SDR data, futures market data) and design a coherent real-time and historical data layer to support execution and analysis</li>\n<li>Identify and size near-term opportunities in execution quality improvement, transaction cost reduction, and flow internalization across credit, rates, and swaps</li>\n<li>Design, build, and operate internal execution algorithms covering the full fixed income liquidity spectrum, from liquid on-the-run Treasuries to illiquid corporate bonds,using RFQ, click-to-trade, and direct connectivity workflows</li>\n<li>Build transaction cost analysis and pre-trade cost models for fixed income instruments;</li>\n<li>Partner with portfolio managers and traders to understand flow characteristics and communicate execution analytics clearly</li>\n<li>Recruit and mentor junior quants and engineers as the platform scales</li>\n</ul>\n<p><strong>Qualifications / Skills Required</strong></p>\n<ul>\n<li>10+ years of relevant experience in fixed income electronic trading, execution, or quantitative research on the buy side or sell side</li>\n<li>Hands-on experience building execution infrastructure for institutional fixed income: RFQ and/or click-to-trade workflows, FIX protocol connectivity, and integration with major electronic venues</li>\n<li>Experience building TCA or cost models for fixed income instruments, including illiquid and sparsely traded securities</li>\n<li>Strong programming skills; experience with data pipelines and market data APIs</li>\n<li>Solid quantitative background; degree in Mathematics, Computer Science, Engineering, Physics, or a related field</li>\n<li>Demonstrated ability to translate data analysis into actionable P&amp;L estimates and communicate findings to non-technical stakeholders</li>\n<li>Experience as a hands-on development lead, with a track record of taking projects from inception to production</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1a20521b-6ce","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Electronic Trading Solutions","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954333818","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["fixed income electronic trading","execution","quantitative research","RFQ and/or click-to-trade workflows","FIX protocol connectivity","integration with major electronic venues","TCA or cost models for fixed income instruments","data pipelines","market data APIs","quantitative background","degree in Mathematics, Computer Science, Engineering, Physics, or a related field"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:00.980Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"fixed income electronic trading, execution, quantitative research, RFQ and/or click-to-trade workflows, FIX protocol connectivity, integration with major electronic venues, TCA or cost models for fixed income instruments, data pipelines, market data APIs, quantitative background, degree in Mathematics, Computer Science, Engineering, Physics, or a related field"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7275ef33-009"},"title":"Staff Data Engineer","description":"<p>At Bayer, we&#39;re seeking a Staff Data Engineer to join our team. As a Staff Data Engineer, you will design and lead the implementation of data flows to connect operational systems, data for analytics and business intelligence (BI) systems. You will recognize opportunities to reuse existing data flows, lead the build of data streaming systems, optimize the code to ensure processes perform optimally, and lead work on database management.</p>\n<p>Communicating Between Technical and Non-Technical Colleagues</p>\n<p>As a Staff Data Engineer, you will communicate effectively with technical and non-technical stakeholders, support and host discussions within a multidisciplinary team, and be an advocate for the team externally.</p>\n<p>Data Analysis and Synthesis</p>\n<p>You will undertake data profiling and source system analysis, present clear insights to colleagues to support the end use of the data.</p>\n<p>Data Development Process</p>\n<p>You will design, build and test data products that are complex or large scale, build teams to complete data integration services.</p>\n<p>Data Innovation</p>\n<p>You will understand the impact on the organization of emerging trends in data tools, analysis techniques and data usage.</p>\n<p>Data Integration Design</p>\n<p>You will select and implement the appropriate technologies to deliver resilient, scalable and future-proofed data solutions and integration pipelines.</p>\n<p>Data Modeling</p>\n<p>You will produce relevant data models across multiple subject areas, explain which models to use for which purpose, understand industry-recognised data modelling patterns and standards, and when to apply them, compare and align different data models.</p>\n<p>Metadata Management</p>\n<p>You will design an appropriate metadata repository and present changes to existing metadata repositories, understand a range of tools for storing and working with metadata, provide oversight and advice to more inexperienced members of the team.</p>\n<p>Problem Resolution</p>\n<p>You will respond to problems in databases, data processes, data products and services as they occur, initiate actions, monitor services and identify trends to resolve problems, determine the appropriate remedy and assist with its implementation, and with preventative measures.</p>\n<p>Programming and Build</p>\n<p>You will use agreed standards and tools to design, code, test, correct and document moderate-to-complex programs and scripts from agreed specifications and subsequent iterations, collaborate with others to review specifications where appropriate.</p>\n<p>Technical Understanding</p>\n<p>You will understand the core technical concepts related to the role, and apply them with guidance.</p>\n<p>Testing</p>\n<p>You will review requirements and specifications, and define test conditions, identify issues and risks associated with work, analyse and report test activities and results.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7275ef33-009","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976928777","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$114,400 to $171,600","x-skills-required":["Proficiency in programming language such as Python or Java","Experience with Big Data technologies such as Hadoop, Spark, and Kafka","Familiarity with ETL processes and tools","Knowledge of SQL and NoSQL databases","Strong understanding of relational databases","Experience with data warehousing solutions","Proficiency with cloud platforms","Expertise in data modeling and design","Experience in designing and building scalable data pipelines","Experience with RESTful APIs and data integration"],"x-skills-preferred":["Relevant certifications (e.g., GCP Certified, AWS Certified, Azure Certified)","Bachelor's degree in Computer Science, Data Engineering, Information Technology, or a related field","Strong analytical and communication skills","Ability to work collaboratively in a team environment","High level of accuracy and attention to detail"],"datePosted":"2026-04-18T22:12:56.654Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"Proficiency in programming language such as Python or Java, Experience with Big Data technologies such as Hadoop, Spark, and Kafka, Familiarity with ETL processes and tools, Knowledge of SQL and NoSQL databases, Strong understanding of relational databases, Experience with data warehousing solutions, Proficiency with cloud platforms, Expertise in data modeling and design, Experience in designing and building scalable data pipelines, Experience with RESTful APIs and data integration, Relevant certifications (e.g., GCP Certified, AWS Certified, Azure Certified), Bachelor's degree in Computer Science, Data Engineering, Information Technology, or a related field, Strong analytical and communication skills, Ability to work collaboratively in a team environment, High level of accuracy and attention to detail","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":114400,"maxValue":171600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f59c4a7f-68e"},"title":"Test Engineering Leader - Evinova","description":"<p>Join Evinova, a health-tech business, in accelerating better health outcomes by advancing digital transformation across the life sciences sector. We&#39;re building AI-native products that reshape how clinical trials are designed, documented, and delivered. As a Test Engineering Leader, you&#39;ll own quality across multiple products, shape the testing culture for cross-functional squads, and pioneer approaches to challenges that simply didn&#39;t exist a few years ago.</p>\n<p><strong>Shape Testing Strategy for AI-Native Products</strong></p>\n<ul>\n<li>Design and own end-to-end test strategies,automated and manual,for products that integrate LLMs, generative AI, and complex data pipelines.</li>\n<li>Develop novel evaluation frameworks for LLM output quality, prompt regression testing, and RAG retrieval accuracy.</li>\n<li>Select, implement, and continuously improve testing tools and frameworks in collaboration with engineering leadership and platform excellence teams.</li>\n<li>Drive a measurable shift from manual to automated testing, with clear metrics to track progress.</li>\n</ul>\n<p><strong>Own Quality and Release Readiness</strong></p>\n<ul>\n<li>Be the single point of accountability for testing across all releases of your two products.</li>\n<li>Build and maintain automation frameworks and scripts that keep pace with rapid release cycles.</li>\n<li>Analyse test results, spot trends, and turn data into actionable improvements,not just reports.</li>\n<li>Prepare release-readiness documentation and quality artifacts that satisfy both internal stakeholders and GxP compliance requirements.</li>\n</ul>\n<p><strong>Lead, Coach, and Build Culture</strong></p>\n<ul>\n<li>Partner with engineering managers, scrum masters, and delivery leads across squads and geographies to establish a shared quality vision.</li>\n<li>Lead external contract test engineering squads by influence,setting standards, mentoring team members, and modelling engineering excellence.</li>\n<li>Champion a quality-engineering mindset: everyone ships quality, not just the test team.</li>\n<li>Stay hands-on,write test cases, debug failures, and pair with engineers when the situation calls for it.</li>\n</ul>\n<p><strong>Essential Skills and Experience</strong></p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Computer Engineering, Mathematics, Information Science, or a related field (or equivalent practical experience).</li>\n<li>10+ years of hands-on software test engineering experience across the full stack (UI, API, data).</li>\n<li>Proven ability to lead test strategy and mentor other test engineers,whether through formal management or technical leadership.</li>\n<li>Strong automation skills with modern frameworks such as Playwright, Selenium, or Cypress, plus scripting fluency in Python or a comparable language.</li>\n<li>Solid experience with REST API testing, database validation (PostgreSQL, MongoDB, or similar), and CI/CD-integrated test pipelines.</li>\n<li>A data-driven approach to quality: you define metrics, instrument dashboards, and use evidence to drive decisions.</li>\n</ul>\n<p><strong>Highly Preferred</strong></p>\n<ul>\n<li>Experience testing AI/ML-powered products,especially LLM evaluation, prompt testing, RAG validation, or output-quality benchmarking.</li>\n<li>Familiarity with GxP software validation, computerised system validation (CSV), or regulated-industry quality practices.</li>\n<li>Background in life sciences, health-tech, or clinical-trial technology.</li>\n<li>Experience working with geographically distributed teams and external vendor squads.</li>\n</ul>\n<p>If you&#39;re looking for a role where deep test engineering craft meets the frontier of AI,and where your work genuinely improves patient outcomes,this is it.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f59c4a7f-68e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Clinical Development Platforms, Evinova","sameAs":"https://www.evinova.com","logo":"https://logos.yubhub.co/evinova.com.png"},"x-apply-url":"https://astrazeneca.eightfold.ai/careers/job/563877689883511","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$165,656.80 - $217,424.55 USD","x-skills-required":["Bachelor's degree in Computer Science, Computer Engineering, Mathematics, Information Science, or a related field","10+ years of hands-on software test engineering experience across the full stack (UI, API, data)","Proven ability to lead test strategy and mentor other test engineers","Strong automation skills with modern frameworks such as Playwright, Selenium, or Cypress","Solid experience with REST API testing, database validation, and CI/CD-integrated test pipelines"],"x-skills-preferred":["Experience testing AI/ML-powered products","Familiarity with GxP software validation, computerised system validation (CSV), or regulated-industry quality practices","Background in life sciences, health-tech, or clinical-trial technology","Experience working with geographically distributed teams and external vendor squads"],"datePosted":"2026-04-18T22:12:50.182Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Gaithersburg, Maryland, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"Bachelor's degree in Computer Science, Computer Engineering, Mathematics, Information Science, or a related field, 10+ years of hands-on software test engineering experience across the full stack (UI, API, data), Proven ability to lead test strategy and mentor other test engineers, Strong automation skills with modern frameworks such as Playwright, Selenium, or Cypress, Solid experience with REST API testing, database validation, and CI/CD-integrated test pipelines, Experience testing AI/ML-powered products, Familiarity with GxP software validation, computerised system validation (CSV), or regulated-industry quality practices, Background in life sciences, health-tech, or clinical-trial technology, Experience working with geographically distributed teams and external vendor squads","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165656.8,"maxValue":217424.55,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_aa5f286d-ad4"},"title":"Senior Genome Editing Digital Pipeline Scientist","description":"<p>At Bayer, we&#39;re seeking a Senior Genome Editing Digital Pipeline Scientist to drive the data vision that powers next-generation gene-edited products. As a Data Strategy &amp; Pipeline Leader in Gene Editing, you will coordinate a holistic data strategy across the editing pipeline so that diverse genomic and biological datasets are connected, accessible, and ready for advanced analytics. You will work closely with multi-functional teams to ensure that data, models, and decision tools are seamlessly integrated into product development workflows, enabling faster, more informed decisions and impactful innovation in gene-edited germplasm.</p>\n<p>Your primary responsibilities will include providing leadership to define and coordinate the data strategy that enables data-driven, model-based analytics for improved gene-edited germplasm, including accelerating data connectivity across the editing pipeline with multi-functional teams. You will also lead cross-functional projects with partners across Crop Science to automate decision making and connect data assets that accelerate development of gene-edited products.</p>\n<p>In addition, you will translate complex business data knowledge, scientific workflows, and product needs into clear technical implementation plans that can be executed by data scientists, data engineers, and developers. You will design and guide the development of robust data systems and analytics pipelines that support a wide variety of genomic and computational biology use cases and can scale with future business needs.</p>\n<p>As a key communicator and integrator between scientific, technical, and business stakeholders, you will align roadmaps, prioritize initiatives, and ensure that data and analytics solutions deliver measurable value. You will also attract, mentor, and develop talent, serving as a coach for peers and colleagues in key areas of expertise to support their professional growth and build a strong data and analytics community.</p>\n<p>Finally, you will champion and support Health, Safety &amp; Environment, Compliance, Business Conduct, and Human Rights policies and culture in all activities and collaborations.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_aa5f286d-ad4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976715204","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$114,400.00 - $171,600.00","x-skills-required":["PhD in Genomics, Computational Biology, Evolution, Quantitative Genetics, or a related scientific field","Minimum of 6 years of relevant experience, or MS with 10+ years of experience","Experience in the analysis of large biological datasets and in developing analytical pipelines using Python, R, or similar software and programming languages","Ability to design and implement data systems and analytical pipelines that can support a broad range of scientific and business use cases","Strong collaboration skills, demonstrated through building cross-functional partnerships and influencing others to drive results and solve complex business problems"],"x-skills-preferred":["Strong understanding of the genomic control of physiological and biochemical pathways in plants or animals","Experience developing data systems and analytical pipelines that leverage genome-wide association (GWA) data, QTL analysis, candidate gene analysis, gene expression analysis, molecular marker development, and pedigree data"],"datePosted":"2026-04-18T22:12:21.373Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Life Sciences","skills":"PhD in Genomics, Computational Biology, Evolution, Quantitative Genetics, or a related scientific field, Minimum of 6 years of relevant experience, or MS with 10+ years of experience, Experience in the analysis of large biological datasets and in developing analytical pipelines using Python, R, or similar software and programming languages, Ability to design and implement data systems and analytical pipelines that can support a broad range of scientific and business use cases, Strong collaboration skills, demonstrated through building cross-functional partnerships and influencing others to drive results and solve complex business problems, Strong understanding of the genomic control of physiological and biochemical pathways in plants or animals, Experience developing data systems and analytical pipelines that leverage genome-wide association (GWA) data, QTL analysis, candidate gene analysis, gene expression analysis, molecular marker development, and pedigree data","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":114400,"maxValue":171600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_efa9d52e-e4e"},"title":"Consultant Specialist","description":"<p>Join HSBC and discover how valued you&#39;ll be in a career where you can make a real impression. We are currently seeking an experienced professional to join our team in the role of Consultant Specialist.</p>\n<p>As a Consultant Specialist, you will be responsible for understanding release cycles within GPE and aligned testing associated with it. You will also manage the release deployment on various environments, manage automation initiatives required for the stable functioning of the environments like healthchecks / monitoring, triage and escalate High priority Incidents to get focused resolutions, reduce overall downtime for end-to-end testing by identifying opportunities in environment upgrades, communicate with regional/country project teams, technical leads and Asset teams, drive root cause analysis with involved teams, analyse issues monthly and run continuous improvement cycles for incidents, raise engagement with partner application / other teams to facilitate the issue resolution, engage and drive resolution where necessary; aiding support to SME&#39;s from application teams to facilitate resolution, report updates for ongoing issues to stakeholders, including exec level management.</p>\n<p>You will also be responsible for infrastructure management activities which comprise of critical vulnerability fixing, OS/DB/MQ patching, certificate renewals etc. Additionally, you will lead and mentor team to achieve above responsibilities successfully.</p>\n<p>Knowledge &amp; Experience / Qualifications:</p>\n<ul>\n<li>Must have strong UNIX and shell scripting experience.</li>\n<li>Basic knowledge on middleware products like WAS/MQ.</li>\n<li>Experience of DevOps tools like Jenkins, GITHUB etc.</li>\n<li>Experience of Control-M, Connect Direct (C:D).</li>\n<li>Experience of deployment / change pipelines like CI / CD.</li>\n<li>Flexible to work in shifts, on weekends, after work hours and on call support as per the need of the project.</li>\n<li>Good understanding of the Payment schemes and e2e flows for US scheme payments.</li>\n<li>Strong communication skills (verbal, written, and presentation of complex information and data).</li>\n<li>Stakeholder Management and working in dynamic environment</li>\n</ul>\n<p>Time management - Ability to prioritize project criticality based on requirements and business needs. Strong analytical skills supported by good decision making and problem solving skills &amp; attitude. Ability to work independently with a hands-on approach. Good project management skills. Knowledge of programming language such Java or Python. Knowledge Automation tools. Knowledge of multiple clearing systems. Hands on experience of CI/D pipelines /51/LP/WX</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_efa9d52e-e4e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC Software Development (GuangDong) Limited","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610678275","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["UNIX","shell scripting","middleware products","DevOps tools","Control-M","Connect Direct","deployment / change pipelines","CI / CD","Payment schemes","e2e flows","US scheme payments","communication skills","stakeholder Management","dynamic environment","project criticality","analytical skills","decision making","problem solving skills","hands-on approach","project management skills","programming language","Java","Python","Automation tools","multiple clearing systems","CI/D pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:11:25.401Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Guangzhou"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"UNIX, shell scripting, middleware products, DevOps tools, Control-M, Connect Direct, deployment / change pipelines, CI / CD, Payment schemes, e2e flows, US scheme payments, communication skills, stakeholder Management, dynamic environment, project criticality, analytical skills, decision making, problem solving skills, hands-on approach, project management skills, programming language, Java, Python, Automation tools, multiple clearing systems, CI/D pipelines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6d7fadcc-6fa"},"title":"Data Scientist Computer Vision","description":"<p>At Bayer, we&#39;re seeking a talented Data Scientist with deep learning and machine learning expertise focused on image-based data to help shape the future of agriculture. In this role, you&#39;ll join a dynamic team that supports the development of Bayer Crop Science next-generation products by applying computer vision to automate critical processes across the Plant Biotechnology organisation.</p>\n<p>The primary responsibilities of this role are to:</p>\n<p>Solve real agricultural problems using deep learning and AI across image and other data modalities, translating complex models into tangible business and scientific impact.</p>\n<p>Design and implement end-to-end machine learning pipelines for computer vision use cases, including segmentation, classification, detection, and multi-task learning.</p>\n<p>Prototype, evaluate, and iterate on cutting-edge architectures such as CNNs, Vision Transformers, foundational and large-scale vision models, ensuring state-of-the-art performance.</p>\n<p>Optimize models for accuracy, robustness, and inference efficiency, including experimentation with hyperparameters, compression, and deployment-oriented optimisations.</p>\n<p>Independently build scalable data pipelines for training, validation, and evaluation, including data ingestion, augmentation strategies, and active learning loops.</p>\n<p>Collaborate cross-functionally with product, data, and software engineering teams to integrate models into production systems and deliver reliable, maintainable solutions.</p>\n<p>Contribute to MLOps practices, including model versioning, deployment, monitoring, and retraining workflows using modern tooling and cloud-based platforms.</p>\n<p>Build strong cross-functional relationships and actively engage with the broader Data Science Community to share best practices, align on standards, and co-create innovative solutions.</p>\n<p>Present clear, compelling, and validated stories about experiments, results, and recommendations to peers, senior management, and internal customers to drive strategic and operational decisions.</p>\n<p>We seek an incumbent who possesses the following:</p>\n<p>M.S. with 2+ years of experience or Ph.D. in Computer Science, Electrical Engineering, or a related field with a focus on machine learning or computer vision.</p>\n<p>Proficiency in Python and experience with deep learning frameworks such as PyTorch or TensorFlow.</p>\n<p>Hands-on experience with modern computer vision architectures including models such as ResNet, UNet, DeepLab, YOLO, SegFormer, SAM, and Vision Transformers.</p>\n<p>Strong background in handling large-scale datasets and creating custom datasets, for example using frameworks such as Hugging Face Datasets.</p>\n<p>Solid understanding of core machine learning concepts including loss functions, regularization, optimisation, and learning rate scheduling.</p>\n<p>Experience developing and deploying models using cloud-based ML platforms such as AWS SageMaker.</p>\n<p>Familiarity with Unix environments, including bash, file systems, and core utilities.</p>\n<p>Strong engineering practices including use of Git, Docker, CI/CD pipelines, modular codebase design, and unit testing.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6d7fadcc-6fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976908666","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$109,370.40 - $164,055.60","x-skills-required":["Python","PyTorch","TensorFlow","ResNet","UNet","DeepLab","YOLO","SegFormer","SAM","Vision Transformers","Hugging Face Datasets","AWS SageMaker","Git","Docker","CI/CD pipelines","modular codebase design","unit testing"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:11:10.602Z","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Manufacturing","skills":"Python, PyTorch, TensorFlow, ResNet, UNet, DeepLab, YOLO, SegFormer, SAM, Vision Transformers, Hugging Face Datasets, AWS SageMaker, Git, Docker, CI/CD pipelines, modular codebase design, unit testing","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":109370.4,"maxValue":164055.6,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3f2cb60f-80a"},"title":"Senior Genome Editing Digital Enablement","description":"<p>At Bayer, we&#39;re seeking a Senior Genome Editing Digital Enablement Scientist to join our team. As a key partner and enabler of multi-disciplinary teams, you will design large-scale data systems and analytical pipelines that power our gene editing efforts. You will develop analytical tools that connect biological and operations data to support more efficient and accurate decisions across the gene editing pipeline. Your expertise in both computational biology and genetics will be essential in driving and coordinating multi-functional teams to enable robust data connectivity and interoperability across the editing pipeline.</p>\n<p>In this role, you will lead cross-functional projects with IT, Data Engineering, Genome Editing, and other partner teams to automate decision making and connect data to accelerate development of gene-edited products. You will translate complex biological processes into scalable digital workflows that support decision making, advancement, and prioritization within the gene editing program. Your strong ability to collaborate and lead in cross-functional, multi-disciplinary teams will be crucial in influencing without authority and aligning diverse stakeholders around shared digital solutions.</p>\n<p>As a member of the Biology and Genome Design community, you will actively build your own acumen and capabilities while sharing best practices with others. You will serve as a key communicator and thought partner on digital enablement strategy, clearly articulating requirements, trade-offs, and opportunities to scientific and non-scientific stakeholders.</p>\n<p>We seek an incumbent who possesses a PhD in Genomics, Computational Biology, Evolution, Quantitative Genetics, or another relevant scientific field with a minimum of 6 years of relevant experience, or an MS with 10+ years of experience developing data systems and analytics pipelines that enable product delivery using genetic and computational biology datasets.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3f2cb60f-80a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Bayer Crop Science","sameAs":"https://talent.bayer.com","logo":"https://logos.yubhub.co/talent.bayer.com.png"},"x-apply-url":"https://talent.bayer.com/careers/job/562949976613783","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$114,400.00 - $171,600.00","x-skills-required":["computational biology","genetics","data systems","analytical pipelines","Python","R","large-scale biological datasets"],"x-skills-preferred":["genome-wide association GWAs data","QTL analysis","candidate gene analysis","gene expression analysis","molecular marker development","pedigree data"],"datePosted":"2026-04-18T22:11:02.858Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Chesterfield"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Manufacturing","skills":"computational biology, genetics, data systems, analytical pipelines, Python, R, large-scale biological datasets, genome-wide association GWAs data, QTL analysis, candidate gene analysis, gene expression analysis, molecular marker development, pedigree data","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":114400,"maxValue":171600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f7aeee90-9b7"},"title":"Technical Specialist (Java, Microservices) / Associate Director, Software Engineering","description":"<p>Join HSBC and help you stand out in your career. We offer opportunities, support and rewards that will take you further. As an Associate Director, Software Engineering, you will lead the development and implementation of Microservices-based solutions using Java. You will also architect and design scalable, distributed systems with high availability, collaborate with cross-functional teams to gather requirements and deliver solutions, ensure code quality through best practices, code reviews, and automated testing, mentor and guide team members in technical aspects and career growth, troubleshoot and resolve complex technical issues in production environments, stay updated with emerging technologies and recommend their adoption, navigate a dynamic ecosystem to deliver change effectively, demonstrating initiative, self-motivation, and drive, and exhibit tenacity and determination to clarify business requirements and deliver solutions in occasionally challenging circumstances.</p>\n<p>To be successful in this role, you should have strong proficiency in Java (Java 21 preferred), hands-on experience with Microservices architecture and frameworks (e.g., Spring Boot, Spring Cloud), expertise in RESTful APIs, messaging systems (e.g., Kafka, Hazelcast), and containerization (e.g., Docker, Kubernetes), solid understanding of cloud platforms (e.g., Kubernetes platform, GCP and AWS), hands-on experience with CI/CD pipelines and DevOps practices, knowledge of database technologies (SQL and NoSQL), payment&#39;s domain experience and clearing scheme experience, excellent problem-solving and communication skills, hands-on experience in both SDLC and Agile methodologies, familiarity with monitoring tools (e.g., Prometheus, Grafana, Splunk), and certifications in Java or cloud technologies are a plus.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f7aeee90-9b7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610662228","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Microservices architecture","Spring Boot","Spring Cloud","RESTful APIs","Kafka","Hazelcast","Docker","Kubernetes","CI/CD pipelines","DevOps practices","database technologies","SQL","NoSQL","payment's domain experience","clearing scheme experience"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:10:55.397Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hyderabad, Telangana, India · Bangalore, Karnataka, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Microservices architecture, Spring Boot, Spring Cloud, RESTful APIs, Kafka, Hazelcast, Docker, Kubernetes, CI/CD pipelines, DevOps practices, database technologies, SQL, NoSQL, payment's domain experience, clearing scheme experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_aee9464f-897"},"title":"Technical Specialist (Java, Microservices) / Associate Director, Software Engineering","description":"<p>We are currently seeking an experienced professional to join our team in the role of a Associate Director, Software Engineering.</p>\n<p>In this role, you will lead the development and implementation of Microservices-based solutions using Java. You will also architect and design scalable, distributed systems with high availability, collaborate with cross-functional teams to gather requirements and deliver solutions, ensure code quality through best practices, code reviews, and automated testing, mentor and guide team members in technical aspects and career growth, troubleshoot and resolve complex technical issues in production environments, stay updated with emerging technologies and recommend their adoption, navigate a dynamic ecosystem to deliver change effectively, demonstrating initiative, self-motivation, and drive, exhibit tenacity and determination to clarify business requirements and deliver solutions in occasionally challenging circumstances.</p>\n<p>To be successful in this role, you should meet the following requirements:</p>\n<ul>\n<li>Strong proficiency in Java (Java 21 preferred).</li>\n<li>Hands-on experience with Microservices architecture and frameworks (e.g., Spring Boot, Spring Cloud).</li>\n<li>Expertise in RESTful APIs, messaging systems (e.g., Kafka, Hazelcast), and containerization (e.g., Docker, Kubernetes).</li>\n<li>Solid understanding of cloud platforms (e.g., Kubernetes platform, GCP and AWS).</li>\n<li>Hands-on experience with CI/CD pipelines and DevOps practices.</li>\n<li>Knowledge of database technologies (SQL and NoSQL).</li>\n<li>Payment&#39;s domain experience and clearing scheme experience.</li>\n<li>Excellent problem-solving and communication skills.</li>\n<li>Hands-on experience in both SDLC and Agile methodologies.</li>\n<li>Familiarity with monitoring tools (e.g., Prometheus, Grafana, Splunk).</li>\n<li>Certifications in Java or cloud technologies are a plus.</li>\n</ul>\n<p>You&#39;ll achieve more when you join HSBC.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_aee9464f-897","directApply":true,"hiringOrganization":{"@type":"Organization","name":"HSBC","sameAs":"https://portal.careers.hsbc.com","logo":"https://logos.yubhub.co/portal.careers.hsbc.com.png"},"x-apply-url":"https://portal.careers.hsbc.com/careers/job/563774610662222","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Microservices","Spring Boot","Spring Cloud","RESTful APIs","Kafka","Hazelcast","Docker","Kubernetes","CI/CD pipelines","DevOps practices","database technologies","SQL","NoSQL","payment's domain experience","clearing scheme experience"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:10:31.133Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, Hyderabad"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Java, Microservices, Spring Boot, Spring Cloud, RESTful APIs, Kafka, Hazelcast, Docker, Kubernetes, CI/CD pipelines, DevOps practices, database technologies, SQL, NoSQL, payment's domain experience, clearing scheme experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6bfc6b4-74f"},"title":"Senior Data Scientist - Marketing (all genders)","description":"<p>Join our Business Intelligence Department, a multidisciplinary group of Data Scientists, Analysts, and Data Engineers. Together, we build machine learning and analytics products that directly influence GMV, conversion, and retention.</p>\n<p>Within the department, we’re building a new Marketing Analytics team and are looking for a Senior Data Scientist to drive its data science initiatives. In this role, you’ll work closely with Analysts, Engineers, and Marketing stakeholders to develop and productionize advanced machine learning, statistical, and predictive models that improve marketing performance and drive measurable company growth.</p>\n<p>As a Senior Data Scientist – Marketing, you’ll take strong ownership of data science initiatives that directly shape our marketing strategy and growth. You will:</p>\n<p>Partner closely with Marketing, Marketing Analytics, and Marketing Technology to identify opportunities and translate business questions into scalable data science solutions.</p>\n<p>Lead the development of high-impact machine learning and statistical models for marketing use cases such as channel allocation, ad bidding, churn prediction, lifetime value, revenue attribution, and business metrics forecasting.</p>\n<p>Work end-to-end - from translating business questions into hypotheses to researching, building, validating, and deploying models.</p>\n<p>Run experiments and iterate in production: design A/B tests, monitor model performance, and continuously improve based on measured impact.</p>\n<p>Advance our MLOps practices with CI/CD pipelines, retraining workflows, lineage tracking, and documentation.</p>\n<p>Help define the team&#39;s roadmap and ways of working as a founding member of Marketing Analytics - your input will help shape this function.</p>\n<p>Act as a senior role model in the team, sharing best practices and helping raise the bar for data science at Holidu.</p>\n<p>We&#39;re looking for someone with 5+ years of experience as a Data Scientist, with clear ownership of projects that delivered measurable business impact. You should have a degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field, and strong expertise in machine learning, statistics, and predictive analytics, with hands-on experience using Python and SQL.</p>\n<p>Experience with marketing data science use cases such as attribution modeling, customer lifetime value prediction, churn modeling, or bid optimization is also required. You should have a solid understanding of marketing concepts across channels (e.g. Performance Marketing, SEO, CRM, Affiliate) and how data science can improve them.</p>\n<p>Additionally, you should have experience working with modern data stacks, ideally including AWS (Redshift, Athena, S3), Airflow, dbt, and Git. A collaborative mindset paired with great communication skills is essential, as you&#39;ll need to work with diverse stakeholders and explain complex topics in a simple way.</p>\n<p>AI proficiency is also a plus, as you&#39;ll be comfortable using AI to enhance coding, planning, and monitoring, and successfully integrating AI tools (such as Claude code, Codex, Copilot, etc.) into your workflow and teaching others to use them efficiently.</p>\n<p>If you&#39;re excited about the opportunity to shape the future of travel with products used by millions of guests and thousands of hosts, apply now!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6bfc6b4-74f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2510157","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Machine Learning","Statistics","Predictive Analytics","Python","SQL","Marketing Data Science","Attribution Modeling","Customer Lifetime Value Prediction","Churn Modeling","Bid Optimization"],"x-skills-preferred":["AI","CI/CD Pipelines","Retraining Workflows","Lineage Tracking","Documentation","Airflow","dbt","Git"],"datePosted":"2026-04-18T22:10:24.739Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, Statistics, Predictive Analytics, Python, SQL, Marketing Data Science, Attribution Modeling, Customer Lifetime Value Prediction, Churn Modeling, Bid Optimization, AI, CI/CD Pipelines, Retraining Workflows, Lineage Tracking, Documentation, Airflow, dbt, Git"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_90b5ac1d-d16"},"title":"Senior Software Engineer, Backend — Frontier Data","description":"<p>The Frontier Data team builds the data and systems that power Scale&#39;s most advanced Frontier AI use cases. We&#39;re looking for a Senior Backend Engineer who thrives in ambiguity, moves fast, and enjoys tackling daunting challenges.</p>\n<p>As a Senior Backend Engineer, you will own major backend systems for frontier agentic data products, driving projects from early exploration through production deployment. You will build scalable services and pipelines that support agent workflows, architect modular, reusable backend systems, and operate in high-ambiguity environments.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Designing and building scalable systems while partnering closely with research, product, operations, and other engineering teams</li>\n<li>Building scalable services and pipelines that support agent workflows</li>\n<li>Architecting modular, reusable backend systems that adapt to evolving product needs</li>\n<li>Operating in high-ambiguity environments and breaking down open-ended problems</li>\n<li>Partnering cross-functionally with product, research/ML, and infrastructure teams</li>\n</ul>\n<p>Ideal experience includes 5+ years of full-time software engineering experience, strong backend engineering fundamentals, and experience building systems that scale.</p>\n<p>Compensation packages at Scale include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors.</p>\n<p>Additional benefits include comprehensive health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_90b5ac1d-d16","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Frontier Data","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4648525005","x-work-arrangement":null,"x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["Distributed systems","API design","Data modeling","Production reliability","Docker","Containerized development/production environments","SQL","Modern database-backed application development"],"x-skills-preferred":["Async processing","Workflow engines","Data pipelines"],"datePosted":"2026-04-18T16:01:34.567Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Distributed systems, API design, Data modeling, Production reliability, Docker, Containerized development/production environments, SQL, Modern database-backed application development, Async processing, Workflow engines, Data pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b40b693d-a0d"},"title":"Senior Software Engineer, Agentic Data Products","description":"<p>We&#39;re forming a new Agentic Data Products team focused on building the next generation of agent-powered tools that ground AI in real operational workflows. Our goal is to help enterprises demystify their data layers and deploy intelligent, agentic systems that can reason over data, take action, and deliver measurable outcomes.</p>\n<p>This is a 0→1 build team. We’re looking for a sharp, product-minded Senior Engineer who thrives in ambiguity, moves quickly, and enjoys building new systems from scratch alongside customers and cross-functional partners. You’ll work closely with product, forward-deployed engineers, data scientists, and applied AI teams to turn real-world problems into scalable, production solutions.</p>\n<p>If you like shipping fast, owning outcomes, and working across the stack,from polished frontends to distributed backends to LLM integrations,this role is for you.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Own major full-stack product areas, driving features from concept and design through production deployment</li>\n<li>Build intuitive, high-performance frontend experiences using React + TypeScript</li>\n<li>Develop reliable backend services in Python, working with distributed systems, data pipelines, and AI/ML infrastructure</li>\n<li>Integrate LLMs, vector databases, and agentic frameworks to power intelligent workflows and decision-making systems</li>\n<li>Ship quickly through tight experimentation loops while maintaining high quality and reliability</li>\n<li>Help define the technical direction and architecture of a brand-new team and product surface</li>\n<li>Adapt across the stack and learn new tools as needed to solve real problems end-to-end</li>\n</ul>\n<p><strong>Ideal Experience</strong></p>\n<ul>\n<li>5+ years of full-time software engineering experience</li>\n<li>0-1 product build experience</li>\n<li>Familiarity with LLMs, embeddings, vector databases, or modern AI data products/tools</li>\n<li>Experience with distributed systems and cloud-based architectures</li>\n<li>Prior experience mentoring or leading team</li>\n</ul>\n<p><strong>What We Value</strong></p>\n<ul>\n<li>Strong product intuition and customer empathy</li>\n<li>Bias toward action and rapid iteration</li>\n<li>Ownership mentality , you see problems through to outcomes</li>\n<li>Comfort collaborating across engineering, product, data science, and applied AI</li>\n<li>Excitement about building agentic systems that make AI genuinely useful in the real world</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b40b693d-a0d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4653827005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["React","TypeScript","Python","Distributed systems","Data pipelines","AI/ML infrastructure","LLMs","Vector databases","Agentic frameworks"],"x-skills-preferred":[],"datePosted":"2026-04-18T16:01:14.176Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, TypeScript, Python, Distributed systems, Data pipelines, AI/ML infrastructure, LLMs, Vector databases, Agentic frameworks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_740da2af-174"},"title":"Security Engineer, Detection & Response","description":"<p>We are seeking a Senior Security Engineer with a specialty in Detection and Incident Response to join our Security Engineering team. This role sits at the intersection of security operations and software engineering, requiring you to investigate incidents and build the systems that detect, contain, and prevent them.</p>\n<p>You will design and ship high-precision detections across cloud services and enterprise SaaS, develop automation that shortens response timelines, and mature the telemetry pipelines that make it all possible. Your ability to write production-quality code is just as important as your ability to triage an alert.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Engineer, test, and deploy detection logic across cloud and enterprise environments, treating detections as software with version control, peer review, and measurable performance.</li>\n</ul>\n<ul>\n<li>Build and maintain incident response automation, runbooks, and tooling that reduce containment timelines without sacrificing developer velocity.</li>\n</ul>\n<ul>\n<li>Mature telemetry pipelines through improved schema design, normalization, enrichment, and quality checks that reduce false positives and increase signal fidelity.</li>\n</ul>\n<ul>\n<li>Perform digital incident investigations to identify and contain potential security breaches.</li>\n</ul>\n<ul>\n<li>Conduct digital forensics and malware analysis to understand attack vectors and adversary methodologies.</li>\n</ul>\n<ul>\n<li>Integrate alerting with messaging and ticketing systems to enable fast, traceable response workflows.</li>\n</ul>\n<ul>\n<li>Partner cross-functionally with IT, security, and engineering teams to harden identity and access patterns, close logging and forensics gaps, and implement maintainable guardrails that scale with the organisation.</li>\n</ul>\n<ul>\n<li>Utilize threat intelligence platforms to improve hunting, detection, and response workflows.</li>\n</ul>\n<ul>\n<li>Clearly explain the significance and impact of incidents, providing actionable recommendations to both technical and non-technical stakeholders.</li>\n</ul>\n<p>Ideal Candidate:</p>\n<ul>\n<li>5+ years of experience in Detection Engineering, Incident Response, or Security Operations, with a strong emphasis on building and shipping security tooling and automation.</li>\n</ul>\n<ul>\n<li>Proficiency in at least one programming language (e.g., Python, Go) and comfort writing production-grade code , not just scripts.</li>\n</ul>\n<ul>\n<li>Hands-on experience designing or improving detection pipelines, SIEM content, and alerting workflows in cloud-native environments.</li>\n</ul>\n<ul>\n<li>Practical experience with SIEM, EDR, and SOAR tools, with a preference for candidates who have built integrations or extended these platforms programmatically.</li>\n</ul>\n<ul>\n<li>Strong understanding of modern cyber threats, common attack techniques, and adversary TTPs.</li>\n</ul>\n<ul>\n<li>Familiarity with digital forensics tools and malware analysis techniques.</li>\n</ul>\n<ul>\n<li>Experience with cloud-native environments (e.g., AWS, GCP, Azure) and the security telemetry those environments generate.</li>\n</ul>\n<ul>\n<li>Exposure to threat intelligence platforms and integrating intel into detection and investigation workflows.</li>\n</ul>\n<ul>\n<li>Strong communication skills, with the ability to translate complex security findings into clear business impact.</li>\n</ul>\n<ul>\n<li>Relevant security certifications (e.g., GCIH, GCFA, GCIA, CISSP, GDSA) are a plus.</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_740da2af-174","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4684073005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$237,600-$297,000 USD","x-skills-required":["Detection Engineering","Incident Response","Security Operations","Cloud Services","Enterprise SaaS","Automation","Telemetry Pipelines","Digital Forensics","Malware Analysis","Threat Intelligence Platforms","SIEM","EDR","SOAR","Cloud-Native Environments","Programming Languages","Python","Go"],"x-skills-preferred":["Hands-on experience designing or improving detection pipelines, SIEM content, and alerting workflows in cloud-native environments","Practical experience with SIEM, EDR, and SOAR tools, with a preference for candidates who have built integrations or extended these platforms programmatically","Strong understanding of modern cyber threats, common attack techniques, and adversary TTPs","Familiarity with digital forensics tools and malware analysis techniques","Experience with cloud-native environments (e.g., AWS, GCP, Azure) and the security telemetry those environments generate","Exposure to threat intelligence platforms and integrating intel into detection and investigation workflows","Strong communication skills, with the ability to translate complex security findings into clear business impact","Relevant security certifications (e.g., GCIH, GCFA, GCIA, CISSP, GDSA)"],"datePosted":"2026-04-18T16:00:14.303Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY; San Francisco, CA; Seattle, WA; Washington, DC"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Detection Engineering, Incident Response, Security Operations, Cloud Services, Enterprise SaaS, Automation, Telemetry Pipelines, Digital Forensics, Malware Analysis, Threat Intelligence Platforms, SIEM, EDR, SOAR, Cloud-Native Environments, Programming Languages, Python, Go, Hands-on experience designing or improving detection pipelines, SIEM content, and alerting workflows in cloud-native environments, Practical experience with SIEM, EDR, and SOAR tools, with a preference for candidates who have built integrations or extended these platforms programmatically, Strong understanding of modern cyber threats, common attack techniques, and adversary TTPs, Familiarity with digital forensics tools and malware analysis techniques, Experience with cloud-native environments (e.g., AWS, GCP, Azure) and the security telemetry those environments generate, Exposure to threat intelligence platforms and integrating intel into detection and investigation workflows, Strong communication skills, with the ability to translate complex security findings into clear business impact, Relevant security certifications (e.g., GCIH, GCFA, GCIA, CISSP, GDSA)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":237600,"maxValue":297000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_56e29c57-cd1"},"title":"Robotics Technician","description":"<p>We&#39;re seeking a Robotics Technician to join our team in Mexico City. As a key contributor, you will partner with cross-functional stakeholders to bring up new robots and productionize the maintenance of robots and collection hardware. You will play a critical role in supporting the day-to-day operations of the factory by bringing up and maintaining robots and collection hardware. You will also provide technical support for data collection operations, manage physical inventory, maintain equipment, and coordinate logistics.</p>\n<p>You will become a subject matter expert on all capabilities of the robotics platforms deployed in the factory. You will develop technical domain expertise in areas of 2D and 3D imaging and annotation, multi-sensor fusion and calibration, GPS/INS navigation systems, computer vision, and other autonomy-adjacent concepts.</p>\n<p>You have a Bachelor&#39;s degree or industry experience, an engineering background, preferably in Computer Science, Mathematics, or other Engineering fields. You have 2+ years of experience developing with Python, C++, Java, and/or other scripting languages. You have 1-3 years of experience in hardware labs or a manufacturing environment. You have experience managing risk and operating robots safely. You have strong project management and interpersonal skills, high attention to detail, and a strong sense of ownership. You have a high level of comfort communicating effectively across internal and external organizations.</p>\n<p>Nice to have: hands-on experience in Robotics, AI, and/or Computer Vision, intellectual curiosity, empathy, and ability to operate with a high degree of autonomy, experience building and/or maintaining lab networks and data pipelines, experience running large-scale data collection and controlled experiments, experience building out facilities, and experience in logistics.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_56e29c57-cd1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4635128005","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","C++","Java","Robotics","AI","Computer Vision","Multi-sensor fusion and calibration","GPS/INS navigation systems"],"x-skills-preferred":["hands-on experience in Robotics, AI, and/or Computer Vision","intellectual curiosity","empathy","ability to operate with a high degree of autonomy","experience building and/or maintaining lab networks and data pipelines","experience running large-scale data collection and controlled experiments","experience building out facilities","experience in logistics"],"datePosted":"2026-04-18T16:00:01.904Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mexico City, MX"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, C++, Java, Robotics, AI, Computer Vision, Multi-sensor fusion and calibration, GPS/INS navigation systems, hands-on experience in Robotics, AI, and/or Computer Vision, intellectual curiosity, empathy, ability to operate with a high degree of autonomy, experience building and/or maintaining lab networks and data pipelines, experience running large-scale data collection and controlled experiments, experience building out facilities, experience in logistics"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_460d00aa-b48"},"title":"Senior / Staff+ Software Engineer, Voice Platform","description":"<p>About the role</p>\n<p>We&#39;re building the infrastructure that lets people talk to Claude,real-time, bidirectional voice conversations that feel natural, responsive, and safe. This is foundational work for how millions of people will interact with AI.</p>\n<p>The Voice Platform team designs and operates the serving systems, streaming pipelines, and APIs that bring Anthropic&#39;s audio models from research into production across Claude.ai, our mobile apps, and the Anthropic API. You&#39;ll work at the intersection of real-time media, low-latency inference, and distributed systems,building infrastructure where every millisecond of latency is felt by the user.</p>\n<p>We partner closely with the Audio research team, who train the speech understanding and generation models, and with product teams shipping voice experiences to users. Your job is to make those models fast, reliable, and delightful to talk to at scale.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Design and build the real-time streaming infrastructure that powers voice conversations with Claude,ingesting microphone audio, orchestrating model inference, and streaming synthesized speech back with minimal latency</li>\n</ul>\n<ul>\n<li>Build low-latency serving systems for speech models, optimizing time-to-first-audio and end-to-end conversational responsiveness</li>\n</ul>\n<ul>\n<li>Develop the public and internal APIs that expose voice capabilities to Claude.ai, mobile clients, and third-party developers</li>\n</ul>\n<ul>\n<li>Own the audio transport layer,codecs, jitter buffers, adaptive bitrate, packet loss recovery,so conversations stay smooth across unreliable networks</li>\n</ul>\n<ul>\n<li>Build observability and quality-measurement systems for voice: latency distributions, audio quality metrics, interruption handling, and turn-taking accuracy</li>\n</ul>\n<ul>\n<li>Partner with Audio research to move new model architectures from experiment to production, and feed real-world performance data back into research</li>\n</ul>\n<ul>\n<li>Collaborate with mobile and product engineering on client-side audio capture, playback, and the end-to-end user experience</li>\n</ul>\n<p>You may be a good fit if you</p>\n<ul>\n<li>Have 6+ years of experience building distributed systems, real-time infrastructure, or platform services at scale</li>\n</ul>\n<ul>\n<li>Have shipped production systems where latency is measured in tens of milliseconds and users notice when you miss</li>\n</ul>\n<ul>\n<li>Are comfortable working across the stack,from transport protocols and serving infrastructure up to the APIs product teams build on</li>\n</ul>\n<ul>\n<li>Are results-oriented, with a bias toward flexibility and impact</li>\n</ul>\n<ul>\n<li>Pick up slack, even if it goes outside your job description</li>\n</ul>\n<ul>\n<li>Enjoy pair programming (we love to pair!)</li>\n</ul>\n<ul>\n<li>Care about the societal impacts of voice AI and want to help shape how these systems are developed responsibly</li>\n</ul>\n<ul>\n<li>Are comfortable with ambiguity,voice is a fast-moving space, and you&#39;ll help define the architecture as we learn what works</li>\n</ul>\n<p>Strong candidates may also have experience with</p>\n<ul>\n<li>Real-time media protocols and stacks: WebRTC, RTP, gRPC bidirectional streaming, or WebSockets at scale</li>\n</ul>\n<ul>\n<li>Audio engineering fundamentals: codecs (Opus, AAC), voice activity detection, echo cancellation, jitter buffering, or audio DSP</li>\n</ul>\n<ul>\n<li>Low-latency ML inference serving, streaming model outputs, or GPU-based serving infrastructure</li>\n</ul>\n<ul>\n<li>Telephony, live streaming, video conferencing, or voice assistant platforms</li>\n</ul>\n<ul>\n<li>Mobile audio pipelines on iOS (AVAudioEngine, AudioUnits) or Android (Oboe, AAudio)</li>\n</ul>\n<ul>\n<li>Working alongside ML researchers to productionize models,speech experience is a plus but not required</li>\n</ul>\n<p>Representative projects</p>\n<ul>\n<li>Driving time-to-first-audio below human perceptual thresholds by co-designing the serving pipeline with the Audio research team</li>\n</ul>\n<ul>\n<li>Building a streaming inference orchestrator that interleaves speech recognition, LLM reasoning, and speech synthesis with overlapping execution</li>\n</ul>\n<ul>\n<li>Designing the voice mode API surface for the Anthropic API so developers can build their own voice agents on Claude</li>\n</ul>\n<ul>\n<li>Implementing graceful barge-in and interruption handling so users can cut Claude off mid-sentence naturally</li>\n</ul>\n<ul>\n<li>Instrumenting end-to-end audio quality metrics and building dashboards that catch regressions before users do</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_460d00aa-b48","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5172245008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$485,000 USD","x-skills-required":["Real-time media protocols and stacks","Audio engineering fundamentals","Low-latency ML inference serving","Distributed systems","Streaming pipelines","APIs"],"x-skills-preferred":["WebRTC","RTP","gRPC bidirectional streaming","WebSockets","Opus","AAC","Voice activity detection","Echo cancellation","Jitter buffering","Audio DSP","GPU-based serving infrastructure","Telephony","Live streaming","Video conferencing","Voice assistant platforms","Mobile audio pipelines on iOS","Android","Working alongside ML researchers"],"datePosted":"2026-04-18T15:59:54.712Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Real-time media protocols and stacks, Audio engineering fundamentals, Low-latency ML inference serving, Distributed systems, Streaming pipelines, APIs, WebRTC, RTP, gRPC bidirectional streaming, WebSockets, Opus, AAC, Voice activity detection, Echo cancellation, Jitter buffering, Audio DSP, GPU-based serving infrastructure, Telephony, Live streaming, Video conferencing, Voice assistant platforms, Mobile audio pipelines on iOS, Android, Working alongside ML researchers","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b68ff4cc-e74"},"title":"Data Engineer, Safeguards","description":"<p><strong>About the role</strong></p>\n<p>Anthropic is looking for a Data Engineer to join the Safeguards team and build the data foundations that keep our AI systems safe. The Safeguards team works to monitor models, prevent misuse, and ensure user well-being.</p>\n<p>You&#39;ll design and build the data pipelines, warehousing solutions, and analytical tooling that power our safety and trust efforts at scale. You&#39;ll work closely with engineers, data scientists, and policy teams to ensure the Safeguards organization has the data it needs to detect abuse patterns, measure the effectiveness of safety interventions, and make informed decisions about model behavior and enforcement.</p>\n<p>This is a high-impact role where your work will directly support Anthropic&#39;s mission to develop AI that is safe and beneficial.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, build, and maintain scalable data pipelines that support safety monitoring, abuse detection, and enforcement workflows</li>\n<li>Develop and optimize data models and warehousing solutions to enable efficient analysis of large-scale usage and safety data</li>\n<li>Build and maintain dashboards and reporting infrastructure that give Safeguards teams visibility into model behavior, misuse patterns, and enforcement outcomes</li>\n<li>Collaborate with engineers to integrate data from multiple sources , including model outputs, user reports, and automated classifiers , into a unified analytical layer</li>\n<li>Implement data quality frameworks, monitoring, and alerting to ensure the reliability of safety-critical data</li>\n<li>Partner with research teams to surface data insights that inform model improvements and safety interventions</li>\n<li>Develop self-service data tooling that enables stakeholders to explore safety data and generate reports independently</li>\n<li>Contribute to data governance practices, including access controls, retention policies, and privacy-compliant data handling</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have 3+ years of experience in data engineering, analytics engineering, or a related role</li>\n<li>Are proficient in SQL and Python, with experience building and maintaining ETL/ELT pipelines</li>\n<li>Have hands-on experience with modern data stack tools such as dbt, Airflow, Spark, or similar orchestration and transformation frameworks</li>\n<li>Have worked with cloud data platforms (BigQuery, Redshift, Snowflake, or similar)</li>\n<li>Are comfortable building dashboards and data visualizations using tools like Looker, Tableau, or Metabase</li>\n<li>Communicate clearly and can translate complex data concepts for both technical and non-technical audiences</li>\n<li>Are results-oriented, flexible, and willing to pick up slack even when it falls outside your job description</li>\n<li>Care about the societal impacts of AI and are motivated by safety work</li>\n</ul>\n<p><strong>Strong candidates may have:</strong></p>\n<ul>\n<li>Experience with trust &amp; safety, integrity, fraud, or abuse detection data systems</li>\n<li>Experience with large-scale event streaming systems (Kafka, Pub/Sub, Kinesis)</li>\n<li>Built data infrastructure that supports ML model monitoring or evaluation</li>\n<li>A background in statistical analysis, or experience collaborating closely with data scientists</li>\n<li>Developed internal tooling or self-service analytics platforms</li>\n</ul>\n<p><strong>Strong candidates need not have:</strong></p>\n<ul>\n<li>A formal degree in Computer Science or a related field , we value practical experience and demonstrated ability over credentials</li>\n<li>Prior experience in AI or machine learning , you&#39;ll learn the domain-specific context on the job</li>\n<li>Previous experience at an AI safety or research organization</li>\n<li>Deep expertise across every tool listed above , familiarity with a subset and a willingness to learn is enough</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices. Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills. The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b68ff4cc-e74","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5156057008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"£170,000-£220,000 GBP","x-skills-required":["SQL","Python","ETL/ELT pipelines","dbt","Airflow","Spark","cloud data platforms","BigQuery","Redshift","Snowflake","Looker","Tableau","Metabase"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:33.960Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, ETL/ELT pipelines, dbt, Airflow, Spark, cloud data platforms, BigQuery, Redshift, Snowflake, Looker, Tableau, Metabase","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":170000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3aedc59f-428"},"title":"Senior Forward Deployed AI Engineer, Enterprise","description":"<p>As a Senior Forward Deployed AI Engineer on our Enterprise team, you&#39;ll be the technical bridge between Scale AI&#39;s cutting-edge AI capabilities and our most strategic customers. You&#39;ll work with enterprise clients to understand their unique challenges, architect custom AI solutions, and ensure successful deployment and adoption of AI systems in production environments.</p>\n<p>This is a hands-on technical role that combines deep engineering expertise with customer-facing problem solving. You&#39;ll work directly with customer engineering teams to integrate AI into their critical workflows.</p>\n<p><strong>Key Responsibilities</strong></p>\n<p><strong>Customer Integration &amp; Deployment</strong></p>\n<ul>\n<li>Partner directly with enterprise customers to understand their technical infrastructure, data pipelines, and business requirements</li>\n<li>Design and implement custom integrations between Scale AI&#39;s platform and customer data environments (cloud platforms, data warehouses, internal APIs)</li>\n<li>Build robust data connectors and ETL pipelines to ingest, process, and prepare customer data for AI workflows</li>\n<li>Deploy and configure AI models and agents within customer security and compliance boundaries</li>\n</ul>\n<p><strong>AI Agent Development</strong></p>\n<ul>\n<li>Develop production-grade AI agents tailored to customer use cases across domains like customer support, data analysis, content generation, and workflow automation</li>\n<li>Architect multi-agent systems that orchestrate between different models, tools, and data sources</li>\n<li>Implement evaluation frameworks to measure agent performance and iterate toward business objectives</li>\n<li>Design human-in-the-loop workflows and feedback mechanisms for continuous agent improvement</li>\n</ul>\n<p><strong>Prompt Engineering &amp; Optimization</strong></p>\n<ul>\n<li>Create sophisticated prompt engineering strategies optimized for customer-specific domains and data</li>\n<li>Build and maintain prompt libraries, templates, and best practices for customer use cases</li>\n<li>Conduct systematic prompt experimentation and A/B testing to improve model outputs</li>\n<li>Implement RAG (Retrieval Augmented Generation) systems and fine-tuning pipelines where appropriate</li>\n</ul>\n<p><strong>Technical Leadership &amp; Collaboration</strong></p>\n<ul>\n<li>Serve as the primary technical point of contact for strategic enterprise accounts</li>\n<li>Collaborate with customer data scientists, ML engineers, and software developers to ensure smooth integration</li>\n<li>Provide technical training and knowledge transfer to customer teams</li>\n<li>Work closely with Scale&#39;s product and engineering teams to translate customer needs into product improvements</li>\n<li>Document technical architectures, integration patterns, and best practices</li>\n</ul>\n<p><strong>Problem Solving &amp; Innovation</strong></p>\n<ul>\n<li>Debug complex technical issues across the entire stack, from data pipelines to model outputs</li>\n<li>Rapidly prototype solutions to unblock customers and prove out new use cases</li>\n<li>Stay current on the latest AI/ML research and tools, bringing innovative approaches to customer problems</li>\n<li>Identify opportunities for productization based on common customer patterns</li>\n</ul>\n<p><strong>Required Qualifications</strong></p>\n<ul>\n<li>4+ years of software engineering experience with strong fundamentals in data structures, algorithms, and system design</li>\n<li>Production Python expertise with experience in modern ML/AI frameworks (e.g., LangChain, LlamaIndex, HuggingFace, OpenAI API)</li>\n<li>Experience with cloud platforms (AWS, GCP, or Azure) and modern data infrastructure</li>\n<li>Strong problem-solving skills with the ability to navigate ambiguous requirements and rapidly iterate toward solutions</li>\n<li>Excellent communication skills with the ability to explain complex technical concepts to both technical and non-technical audiences</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Agent Development Wiz</li>\n<li>Deep understanding of LLMs including prompting techniques, embeddings, and RAG architectures</li>\n<li>Experience building and deploying AI agents or autonomous systems in production</li>\n<li>Knowledge of vector databases and semantic search systems</li>\n<li>Contributions to open-source AI/ML projects</li>\n</ul>\n<ul>\n<li>Infrastructure Guru</li>\n<li>Experience with containerization (Docker, Kubernetes) and CI/CD pipelines</li>\n<li>Experience using Terraform, Bicep, or other Infrastructure as Code (IaC) tools</li>\n<li>Previous work in a devops, platform, or infra role</li>\n</ul>\n<ul>\n<li>Customer Product Whisperer</li>\n<li>Proven ability to work with customers in a technical consulting, solutions engineering, or product engineering role</li>\n<li>Domain expertise in verticals like finance, healthcare, government, or manufacturing</li>\n<li>Experience with technical enablement or teaching programs</li>\n</ul>\n<p><strong>Sample Projects</strong></p>\n<p>The following are some examples of the types of projects we’ve worked on with customers. All of these projects leverage customer data, integrate directly into customers’ existing systems, and are deployed on their infrastructure.</p>\n<ul>\n<li>Deep Research for Due Diligence</li>\n<li>Churn Prediction</li>\n<li>Data Extraction Voice Agent</li>\n</ul>\n<p><strong>Compensation</strong></p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p><strong>Pay Transparency</strong></p>\n<p>For pay transparency purposes, the base salary range for this full-time position in the locations of San Francisco, New York, Seattle is: $216,000-$270,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3aedc59f-428","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4597399005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["Software engineering","Data structures","Algorithms","System design","Python","ML/AI frameworks","Cloud platforms","Modern data infrastructure","Problem-solving","Communication"],"x-skills-preferred":["LLMs","Prompting techniques","Embeddings","RAG architectures","Containerization","CI/CD pipelines","Infrastructure as Code","Devops","Platform","Infra"],"datePosted":"2026-04-18T15:59:30.214Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Software engineering, Data structures, Algorithms, System design, Python, ML/AI frameworks, Cloud platforms, Modern data infrastructure, Problem-solving, Communication, LLMs, Prompting techniques, Embeddings, RAG architectures, Containerization, CI/CD pipelines, Infrastructure as Code, Devops, Platform, Infra","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5b703e8a-47c"},"title":"Robotics Engineer","description":"<p>We&#39;re looking for a talented Robotics Engineer to join our team in San Francisco. As a key contributor, you will work to build out our robotics fleet and software systems for collecting data and performing evaluations.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Developing systems for collecting data from various robotics embodiments and collection modalities</li>\n<li>Designing and building hardware for retrofitting robots and building custom collection modalities</li>\n<li>Contributing to the development of pipelines and tooling to support robotics initiatives</li>\n<li>Owning hardware and software integrations for various robots</li>\n<li>Partnering with cross-functional stakeholders to scale up data services</li>\n<li>Providing technical support for data collection operations and executing on pilots to stand up new workflows</li>\n<li>Becoming a subject matter expert on all capabilities of the robotics labs</li>\n</ul>\n<p>You will have the opportunity to develop technical domain expertise in areas of 2D and 3D imaging and annotation, multi-sensor fusion and calibration, computer vision, machine learning, and other autonomy-adjacent concepts.</p>\n<p>We&#39;re looking for someone with a strong engineering background, preferably in Computer Science, Mathematics, or other Engineering fields. You should have 3+ years of experience developing with Python, C++, Java and/or other scripting language, as well as 1-3 years of experience in hardware labs or a manufacturing environment, 1-3 years of experience in mechanical design and comfort with CAD, hands-on experience in robotics, AI, and computer vision, experience building and/or maintaining lab networks and data pipelines, experience running large-scale data collection and controlled experiments, experience managing risk and operating robots safely, strong project management and interpersonal skills, high attention to detail, and a strong sense of ownership.</p>\n<p>As a Robotics Engineer at Scale, you will have the opportunity to work with a talented team of engineers and researchers to develop cutting-edge robotics solutions. You will be responsible for designing, building, and testing robotics systems, as well as collaborating with cross-functional teams to integrate robotics into our data collection and analysis pipeline.</p>\n<p>We offer a competitive salary range of $208,800-$261,000 USD, as well as a comprehensive benefits package, including health, dental, and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5b703e8a-47c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4655744005","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$208,800-$261,000 USD","x-skills-required":["Python","C++","Java","Mechanical design","CAD","Robotics","AI","Computer vision","Machine learning","Data pipelines","Lab networks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:13.725Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, C++, Java, Mechanical design, CAD, Robotics, AI, Computer vision, Machine learning, Data pipelines, Lab networks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":208800,"maxValue":261000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_60a7e1e6-b51"},"title":"Tech Lead/Manager, Machine Learning Research Scientist- LLM Evals","description":"<p>As the leading data and evaluation partner for frontier AI companies, we&#39;re dedicated to advancing the evaluation and benchmarking of large language models (LLMs). Our Research teams work with the industry&#39;s leading AI labs to provide high-quality data and accelerate progress in GenAI research.</p>\n<p>We&#39;re seeking a Tech Lead Manager to lead a talented team of research scientists and research engineers focused on developing and implementing novel evaluation methodologies, metrics, and benchmarks to assess the capabilities and limitations of our cutting-edge LLMs.</p>\n<p>Key responsibilities:</p>\n<ul>\n<li>Lead a team of highly effective research scientists and research engineers on LLM evals.</li>\n<li>Conduct research on the effectiveness and limitations of existing LLM evaluation techniques.</li>\n<li>Design and develop novel evaluation benchmarks for large language models, covering areas such as instruction following, factuality, robustness, and fairness.</li>\n<li>Communicate, collaborate, and build relationships with clients and peer teams to facilitate cross-functional projects.</li>\n<li>Collaborate with internal teams and external partners to refine metrics and create standardized evaluation protocols.</li>\n<li>Implement scalable and reproducible evaluation pipelines using modern ML frameworks.</li>\n<li>Publish research findings in top-tier AI conferences and contribute to open-source benchmarking initiatives.</li>\n</ul>\n<p>Ideal candidate has 5+ years of hands-on experience in large language model, NLP, and Transformer modeling, in the setting of both research and engineering development. Experience supporting and leading a team of research scientists and research engineers is also required.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_60a7e1e6-b51","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4304790005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$264,800-$331,000 USD","x-skills-required":["large language model","NLP","Transformer modeling","research and engineering development","team leadership","cross-functional collaboration","evaluation methodologies","metrics and benchmarks","scalable and reproducible evaluation pipelines","modern ML frameworks"],"x-skills-preferred":["published research in top-tier AI conferences","open-source benchmarking initiatives","customer-facing role"],"datePosted":"2026-04-18T15:59:10.794Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; Seattle, WA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"large language model, NLP, Transformer modeling, research and engineering development, team leadership, cross-functional collaboration, evaluation methodologies, metrics and benchmarks, scalable and reproducible evaluation pipelines, modern ML frameworks, published research in top-tier AI conferences, open-source benchmarking initiatives, customer-facing role","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":264800,"maxValue":331000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_770c5fe8-cce"},"title":"Staff Security Engineer, Vulnerability Management","description":"<p>We are seeking a Staff Security Engineer to lead the most complex technical work in CoreWeave&#39;s Vulnerability Management program.</p>\n<p>As a Staff Security Engineer, you will design and implement scalable triage, prioritization, and remediation-tracking systems across application, infrastructure, and hardware domains. You will set technical standards, drive high-impact initiatives, and mentor engineers through technical leadership, while partnering with leadership on priorities and execution risks.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Lead high-complexity VM technical initiatives and deliver architecture decisions for assigned program areas</li>\n<li>Design and build scalable triage automation, including integrations, decision logic, and production hardening</li>\n<li>Implement end-to-end workflow components from assessment and detection to ticket routing and remediation tracking</li>\n<li>Provide deep technical leadership on hardware-adjacent vulnerabilities (GPU firmware, DPU firmware/BlueField, and BMC surfaces)</li>\n<li>Act as senior technical responder for embargoed disclosures and zero-day events, coordinating with owner teams that deploy fixes</li>\n<li>Improve prioritization logic, severity models, and exception workflows through code, design reviews, and technical proposals</li>\n<li>Produce actionable technical metrics and risk insights for leadership consumption</li>\n<li>Lead root-cause analysis for high-impact vulnerability incidents and implement durable technical improvements</li>\n<li>Mentor IC3/IC4/IC5 engineers through design guidance, code review, and incident coaching</li>\n<li>Partner with security, engineering, and operational stakeholders to improve workflow reliability and accelerate remediation outcomes</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>9+ years of relevant experience with demonstrated strategic impact in vulnerability management, application security, platform security, or cloud security engineering</li>\n<li>Proven track record building and scaling security automation (SOAR workflows, AI/ML systems, detection pipelines) in production environments</li>\n<li>Deep subject matter expertise with vulnerability management best practices: CVSS, EPSS, CISA KEV, threat intelligence integration, and risk-based prioritization frameworks</li>\n<li>Excellent development background with strong coding skills in Python, Go, or similar languages for building scalable, production-grade security systems</li>\n<li>Significant experience with modern vulnerability management tooling (for example Wiz, Semgrep, Rapid7, Tenable, or equivalent)</li>\n<li>Experience with specialized infrastructure: GPU/DPU environments, firmware security, hardware vulnerabilities, or high-performance computing</li>\n<li>Demonstrated track record mentoring engineers across levels and driving cross-functional technical initiatives at organizational scale</li>\n<li>Strong business acumen and understanding of how security decisions impact engineering velocity, customer trust, and business outcomes</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Practical experience building AI/ML-powered security systems (LLM integration, automated decision-making, human-in-the-loop validation) in production</li>\n<li>Experience managing hardware vendor security partnerships (embargoed disclosures and pre-release collaboration)</li>\n<li>Production experience with security automation platforms such as TINES and serverless frameworks (AWS Lambda, GCP Cloud Functions)</li>\n<li>Strong DevOps, DevSecOps, or SRE background with deep experience in AWS/GCP/Azure cloud services and Infrastructure as Code (Terraform, CloudFormation)</li>\n<li>Deep understanding of Kubernetes security (container scanning, admission controllers, supply chain security, runtime protection)</li>\n<li>Experience leading security programs through rapid hypergrowth (10x+ infrastructure scaling) in startup or cloud-native environments</li>\n<li>Practical experience managing vulnerabilities within a FedRAMP-certified environment or similar regulatory frameworks</li>\n</ul>\n<p>Salary and Benefits: The base salary range for this role is $188,000 to $275,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>Work Environment:</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_770c5fe8-cce","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4653130006","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$188,000 to $275,000","x-skills-required":["vulnerability management","application security","platform security","cloud security engineering","security automation","AI/ML systems","detection pipelines","Python","Go","modern vulnerability management tooling","GPU/DPU environments","firmware security","hardware vulnerabilities","high-performance computing"],"x-skills-preferred":["AI/ML-powered security systems","LLM integration","automated decision-making","human-in-the-loop validation","security automation platforms","TINES","serverless frameworks","AWS Lambda","GCP Cloud Functions","DevOps","DevSecOps","SRE","Kubernetes security","container scanning","admission controllers","supply chain security","runtime protection"],"datePosted":"2026-04-18T15:59:06.360Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"vulnerability management, application security, platform security, cloud security engineering, security automation, AI/ML systems, detection pipelines, Python, Go, modern vulnerability management tooling, GPU/DPU environments, firmware security, hardware vulnerabilities, high-performance computing, AI/ML-powered security systems, LLM integration, automated decision-making, human-in-the-loop validation, security automation platforms, TINES, serverless frameworks, AWS Lambda, GCP Cloud Functions, DevOps, DevSecOps, SRE, Kubernetes security, container scanning, admission controllers, supply chain security, runtime protection","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":188000,"maxValue":275000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_76c9a01c-58a"},"title":"Data Center Portfolio Planning & Execution Lead","description":"<p>We&#39;re looking for a Data Center Portfolio Planning &amp; Execution Lead to drive the planning and framework that ensures every site moves smoothly from the front-end phases through design, construction, equipment delivery, commissioning, and operational readiness.</p>\n<p>This role owns the portfolio-level operating system: translating capacity supply pipeline into integrated project plans that span every phase of delivery, building the tooling and automation that runs it at scale, and maintaining Anthropic&#39;s datacenter capacity catalog , a lifecycle view of our fleet that supports both execution orchestration and steady-state capacity planning.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Manage the integrated master plan for each site across the portfolio , stitching power ramp, design, construction, sourcing, deployment, and operations readiness into a single coordinated schedule with clear milestones and dependencies</li>\n<li>Develop and maintain Anthropic&#39;s datacenter catalog for deployed and in-progress capacity. Manage the portfolio-level view of physical infrastructure &amp; cluster interfaces across all sites and partners to enable planning decisions such as equipment fungibility, accelerator platforms, tech insertion, or workload allocation</li>\n<li>Define and run the stage gates and decision locks for cluster delivery , from lease execution to design lock through procurement, construction, equipment installation, commissioning, and handover</li>\n<li>Drive gate reviews, manage exceptions, and track the downstream impact of deviations across the portfolio</li>\n<li>Manage portfolio reviews and risk tracking for DC Infra leadership and Compute Supply</li>\n</ul>\n<p>Tooling &amp; process:</p>\n<ul>\n<li>Develop tooling and automation to enable cross-functional planning flow-down from datacenter capacity availability dates</li>\n<li>Partner with Design, Supply Chain, Construction, and DC Ops program leads to drive cross-pillar process improvements as portfolio scales</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Are familiar with the full datacenter buildout lifecycle: pipeline → design → sourcing → construction → Cx → deployment</li>\n<li>Have run integrated portfolio or master-schedule planning across a fleet of capital projects (datacenter, energy, fab, or similar) where multiple functional orgs each own a phase</li>\n<li>Have built a stage-gate or decision-lock system from scratch and gotten functional leads to adopt it</li>\n<li>Have re-architected a deployment or delivery process at scale and can point to the cycle-time or throughput result</li>\n<li>Build the tooling yourself using AI-assisted development , stand up planning dashboards, schedule automation, and data pipelines from Smartsheet/P6/partner systems</li>\n<li>Proactively surface schedule risk across functions , comfortable flagging a problem in someone else&#39;s domain before it becomes a slip</li>\n<li>Track record of driving outcomes through influence with cross-functional partners</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Experience building a portfolio planning and execution function from scratch at a hyperscaler or large industrial owner</li>\n<li>Exposure to capacity planning or S&amp;OP processes that connect demand forecast to physical build</li>\n<li>Experience product-managing internal planning, workflow, or scheduling systems</li>\n</ul>\n<p>The annual compensation range for this role is $365,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_76c9a01c-58a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5188939008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$365,000-$485,000 USD","x-skills-required":["data center portfolio planning","execution lead","portfolio-level operating system","capacity supply pipeline","integrated project plans","tooling and automation","datacenter capacity catalog","lifecycle view of fleet","execution orchestration","steady-state capacity planning","stage gates","decision locks","cluster delivery","lease execution","design lock","procurement","construction","equipment installation","commissioning","handover","cross-functional planning","flow-down","datacenter capacity availability dates","cross-pillar process improvements","AI-assisted development","planning dashboards","schedule automation","data pipelines","Smartsheet","P6","partner systems","schedule risk","cross-functional partners","portfolio planning","execution function","hyperscaler","large industrial owner","capacity planning","S&OP processes","demand forecast","physical build","internal planning","workflow","scheduling systems"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:03.702Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-Friendly, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data center portfolio planning, execution lead, portfolio-level operating system, capacity supply pipeline, integrated project plans, tooling and automation, datacenter capacity catalog, lifecycle view of fleet, execution orchestration, steady-state capacity planning, stage gates, decision locks, cluster delivery, lease execution, design lock, procurement, construction, equipment installation, commissioning, handover, cross-functional planning, flow-down, datacenter capacity availability dates, cross-pillar process improvements, AI-assisted development, planning dashboards, schedule automation, data pipelines, Smartsheet, P6, partner systems, schedule risk, cross-functional partners, portfolio planning, execution function, hyperscaler, large industrial owner, capacity planning, S&OP processes, demand forecast, physical build, internal planning, workflow, scheduling systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":365000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f931591c-87a"},"title":"Research Scientist, Frontier Risk Evaluations","description":"<p>As a Research Scientist focused on Frontier Risk Evaluations, you will design and create evaluation measures, harnesses and datasets for measuring the risks posed by frontier AI systems.</p>\n<p>For example, you might do any or all of the following:</p>\n<ul>\n<li>Design and build harnesses to test AI models and systems (including agents) for dangerous capabilities such as security vulnerability exploitation, CBRN uplift, and other high-risk activities;</li>\n</ul>\n<ul>\n<li>Work with government agencies or other labs to collectively scope and design evaluations to measure and mitigate risks posed by advanced AI systems;</li>\n</ul>\n<ul>\n<li>Publish evaluation methodologies and write technical reports for policymakers.</li>\n</ul>\n<p>We are seeking talented researchers to join us in shaping this vision.</p>\n<p>Ideally you&#39;d have:</p>\n<ul>\n<li>Commitment to our mission of promoting safe, secure, and trustworthy AI deployments in the industry as frontier AI capabilities continue to advance;</li>\n</ul>\n<ul>\n<li>Practical experience conducting technical research collaboratively. You should be comfortable building and instrumenting ML pipelines, writing evaluation harnesses, and quickly turning new ideas from the research literature into working prototypes;</li>\n</ul>\n<ul>\n<li>A track record of published research in machine learning, particularly in generative AI;</li>\n</ul>\n<ul>\n<li>At least three years of experience addressing sophisticated ML problems, whether in a research setting or in product development;</li>\n</ul>\n<ul>\n<li>Strong written and verbal communication skills to operate in a cross-functional team.</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience in crafting evaluations and benchmarks, or a background in data science roles related to LLM technologies;</li>\n</ul>\n<ul>\n<li>Experience with red-teaming or adversarial testing of AI systems;</li>\n</ul>\n<ul>\n<li>Familiarity with AI safety policy frameworks (e.g., NIST AI RMF, EU AI Act, Korea AI Basic Act).</li>\n</ul>\n<p>Our research interviews are crafted to assess candidates&#39; skills in practical ML prototyping and debugging, their grasp of research concepts, and their alignment with our organisational culture. We will not ask any LeetCode-style questions. If you’re excited about advancing AI safety and contributing to our mission, we encourage you to apply, even if your experience doesn’t perfectly align with every requirement.</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity-based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f931591c-87a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4677657005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["machine learning","generative AI","ML pipelines","evaluation harnesses","AI safety policy frameworks"],"x-skills-preferred":["crafting evaluations and benchmarks","data science roles related to LLM technologies","red-teaming or adversarial testing of AI systems"],"datePosted":"2026-04-18T15:58:57.212Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"machine learning, generative AI, ML pipelines, evaluation harnesses, AI safety policy frameworks, crafting evaluations and benchmarks, data science roles related to LLM technologies, red-teaming or adversarial testing of AI systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_47de4683-b45"},"title":"Staff+ Software Engineer, Platform","description":"<p>We are looking for experienced software engineers to join our Platform organisation. We build the foundational primitives that accelerate product development across Anthropic, and own infrastructure and systems that teams depend on to ship reliably and at scale.</p>\n<p>As a Staff+ Software Engineer, you will independently scope complex, multi-month projects, drive cross-org alignment through ambiguous problem spaces, and make architectural decisions that shape how Anthropic builds and scales its products. You will partner directly with research to productize cutting-edge capabilities, and will have lasting impact on the platform that hundreds of thousands of companies and internal/external engineers depend on every day.</p>\n<p>Our team is responsible for Platform Acceleration, Service Infra, Multicloud, Auth &amp; Identity, and Connectivity. We work on maximising developer productivity of product engineers at Anthropic, building and maintaining the core infrastructure that powers Anthropic&#39;s engineering organisation, operating across multiple cloud providers, and powering identity and authentication across Anthropic&#39;s product suite.</p>\n<p>You will work on problems where reliability and enterprise trust are the bar: token refresh at scale, admin controls that let IT govern what agents can do, proxy infrastructure that stays up when partner servers don&#39;t. We ship for claude.ai, Claude Code, Cowork, and the API.</p>\n<p>Relevant experience includes OAuth, API gateways, multi-tenant platforms, building for enterprise, and MCP.</p>\n<p>We are looking for someone with 8-10+ years of practical full-stack engineering experience, ideally with 2+ years operating at a Staff or equivalent technical leadership level. You should have led the design and delivery of complex, consumer or B2B user-facing products across the full stack, and take a product-focused approach to building solutions that are robust, scalable, and easy to use.</p>\n<p>Strong candidates may also have served as a technical lead or architect for a foundational platform system, owning both the technical vision and execution end-to-end, or experience designing or scaling billing, payments, or financial infrastructure at high transaction volumes.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_47de4683-b45","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.co/","logo":"https://logos.yubhub.co/anthropic.co.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5157847008","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["OAuth","API gateways","multi-tenant platforms","building for enterprise","MCP"],"x-skills-preferred":["ML training infra","production ML pipelines","backend engineering","finetuning experience"],"datePosted":"2026-04-18T15:58:42.613Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"OAuth, API gateways, multi-tenant platforms, building for enterprise, MCP, ML training infra, production ML pipelines, backend engineering, finetuning experience","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3ba73370-831"},"title":"Internal Audit IT Manager","description":"<p>Ready to be pushed beyond what you think you’re capable of?</p>\n<p>At Coinbase, our mission is to increase economic freedom in the world.</p>\n<p>We’re seeking a very specific candidate who is passionate about our mission and who believes in the power of crypto and blockchain technology to update the financial system.</p>\n<p>As an Internal Audit IT Manager, you will own end-to-end delivery of complex IT and security audits across our cloud infrastructure, security operations, and crypto-native systems.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Owning end-to-end delivery of IT and security audits, from risk assessment and scoping through planning, fieldwork, testing, reporting, and issue validation,covering cloud infrastructure (AWS, GCP), security operations, identity and access management, data protection, IT asset management, vendor/third-party risk, and key in-scope products and services including blockchain infrastructure, centralized and self-hosted wallets, and cold storage.</li>\n</ul>\n<ul>\n<li>Driving AI-enabled audit execution, designing and implementing data analytics, automation, and Generative AI solutions to modernize how we audit (e.g., continuous monitoring, anomaly detection, automated evidence retrieval, AI-assisted workpaper drafting),while maintaining rigorous human-in-the-loop validation to ensure accuracy and audit-quality conclusions.</li>\n</ul>\n<ul>\n<li>Executing audits aligned with the multi-year IT and security audit roadmap, coordinating coverage with co-sourced partners and cross-functional risk initiatives while ensuring alignment with Coinbase&#39;s enterprise risk profile, technology strategy, and regulatory expectations across regions (US, EMEA, APAC).</li>\n</ul>\n<ul>\n<li>Driving high-quality, risk-based findings and executive-level reporting, distilling key themes, emerging risks, and root causes into clear, concise materials for senior management and the Chief Audit Executive,ensuring findings are appropriately documented and supported by evidence.</li>\n</ul>\n<ul>\n<li>Partnering with technology and security leadership across Engineering, Security, Infrastructure, Product, and Operations to build trusted relationships, challenge control design, and advise on pragmatic, risk-based, scalable remediation while maintaining third-line independence.</li>\n</ul>\n<ul>\n<li>Driving disciplined issue management, ensuring timely, risk-based remediation by management, high-quality root cause analysis, and validation of remediation activities,escalating delays or thematic concerns to senior leadership as needed.</li>\n</ul>\n<ul>\n<li>Evaluating and developing talent, assessing candidates and helping build a high-performing, technically credible audit team.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>7+ years of experience in IT/security internal audit, technology risk, or first-line security/engineering roles with significant controls exposure.</li>\n</ul>\n<ul>\n<li>Experience working in a fast-paced, cloud-native, or engineering-driven environment where technology and security practices evolve rapidly.</li>\n</ul>\n<ul>\n<li>Hands-on audit experience with cloud platforms (AWS, GCP), including IAM policies, security configurations, logging/monitoring, and CI/CD pipelines.</li>\n</ul>\n<ul>\n<li>AI-forward mindset with demonstrated experience applying Python, SQL, or AI tools to audit or security work, building workflows rather than just prompting.</li>\n</ul>\n<ul>\n<li>Relevant professional certifications (e.g., CISA, CISSP, CIA, CISM) required; CPA or CFE a plus.</li>\n</ul>\n<ul>\n<li>Working knowledge of key frameworks such as NIST CSF, COBIT, SOC 2, and ITIL.</li>\n</ul>\n<ul>\n<li>High EQ and collaborative style.</li>\n</ul>\n<ul>\n<li>Proven ability to translate complex technical findings into clear, executive-ready narratives for both technical and non-technical audiences.</li>\n</ul>\n<ul>\n<li>Ability to manage multiple audits and initiatives across time zones (EMEA, APAC) with minimal oversight.</li>\n</ul>\n<ul>\n<li>Demonstrated leadership and team-development experience, including mentoring, coaching, and managing direct reports.</li>\n</ul>\n<ul>\n<li>Demonstrates the ability to responsibly use generative AI tools and copilots (e.g., LibreChat, Gemini, Glean) in daily workflows, continuously learn as tools evolve, and apply human-in-the-loop practices to deliver business-ready outputs and drive measurable improvements in efficiency, cost, and quality.</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Experience auditing or building blockchain infrastructure, crypto custody, or wallet systems (hot/cold storage).</li>\n</ul>\n<ul>\n<li>Background in a high-growth or rapidly scaling environment with complex, evolving technology stacks.</li>\n</ul>\n<ul>\n<li>Experience with GRC platforms (Workiva, Archer, AuditBoard) or building custom audit automation tooling.</li>\n</ul>\n<ul>\n<li>Familiarity with DORA, MiCA, or crypto-specific regulatory frameworks.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3ba73370-831","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Coinbase","sameAs":"https://www.coinbase.com/","logo":"https://logos.yubhub.co/coinbase.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coinbase/jobs/7755116","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,345-$195,700 USD","x-skills-required":["IT security","Cloud infrastructure","Security operations","Identity and access management","Data protection","IT asset management","Vendor/third-party risk","Blockchain infrastructure","Centralized and self-hosted wallets","Cold storage","AI-enabled audit execution","Data analytics","Automation","Generative AI","Continuous monitoring","Anomaly detection","Automated evidence retrieval","AI-assisted workpaper drafting","Cloud platforms","IAM policies","Security configurations","Logging/monitoring","CI/CD pipelines","Python","SQL","AI tools","NIST CSF","COBIT","SOC 2","ITIL","CISA","CISSP","CIA","CISM","CPA","CFE"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:31.708Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"IT security, Cloud infrastructure, Security operations, Identity and access management, Data protection, IT asset management, Vendor/third-party risk, Blockchain infrastructure, Centralized and self-hosted wallets, Cold storage, AI-enabled audit execution, Data analytics, Automation, Generative AI, Continuous monitoring, Anomaly detection, Automated evidence retrieval, AI-assisted workpaper drafting, Cloud platforms, IAM policies, Security configurations, Logging/monitoring, CI/CD pipelines, Python, SQL, AI tools, NIST CSF, COBIT, SOC 2, ITIL, CISA, CISSP, CIA, CISM, CPA, CFE","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166345,"maxValue":195700,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bd43aede-675"},"title":"Staff Android Automation Engineer","description":"<p>We&#39;re looking for a Staff Android Automation Engineer to join our Quality Engineering team. As a leader in leveraging AI to redefine and accelerate Quality Engineering, you will drive the strategy towards comprehensive automation coverage for features and releases.</p>\n<p>You will work directly with the product engineering team to develop and maintain our test tools, write and test product code, participate in design reviews to architect testable systems, and guide designs and code to enhance modularity and testability.</p>\n<p>You are eager to understand complex systems top to bottom and thrive working across technologies and codebases. In addition, you excel at working through ambiguity, concept validation, and implementing best-in-class solutions.</p>\n<p>A typical day will involve leveraging AI and tooling to lead the implementation of a test automation strategy, covering the entire testing pyramid (unit, service, integration, and end-to-end testing) to verify feature functionality for customer use cases.</p>\n<p>You will lead building, maintaining, and effective utilization of automated tests, collaborating closely with engineering teams to ensure robust test coverage for features and releases, and actively participating in the continuous improvement of testing processes.</p>\n<p>You will contribute to improving existing automation frameworks to support new functionalities and optimize quality and efficiency.</p>\n<p>Collaboration with CI/CD team to integrate automated testing into CI/CD pipelines, ensuring thorough test coverage at every stage of development.</p>\n<p>Demonstrate excellent troubleshooting abilities, isolate issues, and verify bug fixes.</p>\n<p>Be a player of our high-performance team to ensure code quality, commitment to craft and operational excellence.</p>\n<p>Drive collaboration with cross-functional teams, including product management, development, and other QE teams, in a fast-paced environment with short release cycles.</p>\n<p>Your expertise will be demonstrated through 9+ years of industry experience in software testing and automation, demonstrable knowledge in at least one programming language (e.g., Kotlin, Java) and strong scripting skills.</p>\n<p>Strong knowledge of test automation methodologies, tools, and frameworks, strong hands-on experience with automation frameworks (e.g., Espresso), experience integrating automated tests into CI/CD pipelines (e.g., Buildkite, Spinnaker, Jenkins) and version control systems (Git).</p>\n<p>Excellent communication skills towards facilitating interactions with cross-functional teams, expertise developing solutions to ambiguous problems, and integrations across multiple teams with significant impact.</p>\n<p>Bachelor’s degree in computer science/engineering or equivalent, fluency in English (reading, writing, and speaking) is essential.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bd43aede-675","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7380185","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kotlin","Java","Test automation methodologies","Automation frameworks (e.g., Espresso)","CI/CD pipelines (e.g., Buildkite, Spinnaker, Jenkins)","Version control systems (Git)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:10.751Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Brazil - Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kotlin, Java, Test automation methodologies, Automation frameworks (e.g., Espresso), CI/CD pipelines (e.g., Buildkite, Spinnaker, Jenkins), Version control systems (Git)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b05b9f90-7d3"},"title":"Data Center Engineer, Resource Efficiency – Compute Supply","description":"<p><strong>About the Role</strong></p>\n<p>As a Power &amp; Resource Efficiency Engineer, you&#39;ll sit at the intersection of IT and facilities , building the systems, models, and control loops that optimize how we allocate and consume power, cooling, and physical capacity across our TPU/GPU fleet.</p>\n<p>You&#39;ll own the technical strategy for turning raw data center capacity into reliable, efficient compute, working across power topology, workload scheduling, and real-time telemetry to push utilization as close to the physical envelope as possible while maintaining our availability commitments.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build models that forecast consumption across electrical and mechanical subsystems, informing capacity planning, energy procurement, oversubscription targets and risks, including statistical modeling of cluster utilization, workload profiles, and failure modes.</li>\n</ul>\n<ul>\n<li>Design IT/OT interfaces that bridge compute orchestration with facility controls, enabling real-time telemetry across accelerator hardware, power distribution, cooling, and schedulers.</li>\n</ul>\n<ul>\n<li>Build and operate load management systems that use power and cooling topology to enable load management and power/thermal-aware placement to maximize throughput while meeting SLOs.</li>\n</ul>\n<ul>\n<li>Partner with data center providers to drive design optimizations and hold them accountable to SLA-grade performance standards, providing technical diligence on partner architectures.</li>\n</ul>\n<p><strong>What We&#39;re Looking For</strong></p>\n<ul>\n<li>Deep knowledge of data center power distribution and cooling architectures, and how they interact with IT load profiles. Experience with reliability engineering, SLA development, and failure-mode analysis.</li>\n</ul>\n<ul>\n<li>Proficiency in statistical modeling and simulation for infrastructure capacity or power utilization.</li>\n</ul>\n<ul>\n<li>Familiarity with SCADA/BMS/EPMS, telemetry pipelines, and control systems. Experience building software that bridges IT and OT.</li>\n</ul>\n<ul>\n<li>Exposure to accelerator deployments and their power management interfaces strongly preferred.</li>\n</ul>\n<ul>\n<li>Demand response, grid interaction, or behind-the-meter generation experience is a plus.</li>\n</ul>\n<ul>\n<li>Ability to translate between infrastructure engineering, software teams, and external partners.</li>\n</ul>\n<p><strong>Required Qualifications</strong></p>\n<ul>\n<li>Bachelor&#39;s degree in Electrical Engineering, Mechanical Engineering, Power Systems, Controls Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>5+ years of experience in data center infrastructure or facility engineering.</li>\n</ul>\n<ul>\n<li>Demonstrated experience with data center power distribution and cooling system architectures.</li>\n</ul>\n<ul>\n<li>Experience building or operating software-based power management, load scheduling, or control systems.</li>\n</ul>\n<ul>\n<li>Proficiency in Python or similar languages for statistical modeling, simulation, or automation of data center infrastructure optimizations.</li>\n</ul>\n<ul>\n<li>Familiarity with SCADA, BMS, EPMS, or industrial control systems and associated protocols (Modbus, BACnet, SNMP).</li>\n</ul>\n<ul>\n<li>Track record of cross-functional collaboration across hardware, software, and facilities teams.</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Master&#39;s or PhD in Controls, Power Systems, or related discipline and 3+ years of experience in data center infrastructure or facility engineering.</li>\n</ul>\n<ul>\n<li>Experience with accelerator-class deployments and their power management interfaces.</li>\n</ul>\n<ul>\n<li>Background in control theory, dynamical systems, or cyber-physical systems design.</li>\n</ul>\n<ul>\n<li>Experience with energy storage, microgrid integration, demand response, or behind-the-meter generation.</li>\n</ul>\n<ul>\n<li>Familiarity with reliability engineering methods.</li>\n</ul>\n<ul>\n<li>Experience with SLA development, availability modeling, or service credit frameworks.</li>\n</ul>\n<ul>\n<li>Exposure to ML/optimization techniques applied to infrastructure or energy systems.</li>\n</ul>\n<p><strong>Salary</strong></p>\n<p>The annual compensation range for this role is $320,000-$405,000 USD.</p>\n<p><strong>Benefits</strong></p>\n<p>We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with our team.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b05b9f90-7d3","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5159642008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["data center power distribution","cooling architectures","IT load profiles","reliability engineering","SLA development","failure-mode analysis","statistical modeling","simulation","infrastructure capacity","power utilization","SCADA/BMS/EPMS","telemetry pipelines","control systems","accelerator deployments","power management interfaces","demand response","grid interaction","behind-the-meter generation","Python","automation","data center infrastructure optimizations","SCADA","BMS","EPMS","industrial control systems","Modbus","BACnet","SNMP"],"x-skills-preferred":["accelerator-class deployments","control theory","dynamical systems","cyber-physical systems design","energy storage","microgrid integration","reliability engineering methods","availability modeling","service credit frameworks","ML/optimization techniques"],"datePosted":"2026-04-18T15:58:06.281Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-Friendly, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data center power distribution, cooling architectures, IT load profiles, reliability engineering, SLA development, failure-mode analysis, statistical modeling, simulation, infrastructure capacity, power utilization, SCADA/BMS/EPMS, telemetry pipelines, control systems, accelerator deployments, power management interfaces, demand response, grid interaction, behind-the-meter generation, Python, automation, data center infrastructure optimizations, SCADA, BMS, EPMS, industrial control systems, Modbus, BACnet, SNMP, accelerator-class deployments, control theory, dynamical systems, cyber-physical systems design, energy storage, microgrid integration, reliability engineering methods, availability modeling, service credit frameworks, ML/optimization techniques","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d5f768d1-df6"},"title":"Full-Stack Engineer, AI Data Platform","description":"<p>Shape the Future of AI</p>\n<p>At Labelbox, we&#39;re building the critical infrastructure that powers breakthrough AI models at leading research labs and enterprises. Since 2018, we&#39;ve been pioneering data-centric approaches that are fundamental to AI development, and our work becomes even more essential as AI capabilities expand exponentially.</p>\n<p>We&#39;re the only company offering three integrated solutions for frontier AI development:</p>\n<ul>\n<li>Enterprise Platform &amp; Tools: Advanced annotation tools, workflow automation, and quality control systems that enable teams to produce high-quality training data at scale</li>\n</ul>\n<ul>\n<li>Frontier Data Labeling Service: Specialized data labeling through Alignerr, leveraging subject matter experts for next-generation AI models</li>\n</ul>\n<ul>\n<li>Expert Marketplace: Connecting AI teams with highly skilled annotators and domain experts for flexible scaling</li>\n</ul>\n<p>Why Join Us</p>\n<ul>\n<li>High-Impact Environment: We operate like an early-stage startup, focusing on impact over process. You&#39;ll take on expanded responsibilities quickly, with career growth directly tied to your contributions.</li>\n</ul>\n<ul>\n<li>Technical Excellence: Work at the cutting edge of AI development, collaborating with industry leaders and shaping the future of artificial intelligence.</li>\n</ul>\n<ul>\n<li>Innovation at Speed: We celebrate those who take ownership, move fast, and deliver impact. Our environment rewards high agency and rapid execution.</li>\n</ul>\n<ul>\n<li>Continuous Growth: Every role requires continuous learning and evolution. You&#39;ll be surrounded by curious minds solving complex problems at the frontier of AI.</li>\n</ul>\n<ul>\n<li>Clear Ownership: You&#39;ll know exactly what you&#39;re responsible for and have the autonomy to execute. We empower people to drive results through clear ownership and metrics.</li>\n</ul>\n<p>Role Overview</p>\n<p>We’re looking for a Full-Stack AI Engineer to join our team, where you’ll build the next generation of tools for developing, evaluating, and training state-of-the-art AI systems. You will own features end to end,from user-facing experiences and APIs to backend services, data models, and infrastructure.</p>\n<p>You’ll be at the heart of our applied AI efforts, with a particular focus on human-in-the-loop systems used to generate high-quality training data for Large Language Models (LLMs) and AI agents. This includes building a platform that enables us and our customers to create and evaluate data, as well as systems that leverage LLMs to assist with reviewing, scoring, and improving human submissions.</p>\n<p>Your Impact</p>\n<ul>\n<li>Own End-to-End Product Features</li>\n</ul>\n<p>Design, build, and ship complete workflows spanning frontend UI, APIs, backend services, databases, and production infrastructure.</p>\n<ul>\n<li>Enable Human-in-the-Loop AI Training</li>\n</ul>\n<p>Build systems that allow humans to efficiently create, review, and curate high-quality training and evaluation data used in AI model development.</p>\n<ul>\n<li>Support RLHF and Preference Data Workflows</li>\n</ul>\n<p>Design and implement tooling that supports RLHF-style pipelines, including task generation, human review, scoring, aggregation, and dataset versioning.</p>\n<ul>\n<li>Leverage LLMs in the Review Loop</li>\n</ul>\n<p>Build systems that use LLMs to assist human reviewers,such as automated checks, critiques, ranking suggestions, or quality signals,while maintaining human oversight.</p>\n<ul>\n<li>Advance AI Evaluation</li>\n</ul>\n<p>Design and implement evaluation frameworks and interactive tools for LLMs and AI agents across multiple data modalities (text, images, audio, video).</p>\n<ul>\n<li>Create Intuitive, Reviewer-Focused Interfaces</li>\n</ul>\n<p>Build thoughtful, efficient user interfaces (e.g., in React) optimized for high-throughput human review, quality control, and operational workflows.</p>\n<ul>\n<li>Architect Scalable Data &amp; Service Layers</li>\n</ul>\n<p>Design APIs, backend services, and data schemas that support large-scale data creation, review, and iteration with strong guarantees around correctness and traceability.</p>\n<ul>\n<li>Solve Ambiguous, Real-World Problems</li>\n</ul>\n<p>Translate loosely defined operational and research needs into practical, scalable, end-to-end systems.</p>\n<ul>\n<li>Ensure System Reliability</li>\n</ul>\n<p>Participate in on-call rotations to monitor, troubleshoot, and resolve issues across the full stack.</p>\n<ul>\n<li>Elevate the Team</li>\n</ul>\n<p>Improve engineering practices, development processes, and documentation. Share knowledge through technical writing and design discussions.</p>\n<p>What You Bring</p>\n<ul>\n<li>Bachelor’s degree in Computer Science, Data Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>2+ years of experience in a software or machine learning engineering role.</li>\n</ul>\n<ul>\n<li>A proactive, product-focused mindset and a high degree of ownership, with a passion for building solutions that empower users.</li>\n</ul>\n<ul>\n<li>Experience using frontend frameworks like React/Redux and backend systems and technologies like Python, Java, GraphQL; familiarity with NodeJS and NestJS is a plus.</li>\n</ul>\n<ul>\n<li>Knowledge of designing and managing scalable database systems, including relational databases (e.g., PostgreSQL, MySQL), NoSQL stores (e.g., MongoDB, Cassandra), and cloud-native solutions (e.g., Google Spanner, AWS DynamoDB).</li>\n</ul>\n<ul>\n<li>Familiarity with cloud infrastructure like GCP (GCS, PubSub) and containerization (Kubernetes) is a plus.</li>\n</ul>\n<ul>\n<li>Excellent communication and collaboration skills.</li>\n</ul>\n<ul>\n<li>High proficiency in leveraging AI tools for daily development (e.g., Cursor, GitHub Copilot).</li>\n</ul>\n<ul>\n<li>Comfort and enthusiasm for working in a fast-paced, agile environment where rapid problem-solving is key.</li>\n</ul>\n<p>Bonus Points</p>\n<ul>\n<li>Experience building tools for AI/ML applications, particularly for data annotation, monitoring, or agent evaluation.</li>\n</ul>\n<ul>\n<li>Familiarity with data infrastructure components such as data pipelines, streaming systems, and storage architectures (e.g., Cloud Buckets, Key-Value Stores).</li>\n</ul>\n<ul>\n<li>Previous experience with search engines (e.g., ElasticSearch).</li>\n</ul>\n<ul>\n<li>Experience in optimizing databases for performance (e.g., schema design, indexing, query tuning) and integrating them with broader data workflows.</li>\n</ul>\n<p>Engineering at Labelbox</p>\n<p>At Labelbox Engineering, we&#39;re building a comprehensive platform that powers the future of AI development. Our team combines deep technical expertise with a passion for innovation, working at the intersection of AI infrastructure, data systems, and user experience. We believe in pushing technical boundaries while maintaining high standards of code quality and system reliability. Our engineering culture emphasizes autonomous decision-making, rapid iteration, and collaborative problem-solving. We&#39;ve cultivated an environment where engineers can take ownership of significant challenges, experiment with cutting-edge technologies, and see their solutions directly impact how leading AI labs and enterprises build the next generation of AI systems.</p>\n<p>Our Technology Stack</p>\n<p>Our engineering team works with a modern tech stack designed for scalability, performance, and developer efficiency:</p>\n<ul>\n<li>Frontend: React.js with Redux, TypeScript</li>\n</ul>\n<ul>\n<li>Backend: Node.js, TypeScript, Python, some Java &amp; Kotlin</li>\n</ul>\n<ul>\n<li>APIs: GraphQL</li>\n</ul>\n<ul>\n<li>Cloud &amp; Infrastructure: Google Cloud Platform (GCP), Kubernetes</li>\n</ul>\n<ul>\n<li>Databases: MySQL, Spanner, PostgreSQL</li>\n</ul>\n<ul>\n<li>Queueing / Streaming: Kafka, PubSub</li>\n</ul>\n<p>Labelbox strives to ensure pay parity across the organization and discuss compensation transparently. The expected annual base salary range for United States-based candidates is below. This range is not inclusive of any potential equity packages or additional benefits. Exact compensation varies based on a variety of factors, including skills and competencies, experience, and geographical location.</p>\n<p>Annual base salary range $130,000-$200,000 USD</p>\n<p>Life at Labelbox</p>\n<ul>\n<li>Location: Join our dedicated tech hubs in San Francisco or Wrocław, Poland</li>\n</ul>\n<ul>\n<li>Work Style: Hybrid model with 2 days per week in office, combining collaboration and flexibility</li>\n</ul>\n<ul>\n<li>Environment: Fast-paced and high-intensity, perfect for ambitious individuals who thrive on ownership and quick decision-making</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d5f768d1-df6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Labelbox","sameAs":"https://www.labelbox.com/","logo":"https://logos.yubhub.co/labelbox.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/labelbox/jobs/5019254007","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$130,000-$200,000 USD","x-skills-required":["React","Redux","Node.js","TypeScript","Python","Java","GraphQL","MySQL","PostgreSQL","Spanner","Kafka","PubSub","GCP","Kubernetes","Cloud computing","Containerization","Database management","Cloud infrastructure","API design","Backend services","Data models","Infrastructure"],"x-skills-preferred":["AI tools","Cursor","GitHub Copilot","Data annotation","Monitoring","Agent evaluation","Data infrastructure","Data pipelines","Streaming systems","Storage architectures","Search engines","ElasticSearch","Database optimization","Schema design","Indexing","Query tuning"],"datePosted":"2026-04-18T15:57:55.464Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco Bay Area"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, Redux, Node.js, TypeScript, Python, Java, GraphQL, MySQL, PostgreSQL, Spanner, Kafka, PubSub, GCP, Kubernetes, Cloud computing, Containerization, Database management, Cloud infrastructure, API design, Backend services, Data models, Infrastructure, AI tools, Cursor, GitHub Copilot, Data annotation, Monitoring, Agent evaluation, Data infrastructure, Data pipelines, Streaming systems, Storage architectures, Search engines, ElasticSearch, Database optimization, Schema design, Indexing, Query tuning","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":130000,"maxValue":200000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3238a958-3d9"},"title":"AI Product Manager","description":"<p>We&#39;re looking for an AI Product Manager to own one of the Agent &amp; Reinforcement Learning Environments data verticals, with a focus on Computer Using Agent (CUA) data.</p>\n<p>In this role, you&#39;ll oversee the product roadmap for your data vertical, owning &#39;data as a product&#39;, pipelines for data generation and quality, and researcher-facing tools that help labs train and evaluate intelligent agents in complex environments.</p>\n<p>You&#39;ll work directly with Scale&#39;s most important customers and their leading researchers, representing Scale as the technical expert for your products and influencing both internal and external roadmaps.</p>\n<p>The ideal candidate brings together a strong entrepreneurial &amp; go-to-market mindset, technical depth, and a sense for AI research, enabling them to get in front of technical stakeholders to drive mission-critical outcomes.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Own the roadmap for the Agent &amp; RL Environment Data vertical, setting product direction and driving execution across engineering, operations, and go-to-market teams.</li>\n</ul>\n<ul>\n<li>Build technical partnerships with research teams at leading AI labs, identifying insights that shape new product lines and competitive strategies for your vertical.</li>\n</ul>\n<ul>\n<li>Design, experiment with, and deliver high-quality data pipelines, tooling, and evaluation frameworks that advance RL and agentic model capabilities.</li>\n</ul>\n<ul>\n<li>Scope out and scale the creation of RL environments that simulate real-world use cases.</li>\n</ul>\n<ul>\n<li>Collaborate cross-functionally, influencing business priorities and diving in the weeds of research, operations, and customer interactions.</li>\n</ul>\n<p>Ideally, You&#39;d Have:</p>\n<ul>\n<li>Entrepreneurial mindset: A builder excited by ambiguity and motivated to create new products from the ground up.</li>\n</ul>\n<ul>\n<li>6+ years of experience in product management or a customer-facing role.</li>\n</ul>\n<ul>\n<li>Technical fluency: Software engineering background (a degree in computer science or equivalent experience).</li>\n</ul>\n<ul>\n<li>Understanding of reinforcement learnings, simulation environments, or data pipelines for model training and evaluation</li>\n</ul>\n<ul>\n<li>Strong customer intuition and the ability to translate technical requirements into impactful product decisions.</li>\n</ul>\n<ul>\n<li>Bias for action and comfort wearing multiple hats and operating in fast-moving environments.</li>\n</ul>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3238a958-3d9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4609736005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["reinforcement learnings","simulation environments","data pipelines","model training","evaluation frameworks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:37.306Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"reinforcement learnings, simulation environments, data pipelines, model training, evaluation frameworks","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8871a994-591"},"title":"Machine Learning Engineer, Core Engineering","description":"<p>We&#39;re seeking a talented Machine Learning Engineer to join our Core Engineering team. As a Machine Learning Engineer at Pinterest, you will build cutting-edge technology using the latest advances in deep learning and machine learning to personalize Pinterest. You will partner closely with teams across Pinterest to experiment and improve ML models for various product surfaces, while gaining knowledge of how ML works in different areas.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Build cutting-edge technology using the latest advances in deep learning and machine learning to personalize Pinterest</li>\n<li>Partner closely with teams across Pinterest to experiment and improve ML models for various product surfaces (Homefeed, Ads, Growth, Shopping, and Search), while gaining knowledge of how ML works in different areas</li>\n<li>Use data-driven methods and leverage the unique properties of our data to improve candidate retrieval</li>\n<li>Work in a high-impact environment with quick experimentation and product launches</li>\n<li>Keep up with industry trends in recommendation systems</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>2+ years of industry experience applying machine learning methods (e.g., user modeling, personalization, recommender systems, search, ranking, natural language processing, reinforcement learning, and graph representation learning)</li>\n<li>End-to-end hands-on experience with building data processing pipelines, large-scale machine learning systems, and big data technologies (e.g., Hadoop/Spark)</li>\n<li>Degree in computer science, machine learning, statistics, or related field</li>\n</ul>\n<p>Nice to Have:</p>\n<ul>\n<li>M.S. or PhD in Machine Learning or related areas</li>\n<li>Publications at top ML conferences</li>\n<li>Experience using Cursor, Copilot, Codex, or similar AI coding assistants for development, debugging, testing, and refactoring</li>\n<li>Familiarity with LLM-powered productivity tools for documentation search, experiment analysis, SQL/data exploration, and engineering workflow acceleration</li>\n<li>Expertise in scalable real-time systems that process stream data</li>\n<li>Passion for applied ML and the Pinterest product</li>\n</ul>\n<p>Relocation Statement:</p>\n<p>This position is not eligible for relocation assistance. Visit our PinFlex page to learn more about our working model.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8871a994-591","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/6121450","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$138,905-$285,982 USD","x-skills-required":["machine learning","deep learning","data processing pipelines","large-scale machine learning systems","big data technologies","Hadoop","Spark","natural language processing","reinforcement learning","graph representation learning"],"x-skills-preferred":["Cursor","Copilot","Codex","LLM-powered productivity tools","scalable real-time systems","stream data"],"datePosted":"2026-04-18T15:57:30.186Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, US; Palo Alto, CA, US; Seattle, WA, US; Remote, US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"machine learning, deep learning, data processing pipelines, large-scale machine learning systems, big data technologies, Hadoop, Spark, natural language processing, reinforcement learning, graph representation learning, Cursor, Copilot, Codex, LLM-powered productivity tools, scalable real-time systems, stream data","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":138905,"maxValue":285982,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_95c49f85-a98"},"title":"Staff+ Software Engineer, Observability","description":"<p><strong>About the Role</strong></p>\n<p>Anthropic is seeking talented and experienced Software Engineers to join our Observability team within the Infrastructure organization. The Observability team owns the monitoring and telemetry infrastructure that every engineer and researcher at Anthropic depends on,from metrics and logging pipelines to distributed tracing, error analytics, alerting, and the dashboards and query interfaces that make it all actionable.</p>\n<p>As Anthropic scales its infrastructure across massive GPU, TPU, and Trainium clusters, the volume and complexity of operational data is growing by orders of magnitude. We’re building next-generation observability systems,high-throughput ingest pipelines, cost-efficient columnar storage, unified query layers across signals, and agentic diagnostic tools,to ensure that engineers can detect, diagnose, and resolve issues in minutes rather than hours, even as the systems they operate become exponentially more complex.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and build scalable telemetry ingest and storage pipelines for metrics, logs, traces, and error data across Anthropic’s multi-cluster infrastructure</li>\n</ul>\n<ul>\n<li>Own and evolve core observability platforms, driving migrations and architectural improvements that improve reliability, reduce cost, and scale with organisational growth</li>\n</ul>\n<ul>\n<li>Build instrumentation libraries, SDKs, and integrations that make it easy for engineering teams to emit high-quality telemetry from their services</li>\n</ul>\n<ul>\n<li>Drive alerting and SLO infrastructure that enables teams to define, monitor, and respond to reliability targets with minimal noise</li>\n</ul>\n<ul>\n<li>Reduce mean time to detection and resolution by building cross-signal correlation, unified query interfaces, and AI-assisted diagnostic tooling</li>\n</ul>\n<ul>\n<li>Partner with Research, Inference, Product, and Infrastructure teams to ensure observability solutions meet the unique needs of each organisation</li>\n</ul>\n<p><strong>You May Be a Good Fit If You</strong></p>\n<ul>\n<li>Have 10+ years of relevant industry experience building and operating large-scale observability or monitoring infrastructure</li>\n</ul>\n<ul>\n<li>Have deep experience with at least one observability signal area (metrics, logging, tracing, or error analytics) and familiarity with the others</li>\n</ul>\n<ul>\n<li>Understand high-throughput data pipelines, columnar storage engines, and the tradeoffs involved in ingesting and querying telemetry data at scale</li>\n</ul>\n<ul>\n<li>Have experience operating or building on top of observability platforms such as Prometheus, Grafana, ClickHouse, OpenTelemetry, or similar systems</li>\n</ul>\n<ul>\n<li>Have strong proficiency in at least one of Python, Rust, or Go</li>\n</ul>\n<ul>\n<li>Have excellent communication skills and enjoy partnering with internal teams to improve their operational visibility and incident response capabilities</li>\n</ul>\n<ul>\n<li>Are excited about building foundational infrastructure and are comfortable working independently on ambiguous, high-impact technical challenges</li>\n</ul>\n<p><strong>Strong Candidates May Also Have</strong></p>\n<ul>\n<li>Experience operating metrics systems at very high cardinality (hundreds of millions of active time series or more)</li>\n</ul>\n<ul>\n<li>Experience with log storage migrations or operating columnar databases (ClickHouse, BigQuery, or similar) for analytics workloads</li>\n</ul>\n<ul>\n<li>Experience with OpenTelemetry instrumentation, collector pipelines, and tail-based sampling strategies</li>\n</ul>\n<ul>\n<li>Experience building or operating alerting platforms, on-call tooling, or SLO frameworks at scale</li>\n</ul>\n<ul>\n<li>Experience with Kubernetes-native monitoring, eBPF-based observability, or continuous profiling</li>\n</ul>\n<ul>\n<li>Interest in applying AI/LLMs to operational workflows such as automated root cause analysis, anomaly detection, or intelligent alerting</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>\n</ul>\n<ul>\n<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>\n</ul>\n<ul>\n<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>\n</ul>\n<ul>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n</ul>\n<ul>\n<li>Visa sponsorship: We do sponsor visas! However, we aren’t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We’re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_95c49f85-a98","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5102440008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"£325,000-£390,000 GBP","x-skills-required":["observability","telemetry","metrics","logging","tracing","error analytics","alerting","SLO infrastructure","cross-signal correlation","unified query interfaces","AI-assisted diagnostic tooling","Python","Rust","Go","Prometheus","Grafana","ClickHouse","OpenTelemetry"],"x-skills-preferred":["high-throughput data pipelines","columnar storage engines","Kubernetes-native monitoring","eBPF-based observability","continuous profiling","AI/LLMs","automated root cause analysis","anomaly detection","intelligent alerting"],"datePosted":"2026-04-18T15:57:27.177Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"observability, telemetry, metrics, logging, tracing, error analytics, alerting, SLO infrastructure, cross-signal correlation, unified query interfaces, AI-assisted diagnostic tooling, Python, Rust, Go, Prometheus, Grafana, ClickHouse, OpenTelemetry, high-throughput data pipelines, columnar storage engines, Kubernetes-native monitoring, eBPF-based observability, continuous profiling, AI/LLMs, automated root cause analysis, anomaly detection, intelligent alerting","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":325000,"maxValue":390000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_af586166-0a0"},"title":"Technical Solutions Specialist, Data Operations","description":"<p>In Data Operations on the Strategic Data Partnerships team at Anthropic, you will support a cross-functional team in implementing partnership strategies to improve Anthropic’s products. You’ll ensure data meets our standards and reaches the right teams, build systems to track compliance and data usage across the portfolio, and coordinate across Research, Product, Legal, and external partners to remove barriers and accelerate impact.</p>\n<p>This role requires operational excellence combined with technical hands-on execution, and is a great fit for someone who wants to apply those skills in a high-impact, fast-growth context.</p>\n<p>Responsibilities:</p>\n<p>Data Opportunity Assessment and Processing</p>\n<ul>\n<li>Analyze and review incoming or prospective data to verify it is useful and strategic for Anthropic</li>\n<li>Own and maintain Python-based ETL pipelines that process large partner datasets, applying filtering criteria and deduplicating against existing data</li>\n<li>Write and optimize SQL queries against large relational databases to support filtering and analysis workflows</li>\n<li>Refine processing logic as requirements evolve across new data types and formats</li>\n</ul>\n<p>Data Delivery Infrastructure, Tooling, and Support</p>\n<ul>\n<li>Own end-to-end data delivery workflows, ensuring data moves seamlessly from partners to internal teams to accelerate time-to-impact</li>\n<li>Manage AWS and GCP resources for receiving and organizing partner data deliveries</li>\n<li>Troubleshoot delivery issues and coordinate with partners on formatting and transfer protocols and resolve technical escalations from partners and internal teams</li>\n<li>Build and maintain internal systems, scripts, and automation that support the team’s workflows</li>\n<li>Support occasional research evaluation tasks as needed</li>\n</ul>\n<p>Data Operations and Governance</p>\n<ul>\n<li>Develop and maintain Anthropic&#39;s preferred standards for receiving, consuming and cataloging data, ensuring alignment with Product and Engineering&#39;s evolving needs</li>\n<li>Contribute to systems for monitoring data usage and compliance with partner agreements</li>\n<li>Partner with teammates and cross-functional stakeholders to build out governance practices as the team scales</li>\n</ul>\n<p>You May Be a Good Fit If You</p>\n<ul>\n<li>Bachelor’s degree in Engineering, Computer Science, a related field, or equivalent practical experience</li>\n<li>5-7+ years of experience with data pipelines or data engineering workflows</li>\n<li>Background in solutions engineering, partner engineering or related role at a large tech company</li>\n<li>5+ years of experience in technical troubleshooting or writing code in one or more programming languages</li>\n<li>Proficiency in Python and SQL, including writing, debugging, and optimizing scripts and queries against large datasets</li>\n<li>Hands-on experience with cloud infrastructure (AWS, GCP, or Azure), including managing storage, configuring access, and working from the CLI</li>\n<li>Excellent problem-solving skills with a track record of debugging technical issues, whether at the code level or within a broader system</li>\n<li>Some experience interacting with external third parties delivering data</li>\n</ul>\n<p>Strong Candidates Will Have</p>\n<ul>\n<li>Experience working alongside technical teams (research, engineering, or product) to solve ambiguous problems</li>\n<li>Ability to translate technical concepts into clear, actionable guidance for non-technical stakeholders or external partners</li>\n<li>Experience owning or maintaining a production service or system with uptime expectations</li>\n<li>Familiarity with data governance, compliance, or rights management</li>\n<li>Ability to manage multiple, time-sensitive projects simultaneously and the drive to take a project from an initial idea to full completion</li>\n<li>Experience leveraging AI to automate workflows</li>\n</ul>\n<p>Candidates Need Not Have</p>\n<ul>\n<li>Deep expertise in AI or machine learning</li>\n<li>A pure software engineering background</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_af586166-0a0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5056499008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$205,000-$240,000 USD","x-skills-required":["Python","SQL","Cloud infrastructure (AWS, GCP, or Azure)","Data pipelines","Data engineering workflows","Solutions engineering","Partner engineering"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:08.396Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Cloud infrastructure (AWS, GCP, or Azure), Data pipelines, Data engineering workflows, Solutions engineering, Partner engineering","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":205000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f4cd384f-6ed"},"title":"Senior Software Engineer, Release Engineering","description":"<p>We are seeking a Senior Software Engineer to join our Release Engineering team, focused on building and improving the systems that enable automated, reliable, and scalable software delivery across Temporal&#39;s platform.</p>\n<p>In this role, you will participate in the full software lifecycle , from design and implementation to deployment and long-term operation , and will collaborate with engineering teams to evolve release automation, improve tooling, and reduce manual steps in how we build and ship Temporal.</p>\n<p>Key responsibilities include designing, building, and maintaining tools and systems that support release automation and deployment workflows, writing clean, reliable, and concurrent code that supports distributed systems, collaborating with cross-functional teams to understand and improve release quality and developer productivity, documenting technical designs, deployment practices, and operational procedures, and participating in small-team design reviews and contributing practical engineering solutions.</p>\n<p>As a Senior Software Engineer, you will have the opportunity to explore new ways to use Temporal to power the release and deployment lifecycle, deepen your understanding of Temporal&#39;s architecture and service interactions, and experiment with new automation patterns, testing strategies, and workflow designs that increase release confidence.</p>\n<p>To be successful in this role, you will need strong coding ability, especially in languages used at Temporal (e.g., Go, Java, or similar), a solid understanding of concurrency, distributed systems, and multi-threaded programming, experience contributing to backend systems, tooling, infrastructure, or developer workflows, a track record of solving moderately complex problems with reliable, maintainable solutions, and the ability to collaborate effectively in a remote, fast-paced environment.</p>\n<p>Additionally, you will have familiarity with release automation concepts, CI/CD pipelines, build tools, or deployment orchestration, experience with cloud environments (AWS, GCP) and container tooling, and exposure to distributed systems orchestration, observability tooling, or platform engineering.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f4cd384f-6ed","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Temporal","sameAs":"https://temporal.io/","logo":"https://logos.yubhub.co/temporal.io.png"},"x-apply-url":"https://job-boards.greenhouse.io/temporaltechnologies/jobs/5090613007","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$176,000 - $237,600","x-skills-required":["Go","Java","Concurrency","Distributed Systems","Multi-threaded Programming","Backend Systems","Tooling","Infrastructure","Developer Workflows","Release Automation","CI/CD Pipelines","Build Tools","Deployment Orchestration","Cloud Environments","Container Tooling","Distributed Systems Orchestration","Observability Tooling","Platform Engineering"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:07.513Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States - Remote Opportunity"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Java, Concurrency, Distributed Systems, Multi-threaded Programming, Backend Systems, Tooling, Infrastructure, Developer Workflows, Release Automation, CI/CD Pipelines, Build Tools, Deployment Orchestration, Cloud Environments, Container Tooling, Distributed Systems Orchestration, Observability Tooling, Platform Engineering","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":176000,"maxValue":237600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_16ced9e5-b93"},"title":"3D Tutor","description":"<p>As a 3D Specialist, you will contribute to xAI&#39;s mission by creating high-quality 3D content that supports the development of Grok&#39;s visual understanding capabilities.</p>\n<p>Key to this role is expertise in 3D modeling, lighting, and animation, a track record of producing polished 3D work, and a refined aesthetic judgment in visual composition and technical execution.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Use industry-standard 3D software to create assets, characters, environments, and animations according to project specifications</li>\n<li>Deliver high-quality work that demonstrates strong technical fundamentals and artistic sensibility</li>\n<li>Collaborate with technical staff to understand project requirements and iterate on deliverables efficiently</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>Portfolio displaying excellence in 3D work, such as game assets, animations, architectural visualizations, or VFX shots</li>\n<li>Strong skills in modeling, texturing, lighting, rigging, and animation</li>\n<li>Experience setting up environments and animating cameras</li>\n<li>Ability to produce clean, render-ready outputs to specification</li>\n<li>Strong communication and analytical skills</li>\n<li>Strong written and verbal English skills</li>\n</ul>\n<p>Preferred Skills and Experience:</p>\n<ul>\n<li>Familiarity with PBR workflows, real-time rendering pipelines, and procedural generation techniques</li>\n<li>Python coding skills</li>\n</ul>\n<p>Location and Other Expectations:</p>\n<ul>\n<li>Tutor roles may be offered as full-time, part-time, or contractor positions, depending on role needs and candidate fit</li>\n<li>For contractor positions, hours will vary widely based on project scope and contractor availability, with no fixed commitments required</li>\n<li>Tutor roles may be performed remotely from any location worldwide, subject to legal eligibility, time-zone compatibility, and role specific needs</li>\n<li>For US based candidates, please note we are unable to hire in the states of Wyoming and Illinois at this time</li>\n<li>We are unable to provide visa sponsorship</li>\n<li>For those who will be working from a personal device, your computer must be a Chromebook, Mac with MacOS 11.0 or later, or Windows 10 or later</li>\n</ul>\n<p>Compensation and Benefits:</p>\n<ul>\n<li>US based candidates: $60/hour - $100/hour depending on factors including relevant experience, skills, education, geographic location, and qualifications</li>\n<li>International candidates: Information will be provided to you during the recruitment process</li>\n<li>Benefits vary based on employment type, location and jurisdiction. Benefits for eligible U.S. based positions include health insurance, 401(k) plan, and paid sick leave. Specific details and role specific information will be provided to you during the interview process</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_16ced9e5-b93","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.xai.com/","logo":"https://logos.yubhub.co/xai.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/5045788007","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time|part-time|contract","x-salary-range":"$60/hour - $100/hour","x-skills-required":["3D modeling","lighting","animation","Python coding","PBR workflows","real-time rendering pipelines","procedural generation techniques"],"x-skills-preferred":["familiarity with PBR workflows"],"datePosted":"2026-04-18T15:57:05.388Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"3D modeling, lighting, animation, Python coding, PBR workflows, real-time rendering pipelines, procedural generation techniques, familiarity with PBR workflows"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f24aa64a-8e9"},"title":"DevOps Engineer, GPS","description":"<p>As a DevOps Engineer, you will design and develop core platforms and software systems, while supporting orchestration, data abstraction, data pipelines, identity &amp; access management, security tools, and underlying cloud infrastructure.</p>\n<p>You will:</p>\n<ul>\n<li>Backend Development and System Ownership: Design and implement secure, scalable backend systems for customers using modern, cloud-native AI infrastructure. Own services or systems, define long-term health goals, and improve the health of surrounding components.</li>\n</ul>\n<ul>\n<li>Collaboration and Standards: Collaborate with cross-functional teams to define and execute backend and infrastructure solutions tailored for secure environments. Enhance engineering standards, tooling, and processes to maintain high-quality outputs.</li>\n</ul>\n<ul>\n<li>Infrastructure Automation and Management: Write, maintain, and enhance Infrastructure as Code templates (e.g., Terraform, CloudFormation) for automated provisioning and management. Manage networking architecture, including secure VPCs, VPNs, load balancers, and firewalls, in cloud environments.</li>\n</ul>\n<ul>\n<li>Deployment and Scalability: Design and optimize CI/CD pipelines for efficient testing, building, and deployment processes. Scale and optimize containerized applications using orchestration platforms like Kubernetes to ensure high availability and reliability.</li>\n</ul>\n<ul>\n<li>Disaster Recovery and Hybrid Strategies: Develop and test disaster recovery plans with robust backups and failover mechanisms. Design and implement hybrid and multi-cloud strategies to support workloads across on-premises and multiple cloud providers.</li>\n</ul>\n<p>Our ideal candidate has a strong engineering background, with a Bachelor’s degree in Computer Science, Mathematics, or a related quantitative field (or equivalent practical experience), and 5+ years of post-graduation engineering experience, with a focus on back-end systems and proficiency in at least one of Python, Typescript, Javascript, or C++.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f24aa64a-8e9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://www.scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4613839005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Backend Development","System Ownership","Infrastructure Automation","Deployment and Scalability","Disaster Recovery and Hybrid Strategies","Cloud-Native AI Infrastructure","Terraform","CloudFormation","Kubernetes","Python","Typescript","Javascript","C++"],"x-skills-preferred":["Collaboration and Standards","Networking Architecture","CI/CD Pipelines","Containerized Applications","Orchestration Platforms","Data Abstraction","Data Pipelines","Identity & Access Management","Security Tools"],"datePosted":"2026-04-18T15:56:30.346Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Doha, Qatar"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Backend Development, System Ownership, Infrastructure Automation, Deployment and Scalability, Disaster Recovery and Hybrid Strategies, Cloud-Native AI Infrastructure, Terraform, CloudFormation, Kubernetes, Python, Typescript, Javascript, C++, Collaboration and Standards, Networking Architecture, CI/CD Pipelines, Containerized Applications, Orchestration Platforms, Data Abstraction, Data Pipelines, Identity & Access Management, Security Tools"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_69e8923b-c16"},"title":"Senior Data Scientist","description":"<p>We&#39;re seeking a Senior Data Scientist to join our Research, Analytics &amp; Data Science (RAD) team. Our team uses data and insights to drive evidence-based decision-making, generating actionable insights about our customers, products, and business.</p>\n<p>As a Senior Data Scientist, you&#39;ll partner with product teams to help them identify important questions and answer those questions with data. You&#39;ll work closely with product managers, designers, and engineers to develop key product success metrics, set targets, measure results, and outcomes, and size opportunities.</p>\n<p>You&#39;ll design, build, and update end-to-end data pipelines, working closely with stakeholders to drive the collection of new data and the refinement of existing data sources and tables. You&#39;ll also partner closely with product researchers to build a holistic understanding of our customers, products, and business.</p>\n<p>Increasingly, you&#39;ll use AI-assisted tools to accelerate analysis, coding, and insight generation. You&#39;ll identify opportunities to automate your own workflows and reduce time spent on repetitive tasks. You&#39;ll build scalable data products that enable stakeholders to self-serve insights and raise the bar for how AI is used within RAD.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Partnering with product teams to help them identify important questions and answer those questions with data</li>\n<li>Working closely with product managers, designers, and engineers to develop key product success metrics, set targets, measure results, and outcomes, and size opportunities</li>\n<li>Designing, building, and updating end-to-end data pipelines</li>\n<li>Partnering closely with product researchers to build a holistic understanding of our customers, products, and business</li>\n<li>Using AI-assisted tools to accelerate analysis, coding, and insight generation</li>\n<li>Identifying opportunities to automate your own workflows and reduce time spent on repetitive tasks</li>\n<li>Building scalable data products that enable stakeholders to self-serve insights</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>5+ years of experience working with data to solve problems and drive evidence-based decisions</li>\n<li>Strong SQL skills and solid grounding in statistics</li>\n<li>Experience working closely with product teams</li>\n<li>Proven track record of delivering actionable insights that drive measurable impact with minimal supervision</li>\n<li>Strong product intuition, business acumen, and ability to connect analysis to strategy</li>\n<li>Excellent communication skills (technical and non-technical), with a focus on driving decisions and outcomes</li>\n<li>Strong ownership, curiosity, and growth mindset</li>\n<li>Experience with a scientific computing language (e.g., Python)</li>\n</ul>\n<p>Preferred skills include:</p>\n<ul>\n<li>Experience with data modeling and ETL pipelines (esp. dbt)</li>\n<li>Experience building internal tools, data products, or self-serve analytics capabilities</li>\n<li>Experience leveraging AI across the data workflow - from ideation and coding to analysis and communication</li>\n</ul>\n<p>Benefits include:</p>\n<ul>\n<li>Competitive salary and equity in a fast-growing start-up</li>\n<li>Unlimited access to Claude Code and best-in-class AI tools; experimentation &amp; building is encouraged &amp; celebrated</li>\n<li>We serve lunch every weekday, plus a variety of snack foods and a fully stocked kitchen</li>\n<li>Regular compensation reviews - we reward great work</li>\n<li>Peace of mind with life assurance, as well as comprehensive health and dental insurance for you and your dependents</li>\n<li>Open vacation policy and flexible holidays so you can take time off when you need it</li>\n<li>Paid maternity leave, as well as 6 weeks paternity leave for fathers, to let you spend valuable time with your loved ones</li>\n<li>MacBooks are our standard, but we’re happy to get you whatever equipment helps you get your job done</li>\n</ul>\n<p>Experience Level: Senior Employment Type: Full-time Workplace Type: Hybrid Category: Engineering Industry: Technology Salary Range: Competitive salary and equity in a fast-growing start-up Required Skills: SQL, statistics, experience working with product teams, strong product intuition, business acumen, excellent communication skills, strong ownership, curiosity, and growth mindset, experience with a scientific computing language (e.g., Python) Preferred Skills: data modeling and ETL pipelines (esp. dbt), building internal tools, data products, or self-serve analytics capabilities, leveraging AI across the data workflow</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_69e8923b-c16","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Intercom","sameAs":"https://www.intercom.com/","logo":"https://logos.yubhub.co/intercom.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/intercom/jobs/7749323","x-work-arrangement":null,"x-experience-level":null,"x-job-type":null,"x-salary-range":null,"x-skills-required":["SQL","statistics","experience working with product teams","strong product intuition","business acumen","excellent communication skills","strong ownership","curiosity","growth mindset","experience with a scientific computing language (e.g., Python)"],"x-skills-preferred":["data modeling and ETL pipelines (esp. dbt)","building internal tools","data products","or self-serve analytics capabilities","leveraging AI across the data workflow"],"datePosted":"2026-04-18T15:56:25.055Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, England"}},"skills":"SQL, statistics, experience working with product teams, strong product intuition, business acumen, excellent communication skills, strong ownership, curiosity, growth mindset, experience with a scientific computing language (e.g., Python), data modeling and ETL pipelines (esp. dbt), building internal tools, data products, or self-serve analytics capabilities, leveraging AI across the data workflow"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_192b8eb7-029"},"title":"Staff iOS Engineer - B2C Native Apps","description":"<p>We are looking for a Staff iOS Engineer to join our B2C Native Apps team. As a member of this team, you will be responsible for designing, developing, and maintaining high-quality iOS applications.</p>\n<p>Our team is fast-paced and agile, comprising engineers, a product manager, and designer. We work closely together to deliver innovative solutions that meet the needs of our customers.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and develop high-quality iOS applications using Swift and Objective-C</li>\n<li>Collaborate with the product manager and designer to define and prioritize features</li>\n<li>Work with the engineering team to ensure seamless integration with other components</li>\n<li>Participate in code reviews and contribute to the improvement of our codebase</li>\n<li>Mentor junior engineers and help them grow in their careers</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>8+ years of professional iOS development experience</li>\n<li>Excellent communication and collaboration skills</li>\n<li>Experience building public or internal mobile APIs/SDKs and working with Swift and Objective-C</li>\n<li>Experience with UIKit, SwiftUI, programmatic Auto Layout, and iOS design patterns (MVVM, reactive programming)</li>\n<li>Experience with Unit/UI/integration/performance testing on iOS (Quick, Nimble, XCTest, XCUITest, etc.)</li>\n<li>Experience with Realm database or similar mobile NoSQL solutions</li>\n<li>End-to-end ownership of mobile applications or SDKs</li>\n<li>Experience with mobile CI/CD pipelines (GitHub Actions)</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>1+ years of experience in identity and access management (IAM) domain, particularly with Auth0 Guardian SDK or similar MFA/authentication solutions</li>\n<li>Experience with iOS security best practices, including cryptography (RSA, CommonCrypto), biometric authentication (Face ID/Touch ID), iOS Keychain, Authentication Service framework, and secure data storage</li>\n<li>Experience with reactive programming frameworks (ReactiveSwift, Combine) and migrating legacy architectures to MVVM patterns</li>\n<li>Experience with infrastructure-as-code tools (e.g., Fastlane, Swift Package Manager, Snyk, or Terraform)</li>\n</ul>\n<p>If you are a motivated and experienced iOS engineer looking to join a dynamic team, we encourage you to apply.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_192b8eb7-029","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7598837","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["iOS development","Swift","Objective-C","UIKit","SwiftUI","programmatic Auto Layout","iOS design patterns","MVVM","reactive programming","Unit/UI/integration/performance testing","Realm database","mobile NoSQL solutions","end-to-end ownership","mobile CI/CD pipelines"],"x-skills-preferred":["identity and access management","Auth0 Guardian SDK","MFA/authentication solutions","iOS security best practices","cryptography","biometric authentication","iOS Keychain","Authentication Service framework","secure data storage","reactive programming frameworks","infrastructure-as-code tools"],"datePosted":"2026-04-18T15:56:22.845Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"iOS development, Swift, Objective-C, UIKit, SwiftUI, programmatic Auto Layout, iOS design patterns, MVVM, reactive programming, Unit/UI/integration/performance testing, Realm database, mobile NoSQL solutions, end-to-end ownership, mobile CI/CD pipelines, identity and access management, Auth0 Guardian SDK, MFA/authentication solutions, iOS security best practices, cryptography, biometric authentication, iOS Keychain, Authentication Service framework, secure data storage, reactive programming frameworks, infrastructure-as-code tools"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3057d55e-9f7"},"title":"AI Agent Engineer","description":"<p>Imagine having an enterprise-grade AppStore at work, one that ensures you can easily search, request, and gain access to any app you need, precisely when you need it. No more long waiting times with outstanding IT requests. As an AI Agent Engineer at Lumos, you will build and own core AI features, in addition to helping to ensure the health of our engineering systems. You will work across the stack, focusing on areas you are most excited about and that bring value to customers. Beyond your technical work, you will gain leadership opportunities early on as we grow our engineering team. You&#39;ll be involved in scaling the product, the team, and the entire company.</p>\n<p>Your responsibilities will include leading the development of agent pipelines, architecting composable agent SDKs with built-in safety and robust fallback strategies, designing tracing tools and alert dashboards to ensure agent performance and quality, owning the agent lifecycle, collaborating closely with security and operations teams to ensure agent governance and auditability, mentoring and upleveling teammates on best practices in observability and resilience, and solving challenging technical problems across the stack to develop critical customer-facing features.</p>\n<p>We&#39;re looking for engineers who want to shape the next generation of intelligent agents – people who care deeply about building reliable, modular systems and elevating those around them. If you&#39;re energized by architecting robust agent SDKs, creating tools that ensure safety and observability, and mentoring others, you&#39;ll thrive at Lumos.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3057d55e-9f7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Lumos","sameAs":"https://lumos.com","logo":"https://logos.yubhub.co/lumos.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/lumos/jobs/6629003003","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$175,000 - $300,000","x-skills-required":["AI-driven workflows","Tool-calling systems","Retrieval-augmented generation (RAG) pipelines","Autonomous agentic orchestration","LangChain","LangGraph","API design","System performance","Software architecture","Go","TypeScript","Python","React","Identity and access management systems","SCIM","OAuth2","SAML","IDPs","HRIS tools"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:56:20.231Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Onsite in San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"AI-driven workflows, Tool-calling systems, Retrieval-augmented generation (RAG) pipelines, Autonomous agentic orchestration, LangChain, LangGraph, API design, System performance, Software architecture, Go, TypeScript, Python, React, Identity and access management systems, SCIM, OAuth2, SAML, IDPs, HRIS tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":300000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3e231b3e-949"},"title":"Forward Deployed AI Engineering Manager, Enterprise","description":"<p>As a Forward Deployed AI Engineering Manager on our Enterprise team, you&#39;ll be the technical bridge between Scale AI&#39;s cutting-edge AI capabilities and our most strategic customers.</p>\n<p>You&#39;ll work with enterprise clients to understand their unique challenges, lead a team that architects specific AI solutions, and ensure successful deployment and adoption of AI systems in production environments.</p>\n<p>This is a Management role that combines deep engineering and AI expertise, leading a team, and working on customer-facing problems. You&#39;ll work directly with customer engineering teams to integrate AI into their critical workflows.</p>\n<p><strong>Customer Integration &amp; Deployment</strong></p>\n<p>Partner directly with enterprise customers to understand their technical infrastructure, data pipelines, and business requirements.</p>\n<p>Design and implement custom integrations between Scale AI&#39;s platform and customer data environments (cloud platforms, data warehouses, internal APIs).</p>\n<p>Build robust data connectors and ETL pipelines to ingest, process, and prepare customer data for AI workflows.</p>\n<p>Deploy and configure AI models and agents within customer security and compliance boundaries.</p>\n<p><strong>AI Agent Development</strong></p>\n<p>Develop production-grade AI agents tailored to customer use cases across domains like customer support, data analysis, content generation, and workflow automation.</p>\n<p>Architect multi-agent systems that orchestrate between different models, tools, and data sources.</p>\n<p>Implement evaluation frameworks to measure agent performance and iterate toward business objectives.</p>\n<p>Design human-in-the-loop workflows and feedback mechanisms for continuous agent improvement.</p>\n<p><strong>Prompt Engineering &amp; Optimization</strong></p>\n<p>Create sophisticated prompt engineering strategies optimized for customer-specific domains and data.</p>\n<p>Build and maintain prompt libraries, templates, and best practices for customer use cases.</p>\n<p>Conduct systematic prompt experimentation and A/B testing to improve model outputs.</p>\n<p>Implement RAG (Retrieval Augmented Generation) systems and fine-tuning pipelines where appropriate.</p>\n<p><strong>Leadership &amp; Collaboration</strong></p>\n<p>Serve as the Engineering Manager and technical point of contact for strategic enterprise accounts.</p>\n<p>Lead a team that is collaborating with customer data scientists, ML engineers, and software developers to ensure smooth integration.</p>\n<p>Work closely with Scale&#39;s product and engineering teams to translate customer needs into product improvements.</p>\n<p>Document technical architectures, integration patterns, and best practices.</p>\n<p><strong>Problem Solving &amp; Innovation</strong></p>\n<p>Debug complex technical issues across the entire stack, from data pipelines to model outputs.</p>\n<p>Rapidly prototype solutions to unblock customers and prove out new use cases.</p>\n<p>Stay current on the latest AI/ML research and tools, bringing innovative approaches to customer problems.</p>\n<p>Identify opportunities for productization based on common customer patterns.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3e231b3e-949","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4602177005","x-work-arrangement":"hybrid","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":"$216,000-$270,000 USD","x-skills-required":["Python","Production","Data Structures","Algorithms","System Design","Cloud Platforms","Modern Data Infrastructure","Problem-Solving","Communication"],"x-skills-preferred":["LLMs","Prompting Techniques","Embeddings","RAG Architectures","Vector Databases","Semantic Search Systems","Containerization","CI/CD Pipelines","Terraform","Bicep","Infrastructure as Code"],"datePosted":"2026-04-18T15:56:13.908Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Production, Data Structures, Algorithms, System Design, Cloud Platforms, Modern Data Infrastructure, Problem-Solving, Communication, LLMs, Prompting Techniques, Embeddings, RAG Architectures, Vector Databases, Semantic Search Systems, Containerization, CI/CD Pipelines, Terraform, Bicep, Infrastructure as Code","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216000,"maxValue":270000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_158a429c-4d8"},"title":"Senior Data Scientist - Product Analytics","description":"<p>We are seeking a Senior Data Scientist to join our Research, Analytics &amp; Data Science (RAD) team. The RAD team uses data and insights to drive evidence-based decision-making. We&#39;re a team of data scientists and product researchers who use data to unlock actionable insights about our customers, products, and business.</p>\n<p>As a Senior Data Scientist, you will partner with product teams to help them identify important questions and answer those questions with data. You will work closely with product managers, designers, and engineers to develop key product success metrics, set targets, measure results, and outcomes, and size opportunities.</p>\n<p>Your responsibilities will include designing, building, and updating end-to-end data pipelines, working closely with stakeholders to drive the collection of new data and the refinement of existing data sources and tables. You will also partner closely with product researchers to build a holistic understanding of our customers, products, and business.</p>\n<p>You will influence our product roadmap and product strategy through experimentation, exploratory analysis, and quantitative research. You will build and automate actionable models and dashboards, craft data stories, and share your findings and recommendations across R&amp;D and the broader company.</p>\n<p>You will drive and shape core RAD foundations and help us improve how the RAD org operates.</p>\n<p>We are looking for someone with 5+ years of experience working with data to solve problems and drive evidence-based decisions. You should have excellent SQL skills and experience of applying analytical and statistical approaches to problem-solving. You should also have a proven track record of initiating and delivering actionable analysis and insights that drive tangible impact with minimal supervision.</p>\n<p>Excellent communication skills (technical and non-technical) and a focus on driving impact are essential. A strong growth mindset and sense of ownership, innate passion, and curiosity are also required.</p>\n<p>Experience with a scientific computing language (such as R or Python) is necessary. Experience with BI/Visualization tools like Tableau, Superset, and Looker is a bonus. Experience working with product teams and leveraging AI tools to boost efficiency and creativity across the data science workflow is also desirable.</p>\n<p>We offer a competitive salary and equity in a fast-growing start-up. We serve lunch every weekday, plus a variety of snack foods and a fully stocked kitchen. Regular compensation reviews, life assurance, comprehensive health and dental insurance, open vacation policy, flexible holidays, paid maternity leave, and 6 weeks paternity leave are also part of our benefits package.</p>\n<p>Our working policy is hybrid, with employees expected to be in the office at least three days per week. We have a radically open and accepting culture, avoiding divisive subjects to foster a safe and cohesive work environment for everyone.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_158a429c-4d8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Intercom","sameAs":"https://www.intercom.com/","logo":"https://logos.yubhub.co/intercom.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/intercom/jobs/6317929","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Analytical and statistical approaches","Scientific computing language (R or Python)","BI/Visualization tools (Tableau, Superset, Looker)","Product teams experience"],"x-skills-preferred":["AI tools","Data modeling and ETL pipelines","Communication skills (technical and non-technical)"],"datePosted":"2026-04-18T15:56:11.783Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, England"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Analytical and statistical approaches, Scientific computing language (R or Python), BI/Visualization tools (Tableau, Superset, Looker), Product teams experience, AI tools, Data modeling and ETL pipelines, Communication skills (technical and non-technical)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0b5a4347-f37"},"title":"Sr. Machine Learning Engineer, Monetization Engineering","description":"<p>About this role:</p>\n<p>We&#39;re looking for a Senior Machine Learning Engineer to join our Monetization team. As a key member of the team, you will be responsible for developing and executing a vision for the evolution of the machine learning technology stack within Ads.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Building cutting-edge technology using the latest advances in deep learning and machine learning to personalize Pinterest</li>\n<li>Partnering closely with teams across Pinterest to experiment and improve ML models for various product surfaces (Homefeed, Ads, Growth, Shopping, and Search)</li>\n<li>Using data-driven methods and leveraging the unique properties of our data to improve candidate retrieval</li>\n<li>Working in a high-impact environment with quick experimentation and product launches</li>\n<li>Keeping up with industry trends in recommendation systems</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>2+ years of industry experience applying machine learning methods</li>\n<li>Degree in computer science, statistics, or related field; or equivalent experience</li>\n<li>End-to-end hands-on experience with building data processing pipelines, large-scale machine learning systems, and big data technologies</li>\n<li>Practical knowledge of large-scale recommender systems, or modern ads ranking, retrieval, targeting, marketplace systems</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>M.S. or PhD in Machine Learning or related areas</li>\n<li>Publications at top ML conferences</li>\n<li>Experience using Cursor, Copilot, Codex, or similar AI coding assistants for development, debugging, testing, and refactoring</li>\n<li>Familiarity with LLM-powered productivity tools for documentation search, experiment analysis, SQL/data exploration, and engineering workflow acceleration</li>\n<li>Expertise in scalable real-time systems that process stream data</li>\n<li>Passion for applied ML and the Pinterest product</li>\n<li>Background in computational advertising</li>\n</ul>\n<p>Relocation Statement:</p>\n<p>This position is not eligible for relocation assistance.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0b5a4347-f37","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/6121551","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$189,721-$332,012 USD","x-skills-required":["Machine Learning","Deep Learning","Data Processing Pipelines","Large-Scale Machine Learning Systems","Big Data Technologies","Recommender Systems","Ads Ranking","Retrieval","Targeting","Marketplace Systems"],"x-skills-preferred":["M.S. or PhD in Machine Learning or related areas","Publications at top ML conferences","Experience using Cursor, Copilot, Codex, or similar AI coding assistants","Familiarity with LLM-powered productivity tools","Expertise in scalable real-time systems","Passion for applied ML and the Pinterest product","Background in computational advertising"],"datePosted":"2026-04-18T15:56:06.423Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, US; Palo Alto, CA, US; Seattle, WA, US"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, Deep Learning, Data Processing Pipelines, Large-Scale Machine Learning Systems, Big Data Technologies, Recommender Systems, Ads Ranking, Retrieval, Targeting, Marketplace Systems, M.S. or PhD in Machine Learning or related areas, Publications at top ML conferences, Experience using Cursor, Copilot, Codex, or similar AI coding assistants, Familiarity with LLM-powered productivity tools, Expertise in scalable real-time systems, Passion for applied ML and the Pinterest product, Background in computational advertising","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":189721,"maxValue":332012,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8bd772e4-c1c"},"title":"Manager, Web Experience","description":"<p>At Scale, we&#39;re looking for a Manager, Web Experience to lead our small team of developers and designers. As a key member of the Brand Experience team, you&#39;ll be responsible for managing all aspects of Scale&#39;s digital presence, including the company&#39;s front door scale.com. This role sits at the intersection of brand, engineering, marketing, and design.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Leading the team: owning the web roadmap, prioritizing and communicating clearly, and managing a lean team of web developers and designers</li>\n<li>Managing quality and reliability: owning the quality of everything that ships, designing and running planning around urgent issues, and building systems to keep marketing sites secure, performant, and compliant</li>\n<li>Managing our stack: overseeing Scale&#39;s integrations on web properties, managing relationships with web vendors, internal security, and IT teams, and regularly auditing the AI tooling</li>\n</ul>\n<p>Ideal candidate will have:</p>\n<ul>\n<li>4+ years in web management, digital operations, or a related field, including direct people management</li>\n<li>Technical fluency: comfortable in conversations about architecture, deployment pipelines, or front-end frameworks</li>\n<li>Experience managing AI integrations or APIs: genuine curiosity about how the underlying models work</li>\n<li>Operational instinct: keeps clean documentation, runs tight sprint cycles, and treats QA as a feature</li>\n<li>Calm under pressure: steadies the team when something breaks</li>\n<li>Ambitious and meticulous: motivated by achieving broader business results, and sweats the small details that compound over time into a reputation for quality</li>\n</ul>\n<p>Please note that our policy requires a 90-day waiting period before reconsidering candidates for the same role.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8bd772e4-c1c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale","sameAs":"https://scale.com","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4676261005","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["web management","digital operations","AI integrations","APIs","front-end frameworks","architecture","deployment pipelines","QA","technical leadership"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:55:43.784Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"web management, digital operations, AI integrations, APIs, front-end frameworks, architecture, deployment pipelines, QA, technical leadership"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8317ba42-502"},"title":"Senior Technical Solutions Engineer (Platform)","description":"<p>We are seeking a highly skilled Frontline Senior Technical Solutions Engineer with over 7+ years of experience to join our Platform Support team.</p>\n<p>This role is pivotal in delivering exceptional support for our Databricks Data Intelligence platform, addressing complex technical challenges, and ensuring the seamless operation of our data solutions.</p>\n<p>As a frontline engineer, you will be the primary point of contact for critical issues, working closely with both internal teams and customers to resolve high-impact problems and drive platform improvements.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Frontline Support: Serve as the primary technical point of contact for escalated issues related to the Databricks Data Intelligence platform. Provide expert-level troubleshooting, diagnostics, and resolution for complex problems affecting system performance and reliability.</li>\n</ul>\n<ul>\n<li>Customer Interaction: Engage with customers directly to understand their technical issues and requirements. Provide timely, clear, and actionable solutions to ensure high levels of customer satisfaction.</li>\n</ul>\n<ul>\n<li>Incident Management: Lead the resolution of high-priority incidents, coordinating with various teams to address and mitigate issues swiftly. Conduct thorough root cause analyses and develop preventive measures to avoid recurrence.</li>\n</ul>\n<ul>\n<li>Collaboration: Work closely with engineering, product management, and DevOps teams to share insights, identify recurring issues, and drive improvements to the Databricks Data Intelligence platform.</li>\n</ul>\n<ul>\n<li>Documentation and Knowledge Sharing: Create and maintain detailed documentation on support procedures, known issues, and solutions. Contribute to internal knowledge bases and create training materials to assist other support engineers.</li>\n</ul>\n<ul>\n<li>Performance Monitoring: Monitor and analyze platform performance metrics to identify potential issues before they impact customers. Implement optimizations and enhancements to improve platform stability and efficiency.</li>\n</ul>\n<ul>\n<li>Platform Upgrades: Manage and oversee the deployment of Databricks Data Intelligence platform upgrades and patches, ensuring minimal disruption to services and maintaining system integrity.</li>\n</ul>\n<ul>\n<li>Innovation and Improvement: Stay abreast of industry trends and advancements in Databricks technology. Propose and drive initiatives to enhance platform capabilities and support processes.</li>\n</ul>\n<ul>\n<li>Customer Feedback: Collect and analyze customer feedback to drive continuous improvement in support processes and platform features.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Experience: Minimum of 7+ years of hands-on experience in a technical support or engineering role related to Databricks Data Intelligence platform, cloud data platforms, or big data technologies.</li>\n</ul>\n<ul>\n<li>Technical Skills: A deep understanding of Databricks architecture and Apache Spark, along with experience in cloud platforms like AWS, Azure, or GCP, is essential. Strong capabilities in designing and managing data pipelines, distributed computing are required. Proficiency in Unix/Linux administration, familiarity with DevOps practices, and skills in log analysis and monitoring tools are also crucial for effective troubleshooting and system optimization.</li>\n</ul>\n<ul>\n<li>Problem-Solving: Demonstrated ability to diagnose and resolve complex technical issues with a strong analytical and methodical approach.</li>\n</ul>\n<ul>\n<li>Communication: Exceptional verbal and written communication skills, with the ability to effectively convey technical information to both technical and non-technical stakeholders.</li>\n</ul>\n<ul>\n<li>Customer Focus: Proven experience in managing high-impact customer interactions and ensuring a positive customer experience.</li>\n</ul>\n<ul>\n<li>Collaboration: Ability to work effectively in a team environment, collaborating with engineering, product, and customer-facing teams.</li>\n</ul>\n<ul>\n<li>Education: Bachelor’s degree in Computer Science, Engineering, or a related field. Advanced degree or relevant certifications are highly desirable.</li>\n</ul>\n<p>Preferred Skills:</p>\n<ul>\n<li>Experience with additional big data tools and technologies such as Hadoop, Kafka, or NoSQL databases.</li>\n</ul>\n<ul>\n<li>Familiarity with automation tools and CI/CD pipelines.</li>\n</ul>\n<ul>\n<li>Understanding of data governance and compliance requirements.</li>\n</ul>\n<p>Why Join Us?</p>\n<ul>\n<li>Innovative Environment: Work with cutting-edge technology in a fast-paced, innovative company.</li>\n</ul>\n<ul>\n<li>Career Growth: Opportunities for professional development and career advancement.</li>\n</ul>\n<ul>\n<li>Team Culture: Collaborate with a talented and motivated team dedicated to excellence and continuous improvement.</li>\n</ul>\n<p>PLEASE NOTE: THE ROLE INVOLVES WORKING IN THE EMEA TIMEZONE</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8317ba42-502","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8041698002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Databricks architecture","Apache Spark","AWS","Azure","GCP","Unix/Linux administration","DevOps practices","log analysis and monitoring tools"],"x-skills-preferred":["Hadoop","Kafka","NoSQL databases","automation tools","CI/CD pipelines","data governance and compliance requirements"],"datePosted":"2026-04-18T15:55:32.901Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Databricks architecture, Apache Spark, AWS, Azure, GCP, Unix/Linux administration, DevOps practices, log analysis and monitoring tools, Hadoop, Kafka, NoSQL databases, automation tools, CI/CD pipelines, data governance and compliance requirements"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d3d37bf3-6e8"},"title":"Staff Software Engineer, Backend (Consumer- Retail Cash)","description":"<p>Ready to be pushed beyond what you think you’re capable of?</p>\n<p>At Coinbase, our mission is to increase economic freedom in the world.</p>\n<p>We&#39;re seeking a Staff Software Engineer to join our Consumer Cash team, which provides the foundational cash layer for Coinbase’s Consumer business.</p>\n<p>As a Staff Engineer, you will be the technical anchor for Cash services, defining the architecture and roadmap for core cash capabilities.</p>\n<p>You will be part of the vision to build a compelling and trusted single cash balance that serves Everything Exchange users’ risk-off needs.</p>\n<p>This role is for an engineer who thrives on tackling complex, high-impact distributed systems that require high reliability and performance,especially in a trading and financial technology context.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Serve as the technical leader and strategist for the Consumer Cash team, defining multi-quarter technical strategies that intersect multiple financial products.</li>\n</ul>\n<ul>\n<li>Architect, develop, and own distributed systems that power low-latency APIs and event-driven pipelines that process large volumes of cash transactions with strong correctness guarantees.</li>\n</ul>\n<ul>\n<li>Provide technical structure and partner closely with management and stakeholders to translate business goals into a defined strategic roadmap.</li>\n</ul>\n<ul>\n<li>Design and implement foundational, high-performance infrastructure components, leveraging tools like Kafka and Clickhouse in an event-sourced architecture.</li>\n</ul>\n<ul>\n<li>Manage individual project priorities, deadlines, and deliverables with strong technical expertise.</li>\n</ul>\n<ul>\n<li>Mentor and coach other team members on advanced design techniques, coding standards, and best practices for building robust value-add products.</li>\n</ul>\n<ul>\n<li>Leverage our modern, diverse tech stack to write high-quality, production-ready code that is thoroughly tested and delivers a critical product to market.</li>\n</ul>\n<p>What we look for in you:</p>\n<ul>\n<li>8+ years of experience in software engineering, with significant experience architecting and developing solutions to ambiguous, high-impact problems.</li>\n</ul>\n<ul>\n<li>Demonstrated experience with low-latency, event-driven, or distributed systems.</li>\n</ul>\n<ul>\n<li>A strong signal if you have a background in building consumer facing trading products or any application that handles large amounts of streaming data.</li>\n</ul>\n<ul>\n<li>Passion for building an open financial system that brings the world together.</li>\n</ul>\n<ul>\n<li>Intellectual curiosity, openness, and a passion for building a culture of positive energy and blameless truth-seeking.</li>\n</ul>\n<p>Nice to haves:</p>\n<ul>\n<li>Experience in payments, banking, wallets, or trading systems, especially transaction processing or ledgering.</li>\n</ul>\n<ul>\n<li>Familiarity with the tech stack, including Golang, Clickhouse, Kafka, Redis, MongoDB.</li>\n</ul>\n<ul>\n<li>Experience building financial, high reliability, or security systems.</li>\n</ul>\n<ul>\n<li>Background in Blockchains (such as Bitcoin, Ethereum) or crypto-forward experience (e.g., interacting with Ethereum addresses, ENS, dApps).</li>\n</ul>\n<ul>\n<li>Experience with a company going through rapid growth (from 10 to 100s of engineers).</li>\n</ul>\n<p>Job #: 75913</p>\n<p>#LI-Remote</p>\n<p>Pay Transparency Notice: The target annual base salary for this position can range as detailed below. Total compensation may also include equity and bonus eligibility and benefits (including medical, dental, and vision).</p>\n<p>Annual base salary range (excluding equity and bonus):</p>\n<p>$217,900-$217,900 CAD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d3d37bf3-6e8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Coinbase","sameAs":"https://www.coinbase.com/","logo":"https://logos.yubhub.co/coinbase.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coinbase/jobs/7659458","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$217,900-$217,900 CAD","x-skills-required":["software engineering","distributed systems","low-latency APIs","event-driven pipelines","Kafka","Clickhouse","Golang","MongoDB","Redis"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:55:27.782Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"software engineering, distributed systems, low-latency APIs, event-driven pipelines, Kafka, Clickhouse, Golang, MongoDB, Redis","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":217900,"maxValue":217900,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_946354fd-05b"},"title":"Specialist Solutions Architect - AI Tooling & Platform Management","description":"<p>As a Specialist Solutions Architect (SSA),AI Tooling &amp; System Management, you will build and manage the AI tooling stack and system infrastructure that empowers Field Engineering to deliver customer outcomes with higher velocity.</p>\n<p>These capabilities will be utilized by our Go-To-Market teams, including Solutions Architects and Account Executives, to accelerate technical demos, proofs of concept, and customer engagements.</p>\n<p>You will bring consistency to our internal AI tooling stack, establish standards for AI-driven development practices, and scale these capabilities across the department.</p>\n<p>A critical aspect of this role is building the infrastructure that enables agent networks to perform with high quality and reliability,including context management systems, data integrations, and supporting tooling.</p>\n<p>Additionally, you will develop internal applications and technical tools that enhance the overall lifecycle, track adoption metrics to measure impact, and partner with stakeholders to drive continuous improvement through intelligent automation and AI-augmented workflows.</p>\n<p>The impact you will have:</p>\n<ul>\n<li>Architect production-level AI tooling deployments that meet security, networking, and data integration requirements</li>\n</ul>\n<ul>\n<li>Build and maintain internal AI tooling infrastructure for demos, learning, building POCs, and production workflows across platforms, including AI-assisted development environments, Databricks environments, and cloud-based tooling</li>\n</ul>\n<ul>\n<li>Establish consistency in the AI tooling stack by defining standards, best practices, and reusable patterns that enable Field Engineering to build with AI efficiently and reliably at scale</li>\n</ul>\n<ul>\n<li>Build context management infrastructure for agent networks, including vector databases, knowledge bases, and retrieval systems that ensure AI agents have access to the right information at the right time</li>\n</ul>\n<ul>\n<li>Design and implement system integrations to bring data from enterprise sources into AI applications, ensuring secure, scalable, and reliable data flows</li>\n</ul>\n<ul>\n<li>Develop internal applications to streamline Field Engineering workflows, improve demo and builder environments, and accelerate customer engagement velocity</li>\n</ul>\n<ul>\n<li>Track adoption metrics and tooling effectiveness by instrumenting the AI tooling stack, building dashboards, and providing data-driven insights to leadership on adoption rates, productivity gains, and ROI</li>\n</ul>\n<ul>\n<li>Manage AI tooling infrastructure and spend by overseeing cloud costs, monitoring consumption as teams scale, resolving capacity issues, and deploying automation to reduce operational overhead</li>\n</ul>\n<ul>\n<li>Partner with Scale and Technical Enablement teams to develop documentation, AI-powered development patterns, and training materials</li>\n</ul>\n<ul>\n<li>Support Solution Architects with custom proof of concept environments, AI tooling configurations, and technical guidance for customer engagements</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_946354fd-05b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8409019002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,000-$247,500 USD","x-skills-required":["Cloud Platforms & Architecture","AI Tooling","Context Management & Agent Networks","Application Development","Metrics & Analytics","System Integration & Data Pipelines","Security & Platform Administration","Infrastructure Automation & DevOps"],"x-skills-preferred":["Security","System Integrations & Application Deployment","Developer Experience & AI Tooling"],"datePosted":"2026-04-18T15:55:11.227Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Northeast - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Cloud Platforms & Architecture, AI Tooling, Context Management & Agent Networks, Application Development, Metrics & Analytics, System Integration & Data Pipelines, Security & Platform Administration, Infrastructure Automation & DevOps, Security, System Integrations & Application Deployment, Developer Experience & AI Tooling","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":247500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ce9f3d34-c8a"},"title":"Senior / Staff+ Software Engineer, Voice Platform","description":"<p>We&#39;re building the infrastructure that lets people talk to Claude,real-time, bidirectional voice conversations that feel natural, responsive, and safe. This is foundational work for how millions of people will interact with AI.</p>\n<p>The Voice Platform team designs and operates the serving systems, streaming pipelines, and APIs that bring Anthropic&#39;s audio models from research into production across Claude.ai, our mobile apps, and the Anthropic API. You&#39;ll work at the intersection of real-time media, low-latency inference, and distributed systems,building infrastructure where every millisecond of latency is felt by the user.</p>\n<p>We partner closely with the Audio research team, who train the speech understanding and generation models, and with product teams shipping voice experiences to users. Your job is to make those models fast, reliable, and delightful to talk to at scale.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build the real-time streaming infrastructure that powers voice conversations with Claude,ingesting microphone audio, orchestrating model inference, and streaming synthesized speech back with minimal latency</li>\n</ul>\n<ul>\n<li>Build low-latency serving systems for speech models, optimizing time-to-first-audio and end-to-end conversational responsiveness</li>\n</ul>\n<ul>\n<li>Develop the public and internal APIs that expose voice capabilities to Claude.ai, mobile clients, and third-party developers</li>\n</ul>\n<ul>\n<li>Own the audio transport layer,codecs, jitter buffers, adaptive bitrate, packet loss recovery,so conversations stay smooth across unreliable networks</li>\n</ul>\n<ul>\n<li>Build observability and quality-measurement systems for voice: latency distributions, audio quality metrics, interruption handling, and turn-taking accuracy</li>\n</ul>\n<ul>\n<li>Partner with Audio research to move new model architectures from experiment to production, and feed real-world performance data back into research</li>\n</ul>\n<ul>\n<li>Collaborate with mobile and product engineering on client-side audio capture, playback, and the end-to-end user experience</li>\n</ul>\n<p>You may be a good fit if you</p>\n<ul>\n<li>Have 6+ years of experience building distributed systems, real-time infrastructure, or platform services at scale</li>\n</ul>\n<ul>\n<li>Have shipped production systems where latency is measured in tens of milliseconds and users notice when you miss</li>\n</ul>\n<ul>\n<li>Are comfortable working across the stack,from transport protocols and serving infrastructure up to the APIs product teams build on</li>\n</ul>\n<ul>\n<li>Are results-oriented, with a bias toward flexibility and impact</li>\n</ul>\n<ul>\n<li>Pick up slack, even if it goes outside your job description</li>\n</ul>\n<ul>\n<li>Enjoy pair programming (we love to pair!)</li>\n</ul>\n<ul>\n<li>Care about the societal impacts of voice AI and want to help shape how these systems are developed responsibly</li>\n</ul>\n<ul>\n<li>Are comfortable with ambiguity,voice is a fast-moving space, and you&#39;ll help define the architecture as we learn what works</li>\n</ul>\n<p>Strong candidates may also have experience with</p>\n<ul>\n<li>Real-time media protocols and stacks: WebRTC, RTP, gRPC bidirectional streaming, or WebSockets at scale</li>\n</ul>\n<ul>\n<li>Audio engineering fundamentals: codecs (Opus, AAC), voice activity detection, echo cancellation, jitter buffering, or audio DSP</li>\n</ul>\n<ul>\n<li>Low-latency ML inference serving, streaming model outputs, or GPU-based serving infrastructure</li>\n</ul>\n<ul>\n<li>Telephony, live streaming, video conferencing, or voice assistant platforms</li>\n</ul>\n<ul>\n<li>Mobile audio pipelines on iOS (AVAudioEngine, AudioUnits) or Android (Oboe, AAudio)</li>\n</ul>\n<ul>\n<li>Working alongside ML researchers to productionize models,speech experience is a plus but not required</li>\n</ul>\n<p>Representative projects</p>\n<ul>\n<li>Driving time-to-first-audio below human perceptual thresholds by co-designing the serving pipeline with the Audio research team</li>\n</ul>\n<ul>\n<li>Building a streaming inference orchestrator that interleaves speech recognition, LLM reasoning, and speech synthesis with overlapping execution</li>\n</ul>\n<ul>\n<li>Designing the voice mode API surface for the Anthropic API so developers can build their own voice agents on Claude</li>\n</ul>\n<ul>\n<li>Implementing graceful barge-in and interruption handling so users can cut Claude off mid-sentence naturally</li>\n</ul>\n<ul>\n<li>Instrumenting end-to-end audio quality metrics and building dashboards that catch regressions before users do</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ce9f3d34-c8a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5172245008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$485,000 USD","x-skills-required":["Real-time media protocols and stacks","Audio engineering fundamentals","Low-latency ML inference serving","Distributed systems","API design"],"x-skills-preferred":["WebRTC","RTP","gRPC bidirectional streaming","WebSockets","Opus","AAC","voice activity detection","echo cancellation","jitter buffering","audio DSP","GPU-based serving infrastructure","telephony","live streaming","video conferencing","voice assistant platforms","mobile audio pipelines on iOS","Android","pair programming"],"datePosted":"2026-04-18T15:55:09.622Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Real-time media protocols and stacks, Audio engineering fundamentals, Low-latency ML inference serving, Distributed systems, API design, WebRTC, RTP, gRPC bidirectional streaming, WebSockets, Opus, AAC, voice activity detection, echo cancellation, jitter buffering, audio DSP, GPU-based serving infrastructure, telephony, live streaming, video conferencing, voice assistant platforms, mobile audio pipelines on iOS, Android, pair programming","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6a946fed-007"},"title":"Software Engineering Intern","description":"<p>Job Description:</p>\n<p>About Starburst</p>\n<p>Starburst is the data platform for analytics, applications, and AI, unifying data across clouds and on-premises to accelerate AI innovation.</p>\n<p>About the role</p>\n<p>Want to work on systems that actually matter? At Starburst, our backend is the engine behind fast, scalable data access , and as an intern, you won’t be on the sidelines.</p>\n<p>You’ll join the team building Starburst Galaxy or Enterprise and start contributing from day one. Real code, real impact, real ownership.</p>\n<p>You’ll work on distributed systems, APIs, and data processing pipelines, learning directly from experienced engineers while solving problems that show up in production , not just in theory.</p>\n<p>We’ll support your growth, but we’ll also expect you to take initiative, move fast, and ship.</p>\n<p>This is a paid, 10-week internship (June 1st – August 28th, 2026).</p>\n<p>Responsibilities</p>\n<ul>\n<li>Build backend features used by real customers</li>\n</ul>\n<ul>\n<li>Work on scalable, distributed systems</li>\n</ul>\n<ul>\n<li>Ship code early and often</li>\n</ul>\n<ul>\n<li>Learn modern backend technologies in practice</li>\n</ul>\n<ul>\n<li>Collaborate with engineers across different time zones</li>\n</ul>\n<ul>\n<li>Tackle real-world performance and reliability challenges</li>\n</ul>\n<p>What we’re looking for</p>\n<ul>\n<li>You’re excited about backend engineering</li>\n</ul>\n<ul>\n<li>You’ve coded in Java</li>\n</ul>\n<ul>\n<li>You like solving hard problems</li>\n</ul>\n<ul>\n<li>You take ownership and get things done</li>\n</ul>\n<ul>\n<li>You’re curious and learn fast</li>\n</ul>\n<ul>\n<li>You’re a rising Senior or recent graduate</li>\n</ul>\n<p>Benefits</p>\n<p>All-Stars have the opportunity and freedom to realize their true potential. By building alongside top talent, we’re empowered to take ownership of our careers and drive meaningful change. Anchored in industry-proven technology and unprecedented success, All-Stars are taking on the challenge everyday to disrupt our industry – and the future.</p>\n<p>Our global workforce is supported by a competitive Total Rewards program that reflects our commitment to a rewarding and supportive work environment. This includes a variety of benefits like competitive pay, attractive stock grants, flexible paid time off, and more.</p>\n<p>We are committed to fostering an intentional, inclusive, and diverse culture that drives deep engagement, authentic belonging, and an exceptional All-Star experience. We believe that diversity of thought, perspective, background and experience will enable us to own what we do, drive our success and empower our All-Stars to show up authentically.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6a946fed-007","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Starburst","sameAs":"https://www.starburst.io/","logo":"https://logos.yubhub.co/starburst.io.png"},"x-apply-url":"https://job-boards.greenhouse.io/starburst/jobs/5129429008","x-work-arrangement":"hybrid","x-experience-level":"entry","x-job-type":"internship","x-salary-range":"8 000 zł-12 000 zł PLN","x-skills-required":["Java","backend engineering","distributed systems","APIs","data processing pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:55:02.342Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Poland"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, backend engineering, distributed systems, APIs, data processing pipelines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f95ac4b6-a7c"},"title":"Software Engineer - Delivery Platform","description":"<p>At Squarespace, we&#39;re reimagining how people bring their ideas to life online. Our Infrastructure Engineering teams are at the heart of that mission --- building the platforms and tooling that let every engineer ship with speed and confidence.</p>\n<p>As a Software Engineer on the Delivery team, you&#39;ll work on the systems that sit between GitHub and production. These systems include nearly every Squarespace service, such as CI/CD pipelines, GitOps workflows, and the deployment platform that spans our Kubernetes clusters and regions. If you&#39;re passionate about developer experience, modern deployment tooling, and making other engineers more productive, we want to hear from you.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build and evolve the platform that ships Squarespace services to production --- CI/CD pipelines, GitOps workflows, and deployment tooling across Kubernetes clusters.</li>\n<li>Increase adoption of modern deployment tooling across high-traffic services</li>\n<li>Design reusable Helm charts, GitOps templates, and standardized rollout/rollback patterns for engineering teams.</li>\n<li>Identify improvements to CI pipeline performance and reliability across the organization.</li>\n<li>Contribute to AI-assisted delivery tooling that helps engineers self-serve and diagnose build failures.</li>\n<li>Develop technical documentation to ensure knowledge sharing and reusability.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>3+ years of backend or platform engineering experience.</li>\n<li>Experience building or improving CI/CD pipelines (e.g., Drone, Jenkins, GitHub Actions, Harness).</li>\n<li>Knowledge of Docker and Kubernetes.</li>\n<li>Familiarity with GitOps tooling such as Argo CD or Flux.</li>\n<li>Proficiency in Go, Python, or Java.</li>\n<li>Experience with Google Cloud, AWS, or Azure.</li>\n<li>Comfortable with Agile methodologies and Git.</li>\n<li>Experience troubleshooting issues with users.</li>\n</ul>\n<p><strong>Benefits &amp; Perks</strong></p>\n<ul>\n<li>A choice between medical plans with an option for 100% covered premiums</li>\n<li>Fertility and adoption benefits</li>\n<li>Access to supplemental insurance plans for additional coverage</li>\n<li>Headspace mindfulness app subscription</li>\n<li>Global Employee Assistance Program</li>\n<li>Retirement benefits with employer match</li>\n<li>Flexible paid time off</li>\n<li>12 weeks paid parental leave and family care leave</li>\n<li>Pretax commuter benefit</li>\n<li>Education reimbursement</li>\n<li>Employee donation match to community organizations</li>\n<li>7 Global Employee Resource Groups (ERGs)</li>\n<li>Dog-friendly workplace</li>\n<li>Free lunch and snacks</li>\n<li>Private rooftop</li>\n<li>Hack week twice per year</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f95ac4b6-a7c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Squarespace","sameAs":"https://www.squarespace.com/about/careers","logo":"https://logos.yubhub.co/squarespace.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/squarespace/jobs/7789058","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$110,500 - $178,250 USD","x-skills-required":["backend or platform engineering experience","CI/CD pipelines","Docker","Kubernetes","GitOps tooling","Go","Python","Java","Google Cloud","AWS","Azure","Agile methodologies","Git"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:49.772Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend or platform engineering experience, CI/CD pipelines, Docker, Kubernetes, GitOps tooling, Go, Python, Java, Google Cloud, AWS, Azure, Agile methodologies, Git","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":110500,"maxValue":178250,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3beddc8f-183"},"title":"Staff Data Systems Analyst","description":"<p>At ZoomInfo, we&#39;re looking for a Senior Data Systems Analyst to join our team. As a key member of our data operations team, you&#39;ll be responsible for building deep expertise in our company data pipeline, which ingests, processes, and profiles millions of company records. Your primary focus will be on mastering our pipeline architecture, contributing to our infrastructure transition, and leading strategic data improvement initiatives.</p>\n<p>In your first 6-12 months, you&#39;ll work alongside other analysts who have context on our systems, learning the architecture while bringing fresh perspectives and technical depth. As you gain mastery and systems stabilize, you&#39;ll increasingly own pipeline architecture decisions and lead strategic data improvement initiatives.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Mastering our company data pipeline architecture, including how data flows from ingestion through profiling, what transforms are applied at each stage, and how components interconnect</li>\n<li>Reading and analyzing production code to understand data transformations, trace data lineage, and assess how proposed changes would impact the system</li>\n<li>Developing frameworks for evaluating tradeoffs between technical complexity, implementation effort, and customer impact</li>\n<li>Creating clear documentation, system maps, and knowledge resources that capture architecture decisions, dependencies, and design rationale</li>\n</ul>\n<ul>\n<li>Contributing to pipeline evolution and infrastructure improvements by participating in design conversations with Engineering and Product, validating pipeline improvements through rigorous testing, and translating data quality investigations and emerging requirements into system-level improvement opportunities</li>\n</ul>\n<ul>\n<li>Solving complex, ambiguous data challenges by leading or contributing to data improvement initiatives that require both systems thinking and creative problem-solving</li>\n</ul>\n<ul>\n<li>Building partnerships and institutional knowledge by developing strong working relationships with Data Acquisition, Product, Engineering, and fellow data analysts, conducting impact analyses and validation studies, and documenting your learning, approaches, and insights</li>\n</ul>\n<p>We&#39;re looking for a highly skilled individual with a strong background in data analytics, data engineering, or related technical roles. You should have experience working with data pipelines, ETL systems, or data processing infrastructure, and be able to read and understand code (Python, Java, SQL, or similar) to analyze data transformations, understand system logic, and assess technical feasibility.</p>\n<p>Required qualifications include:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, Engineering, Mathematics, Statistics, or related quantitative field</li>\n<li>5+ years of experience in data analytics, data engineering, or related technical roles</li>\n<li>Experience working with data pipelines, ETL systems, or data processing infrastructure</li>\n<li>Ability to read and understand code (Python, Java, SQL, or similar)</li>\n<li>Strong programming skills in Python and SQL for data analysis and manipulation</li>\n<li>Experience solving ambiguous, multi-faceted data problems that required figuring out the approach, not just executing a well-defined analysis</li>\n<li>Demonstrated ability to work effectively with Engineering and/or Product teams, translating between technical implementation and business/customer needs</li>\n<li>Strong analytical skills with ability to investigate complex issues systematically</li>\n<li>Excellent communication skills,able to explain technical concepts clearly to diverse audiences</li>\n<li>Self-directed with strong ownership mentality,you drive your work forward and know when to seek input</li>\n</ul>\n<p>Preferred qualifications include experience with company data, business data, web data acquisition, or data quality initiatives, as well as experience with data profiling, entity resolution, record linkage, or data matching systems.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3beddc8f-183","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8408622002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data analytics","data engineering","data pipelines","ETL systems","data processing infrastructure","Python","Java","SQL","data transformation","system logic","technical feasibility"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:46.937Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver, Washington, United States; Waltham, Massachusetts, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data analytics, data engineering, data pipelines, ETL systems, data processing infrastructure, Python, Java, SQL, data transformation, system logic, technical feasibility"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3917fb4f-2ab"},"title":"Full Stack Software Engineer","description":"<p>We are looking for a talented full stack software engineer to join our growing team at Anduril Labs in Washington, DC.</p>\n<p>As a full stack software engineer in Anduril Labs, you will help bring innovative, next-generation concepts to life through proof-of-concept development and rapid prototyping using bleeding edge technologies.</p>\n<p>The ideal candidate has exceptional software development and creative problem-solving skills, is a self-starter, and can quickly grasp complex concepts.</p>\n<p>As a full stack software engineer, you possess the skills to architect, develop, and deploy distributed applications and services, including both front-end and back-end components.</p>\n<p>You have experience with agile, end-to-end software development lifecycle and are comfortable developing and deploying code across Windows and Linux-based systems (including standalone bare-metal hardware, virtualized environments, and cloud-hosted platforms).</p>\n<p>Embedded software development experience is a plus.</p>\n<p>You are also proficient in integrating legacy code and systems, leveraging open-source technologies, and developing and utilizing APIs.</p>\n<p>Additionally, you have a solid understanding of AI/ML core concepts (e.g., feature extraction, supervised vs. unsupervised learning, regression, classification, clustering, deep learning neural networks, NLP, LLMs, SLMs, model fine-tuning, prompt engineering, RAG) and hands-on experience developing (Gen)AI-enhanced applications or services.</p>\n<p>We also expect candidates to have familiarity with database technologies (e.g., SQL, NoSQL, Graph DB, Vector DB) and experience with data modeling, data wrangling, analytics, and visualization.</p>\n<p>Since Anduril Labs supports all Anduril businesses and product lines, you will have the unique opportunity to work closely with multi-disciplinary engineering and product development teams across the entire company.</p>\n<p>This means you will get to directly contribute to the development of Anduril’s next-generation products and services.</p>\n<p>So if you thrive in a dynamic environment that values creative problem-solving, love writing code, excel as both an individual contributor and team player, are eager to learn, and bring a can-do attitude, this role is for you.</p>\n<p><strong>Key Responsibilities:</strong></p>\n<ul>\n<li>Lead the development of prototypes to demonstrate advanced concepts in areas like autonomous and multi-agent systems, GenAI, advanced data analytics, quantum computing/sensing/networking/comms/machine learning, modeling, simulation, optimization, visualization, next-gen human-machine interfaces, heterogenous computing, and cybersecurity.</li>\n</ul>\n<ul>\n<li>Own the entire Software Development Lifecycle from inception through development, testing, deployment, and documentation for Anduril Labs-developed software prototypes.</li>\n</ul>\n<ul>\n<li>Interface and collaborate with other Anduril and customer engineering teams, and strategic partners.</li>\n</ul>\n<ul>\n<li>Support Anduril- and customer-funded R&amp;D efforts.</li>\n</ul>\n<ul>\n<li>Participate in field experiments and technology demonstrations.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>3+ years of programming with Python, C++, Java, Rust, Go, or JavaScript/TypeScript.</li>\n</ul>\n<ul>\n<li>Proven software architecture and design skills.</li>\n</ul>\n<ul>\n<li>Ability to quickly understand and navigate complex systems and established codebases.</li>\n</ul>\n<ul>\n<li>AI/ML development using commercial and open-source AI frameworks, models, and tools (e.g., Jupyter Notebook, PyTorch, TensorFlow, Scikit-learn, OpenAI, Claude, Gemini, Llama, LangChain, YOLO, AWS Sagemaker, Bedrock, Azure AI, RAG).</li>\n</ul>\n<ul>\n<li>Web app development (e.g., React, Angular, or Vue).</li>\n</ul>\n<ul>\n<li>Cloud development (e.g., AWS, Azure, or GCP).</li>\n</ul>\n<ul>\n<li>Data modeling and wrangling.</li>\n</ul>\n<ul>\n<li>Networking basics (e.g., DNS, TCP/IP vs. UDP, socket communications, LDAP, Active Directory).</li>\n</ul>\n<ul>\n<li>Database technologies (e.g., SQL, NoSQL, Graph DB, Vector DB).</li>\n</ul>\n<ul>\n<li>API development and integration (e.g., REST, GraphQL).</li>\n</ul>\n<ul>\n<li>Containerization technologies (e.g., Docker, Kubernetes).</li>\n</ul>\n<ul>\n<li>Software development on Linux and Windows.</li>\n</ul>\n<ul>\n<li>Demonstrable hands-on experience using GenAI tools (e.g., OpenAI Codex, Claude Code, Gemini Code Assist, GitHub Copilot, Amazon CodeWhisperer, or similar) for software development, code generation, debugging, and algorithmic exploration.</li>\n</ul>\n<ul>\n<li>Experience with Git version control, build tools, and CI/CD pipelines.</li>\n</ul>\n<ul>\n<li>Demonstrated understanding and application of software testing principles and practices, including unit testing, integration testing, and end-to-end testing.</li>\n</ul>\n<ul>\n<li>Strong problem-solving skills, meticulous attention to detail, and the ability to work effectively in a collaborative team environment.</li>\n</ul>\n<ul>\n<li>Excellent communication and interpersonal skills, with the ability to effectively articulate complex technical concepts to diverse audiences.</li>\n</ul>\n<ul>\n<li>Eligible to obtain and maintain an active U.S. Top Secret SCI security clearance.</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>BS in Computer Science, Engineering, or similar field.</li>\n</ul>\n<ul>\n<li>Distributed applications development (e.g., client/server, microservices, multi-agent solutions).</li>\n</ul>\n<ul>\n<li>High performance computing (HPC) and big data technologies (e.g., Apache Spark, Hadoop).</li>\n</ul>\n<ul>\n<li>Mobile app development (e.g., iOS or Android).</li>\n</ul>\n<ul>\n<li>Embedded software development experience.</li>\n</ul>\n<ul>\n<li>Willingness to travel up to approximately 10% US</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3917fb4f-2ab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5089044007","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$132,000-$198,000 USD","x-skills-required":["Python","C++","Java","Rust","Go","JavaScript/TypeScript","Software Architecture","AI/ML","Web App Development","Cloud Development","Data Modeling","Networking","Database Technologies","API Development","Containerization","Git Version Control","Build Tools","CI/CD Pipelines","Unit Testing","Integration Testing","End-to-End Testing"],"x-skills-preferred":["Distributed Applications Development","High Performance Computing","Big Data Technologies","Mobile App Development","Embedded Software Development"],"datePosted":"2026-04-18T15:54:45.879Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, District of Columbia, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, C++, Java, Rust, Go, JavaScript/TypeScript, Software Architecture, AI/ML, Web App Development, Cloud Development, Data Modeling, Networking, Database Technologies, API Development, Containerization, Git Version Control, Build Tools, CI/CD Pipelines, Unit Testing, Integration Testing, End-to-End Testing, Distributed Applications Development, High Performance Computing, Big Data Technologies, Mobile App Development, Embedded Software Development","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":132000,"maxValue":198000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_45dbbd5c-38c"},"title":"Director, Technical Account Management","description":"<p>As the Director of Technical Account Management at Airtable, you will lead and scale a high-impact team that owns the persistent technical relationship with our most strategic Premium Support customers.</p>\n<p>This role requires deep experience in platform architecture and integration, hands-on fluency with AI agent capabilities, and a clear-eyed understanding of what enterprise customers need to run Airtable as mission-critical infrastructure.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Lead and scale a high-performing team of Technical Account Managers who serve as the persistent technical authority for Premium accounts , ensuring customer environments are built to fully leverage Airtable&#39;s platform, including Field Agents, Omni, automation architecture, and the connected data structures that make intelligent workflows perform at scale.</li>\n</ul>\n<ul>\n<li>Own the team&#39;s technical depth across Airtable&#39;s agent capabilities , including Field Agent configuration, data semantics, schema design, MCP connectivity, and automation architecture , so TAMs can guide customers through key architectural decisions and implementation.</li>\n</ul>\n<ul>\n<li>Coach and mentor Managers and ICs, building architectural judgment and platform fluency across the team. Foster a culture of ownership and continuous learning that keeps pace with Airtable&#39;s rapid product evolution.</li>\n</ul>\n<ul>\n<li>Establish and evolve frameworks for how TAMs assess and improve the technical health of Premium accounts , evaluating agent configurations, data semantics, integration coverage, and automation architecture against the full capability of the platform.</li>\n</ul>\n<ul>\n<li>Engage directly with customers during critical technical projects or escalations, diagnosing root cause, proposing structural remediation, and representing Airtable as a calm, expert partner.</li>\n</ul>\n<ul>\n<li>Partner across Sales, Customer Success, and Support to maintain clear ownership boundaries and identify high-value accounts for Premium Support , articulating the TAM value proposition in terms of architectural depth, agent reliability, and long-term technical health.</li>\n</ul>\n<ul>\n<li>Drive program development and influence product direction by iterating on delivery models and surfacing patterns around friction, gaps, or constraints that limit how customers realise value from Airtable&#39;s capabilities.</li>\n</ul>\n<ul>\n<li>Leverage data and KPIs (e.g., technical health scores, automation adoption, integration depth, CSAT) to inform decisions, measure success, and prioritise team focus.</li>\n</ul>\n<p>Who you are:</p>\n<ul>\n<li>You have 10+ years in technical support, solution architecture, or technical account management roles, including at least 5+ years leading enterprise-facing technical teams.</li>\n</ul>\n<ul>\n<li>You bring a solutions-architect mindset, with the ability to evaluate a customer&#39;s existing build, identify structural risk, and prescribe scalable improvements , translating complex technical requirements into concrete, actionable plans. You&#39;ve done this in platform or integration-heavy SaaS environments where customers require ongoing architectural guidance to realise full product value.</li>\n</ul>\n<ul>\n<li>You use AI heavily in your own work , not experimentally, but as a core part of how you operate. You have strong intuition for which tools and approaches extract real value, and you build that thinking into the workflows, playbooks, and frameworks you create for your team.</li>\n</ul>\n<ul>\n<li>You have working fluency in AI architecture concepts relevant to enterprise customers: agent frameworks, MCP connectivity, automation pipelines, and schema design that supports AI-powered workflows.</li>\n</ul>\n<ul>\n<li>You&#39;re a strategic leader and strong operator, known for building scalable frameworks that allow your team to deliver consistent technical value across a complex account portfolio , and for developing the technical depth and architectural judgment of the people around you.</li>\n</ul>\n<ul>\n<li>You are calm and confident under pressure, especially in high-stakes technical escalations, and you balance immediate resolution with long-term architectural remediation.</li>\n</ul>\n<ul>\n<li>You possess exceptional written and verbal communication skills, with the ability to make complex architectural trade-offs legible to audiences ranging from developers and data architects to leadership and executive sponsors.</li>\n</ul>\n<ul>\n<li>You&#39;re analytical and comfortable making data-informed decisions, using technical health signals and program metrics to prioritise resources and identify opportunities for evolution.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_45dbbd5c-38c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airtable","sameAs":"https://airtable.com/","logo":"https://logos.yubhub.co/airtable.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airtable/jobs/8485839002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Technical Account Management","Platform Architecture","Integration","AI Agent Capabilities","Agent Frameworks","MCP Connectivity","Automation Pipelines","Schema Design","Field Agent Configuration","Data Semantics","Automation Architecture"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:30.916Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - US; Remote - Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Technical Account Management, Platform Architecture, Integration, AI Agent Capabilities, Agent Frameworks, MCP Connectivity, Automation Pipelines, Schema Design, Field Agent Configuration, Data Semantics, Automation Architecture"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b2637f59-e14"},"title":"Full-Stack Software Engineer, Reinforcement Learning","description":"<p>As a Full-Stack Software Engineer in RL, you&#39;ll build the platforms, tools, and interfaces that power environment creation, data collection, and training observability. The quality of Claude&#39;s next generation depends on the quality of the data we train it on , and the systems you build are what make that data possible. You&#39;ll own product surfaces end-to-end , from backend services and APIs to the web UIs that researchers, external vendors, and thousands of data labelers use every day.\\n\\nYou don&#39;t need a background in ML research. What matters is that you can take an ambiguous, high-stakes problem and ship a polished, reliable product against it, fast. This team moves very quickly. Claude writes a lot of the code we commit, which means the bottleneck isn&#39;t typing , it&#39;s judgment, taste, and the ability to react to what researchers need next.\\n\\nYou&#39;ll iterate on data collection strategies to distill the knowledge of thousands of human experts around the world into our models, and you&#39;ll do it in a loop that closes in hours and days, not quarters or months.\\n\\nAnthropic&#39;s Reinforcement Learning organization leads the research and development that trains Claude to be capable, reliable, and safe. We&#39;ve contributed to every Claude model, with significant impact on the autonomy and coding capabilities of our most advanced models.\\n\\nOur work spans teaching models to use computers effectively, advancing code generation through RL, pioneering fundamental RL research for large language models, and building the scalable training methodologies behind our frontier production models.\\n\\nThe RL org is organized around four goals: solving the science of long-horizon tasks and continual learning, scaling RL data and environments to be comprehensive and diverse, automating software engineering end-to-end, and training the frontier production model.\\n\\nOur engineering teams build the environments, evaluation systems, data pipelines, and tooling that make all of this possible , from realistic agentic training environments and scalable code data generation to human data collection platforms and production training operations.\\n\\n### Responsibilities\\n\\n<em>   Build and extend web platforms for RL environment creation, management, and quality review , including environment configuration, versioning, and validation workflows\\n</em>   Develop vendor-facing interfaces and tooling that let external partners create, submit, and iterate on training environments with minimal friction\\n<em>   Design and implement platforms for human data collection at scale, including labeling workflows, quality assurance systems, and feedback mechanisms that surface reward signal integrity issues early\\n</em>   Build evaluation dashboards and observability UIs that give researchers real-time insight into environment quality, training run health, and reward hacking\\n<em>   Create backend services and APIs that connect environment authoring tools, data collection systems, and RL training infrastructure\\n</em>   Build and expand scalable code data generation pipelines, producing diverse programming tasks with robust reward signals across languages and difficulty levels\\n<em>   Develop onboarding automation and documentation tooling so new vendors and internal users ramp up in hours, not weeks\\n</em>   Partner closely with RL researchers, data operations, and vendor management to translate ambiguous requirements into well-scoped, well-designed products\\n\\n### Requirements\\n\\n<em>   Strong software engineering fundamentals and real full-stack range , you&#39;re comfortable owning a surface from database schema to frontend\\n</em>   Proficient in Python and a modern web stack (React, TypeScript, or similar)\\n<em>   Track record of shipping systems that solved a hard problem, not just shipped on time , e.g. you built the thing that made your team 10x faster, or the internal tool nobody thought was possible\\n</em>   Operate with high agency: you identify what needs to be done and drive it forward without waiting for a ticket\\n<em>   Found yourself wondering &quot;why isn&#39;t this moving faster?&quot; in previous roles , and then have done something about it\\n</em>   Care about UX and can build interfaces that are intuitive for both technical researchers and non-technical labelers\\n<em>   Communicate clearly with researchers, operations teams, and engineers, and can turn vague asks into well-scoped work\\n</em>   Thrive in a fast-moving environment where priorities shift, Claude is your pair programmer, and the next problem is often one nobody has solved before\\n<em>   Care about Anthropic&#39;s mission to build safe, beneficial AI and want your work to contribute directly to it\\n\\n### Nice to Have\\n\\n</em>   Built data collection, labeling, or annotation platforms , ideally ones that had to scale across many vendors or many task types\\n<em>   Background building multi-tenant platforms with role-based access, audit trails, and vendor management workflows\\n</em>   Experience with cloud infrastructure (GCP or AWS), Docker, and CI/CD pipelines\\n<em>   Familiarity with LLM training, fine-tuning, or evaluation workflows\\n</em>   Experience with async Python (Trio, asyncio) or high-throughput API design\\n<em>   Background in dashboards, monitoring, or observability tooling\\n</em>   Experience working directly with external vendors or partners on technical integrations\\n<em>   A background that isn&#39;t a straight line , e.g. math or physics into SWE, competitive programming, research into engineering, or a side project that outgrew its scope\\n\\n### Representative Projects\\n\\n</em>   Building a unified platform for human data collection that integrates labeling workflows, vendor management, and QA for complex agentic tasks\\n<em>   Developing vendor onboarding automation that handles Docker registry access, API token management, and environment validation\\n</em>   Creating evaluation and observability dashboards that catch reward hacks, measure environment difficulty, and give real-time feedback during production training\\n<em>   Building environment quality review workflows that let researchers browse, grade, and provide feedback on training environments\\n</em>   Developing automated environment quality pipelines that validate correctness and difficulty calibration before environments hit production training\\n*   Building internal tools for browsing and analyzing training run results, environment statistics, and data collection progress</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b2637f59-e14","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5186067008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$300,000-$405,000 USD","x-skills-required":["Python","Modern web stack","React","TypeScript","Strong software engineering fundamentals","Full-stack range","Database schema","Frontend","Cloud infrastructure","Docker","CI/CD pipelines","LLM training","Fine-tuning","Evaluation workflows","Async Python","High-throughput API design","Dashboards","Monitoring","Observability tooling"],"x-skills-preferred":["Data collection","Labeling","Annotation platforms","Multi-tenant platforms","Role-based access","Audit trails","Vendor management workflows"],"datePosted":"2026-04-18T15:54:27.784Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Modern web stack, React, TypeScript, Strong software engineering fundamentals, Full-stack range, Database schema, Frontend, Cloud infrastructure, Docker, CI/CD pipelines, LLM training, Fine-tuning, Evaluation workflows, Async Python, High-throughput API design, Dashboards, Monitoring, Observability tooling, Data collection, Labeling, Annotation platforms, Multi-tenant platforms, Role-based access, Audit trails, Vendor management workflows","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":300000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c0df50e1-9cd"},"title":"Consultant, Developer Platform","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>As a Cloud Engineer for Developer Platform, you are an individual contributor working in the post-sales landscape, responsible for the technical execution of solutions and guidance to our customers, following a consultative approach, to get the most value possible from their Cloudflare investment.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Plan and deliver timely and organized services for customers, ensure customers see the full value in Cloudflare’s products and advice on product best practices.</li>\n</ul>\n<ul>\n<li>Gather business and technical requirements, use cases and any other information required to build, migrate and deliver a solution on behalf of the customer and transition the Cloudflare working environment to the customer.</li>\n</ul>\n<ul>\n<li>Produce a Solution Design, HLD, LLD, databuilds, procedures, scripts, test plans, drawings, deployment plan, migration plan, as-builts, and any other artifacts necessary to deliver the solution and transition smoothly into the customer’s technical teams.</li>\n</ul>\n<ul>\n<li>Implement changes on behalf of the customer in the Cloudflare environment following the customer’s change management process.</li>\n</ul>\n<ul>\n<li>Troubleshoot implementation issues and collaborate with Customer Support, Engineering and other teams to assist technical escalations.</li>\n</ul>\n<ul>\n<li>Contribute towards the success of the organization through knowledge sharing activities such as contributing to internal and external documentation, answering technical Q&amp;A, and helping to iterate on best practices.</li>\n</ul>\n<p>Support building operational assets like templates, automation scripts, procedures, workflows, etc.</p>\n<p>Requirements:</p>\n<ul>\n<li>3+ years of experience in a customer facing position as a Consultant delivering services.</li>\n</ul>\n<ul>\n<li>Demonstrated experience with:</li>\n</ul>\n<ul>\n<li>Developing serverless code in a CI/CD pipeline using an Agile methodology.</li>\n</ul>\n<ul>\n<li>Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP.</li>\n</ul>\n<ul>\n<li>Scripting languages.</li>\n</ul>\n<ul>\n<li>A scripting language (e.g. Python, JavaScript, Bash) and a desire to expand those skills.</li>\n</ul>\n<ul>\n<li>Infrastructure as code tools like Terraform.</li>\n</ul>\n<ul>\n<li>Strong experience with APIs.</li>\n</ul>\n<ul>\n<li>CI/CD pipelines using Azure DevOps or Git.</li>\n</ul>\n<ul>\n<li>Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc.</li>\n</ul>\n<ul>\n<li>Good understanding and knowledge of:</li>\n</ul>\n<ul>\n<li>Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs.</li>\n</ul>\n<ul>\n<li>Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP.</li>\n</ul>\n<ul>\n<li>Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>You have worked with a Cybersecurity company or products and have performed migrations using migration tools.</li>\n</ul>\n<ul>\n<li>You have developed application security and performance capabilities.</li>\n</ul>\n<ul>\n<li>Ability to manage a project, work to deadlines, prioritize between competing demands and manage uncertainty.</li>\n</ul>\n<ul>\n<li>The work will be performed in English. Fluency in a second regional European language is a strong advantage.</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law. We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c0df50e1-9cd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7383015","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Developing serverless code in a CI/CD pipeline using an Agile methodology","Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP","Scripting languages","Infrastructure as code tools like Terraform","Strong experience with APIs","CI/CD pipelines using Azure DevOps or Git","Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc","Good understanding and knowledge of Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs","Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP","Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3"],"x-skills-preferred":["You have worked with a Cybersecurity company or products and have performed migrations using migration tools","You have developed application security and performance capabilities","Ability to manage a project, work to deadlines, prioritize between competing demands and manage uncertainty","The work will be performed in English. Fluency in a second regional European language is a strong advantage"],"datePosted":"2026-04-18T15:54:26.532Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Developing serverless code in a CI/CD pipeline using an Agile methodology, Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP, Scripting languages, Infrastructure as code tools like Terraform, Strong experience with APIs, CI/CD pipelines using Azure DevOps or Git, Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc, Good understanding and knowledge of Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs, Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP, Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3, You have worked with a Cybersecurity company or products and have performed migrations using migration tools, You have developed application security and performance capabilities, Ability to manage a project, work to deadlines, prioritize between competing demands and manage uncertainty, The work will be performed in English. Fluency in a second regional European language is a strong advantage"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_dd290e64-a85"},"title":"Quantum Software Engineer","description":"<p>We are seeking a talented and innovative Quantum Software Engineer to join our forward-looking team at Anduril Labs. In this role, you will be instrumental in building and delivering impactful quantum solutions for both Anduril-internal use cases and external customer applications.</p>\n<p>You will work closely with delivery leads, application developers, and other solutions architects, as well as internal and external partners to design, implement, and deliver bleeding edge quantum solutions on state-of-the-art quantum-inspired, quantum annealing, and quantum gate platforms for real-world defense and national security challenges.</p>\n<p>The ideal candidate will combine a strong foundation in quantum computing principles with hands-on classical and quantum software development expertise. You will leverage your skills to translate complex problems into (hybrid) quantum algorithms, applications, and services.</p>\n<p>This includes developing robust software implementations, and integrate quantum-enhanced solutions into existing and new defense systems.</p>\n<p>If you are passionate about applying theoretical quantum concepts to deliver tangible, high-impact results, and thrive in an environment that values innovation, collaboration, and rapid prototyping, we encourage you to apply.</p>\n<p><strong>Key Responsibilities:</strong> Be a key contributor to the development of next-generation quantum-enhanced Anduril offerings and lead the design, development, and deployment of novel quantum-enhanced applications and services in the defense and national security domain. Develop impactful hybrid quantum algorithms and applications that promise significant decision advantages and focus on practical scalability and real-world applicability. Contribute knowledge of classical and quantum optimization algorithms and tools, evaluating, and communicating their pros and cons, current state-of-the-art, scaling behaviors, trade-offs, and cross-over points. Participate in the full (hybrid) quantum software development lifecycle, from concept and design to testing, deployment, and ongoing maintenance.</p>\n<p><strong>Requirements:</strong> Bachelor&#39;s degree in Computer Science, Quantum Information Science, Physics, Mathematics, or a closely related technical field. 3+ years of hands-on, professional software development experience with C, C++, Python, or another general-purpose compiled programming language. Practical experience in quantum computing, including programming quantum applications, or quantum circuit compilation. Proficiency with one or more leading quantum programming languages, SDKs, or APIs such as Qiskit, CUDA-Q, Q#, Cirq, PennyLane, or similar. Expertise in key mathematical techniques foundational to quantum computing, including linear algebra, matrix decompositions, probability theory, group theory, symmetry, and computational complexity. Proficient with database systems and SQL, with hands-on experience working with relational databases (e.g., PostgreSQL, Oracle, MySQL). Experience with Git version control, build tools, and CI/CD pipelines. Demonstrated understanding and application of software testing principles and practices, including unit testing, integration testing, and end-to-end testing. Strong problem-solving skills, meticulous attention to detail, and the ability to work effectively in a collaborative team environment. Excellent communication and interpersonal skills, with the ability to effectively articulate complex technical concepts to diverse audiences. Eligible to obtain and maintain an active U.S. Top Secret SCI security clearance. Demonstrable hands-on experience using GenAI tools (e.g., OpenAI Codex, Claude Code, Gemini Code Assist, GitHub Copilot, Amazon CodeWhisperer, or similar) for software development, code generation, debugging, and algorithmic exploration.</p>\n<p><strong>Preferred Qualifications:</strong> Master&#39;s or Ph.D. in Quantum Information Science, Physics, Computer Science, or a related quantitative field. Familiarity with leading classical optimization tools and solvers (e.g., CPLEX, Gurobi, OR-Tools) and knowledge of mathematical modeling and classical optimization solution techniques. Experience building and deploying applications to solve complex business or defense problems for customers. Proven record of successful on-time delivery of complex software projects with a high degree of predictability and quality. Experience with deployment of code in distributed environments, cloud application development (e.g., AWS, Azure, GCP), and RESTful API-driven architectures. Experience with high-performance computing (HPC) environments or parallel programming. Familiarity with quantum hardware platforms and their unique characteristics. Prior experience in defense, aerospace, or related industries applying advanced technologies. Willingness to travel up to approximately 10%.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_dd290e64-a85","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5089054007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$132,000-$198,000 USD","x-skills-required":["C","C++","Python","Qiskit","CUDA-Q","Q#","Cirq","PennyLane","Linear Algebra","Matrix Decompositions","Probability Theory","Group Theory","Symmetry","Computational Complexity","Database Systems","SQL","Git","Build Tools","CI/CD Pipelines","Software Testing Principles","Unit Testing","Integration Testing","End-to-End Testing","GenAI Tools"],"x-skills-preferred":["Master's or Ph.D. in Quantum Information Science, Physics, Computer Science, or a related quantitative field","Familiarity with leading classical optimization tools and solvers","Experience building and deploying applications to solve complex business or defense problems for customers","Proven record of successful on-time delivery of complex software projects with a high degree of predictability and quality","Experience with deployment of code in distributed environments, cloud application development, and RESTful API-driven architectures","Experience with high-performance computing (HPC) environments or parallel programming","Familiarity with quantum hardware platforms and their unique characteristics","Prior experience in defense, aerospace, or related industries applying advanced technologies"],"datePosted":"2026-04-18T15:54:19.846Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Washington, District of Columbia, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"C, C++, Python, Qiskit, CUDA-Q, Q#, Cirq, PennyLane, Linear Algebra, Matrix Decompositions, Probability Theory, Group Theory, Symmetry, Computational Complexity, Database Systems, SQL, Git, Build Tools, CI/CD Pipelines, Software Testing Principles, Unit Testing, Integration Testing, End-to-End Testing, GenAI Tools, Master's or Ph.D. in Quantum Information Science, Physics, Computer Science, or a related quantitative field, Familiarity with leading classical optimization tools and solvers, Experience building and deploying applications to solve complex business or defense problems for customers, Proven record of successful on-time delivery of complex software projects with a high degree of predictability and quality, Experience with deployment of code in distributed environments, cloud application development, and RESTful API-driven architectures, Experience with high-performance computing (HPC) environments or parallel programming, Familiarity with quantum hardware platforms and their unique characteristics, Prior experience in defense, aerospace, or related industries applying advanced technologies","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":132000,"maxValue":198000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2a2d718a-f65"},"title":"Senior Software Engineer, AI Platform and Enablement","description":"<p><strong>About the Role</strong></p>\n<p>We&#39;re building a next-generation AI-powered platform and web application for creating audio and video content quickly and easily. This involves developing a revolutionary way to record, transcribe, edit, and mix audio and video on the web using state-of-the-art AI models,a challenge that requires solving complex technical problems. We&#39;re hiring a senior engineer to join our AI Platform and Enablement team. The ideal candidate thrives in a fast-moving, high-ownership environment and is comfortable navigating the ambiguity of bringing research work into an established product.</p>\n<p><strong>About the Team</strong></p>\n<p>The team’s objective is to support integrating cutting-edge first-party models (developed by our in-house AI Research team) and third-party/open source AI models into the Descript product.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Build, maintain, and standardize third-party model integrations, including consulting for other engineering teams with AI model integration needs</li>\n</ul>\n<ul>\n<li>Design, implement, and maintain our AI infrastructure supporting our machine learning life cycle, including data ingestion pipelines, training developer experience and infrastructure, evaluation frameworks, and deployments / GPU infrastructure</li>\n</ul>\n<ul>\n<li>Collaborate with Product Managers, Research Engineers, and AI Researchers to understand their infrastructure needs and ensure our AI systems are robust, scalable, and efficient</li>\n</ul>\n<ul>\n<li>Optimise and scale our models and algorithms for efficient inference</li>\n</ul>\n<ul>\n<li>Deploy, monitor, and manage AI models in production</li>\n</ul>\n<p><strong>What You Bring</strong></p>\n<ul>\n<li>Experience in deploying and managing AI models in production</li>\n</ul>\n<ul>\n<li>Experience with the tools of large volume data pipelines like spark, flume, dask, etc.</li>\n</ul>\n<ul>\n<li>Familiarity with cloud platforms (AWS, Google Cloud, Azure) and container technologies (Docker, Kubernetes).</li>\n</ul>\n<ul>\n<li>Knowledge of DevOps and MLOps best practices</li>\n</ul>\n<ul>\n<li>Strong problem-solving abilities and excellent communication skills.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Generous healthcare package</li>\n</ul>\n<ul>\n<li>401k matching program</li>\n</ul>\n<ul>\n<li>Catered lunches</li>\n</ul>\n<ul>\n<li>Flexible vacation time</li>\n</ul>\n<p><strong>Fun fact about me: I love pineapple on pizza.</strong></p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2a2d718a-f65","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Descript","sameAs":"https://descript.com/","logo":"https://logos.yubhub.co/descript.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/descript/jobs/7580335003","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,000 - $286,000/year","x-skills-required":["Experience in deploying and managing AI models in production","Experience with the tools of large volume data pipelines like spark, flume, dask, etc.","Familiarity with cloud platforms (AWS, Google Cloud, Azure) and container technologies (Docker, Kubernetes)","Knowledge of DevOps and MLOps best practices","Strong problem-solving abilities and excellent communication skills"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:54:12.258Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Experience in deploying and managing AI models in production, Experience with the tools of large volume data pipelines like spark, flume, dask, etc., Familiarity with cloud platforms (AWS, Google Cloud, Azure) and container technologies (Docker, Kubernetes), Knowledge of DevOps and MLOps best practices, Strong problem-solving abilities and excellent communication skills","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":286000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9be280f4-cbc"},"title":"Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for an engineer to join our small, high-impact team responsible for architecting and scaling the core infrastructure behind distributed training pipelines, multimodal data catalogs, and intelligent processing systems that operate over petabytes of data.</p>\n<p>As a software engineer on our data infrastructure team, you&#39;ll design, build, and operate scalable, fault-tolerant infrastructure for LLM Research: distributed compute, data orchestration, and storage across modalities. You&#39;ll develop high-throughput systems for data ingestion, processing, and transformation , including training data catalogs, deduplication, quality checks, and search. You&#39;ll also build systems for traceability, reproducibility, and robust quality control at every stage of the data lifecycle.</p>\n<p>You&#39;ll collaborate with research teams to unlock new features, improve data quality, and accelerate training cycles. You&#39;ll implement and maintain monitoring and alerting to support platform reliability and performance.</p>\n<p>If you&#39;re excited by distributed systems, large-scale data mining, open-source tools like Spark, Kafka, Beam, Ray, and Delta Lake, and enjoy building from the ground up, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9be280f4-cbc","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Thinking Machines Lab","sameAs":"https://thinkingmachines.ai/","logo":"https://logos.yubhub.co/thinkingmachines.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/thinkingmachines/jobs/5013919008","x-work-arrangement":"onsite","x-experience-level":"entry|mid|senior","x-job-type":"full-time","x-salary-range":"$350,000 - $475,000 USD","x-skills-required":["backend language (Python or Rust)","distributed compute frameworks (Apache Spark or Ray)","cloud infrastructure","data lake architectures","batch and streaming pipelines"],"x-skills-preferred":["Kafka","dbt","Terraform","Airflow","web crawler","deduplication","data mining","search","file formats and storage systems"],"datePosted":"2026-04-18T15:54:00.309Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend language (Python or Rust), distributed compute frameworks (Apache Spark or Ray), cloud infrastructure, data lake architectures, batch and streaming pipelines, Kafka, dbt, Terraform, Airflow, web crawler, deduplication, data mining, search, file formats and storage systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":350000,"maxValue":475000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_307c2f1c-d78"},"title":"Senior SDET - Tooling Engineer","description":"<p>We are looking for a highly skilled Senior Software Quality Engineer (SDET) to lead our end-to-end quality engineering initiatives across mobile, web, backend, and data platforms. This role combines deep technical expertise with a forward-thinking, AI-first mindset, driving innovation, scalability, and reliability through advanced automation and intelligent testing strategies.</p>\n<p>As a senior member of the team, you will champion modern, AI-enhanced quality practices and help build a culture where continuous improvement, automation-first thinking, and data-driven decisions are embedded at every stage of product development. This is a hybrid position in Mountain View (Headquarters) and will require in-office work 2 days a week.</p>\n<p>The base salary range for this full-time position is $210,000 to $257,000, plus equity and benefits. Our salary ranges are determined by role, level, and location. EarnIn provides excellent benefits for our employees, including healthcare, internet/cell phone reimbursement, and a learning and development stipend.</p>\n<p><strong>Quality Engineering &amp; Test</strong></p>\n<p>Own end-to-end quality across iOS and Android applications and their supporting backend services, ensuring high confidence in weekly (or faster) releases. Design and implement comprehensive test strategies covering:</p>\n<ul>\n<li>Native mobile applications (iOS &amp; Android)</li>\n<li>Mobile-to-backend integrations (REST APIs, auth flows, event-driven systems)</li>\n<li>Microservices and distributed systems</li>\n<li>Critical web workflows that intersect with mobile journeys</li>\n<li>Device, OS, browser, and network variability</li>\n<li>App lifecycle events, offline behavior, retries, and edge cases</li>\n</ul>\n<p>Ensure critical user journeys are validated across mobile UI → API → backend → web touchpoints, preventing production escapes in high-impact flows. Partner with engineering teams to embed quality gates into the mobile release lifecycle, including pre-merge validation, release candidate verification, and post-deploy smoke testing.</p>\n<p>Drive improvements in testability by introducing better logging, API contracts, observability hooks, feature flags, and deterministic state management. Establish meaningful quality metrics (crash analytics, defect trends, flaky tests, API reliability, release risk scoring) and surface actionable insights to engineering stakeholders.</p>\n<p>Champion shift-left quality by influencing design reviews, API schema discussions, and acceptance criteria early in development.</p>\n<p><strong>AI-Driven Quality and Automation</strong></p>\n<p>Leverage AI to enhance mobile, backend, and web testing effectiveness, including:</p>\n<ul>\n<li>AI-assisted test case and test data generation</li>\n<li>Intelligent regression suite prioritization based on code changes</li>\n<li>Predictive defect detection and risk-based testing</li>\n<li>Flaky test detection and automated stabilization insights</li>\n</ul>\n<p>Integrate AI-powered log intelligence, crash clustering, and anomaly detection into quality workflows. Continuously evaluate and experiment with AI-driven QA tools to increase coverage, reduce maintenance overhead, and accelerate release cycles.</p>\n<p>Contribute to building an AI-augmented quality ecosystem that improves speed without compromising reliability.</p>\n<p><strong>Automation Excellence</strong></p>\n<p>Design, build, and scale robust automation frameworks using:</p>\n<ul>\n<li>XCUITest, Espresso, Appium (mobile automation)</li>\n<li>Playwright (web and mobile web validation)</li>\n<li>REST Assured or similar tools for API and service validation</li>\n</ul>\n<p>Ensure frameworks are modular, maintainable, and optimized for scale across multiple teams. Integrate automated validation into CI/CD pipelines (Jenkins, GitHub Actions, etc.) to enable:</p>\n<ul>\n<li>Pre-merge quality gates</li>\n<li>Parallelized execution</li>\n<li>Environment-aware test runs</li>\n<li>Post-deployment smoke and regression coverage</li>\n</ul>\n<p>Build developer-friendly tooling that enables:</p>\n<ul>\n<li>Self-service test execution</li>\n<li>Real-time reporting and dashboards</li>\n<li>Faster debugging and failure triage</li>\n<li>Scalable test data and environment management</li>\n</ul>\n<p>Continuously reduce flakiness, improve signal quality, and optimize execution time across mobile and backend suites.</p>\n<p><strong>Performance, Scalability &amp; Reliability</strong></p>\n<p>Design and execute performance validation across:</p>\n<ul>\n<li>Mobile app startup time and responsiveness</li>\n<li>API latency, throughput, and reliability</li>\n<li>Backend load and stress conditions</li>\n<li>Web performance for critical flows</li>\n</ul>\n<p>Partner with engineering teams to analyze production logs, crash reports, browser telemetry, and service metrics. Lead root-cause analysis of complex cross-layer defects spanning mobile UI, APIs, backend services, and web surfaces.</p>\n<p>Ensure reliability validation is embedded directly into release workflows.</p>\n<p><strong>Cross-Functional Collaboration and Leadership</strong></p>\n<p>Collaborate closely with mobile engineers, backend developers, web engineers, product managers, DevOps teams, and release managers to define clear, testable requirements and release criteria. Actively participate in sprint grooming, planning, stand-ups, and retrospectives.</p>\n<p>Influence best practices around mobile-first design, API contracts, and release readiness. Support mobile app release activities, including release candidate validation, go/no-go recommendations, and post-release monitoring.</p>\n<p>Mentor junior QA engineers and contribute to raising the technical bar in automation and cross-platform validation. Work effectively with globally distributed teams to coordinate testing across time zones.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_307c2f1c-d78","directApply":true,"hiringOrganization":{"@type":"Organization","name":"EarnIn","sameAs":"https://www.earnin.com/","logo":"https://logos.yubhub.co/earnin.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/earnin/jobs/7403324","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$210,000 to $257,000.Minus equity and benefits","x-skills-required":["XCUITest","Espresso","Appium","Playwright","REST Assured","API contracts","Feature flags","Deterministic state management","AI-assisted test case and test data generation","Intelligent regression suite prioritization","Predictive defect detection","Risk-based testing","Flaky test detection","Automated stabilization insights","Log intelligence","Crash clustering","Anomaly detection","CI/CD pipelines","Pre-merge quality gates","Parallelized execution","Environment-aware test runs","Post-deployment smoke and regression coverage","Self-service test execution","Real-time reporting and dashboards","Faster debugging and failure triage","Scalable test data and environment management"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:53:59.311Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View, US"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"XCUITest, Espresso, Appium, Playwright, REST Assured, API contracts, Feature flags, Deterministic state management, AI-assisted test case and test data generation, Intelligent regression suite prioritization, Predictive defect detection, Risk-based testing, Flaky test detection, Automated stabilization insights, Log intelligence, Crash clustering, Anomaly detection, CI/CD pipelines, Pre-merge quality gates, Parallelized execution, Environment-aware test runs, Post-deployment smoke and regression coverage, Self-service test execution, Real-time reporting and dashboards, Faster debugging and failure triage, Scalable test data and environment management","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":210000,"maxValue":257000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bec4e006-74f"},"title":"Consultant, Developer Platform","description":"<p>About the role: Cloudflare provides advisory and hands-on-keyboard implementation and migration services for enterprise customers. As a Consultant for Developer Platform, you are an individual contributor working in the post-sales landscape, responsible for the technical execution of solutions and guidance to our customers, following a consultative approach, to get the most value possible from their Cloudflare investment.</p>\n<p>You are an expert in Developer Platform products or equivalent and will focus on building and deploying serverless applications with scale, performance, security and reliability leveraging: Workers, Workers KV, Workers AI, D1, R2, Images, and many other products.</p>\n<p>This position has working hours Monday to Friday 09:00 a.m. to 06:00 p.m. Occasionally, we support our customers during the weekends for specific changes that need to be done outside of their business hours. Travel is expected to be around 40%.</p>\n<p>Experience might include a combination of the skills below:</p>\n<ul>\n<li>Plan and deliver timely and organized services for customers, ensure customers see the full value in Cloudflare’s products and advice on product best practices.</li>\n<li>Gather business and technical requirements, use cases and any other information required to build, migrate and deliver a solution on behalf of the customer and transition the Cloudflare working environment to the customer.</li>\n<li>Produce a Solution Design, HLD, LLD, databuilds, procedures, scripts, test plans, drawings, deployment plan, migration plan, as-builts, and any other artifacts necessary to deliver the solution and transition smoothly into the customer’s technical teams.</li>\n<li>Implement changes on behalf of the customer in the Cloudflare environment following the customer’s change management process.</li>\n<li>Proven experience with Cloudflare or similar with Workers, Javascript/Typescript and Workers APIs.</li>\n<li>Troubleshoot implementation issues and collaborate with Customer Support, Engineering and other teams to assist technical escalations.</li>\n<li>Contribute towards the success of the organization through knowledge sharing activities such as contributing to internal and external documentation, answering technical Q&amp;A, and helping to iterate on best practices.</li>\n</ul>\n<p>Support building operational assets like templates, automation scripts, procedures, workflows, etc.</p>\n<p>Experience might include a combination of the skills below:</p>\n<ul>\n<li>3+ years of experience in a customer facing position as a Consultant delivering services.</li>\n<li>Demonstrated experience with:</li>\n</ul>\n<p>Developing serverless code in a CI/CD pipeline using an Agile methodology. Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP Scripting languages A scripting language (e.g. Python, JavaScript, Bash) and a desire to expand those skills. Infrastructure as code tools like Terraform. Strong experience with APIs. CI/CD pipelines using Azure DevOps or Git. Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc. Good understanding and knowledge of:</p>\n<p>Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs. Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP. Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3.</p>\n<p>Strong advantage if:</p>\n<p>You have worked with a Cybersecurity company or products and have performed migrations using migration tools. You have developed application security and performance capabilities. Ability to manage a project, work to deadlines, prioritize between competing demands and manage uncertainty.</p>\n<p>The work will be performed in English. Fluency in a second regional European language is a strong advantage.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bec4e006-74f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7383013","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Developing serverless code in a CI/CD pipeline using an Agile methodology","Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP","Scripting languages","Infrastructure as code tools like Terraform","Strong experience with APIs","CI/CD pipelines using Azure DevOps or Git","Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc","Good understanding and knowledge of Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs","Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP","Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:53:29.137Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Developing serverless code in a CI/CD pipeline using an Agile methodology, Layers and protocols of the OSI model, such as TCP/IP, TLS, DNS, HTTP, Scripting languages, Infrastructure as code tools like Terraform, Strong experience with APIs, CI/CD pipelines using Azure DevOps or Git, Implementation and troubleshooting experience, knowledge of tools to troubleshoot, observability, logs, etc, Good understanding and knowledge of Internet and Security technologies such as DDoS, Web Application Firewall, Certificates, DNS, CDN, Analytics and Logs, Security aspects of an internet property, such as DNS, WAFs, Bot Management, Rate Limiting, (M)TLS, certificates, OWASP, Performance aspects of an internet property, such as Speed, Latency, Caching, HTTP/3, TLSv1.3"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_62b851a9-660"},"title":"Data Scientist","description":"<p>Job Title: Data Scientist</p>\n<p><strong>About the Position</strong></p>\n<p>As a Data Scientist on the platform prediction team, you will translate our probability of success predictions into measurable portfolio-level outcomes. You will architect core systems that let us rigorously evaluate signals from our AI-driven predictions in public and private equities and our internal portfolio.</p>\n<p>This role sits at the intersection of quantitative finance, healthcare data, and AI-driven drug development. If you&#39;re excited about applying portfolio construction and risk management fundamentals to one of the most consequential prediction problems in healthcare, this is the role.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Work with the team to implement and maintain core portfolio engine: order management system, execution simulation layer, portfolio construction service, and performance tracking</li>\n<li>Design risk frameworks that quantify exposure across a portfolio of drug development bets with radically different risk profiles, timelines, and failure modes</li>\n<li>Run rigorous backtesting experiments with strict temporal constraints to evaluate Formation strategies against baseline approaches and measure marginal signal from new evidence sources</li>\n<li>Coordinate across the organization to integrate internal Formation data sources (clinical trial data, genomic evidence, real-world data) and proprietary tooling into portfolio analytics pipelines</li>\n<li>Work with product and engineering teams to build dashboards and reporting that communicate portfolio performance, risk metrics, and strategy comparisons to both technical and executive stakeholders</li>\n<li>Collaborate with the broader data science team to ensure portfolio-level evaluation feeds back into model improvement and evidence prioritization</li>\n</ul>\n<p><strong>About You</strong></p>\n<p>We are looking for a highly motivated and experienced Data Scientist to join our team. The ideal candidate will have a strong background in data science, machine learning, and software development, with a proven track record of delivering high-quality results in a fast-paced environment.</p>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>MS or PhD in a quantitative field (statistics, finance, physics, computational science, engineering, or related)</li>\n<li>1-3 years in a quantitative research, data science, or analytics role , finance, healthcare, academic research, or consulting all count; substantive internships qualify</li>\n<li>Strong Python programming skills with experience in data-intensive workflows (pandas, numpy, scipy)</li>\n<li>Solid grasp of core portfolio construction and risk concepts: position sizing, rebalancing, Sharpe ratio, drawdown, volatility, benchmark comparison</li>\n<li>Demonstrated ability to work with messy, real-world datasets , comfortable with data wrangling, deduplication, and quality assessment</li>\n<li>Clear communicator who can present quantitative results to both technical peers and business stakeholders</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Experience with backtesting frameworks or portfolio simulation (vectorbt, Backtrader, or custom implementations)</li>\n<li>Exposure to healthcare, pharma, or biotech data (clinical trials, claims data, -omics, real-world evidence)</li>\n<li>Familiarity with alternative data in a research or investment context</li>\n<li>Experience with probability-of-success modeling, drug development decision analysis, or health economics</li>\n<li>Comfort with LLMs or AI/ML pipelines in a production or research setting</li>\n<li>Familiarity with dashboard/visualization tools (Streamlit, Plotly, Dash) and pipeline orchestration (Dagster, Airflow)</li>\n</ul>\n<p><strong>Total Compensation Range:</strong> $154,500 - $202,000</p>\n<p>**Compensation Individual compensation is determined by several factors, including role scope, geographic location, and skills &amp; experience. Your offer will reflect where you fall within the range based on these considerations. In addition to base salary, we offer equity, comprehensive benefits, and generous perks. If the posted range doesn&#39;t match your expectations, we still encourage you to apply!</p>\n<p>**Where We Hire Formation Bio is prioritizing hiring in key hubs, primarily the New York City and Boston metro areas, with a hybrid model requiring 3 days per week in office. Applicants from the Research Triangle (NC) and San Francisco Bay Area may also be considered. Please apply only if you reside in these locations or are willing to relocate.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_62b851a9-660","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Formation Bio","sameAs":"https://www.formation.bio/","logo":"https://logos.yubhub.co/formation.bio.png"},"x-apply-url":"https://job-boards.greenhouse.io/formationbio/jobs/7757667","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$154,500 - $202,000","x-skills-required":["Python","pandas","numpy","scipy","portfolio construction","risk management","backtesting","data wrangling","data visualization"],"x-skills-preferred":["backtesting frameworks","portfolio simulation","healthcare data","alternative data","probability-of-success modeling","drug development decision analysis","health economics","LLMs","AI/ML pipelines","dashboard/visualization tools","pipeline orchestration"],"datePosted":"2026-04-18T15:53:29.085Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY; Boston, MA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Healthcare","skills":"Python, pandas, numpy, scipy, portfolio construction, risk management, backtesting, data wrangling, data visualization, backtesting frameworks, portfolio simulation, healthcare data, alternative data, probability-of-success modeling, drug development decision analysis, health economics, LLMs, AI/ML pipelines, dashboard/visualization tools, pipeline orchestration","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":154500,"maxValue":202000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3e12d6b2-155"},"title":"Capital Solutions Manager (Data OS, Insight OS)","description":"<p>Behind many of life&#39;s most important transactions , buying a house, applying for a mortgage, getting a small business loan, or refinancing a credit card , is a network of credit relationships. Setpoint provides critical operational infrastructure for relationships between the world&#39;s largest banks, credit funds and capital markets counterparties. We&#39;re building trust in this system of credit.</p>\n<p>We&#39;re looking for a Capital Solutions Manager to join our team and serve as the bridge between our clients and our engineering organisation. You&#39;ll take ownership of live client portfolios across Data OS and Insight OS, our data management and analytics platforms.</p>\n<p>This isn&#39;t a back-office analytics role. You&#39;ll be client-facing from day one, owning deal relationships, translating complex structured finance requirements into engineering specs, and ensuring that every dashboard, data pipeline, and export meets institutional-grade standards.</p>\n<p>This is an opportunity to be an early owner of a fast-growing product lines (Data OS and Insight OS) at a fast-growing platform. You&#39;ll be a co-owner of creating tech solutions for lenders and borrowers in asset backed credit. We have a strong ethos of promoting from within, and you&#39;ll be given ample opportunities for career development and advancement.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Own client portfolios end-to-end. Serve as the primary point of contact for assigned clients across Data OS and Insight OS , managing onboarding, success, and supporting growth.</li>\n</ul>\n<ul>\n<li>Translate our customer&#39;s structured finance needs for Engineering. Act as the bridge between our clients and our product/engineering organisation for Data OS and Insight OS. Define what needs to be built, flag what may be custom work, write the specs, review the output, and validate that dashboards and data pipelines match analytical intent.</li>\n</ul>\n<ul>\n<li>Own the accuracy of Setpoint&#39;s data layer across your assigned portfolio. Lead data quality assessments on incoming loan tapes and client deliverables , identifying anomalies, missing fields, and population gaps before they reach production.</li>\n</ul>\n<ul>\n<li>Supervise offshore implementation resources. Directly manage a team of offshore analysts supporting data ingestion, validation, and reporting workflows. Set priorities, review work product, and ensure delivery standards are met.</li>\n</ul>\n<ul>\n<li>Leverage AI-powered workflows and internal tooling. Use and help refine our internal AI-assisted deal workflows , from automated data quality checks to metric design and schema mapping , to accelerate delivery and improve consistency across client portfolios.</li>\n</ul>\n<ul>\n<li>Make us better. Contribute to product priorities, onboarding playbooks, sector templates, and process documentation that make our delivery engine repeatable as the portfolio grows.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>3–6 years in private credit, asset-backed lending, or structured finance. You&#39;ve worked with loan tapes, servicer reports, borrowing base certificates, or compliance packages , not just in theory, but hands-on. Experience across multiple asset classes (consumer, auto, fund finance, CRE) is a strong plus.</li>\n</ul>\n<ul>\n<li>Demonstrated client management experience. You&#39;ve owned client relationships , running calls, managing expectations, resolving issues , in a professional services, advisory, or platform context. You&#39;re comfortable being the face of the company to institutional investors and lenders.</li>\n</ul>\n<ul>\n<li>Strong analytical and data skills. Expert-level Excel is baseline. Comfort with SQL, Python, data pipelines, or business intelligence tools (Metabase, Tableau, etc.) is highly valued. You don&#39;t need to write production code, but you should be able to read a data schema, trace a metric back to its source field, and spot when something doesn&#39;t add up.</li>\n</ul>\n<ul>\n<li>Experience writing technical specifications or engineering handoff documents. You&#39;ve translated business requirements into structured artifacts , field mappings, data dictionaries, logic definitions, or acceptance criteria , that a technical team can execute against.</li>\n</ul>\n<ul>\n<li>Comfort with AI/LLM tooling and automation. You don&#39;t need to be an AI engineer, but you should be excited about using AI-assisted workflows to accelerate data analysis, quality checks, and specification writing. Familiarity with prompt engineering or AI copilot tools is a plus.</li>\n</ul>\n<ul>\n<li>Team supervision experience. You&#39;ve managed or coordinated the work of junior analysts, offshore teams, or cross-functional workstreams. You can set priorities, review deliverables, and maintain quality without micromanaging.</li>\n</ul>\n<ul>\n<li>Ability to operate independently in ambiguous environments. You can take a vague client request, figure out what&#39;s actually needed, scope the work, and deliver , without someone laying out every step. You handle tight timelines and competing priorities without losing quality.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<p>We offer a comprehensive benefits package that includes competitive salaries, stock options, medical, dental, and vision coverage, 401(k), short term and long term disability coverage, and flexible vacation. We have offices in Austin, TX, New York City, NY, and Salt Lake City, UT with hybrid roles based in these locations and an expectation of two days a week in office (Tuesdays and Thursdays).</p>\n<p><strong>Compensation</strong></p>\n<p>$140,000 - $160,000 dependent on multiple factors, which may include the successful candidate&#39;s skills, experience and other qualifications.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3e12d6b2-155","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Setpoint","sameAs":"https://setpoint.com/","logo":"https://logos.yubhub.co/setpoint.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/setpoint/jobs/5106278007","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$140,000 - $160,000","x-skills-required":["structured finance","data management","analytics","client relationship management","technical specifications","engineering handoff documents","AI/LLM tooling","automation","team supervision","independent problem-solving"],"x-skills-preferred":["SQL","Python","data pipelines","business intelligence tools","prompt engineering","AI copilot tools"],"datePosted":"2026-04-18T15:53:12.581Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Austin or New York (Hybrid)"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"structured finance, data management, analytics, client relationship management, technical specifications, engineering handoff documents, AI/LLM tooling, automation, team supervision, independent problem-solving, SQL, Python, data pipelines, business intelligence tools, prompt engineering, AI copilot tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":140000,"maxValue":160000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2a04373f-0ca"},"title":"Engineering Manager (Integrations)","description":"<p>About Dialpad</p>\n<p>Dialpad is the AI-native business communications platform. We unify calling, messaging, meetings, and contact center on a single platform - powered by AI that understands every conversation in real time.</p>\n<p>More than 70,000 companies around the globe, including WeWork, Asana, NASDAQ, AAA Insurance, COMPASS Realty, Uber, Randstad, and Tractor Supply, rely on Dialpad to build stronger customer connections using real-time, AI-driven insights.</p>\n<p>We’re now leading the shift to Agentic AI: intelligent agents that don’t just analyse conversations but take action by automating workflows, resolving customer issues, and accelerating revenue in real time.</p>\n<p>Our DAART initiative (Dialpad Agentic AI in Real Time) is redefining what a communications platform can do. Visit dialpad.com to learn more.</p>\n<p>Being a Dialer</p>\n<p>At Dialpad, AI isn’t just a feature; it’s how our teams do their best work every day. We put powerful AI tools in every employee’s hands so they can move faster, think bigger, and achieve more.</p>\n<p>We believe every conversation matters. And we’ve built the platform that turns those conversations into insight and action, for our customers and ourselves.</p>\n<p>We look for people who are intensely curious and hold themselves to a high bar. Our ambition is significant, and achieving it requires a team that operates at the highest level.</p>\n<p>We seek individuals who embody our core traits: Scrappy, Curious, Optimistic, Persistent, and Empathetic.</p>\n<p>About the team</p>\n<p>Dialpad’s Salesforce Integrations team plays an essential role in developing a robust layer of integrations that seamlessly connect Dialpad&#39;s products with external services, in particular Salesforce.</p>\n<p>Our teams are highly collaborative and comprise cross-disciplinary professionals, including Product Managers, Designers, QA specialists, as well as Engineers specialising in Full-Stack Engineering, Data Engineering, Data Science, and Telephony.</p>\n<p>Additionally, the integrations team collaborates closely with Dialpad’s Agentic AI organisation to help expand the ecosystem powering Dialpad’s AI agents.</p>\n<p>Your role</p>\n<p>As an Engineering Manager of the Salesforce Integrations team, you will lead a team of 6+ mid-senior full-stack engineers based in London &amp; India.</p>\n<p>Your role will involve closely collaborating with other engineering managers &amp; teams also focusing on integrations to achieve alignment &amp; efficiency whilst delivering multiple simultaneous projects with cross-functional stakeholders.</p>\n<p>Although this is primarily a leadership position, given the current team size technical IC tasks will also be performed including system design, architecture &amp; code reviews &amp; AI-driven development.</p>\n<p>This team is expected to grow in both size, scope &amp; impact with strong potential for additional career opportunities &amp; responsibilities.</p>\n<p>This position reports to our Director of Engineering who is based in Canada.</p>\n<p>Candidates are expected to be flexible with their working hours, ensuring overlap with IST and PST timezones for team meetings, discussions &amp; escalations.</p>\n<p>What you’ll do</p>\n<ul>\n<li>You’ll have direct reports ranging from mid-level to highly experienced engineers, and will support their performance and career growth through regular one-on-ones, performance reviews, coaching, and mentoring.</li>\n</ul>\n<ul>\n<li>Help to define a 1-3 year roadmap &amp; vision for the Salesforce integrations team</li>\n</ul>\n<ul>\n<li>Consistently work with your direct reports to support their career growth</li>\n</ul>\n<ul>\n<li>Assist in evaluating technical design &amp; architecture documents and proposals on an ongoing basis, in anticipation of increased scale and ever-evolving technology to meet the demands of rapidly growing business needs</li>\n</ul>\n<ul>\n<li>Work with geographically distributed peers including engineering managers, technical leaders, product managers, designers, support engineers and other stakeholders in order to align on engineering-wide priorities</li>\n</ul>\n<ul>\n<li>Own large projects end-to-end including requirements gathering, planning, resource allocation, and sometimes execution</li>\n</ul>\n<ul>\n<li>Drive effective engineering processes and policies</li>\n</ul>\n<ul>\n<li>Scale the team by recruiting candidates from diverse backgrounds</li>\n</ul>\n<ul>\n<li>Get hands-on when necessary and assist with technical implementations</li>\n</ul>\n<ul>\n<li>Assist with emerging Agentic AI technologies &amp; initiatives</li>\n</ul>\n<ul>\n<li>Use AI coding tools for active development, reviews, testing, etc.</li>\n</ul>\n<p>What we’re looking for</p>\n<ul>\n<li>3+ years experience leading a high-performing team of engineers, including managing and shipping cross-functional or multi-team projects</li>\n</ul>\n<ul>\n<li>10+ years professional experience as an engineer or engineering leader</li>\n</ul>\n<ul>\n<li>Experience with the development of integrations &amp; APIs</li>\n</ul>\n<ul>\n<li>Hiring &amp; interviewing skills</li>\n</ul>\n<ul>\n<li>Onboarding &amp; mentorship experience</li>\n</ul>\n<ul>\n<li>Agentic AI experience</li>\n</ul>\n<ul>\n<li>Proficiency with ETL data pipelines</li>\n</ul>\n<p>Why Join Dialpad</p>\n<ul>\n<li>Work at the center of the AI transformation in business communications</li>\n</ul>\n<ul>\n<li>Build and ship agentic AI products that are redefining how companies operate</li>\n</ul>\n<ul>\n<li>Join a team where AI amplifies every employee’s impact</li>\n</ul>\n<ul>\n<li>Competitive salary, comprehensive benefits, and real opportunities for growth</li>\n</ul>\n<p>We believe in investing in our people. Dialpad offers competitive benefits and perks, cutting-edge AI tools, and a robust training program that help you reach your full potential.</p>\n<p>We have designed our offices to be inclusive, offering a vibrant environment to cultivate collaboration and connection.</p>\n<p>Our exceptional culture, repeatedly recognised as a Great Place to Work, ensures that every employee feels valued and empowered to contribute to our collective success.</p>\n<p>Don’t meet every single requirement? If you’re excited about this role and possess the fundamental traits, drive, and strong ambition we seek, but your experience doesn’t meet every qualification, we encourage you to apply.</p>\n<p>Dialpad is an equal-opportunity employer. We are dedicated to creating a community of inclusion and an environment free from discrimination or harassment.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2a04373f-0ca","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dialpad","sameAs":"https://dialpad.com","logo":"https://logos.yubhub.co/dialpad.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dialpad/jobs/8421276002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Agentic AI","APIs","ETL data pipelines","Full-stack engineering","Leadership"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:53:10.903Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Agentic AI, APIs, ETL data pipelines, Full-stack engineering, Leadership"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a966b1bf-e76"},"title":"Staff Software Engineer, Compute Infrastructure","description":"<p>As a Staff Software Engineer, you will shape the backbone of our GPU-driven data centers,powering some of the most advanced workloads in AI and large-scale computing. This isn&#39;t just about keeping the lights on; it&#39;s about architecting the next generation of reliable, secure, and massively scalable infrastructure.</p>\n<p>The METALDEV team builds and operates a suite of Go-based services that power large-scale datacenter deployments. These platforms automate complex workflows while providing deep observability and monitoring for tens of thousands of GPU servers and diverse infrastructure components,including CDUs, PDUs, and NVLink switches. Our tooling is designed for next-generation rack systems like NVIDIA GB200 and GB300, as well as a broad range of GPU server platforms.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Providing technical leadership in designing, architecting, and operating large-scale infrastructure services for GPU servers, with a focus on security, reliability, and scalability.</li>\n<li>Building and enhancing infrastructure services and automation, including inventory management systems and lifecycle management solutions using open source technologies.</li>\n<li>Driving strategic direction for infrastructure automation, lifecycle management, and service orchestration, making MetalDev core services more scalable and resilient.</li>\n<li>Defining best practices for API development (REST/gRPC), distributed databases, and Kubernetes orchestration,while mentoring engineers to follow your lead.</li>\n<li>Partnering with hardware, software, and operations teams to align infrastructure with business impact.</li>\n<li>Contributing to open source communities (e.g., Go, Redfish) through collaboration and technical thought leadership.</li>\n<li>Leading and improving CI/CD pipelines for hardware compliance, firmware management, and data systems.</li>\n<li>Championing reliability and operational excellence by driving observability (Prometheus/Grafana), production incident response, and continuous service improvement.</li>\n</ul>\n<p>We&#39;re looking for someone with a strong background in software engineering, particularly in infrastructure, cloud engineering, and distributed databases. You should have experience with Go and a proven track record of building REST/gRPC APIs for mission-critical platforms. Additionally, you should be familiar with architecting and scaling cloud-native Kubernetes infrastructure and distributed services.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a966b1bf-e76","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4603505006","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$188,000 to $275,000","x-skills-required":["Go","REST/gRPC","Distributed databases","Kubernetes orchestration","API development","Infrastructure services","Automation","Inventory management","Lifecycle management","CI/CD pipelines","Hardware compliance","Firmware management","Data systems","Observability","Production incident response","Continuous service improvement"],"x-skills-preferred":["Kafka","ClickHouse","CRDB","DMTF","RedFish APIs","GPU servers"],"datePosted":"2026-04-18T15:53:06.173Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Manhattan, NY / Sunnyvale, CA / Bellevue, WA / Livingston, NJ"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, REST/gRPC, Distributed databases, Kubernetes orchestration, API development, Infrastructure services, Automation, Inventory management, Lifecycle management, CI/CD pipelines, Hardware compliance, Firmware management, Data systems, Observability, Production incident response, Continuous service improvement, Kafka, ClickHouse, CRDB, DMTF, RedFish APIs, GPU servers","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":188000,"maxValue":275000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d70a8194-b84"},"title":"Software Engineer, Machine Learning","description":"<p>We are seeking a versatile and experienced Machine Learning / AI Engineer to join our growing AI team, working at the intersection of applied machine learning, infrastructure, and product innovation. Your work will drive user productivity, shape new product experiences, and advance the state of AI at Figma.</p>\n<p>As a Machine Learning / AI Engineer, you will design, build, and productionize ML models for Search, Discovery, Ranking, Retrieval-Augmented Generation (RAG), and generative AI features. You will also build and maintain scalable data pipelines to collect high-quality training and evaluation datasets, including annotation systems and human-in-the-loop workflows.</p>\n<p>You will collaborate closely with engineers, researchers, designers, and product managers across multiple teams to deliver high-quality ML-driven features and infrastructure. This is a high-impact, cross-functional role where you will shape both foundational systems and user-facing capabilities.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Design, build, and productionize ML models for Search, Discovery, Ranking, Retrieval-Augmented Generation (RAG), and generative AI features.</li>\n<li>Build and maintain scalable data pipelines to collect high-quality training and evaluation datasets, including annotation systems and human-in-the-loop workflows.</li>\n<li>Collaborate with AI researchers to iterate on datasets, evaluation metrics, and model architectures to improve quality and relevance.</li>\n<li>Work with product engineers to define and deliver impactful AI features across Figma&#39;s platform.</li>\n<li>Partner with infrastructure engineers to develop and optimize systems for training, inference, monitoring, and deployment.</li>\n<li>Explore new ideas at the edge of what&#39;s technically possible and help shape the long-term AI vision at Figma.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>5+ years of industry experience in software engineering, with 3+ years focused on applied machine learning or AI.</li>\n<li>Strong experience with end-to-end ML model development, including training, evaluation, deployment, and monitoring.</li>\n<li>Proficiency in Python and familiarity with ML libraries like PyTorch, TensorFlow, Scikit-learn, Spark MLlib, or XGBoost.</li>\n<li>Experience designing and building scalable data and annotation pipelines, as well as evaluation systems for AI model quality.</li>\n<li>Experience mentoring or leading others and contributing to a culture of technical excellence and innovation.</li>\n</ul>\n<p>Preferred qualifications include:</p>\n<ul>\n<li>Familiarity with search relevance, ranking, NLP, or RAG systems.</li>\n<li>Experience with AI infrastructure and MLOps, including observability, CI/CD, and automation for ML workflows.</li>\n<li>Experience working on creative or design-focused ML applications.</li>\n<li>Knowledge of additional languages such as C++ or Go is a plus, but not required.</li>\n<li>A product mindset with the ability to tie technical work to user outcomes and business impact.</li>\n<li>Strong collaboration and communication skills, especially when working across functions (engineering, product, research).</li>\n</ul>\n<p>At Figma, one of our values is Grow as you go. We believe in hiring smart, curious people who are excited to learn and develop their skills. If you&#39;re excited about this role but your past experience doesn&#39;t align perfectly with the points outlined in the job description, we encourage you to apply anyways. You may be just the right candidate for this or other roles.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d70a8194-b84","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Figma","sameAs":"https://www.figma.com/","logo":"https://logos.yubhub.co/figma.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/figma/jobs/5551532004","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$153,000-$376,000 USD","x-skills-required":["Machine Learning","AI","Python","PyTorch","TensorFlow","Scikit-learn","Spark MLlib","XGBoost","Data Pipelines","Annotation Systems","Human-in-the-loop Workflows"],"x-skills-preferred":["Search Relevance","Ranking","NLP","RAG Systems","AI Infrastructure","MLOps","Observability","CI/CD","Automation","Creative or Design-Focused ML Applications"],"datePosted":"2026-04-18T15:53:04.257Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA • New York, NY • United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, AI, Python, PyTorch, TensorFlow, Scikit-learn, Spark MLlib, XGBoost, Data Pipelines, Annotation Systems, Human-in-the-loop Workflows, Search Relevance, Ranking, NLP, RAG Systems, AI Infrastructure, MLOps, Observability, CI/CD, Automation, Creative or Design-Focused ML Applications","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":153000,"maxValue":376000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_74be15a1-bce"},"title":"Software Engineer, Inference Deployment","description":"<p>Our mandate is to make inference deployment boring and unattended. We serve Claude to millions of users across GPUs, TPUs, and Trainium , and every model update must reach production safely, quickly, and without disrupting service. As a Software Engineer on the Launch Engineering team, you&#39;ll design and build the deployment infrastructure that moves inference code from merge to production.</p>\n<p>This is a resource-constrained optimization problem at its core: validation and deployment consume the same accelerator chips that serve customer traffic , your deploys compete with live user requests for the same hardware. Every model brings different fleet sizes, startup times, and correctness requirements, so the system must adapt continuously. You&#39;ll build systems that navigate these constraints , orchestrating validation, scheduling deployments intelligently, and driving down cycle time from merge to production.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Own deployment orchestration that continuously moves validated inference builds into production across GPU, TPU, and Trainium fleets, unattended under normal conditions</li>\n</ul>\n<ul>\n<li>Improve capacity-aware deployment scheduling to maximize deployment throughput against constrained accelerator budgets and variable fleet sizes</li>\n</ul>\n<ul>\n<li>Extend deployment observability , dashboards and tooling that answer &quot;what code is running in production,&quot; &quot;where is my commit,&quot; and &quot;what validation passed for this deploy&quot;</li>\n</ul>\n<ul>\n<li>Drive down cycle time from code merge to production with pipeline architectures that minimize serial dependencies and maximize parallelism</li>\n</ul>\n<ul>\n<li>Optimize fleet rollout strategies for large-scale deployments across thousands of GPU, TPU, and Trainium chips, minimizing disruption to serving capacity</li>\n</ul>\n<ul>\n<li>Evolve self-service model onboarding so that new models can be added to the continuous deployment pipeline without Launch Engineering involvement</li>\n</ul>\n<ul>\n<li>Partner across the Inference organization with teams owning validation, autoscaling, and model routing to integrate deployment automation with their systems</li>\n</ul>\n<p>You May Be a Good Fit If You Have:</p>\n<ul>\n<li>5+ years of experience building deployment, release, or delivery infrastructure at scale</li>\n</ul>\n<ul>\n<li>Strong software engineering skills with experience designing systems that manage complex state machines and multi-stage pipelines</li>\n</ul>\n<ul>\n<li>Experience with deployment systems where resource constraints shape the design , whether that&#39;s fleet capacity, network bandwidth, hardware availability, or coordinated rollout windows</li>\n</ul>\n<ul>\n<li>A track record of building automation that measurably improves deployment velocity and reliability</li>\n</ul>\n<ul>\n<li>Proficiency with Kubernetes-based deployments, rolling update mechanics, and container orchestration</li>\n</ul>\n<ul>\n<li>Comfort working across the stack , from backend services and databases to CLI tools and web UIs</li>\n</ul>\n<ul>\n<li>Strong communication skills and the ability to work closely with oncall engineers, model teams, and infrastructure partners</li>\n</ul>\n<p>Strong Candidates May Also Have:</p>\n<ul>\n<li>Experience with ML inference or training infrastructure deployment, particularly across multiple accelerator types (GPU, TPU, Trainium)</li>\n</ul>\n<ul>\n<li>Background in capacity planning or resource-constrained scheduling (e.g., bin-packing, fleet management, job scheduling with hardware affinity)</li>\n</ul>\n<ul>\n<li>Experience with progressive delivery in systems with long validation cycles: canary/soak testing, blue-green deployments, traffic shifting, automated rollback</li>\n</ul>\n<ul>\n<li>Experience at companies with large-scale release engineering challenges (mobile release trains, monorepo deployments, multi-datacenter rollouts)</li>\n</ul>\n<ul>\n<li>Experience with Python and/or Rust in production systems</li>\n</ul>\n<p>The annual compensation range for this role is $320,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_74be15a1-bce","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5111745008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$485,000 USD","x-skills-required":["deployment infrastructure","software engineering","complex state machines","multi-stage pipelines","Kubernetes-based deployments","container orchestration","backend services","databases","CLI tools","web UIs"],"x-skills-preferred":["ML inference","training infrastructure deployment","capacity planning","resource-constrained scheduling"," deployments","progressive delivery","Python","Rust"],"datePosted":"2026-04-18T15:53:04.252Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"deployment infrastructure, software engineering, complex state machines, multi-stage pipelines, Kubernetes-based deployments, container orchestration, backend services, databases, CLI tools, web UIs, ML inference, training infrastructure deployment, capacity planning, resource-constrained scheduling,  deployments, progressive delivery, Python, Rust","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_065005a6-23e"},"title":"OIC Developer","description":"<p>We are looking for an expert Oracle Integration Developer to join our Arsenal (Enterprise Systems) team. Your immediate mission: take ownership of our critical enterprise integrations connecting Oracle Fusion ERP with our upstream and downstream systems. These integrations, built on Oracle Integration Cloud, form the digital backbone that governs how we manage our business operations, from product data and procurement to manufacturing and financial processes.</p>\n<p>As the world enters an era of strategic competition, Anduril is committed to bringing cutting-edge autonomy, AI, computer vision, sensor fusion, and networking technology to the military in months, not years.</p>\n<p>Long-term, you will be the subject matter expert responsible for architecting and scaling our enterprise integration landscape. This is a high-impact role for someone who thrives on solving complex data challenges and wants to build the operational foundation that enables Anduril to scale its mission.</p>\n<p><strong>Key Responsibilities:</strong></p>\n<ul>\n<li>Stabilize &amp; Optimize: Dive deep into existing Oracle Fusion ERP integrations across manufacturing, supply chain, finance, and engineering systems. Diagnose root causes of instability, re-architect weak points, and implement robust error handling and monitoring to achieve mission-critical reliability.</li>\n</ul>\n<ul>\n<li>Architect &amp; Build: Design and develop new, scalable enterprise integrations using Oracle Integration Cloud (OIC). Translate complex business requirements for product data, multi-level Bills of Material (BOMs), procurement, inventory, work orders, and financial transactions into resilient and efficient integration flows.</li>\n</ul>\n<ul>\n<li>Own the Integration Lifecycle: Manage the end-to-end process from design and development through testing (unit, SIT, UAT) and deployment, utilizing CI/CD best practices. Proactively tune and maintain integrations to ensure peak performance as data volumes grow.</li>\n</ul>\n<ul>\n<li>Ensure Data Integrity: Become the trusted expert on data transformation and mapping between systems. Implement rigorous validation and reconciliation logic to guarantee that our enterprise data is flawless across all systems.</li>\n</ul>\n<ul>\n<li>Collaborate &amp; Influence: Act as the key technical partner to our ERP, Manufacturing, Supply Chain, and Finance teams. Clearly articulate technical designs, trade-offs, and progress to both engineering peers and business stakeholders, guiding them toward best-practice integration patterns.</li>\n</ul>\n<ul>\n<li>Leverage Modern Oracle Cloud Tools: Utilize Oracle Visual Builder Cloud Service (VBCS) where appropriate to build lightweight user interfaces that enhance integration workflows, data validation, or operational dashboards.</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>5+ years of hands-on experience developing complex integrations with deep specialization in Oracle Integration Cloud (OIC), specifically Oracle Integration 3.</li>\n</ul>\n<ul>\n<li>Proven experience integrating Oracle Fusion Cloud ERP with upstream and downstream enterprise systems (e.g., PLM, MES, WMS, CRM, third-party applications), including deep familiarity with ERP data objects such as Items, BOMs, Suppliers, Purchase Orders, Work Orders, Inventory Transactions, and Financial data.</li>\n</ul>\n<ul>\n<li>Expert-level proficiency in OIC 3 components: Application and Tech Adapters (REST, SOAP, File, FTP, Oracle SaaS, Database), Connections, Mappings, Lookups, Error Handling, and JavaScript.</li>\n</ul>\n<ul>\n<li>Strong command of XSLT, XPath, and complex data mapping for transforming large and nested XML/JSON payloads.</li>\n</ul>\n<ul>\n<li>Demonstrable experience building, securing, and consuming RESTful APIs and SOAP web services.</li>\n</ul>\n<ul>\n<li>Excellent SQL skills and a solid understanding of relational database concepts.</li>\n</ul>\n<ul>\n<li>Experience with Oracle Fusion ERP modules such as SCM (Supply Chain Management), Manufacturing, Procurement, or Financials.</li>\n</ul>\n<ul>\n<li>A tenacious problem-solver with a track record of troubleshooting, debugging, and stabilizing complex, business-critical systems.</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>Hands-on experience with Oracle Visual Builder Cloud Service (VBCS) for building user interfaces and extensions.</li>\n</ul>\n<ul>\n<li>Experience with Oracle Business Intelligence Cloud Connector (BICC) for high-volume data extraction from Fusion ERP.</li>\n</ul>\n<ul>\n<li>Experience with Oracle Cloud Infrastructure (OCI) services (e.g., Functions, API Gateway, Object Storage, Logging, Autonomous Database).</li>\n</ul>\n<ul>\n<li>Experience integrating PLM systems (e.g., Teamcenter, Windchill, Arena) with Oracle Fusion ERP.</li>\n</ul>\n<ul>\n<li>Familiarity with Git-based source control and CI/CD pipelines for integration deployments.</li>\n</ul>\n<ul>\n<li>Experience in a discrete manufacturing environment.</li>\n</ul>\n<ul>\n<li>Knowledge of other programming languages (e.g., Python, Groovy, Java).</li>\n</ul>\n<ul>\n<li>Relevant Oracle Cloud Certifications (e.g., OIC 3 Application Integration Professional, Oracle Fusion Cloud certifications).</li>\n</ul>\n<p><strong>Salary Range:</strong> $129,000-$171,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_065005a6-23e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5061434007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$129,000-$171,000 USD","x-skills-required":["Oracle Integration Cloud","Oracle Fusion ERP","APIs","SQL","XSLT","XPath","RESTful APIs","SOAP web services","JavaScript","CI/CD pipelines","Git-based source control"],"x-skills-preferred":["Oracle Visual Builder Cloud Service","Oracle Business Intelligence Cloud Connector","Oracle Cloud Infrastructure","PLM systems","Python","Groovy","Java","Oracle Cloud Certifications"],"datePosted":"2026-04-18T15:53:00.850Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Boston, Massachusetts, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Oracle Integration Cloud, Oracle Fusion ERP, APIs, SQL, XSLT, XPath, RESTful APIs, SOAP web services, JavaScript, CI/CD pipelines, Git-based source control, Oracle Visual Builder Cloud Service, Oracle Business Intelligence Cloud Connector, Oracle Cloud Infrastructure, PLM systems, Python, Groovy, Java, Oracle Cloud Certifications","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":129000,"maxValue":171000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9d52173c-23a"},"title":"Sr. Software Development Engineer in Test","description":"<p>About Dialpad</p>\n<p>Dialpad is the AI-native business communications platform. We unify calling, messaging, meetings, and contact center on a single platform - powered by AI that understands every conversation in real time.</p>\n<p>More than 70,000 companies around the globe, including WeWork, Asana, NASDAQ, AAA Insurance, COMPASS Realty, Uber, Randstad, and Tractor Supply, rely on Dialpad to build stronger customer connections using real-time, AI-driven insights.</p>\n<p>We’re now leading the shift to Agentic AI: intelligent agents that don’t just analyse conversations but take action by automating workflows, resolving customer issues, and accelerating revenue in real time.</p>\n<p>Our DAART initiative (Dialpad Agentic AI in Real Time) is redefining what a communications platform can do. Visit dialpad.com to learn more.</p>\n<p>Being a Dialer</p>\n<p>At Dialpad, AI isn’t just a feature; it’s how our teams do their best work every day. We put powerful AI tools in every employee’s hands so they can move faster, think bigger, and achieve more.</p>\n<p>We believe every conversation matters. And we’ve built the platform that turns those conversations into insight and action, for our customers and ourselves.</p>\n<p>We look for people who are intensely curious and hold themselves to a high bar. Our ambition is significant, and achieving it requires a team that operates at the highest level.</p>\n<p>We seek individuals who embody our core traits: Scrappy, Curious, Optimistic, Persistent, and Empathetic.</p>\n<p>The Role</p>\n<p>As a Senior SDET, you are a strong software engineer with deep expertise in test automation and quality engineering. You design and own scalable test frameworks, drive quality strategy for your domain, and proactively identify risk in complex, cloud-native systems.</p>\n<p>This position requires a hybrid work arrangement, with three days in the office. The role reports to the Manager of Quality Assurance, who is located in Bangalore.</p>\n<p>This role is hands-on, technical, and impact-driven, with clear expectations around ownership, influence, and mentorship.</p>\n<p>What You’ll Do</p>\n<ul>\n<li>Design, develop, and maintain scalable automated testing frameworks for APIs, microservices, and web integrations.</li>\n</ul>\n<ul>\n<li>Perform deep-dive testing of the Connect platform, with a strong focus on asynchronous workflows, data consistency, resiliency, and latency.</li>\n</ul>\n<ul>\n<li>Own and evolve quality gates within the CI/CD pipeline, ensuring fast, actionable feedback on every pull request.</li>\n</ul>\n<ul>\n<li>Plan and execute functional, regression, and end-to-end test coverage across UI, API, and database layers.</li>\n</ul>\n<ul>\n<li>Build internal tools and utilities to help reproduce, debug, and isolate complex production issues.</li>\n</ul>\n<ul>\n<li>Set up, execute, and continuously improve automated test suites; derive meaningful quality KPIs and clearly communicate results.</li>\n</ul>\n<ul>\n<li>Provide detailed failure analysis to enable rapid diagnosis and resolution of product or test defects.</li>\n</ul>\n<ul>\n<li>Design and execute load, stress, and performance tests across services and critical user workflows.</li>\n</ul>\n<ul>\n<li>Participate actively in architecture and design reviews, advocating for testability and appropriate test hooks.</li>\n</ul>\n<ul>\n<li>Define and execute comprehensive test strategies aligned with product and platform goals.</li>\n</ul>\n<ul>\n<li>Write clean, reliable, and maintainable code, solving complex problems with scalable solutions.</li>\n</ul>\n<ul>\n<li>Mentor junior and mid-level engineers, while staying current on modern testing and software development best practices.</li>\n</ul>\n<p>What You’ll Bring</p>\n<ul>\n<li>6+ years of professional software development experience, with strong emphasis on test automation for large-scale systems.</li>\n</ul>\n<ul>\n<li>Strong coding skills in Python, Java, or JavaScript.</li>\n</ul>\n<ul>\n<li>Proven experience designing test frameworks, test strategies, and reviewing system designs.</li>\n</ul>\n<ul>\n<li>Solid understanding of testing methodologies: regression, integration, end-to-end, load, and performance testing.</li>\n</ul>\n<ul>\n<li>Hands-on experience with API and integration testing; strong knowledge of RESTful services.</li>\n</ul>\n<ul>\n<li>Experience working in cloud-native, distributed environments.</li>\n</ul>\n<ul>\n<li>Experience building and maintaining CI/CD pipelines using Jenkins and GitHub.</li>\n</ul>\n<ul>\n<li>Strong written and verbal communication skills; comfortable collaborating across teams and geographies.</li>\n</ul>\n<ul>\n<li>Demonstrated ownership, proactive problem-solving, and ability to operate independently.</li>\n</ul>\n<p>Technologies Utilised</p>\n<ul>\n<li>GCP: App Engine (GAE), Kubernetes (GKE), Compute Engine (GCE)</li>\n</ul>\n<ul>\n<li>Languages &amp; Tools: Python, Vue, Git, GitHub, Jira</li>\n</ul>\n<p>Why Join Dialpad</p>\n<ul>\n<li>Work at the center of the AI transformation in business communications</li>\n</ul>\n<ul>\n<li>Build and ship agentic AI products that are redefining how companies operate</li>\n</ul>\n<ul>\n<li>Join a team where AI amplifies every employee’s impact</li>\n</ul>\n<ul>\n<li>Competitive salary, comprehensive benefits, and real opportunities for growth</li>\n</ul>\n<p>We believe in investing in our people. Dialpad offers competitive benefits and perks, cutting-edge AI tools, and a robust training program that help you reach your full potential.</p>\n<p>We have designed our offices to be inclusive, offering a vibrant environment to cultivate collaboration and connection.</p>\n<p>Our exceptional culture, repeatedly recognised as a Great Place to Work, ensures that every employee feels valued and empowered to contribute to our collective success.</p>\n<p>Don’t meet every single requirement? If you’re excited about this role and possess the fundamental traits, drive, and strong ambition we seek, but your experience doesn’t meet every qualification, we encourage you to apply.</p>\n<p>Dialpad is an equal-opportunity employer. We are dedicated to creating a community of inclusion and an environment free from discrimination or harassment.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9d52173c-23a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dialpad","sameAs":"https://dialpad.com","logo":"https://logos.yubhub.co/dialpad.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dialpad/jobs/8407069002","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","JavaScript","Test automation","Quality engineering","Cloud-native systems","APIs","Microservices","Web integrations","CI/CD pipelines","Jenkins","GitHub","RESTful services","GCP","App Engine","Kubernetes","Compute Engine"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:48.221Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, JavaScript, Test automation, Quality engineering, Cloud-native systems, APIs, Microservices, Web integrations, CI/CD pipelines, Jenkins, GitHub, RESTful services, GCP, App Engine, Kubernetes, Compute Engine"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1aad838f-387"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability.</p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li>Data Governance &amp; Access Control: Design and implement robust access control systems ensuring only authorized users can access sensitive data.</li>\n<li>Financial Data Infrastructure: Build and maintain data pipelines and warehouses powering business-critical reporting.</li>\n<li>Cloud Storage &amp; Reliability: Architect disaster recovery, backup, and replication systems for petabyte-scale data.</li>\n<li>Data Platform &amp; Tooling: Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark.</li>\n</ul>\n<p>You&#39;ll work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p>To be successful in this role, you&#39;ll need:</p>\n<ul>\n<li>10+ years of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems.</li>\n<li>3+ years of experience leading large scale, complex projects or teams as an engineer or tech lead.</li>\n<li>Deep experience with at least one of:</li>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar.</li>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS).</li>\n<li>Can navigate complex technical tradeoffs between performance, cost, security, and maintainability.</li>\n<li>Have excellent collaboration skills - you work well with both technical and non-technical stakeholders.</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure.</li>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures.</li>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale.</li>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks.</li>\n<li>Experience working in fintech, financial services, or highly regulated environments.</li>\n<li>Security engineering background with focus on data protection and access controls.</li>\n</ul>\n<p>Technologies We Use:</p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran.</li>\n<li>Storage: GCS, S3.</li>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS.</li>\n<li>Languages: Python, Go, SQL.</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1aad838f-387","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures"],"x-skills-preferred":["data warehousing","ETL/ELT pipelines","analytics infrastructure","data reliability","availability","cost efficiency","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"datePosted":"2026-04-18T15:52:47.297Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, data reliability, availability, cost efficiency, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_78ab6fa5-133"},"title":"Staff Security Engineer, Defensive Cyber Engineering","description":"<p>Secure Every Identity, from AI to Human Identity is the key to unlocking the potential of AI. Okta secures AI by building the trusted, neutral infrastructure that enables organisations to safely embrace this new era.</p>\n<p>This work requires a relentless drive to solve complex challenges with real-world stakes. We are looking for builders and owners who operate with speed and urgency and execute with excellence. This is an opportunity to do career-defining work. We&#39;re all in on this mission. If you are too, let&#39;s talk.</p>\n<p>Join Okta’s Defensive Cyber Engineering team as a Staff Engineer responsible for safeguarding Okta’s environments. You’ll work closely with the Security, Business Technology Engineering and Product teams to implement and manage security solutions and ensure that core infrastructure applications are protecting our workforce, endpoints, and corporate data.</p>\n<p>A strong desire to make tools and people work together to solve complex security problems is central to this role. This approach mandates an engineering-first approach: maximising the utility of existing security tools before strategically building or buying new solutions to address any remaining security gaps.</p>\n<p>To execute this vision, you will combine your enterprise security expertise with your hands-on engineering skills, leveraging automation, policy-as-code, and cloud-native technologies to deliver scalable, resilient, and secure solutions. Your work will ultimately set standards for security best practices across the organisation and influence the architecture of business-critical systems.</p>\n<p>What you bring:</p>\n<ul>\n<li>Hands on experience with enterprise security tools such as Okta, Crowdstrike and Palo Alto suite covering EDR (Endpoint Detection and Response), CASB (Cloud Access Security Broker), DLP (Data Loss Prevention), MDM (Mobile Device Management), SASE (Secure Access Service Edge), and SSPM (SaaS Secure Posture Management) capabilities.</li>\n</ul>\n<ul>\n<li>Strong coding and scripting skills are required for building automation and custom tooling. Python experience is preferred, but proficiency in other languages (e.g., Bash, PowerShell, Go) is a plus.</li>\n</ul>\n<ul>\n<li>Proven track record automating security controls and workflows using a cloud-first approach</li>\n</ul>\n<ul>\n<li>Experience with Terraform and other infrastructure-as-code tools to orchestrate security infrastructure</li>\n</ul>\n<ul>\n<li>Familiarity with CI/CD pipelines for security automation and drift management</li>\n</ul>\n<ul>\n<li>Strong communication skills across technical staff, support teams, executive leadership, and external vendors.</li>\n</ul>\n<p>What you’ll be doing:</p>\n<ul>\n<li>Serve as a security subject matter expert (SME) for solution engineering, architecture reviews, security assessment, and vulnerability mitigation</li>\n</ul>\n<ul>\n<li>Lead technical efforts evaluating, designing, and implementing new enterprise security systems and feature enhancements</li>\n</ul>\n<ul>\n<li>Build, maintain, and enhance custom automation and cloud infrastructure using Terraform or similar tools to support team workflows and the enforcement of security controls</li>\n</ul>\n<ul>\n<li>Develop integrations with APIs, cloud platforms (AWS, GCP, Azure), and security infrastructure to improve detection, response, and remediation</li>\n</ul>\n<ul>\n<li>Collaborate with cross-functional teams to tackle global technology and security challenges</li>\n</ul>\n<ul>\n<li>Write and maintain scripts and automation to streamline security operations, with an emphasis on Python-based solutions</li>\n</ul>\n<ul>\n<li>Establish monitoring and alerting for security posture, misconfigurations, and threats across endpoints, SaaS, and cloud workloads</li>\n</ul>\n<ul>\n<li>Proactively identify and remediate security gaps; stay updated on emerging threats, solutions, and tooling across the industry</li>\n</ul>\n<p>And extra credit if you have experience in any of the following!</p>\n<ul>\n<li>Working with advanced identity management technologies (MFA, SAML, OAuth, OIDC, WebAuthn)</li>\n</ul>\n<ul>\n<li>Deep understanding of Okta&#39;s ecosystem, including advanced configuration and integrations</li>\n</ul>\n<ul>\n<li>Experience with continuous compliance solutions (e.g., policy-as-code, automated evidence gathering)</li>\n</ul>\n<p>What you can look forward to as an Full-Time Okta employee!</p>\n<p>World-class benefits, flexibility, and growth opportunities</p>\n<p>The chance to shape the security posture of a global leader in identity</p>\n<p>Opportunities to make a social Impact through technology and innovation</p>\n<p>Ready to join Okta and make security the foundation of our innovation? Apply today!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_78ab6fa5-133","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Okta","sameAs":"https://www.okta.com/","logo":"https://logos.yubhub.co/okta.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/okta/jobs/7476261","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$141,000-$211,000 CAD","x-skills-required":["Enterprise security tools","Okta","Crowdstrike","Palo Alto suite","EDR","CASB","DLP","MDM","SASE","SSPM","Python","Bash","PowerShell","Go","Terraform","Infrastructure-as-code tools","CI/CD pipelines","Security automation","Drift management"],"x-skills-preferred":["Advanced identity management technologies","Okta's ecosystem","Continuous compliance solutions"],"datePosted":"2026-04-18T15:52:38.855Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Toronto, Ontario, Canada; Vancouver, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Enterprise security tools, Okta, Crowdstrike, Palo Alto suite, EDR, CASB, DLP, MDM, SASE, SSPM, Python, Bash, PowerShell, Go, Terraform, Infrastructure-as-code tools, CI/CD pipelines, Security automation, Drift management, Advanced identity management technologies, Okta's ecosystem, Continuous compliance solutions","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":141000,"maxValue":211000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_fdfef6df-396"},"title":"AI Solutions Engineer","description":"<p>About Pinterest:</p>\n<p>Millions of people around the world come to our platform to find creative ideas, dream about new possibilities and plan for memories that will last a lifetime.</p>\n<p>At Pinterest, we’re on a mission to bring everyone the inspiration to create a life they love, and that starts with the people behind the product.</p>\n<p>Discover a career where you ignite innovation for millions, transform passion into growth opportunities, celebrate each other’s unique experiences and embrace the flexibility to do your best work.</p>\n<p>Creating a career you love? It’s Possible.</p>\n<p>At Pinterest, AI isn&#39;t just a feature, it&#39;s a powerful partner that augments our creativity and amplifies our impact, and we’re looking for candidates who are excited to be a part of that.</p>\n<p>To get a complete picture of your experience and abilities, we’ll explore your foundational skills and how you collaborate with AI.</p>\n<p>Through our interview process, what matters most is that you can always explain your approach, showing us not just what you know, but how you think.</p>\n<p>You can read more about our AI interview philosophy and how we use AI in our recruiting process here.</p>\n<p>We&#39;re building a new capability at Pinterest: embedding AI-native engineering directly inside our business functions. The AI Solutions Engineer will partner with teams across Marketing, Finance, Sales, HR, Legal, and other functions to surface high-value automation opportunities, then design and ship the AI-powered tools that bring those opportunities to life.</p>\n<p>This is a hands-on, mid-level software engineering role for someone who is equally comfortable reading a business process flowchart and writing production-grade Python.</p>\n<p>You&#39;ll work end-to-end , from discovery and scoping through prototyping, launch, and iteration , using the latest agentic frameworks, tool-calling patterns, and responsible AI practices.</p>\n<p>What you&#39;ll do:</p>\n<ul>\n<li>Discover and scope AI opportunities: Partner with internal teams across corporate functions to understand their workflows, pain points, and goals, and identify high‑value AI/automation opportunities.</li>\n<li>Map and improve business processes: document current workflows, identify bottlenecks, and propose AI‑enabled changes that deliver clear business outcomes (e.g., time or cost savings, improved quality or compliance).</li>\n<li>Design end-to-end AI solutions: Design and implement AI‑enabled tools and workflows that integrate with existing systems and data sources, and that are intuitive for non‑technical users.</li>\n<li>Build and ship production-quality software: Write clean, maintainable code and tests.</li>\n<li>Pilot, rollout, and drive adoption: Pilot, roll out, and drive adoption of solutions by working closely with end‑users, gathering feedback, and iterating based on real‑world usage.</li>\n<li>Champion for responsible AI: Ensure solutions follow privacy, security, and compliance expectations, especially when working with sensitive or regulated data.</li>\n<li>Build for reuse: Create and share reusable patterns, components, and documentation to make future AI/automation work faster and more consistent across teams.</li>\n<li>Accelerate Workflows with Generative AI and Automation: Leverage AI to accelerate execution (e.g., draft, prototype, outline), explore alternative solutions using AI (iterate, compare approaches), synthesize information with AI (summarize, distill key themes), automate repeatable work (documentation, reporting, QA checks)</li>\n</ul>\n<p>What we&#39;re looking for:</p>\n<p>We&#39;re looking for mid-level engineers who have already shipped something real with AI , and who can work as a peer with non-technical business partners, not just as an order-taker.</p>\n<p>Specifically, you bring:</p>\n<ul>\n<li>Software engineering foundation. A CS, Engineering, Data Science, or related degree (or equivalent experience), with demonstrated ability to build and operate production systems , backend services, internal tools, integrations, or data applications.</li>\n<li>Hands-on AI and automation delivery. You&#39;ve shipped AI-powered or automation-driven solutions in a real environment.</li>\n<li>Agentic AI literacy. You understand how modern agentic systems are constructed , the difference between local and remote agents, how MCP (Model Context Protocol) works, what Agent Skills and Hooks are for, and how A2A (Agent-to-Agent) coordination is structured.</li>\n<li>System design and architecture thinking. You can sketch a data flow, reason about integration points, evaluate trade-offs between approaches, and design for failure , including fallbacks, retry logic, timeouts, and human escalation paths.</li>\n<li>Data and security judgment. You understand data access controls, the risks of giving AI broad access to sensitive information, PII minimization, audit logging, and what responsible data handling looks like in an enterprise environment.</li>\n<li>Business function acumen. You can engage credibly with stakeholders in Marketing, Finance, Sales, HR, Legal, or Operations , understanding their workflows, KPIs, and constraints well enough to scope solutions that fit their real needs.</li>\n<li>Clear, collaborative communication. You can explain architecture trade-offs to a Finance Manager and debug a prompt failure with an engineer in the same afternoon.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience working embedded with or alongside corporate / G&amp;A functions (Finance, Legal, HR, Marketing, Sales Operations, or similar).</li>\n<li>Practical experience with agentic frameworks such as LangGraph, Claude Agent SDK, or comparable tooling.</li>\n<li>Familiarity with MCP server design , including building, deploying, and securing MCP-compliant tool servers.</li>\n<li>Experience designing and evaluating AI outputs at scale: eval sets, sampling pipelines, human-in-the-loop review queues, or A/B testing of AI-powered features.</li>\n<li>Exposure to responsible AI frameworks: data minimization, differential privacy concepts, model output auditing, or working in PII-sensitive / regulated domains.</li>\n<li>Experience with RAG (Retrieval-Augmented Generation) pipelines, vector databases, or enterprise search integrations.</li>\n<li>Familiarity with CI/CD for AI: prompt versioning, model version pinning, regression testing for LLM-powered features.</li>\n</ul>\n<p>Relocation Statement:</p>\n<p>This position is not eligible for relocation assistance.</p>\n<p>Visit our PinFlex page to learn more about our working model.</p>\n<p>In-Office Requirement Statement:</p>\n<p>This role will need to be in the office for in-person collaboration 1-2 times every 6-months and therefore can be situated anywhere in the country.</p>\n<p>#LI-REMOTE #LI-KBF</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_fdfef6df-396","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Pinterest","sameAs":"https://www.pinterest.com/","logo":"https://logos.yubhub.co/pinterest.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/pinterest/jobs/7714127","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Software engineering foundation","Hands-on AI and automation delivery","Agentic AI literacy","System design and architecture thinking","Data and security judgment","Business function acumen","Clear, collaborative communication"],"x-skills-preferred":["Experience working embedded with or alongside corporate / G&A functions","Practical experience with agentic frameworks","Familiarity with MCP server design","Experience designing and evaluating AI outputs at scale","Exposure to responsible AI frameworks","Experience with RAG (Retrieval-Augmented Generation) pipelines","Familiarity with CI/CD for AI"],"datePosted":"2026-04-18T15:52:21.923Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, Seattle, WA, US; Remote, US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Software engineering foundation, Hands-on AI and automation delivery, Agentic AI literacy, System design and architecture thinking, Data and security judgment, Business function acumen, Clear, collaborative communication, Experience working embedded with or alongside corporate / G&A functions, Practical experience with agentic frameworks, Familiarity with MCP server design, Experience designing and evaluating AI outputs at scale, Exposure to responsible AI frameworks, Experience with RAG (Retrieval-Augmented Generation) pipelines, Familiarity with CI/CD for AI"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_23e0d317-de2"},"title":"GTM Systems Engineer","description":"<p>We&#39;re looking for a GTM Systems Engineer to build the nervous system of our revenue operations. You&#39;ll design systems that handle complex usage-based API pricing, enterprise licenses, self-serve Playground flows, and dual US/German entities. Your focus will be on building the technical architecture of our business, including billing infrastructure, enterprise GTM systems, and analytics infrastructure.</p>\n<p>As a GTM Systems Engineer, you&#39;ll work on building the systems that turn API calls into revenue, transform messy multi-jurisdictional data into clarity, and automate what currently requires 10 people to do manually. You&#39;ll architect integrations between CRM, billing, contracts, and finance, building the workflows that turn enterprise sales from a manual slog into something elegant.</p>\n<p>You&#39;ll be responsible for building the data pipelines and dashboards that show us what&#39;s actually happening: consumption patterns, churn signals, expansion opportunities. Not vanity metrics,the kind of real-time visibility that changes how we make decisions.</p>\n<p>We&#39;re looking for someone with 3-5+ years of experience as a Software Engineer, Systems Engineer, or RevOps Engineer at a B2B AI or SaaS company. You should have programming proficiency in Python, JavaScript/TypeScript, and SQL, as well as CRM expertise and experience building with integration platforms such as Workato, Zapier, or Tray.io.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_23e0d317-de2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Black Forest Labs","sameAs":"https://www.blackforestlabs.com/","logo":"https://logos.yubhub.co/blackforestlabs.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/blackforestlabs/jobs/5045195008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","JavaScript/TypeScript","SQL","CRM expertise","Integration platforms","API development","Data pipelines","Billing systems"],"x-skills-preferred":["Experience at a high-growth startup scaling from $1M to $100M+ ARR","Familiarity with API-first or usage-based products","Experience with data visualization tools"],"datePosted":"2026-04-18T15:52:19.665Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Freiburg (Germany)"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, JavaScript/TypeScript, SQL, CRM expertise, Integration platforms, API development, Data pipelines, Billing systems, Experience at a high-growth startup scaling from $1M to $100M+ ARR, Familiarity with API-first or usage-based products, Experience with data visualization tools"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ea9aa5d2-721"},"title":"Data Engineer Intern (Summer 2026)","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. We run one of the world&#39;s largest networks that powers millions of websites and other Internet properties.</p>\n<p>This internship is targeting students with experience and interest in Data Engineering. The Data Engineer Intern delivers full-stack data solutions across the entire data processing pipeline. This role relies on systems engineering principles to design and implement solutions that span the data lifecycle - collect, ingest, process, store, persist, access, and deliver data at scale and at speed.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Work through all stages of a data solution lifecycle – analyse / profile data, create conceptual, logical and physical data model designs, architect and design ETL, reporting and analytics</li>\n<li>Knowledge of modern enterprise data architectures, design patterns, and data tool sets and the ability to apply them</li>\n<li>Identify key metrics and build exec-facing dashboards to track progress of the business and its highest priority initiatives</li>\n<li>Identify key business levers, establish cause &amp; effect, perform analysis, and communicate key findings to various stakeholders to facilitate data driven decision-making</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>Currently enrolled in M.S in Computer Science, Engineering or related STEM field</li>\n<li>Experience working with Go, Python, SQL, Java, or equivalent programming languages</li>\n<li>Experience working with distributed systems (Spark etc.)</li>\n<li>Hands-on experience in data pipelines/ frameworks development</li>\n<li>Ability and interest to learn new skills and technologies quickly</li>\n<li>Excellent communication and problem-solving skills</li>\n<li>Ability to commit to a 12 week summer internship</li>\n</ul>\n<p>Bonus Points</p>\n<ul>\n<li>Familiarity with container based deployments such as Docker and Kubernetes</li>\n<li>Experience with JavaScript, Typescript, and React</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We&#39;re not just a highly ambitious, large-scale technology company. We&#39;re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organisations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law.</p>\n<p>We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals with disabilities. Please tell us if you require a reasonable accommodation to apply for a job.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ea9aa5d2-721","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7374706","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"internship","x-salary-range":null,"x-skills-required":["Go","Python","SQL","Java","Distributed systems","Data pipelines","Frameworks development"],"x-skills-preferred":["JavaScript","Typescript","React","Docker","Kubernetes"],"datePosted":"2026-04-18T15:52:05.982Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"In-Office"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Python, SQL, Java, Distributed systems, Data pipelines, Frameworks development, JavaScript, Typescript, React, Docker, Kubernetes"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a1ba5c28-9ce"},"title":"Senior Software Engineer, Observability","description":"<p>Join CoreWeave&#39;s Observability team, responsible for building the systems that give our customers and internal teams unparalleled visibility into complex AI workloads.</p>\n<p>Our team empowers engineers to understand, troubleshoot, and optimize high-performance infrastructure at massive scale.</p>\n<p>As a Senior Software Engineer on the Observability team, you will design, build, and maintain core observability infrastructure spanning metrics, logging, tracing, and telemetry pipelines.</p>\n<p>Your day-to-day will involve developing highly reliable and scalable systems, collaborating with internal engineering teams to embed observability best practices, and tackling performance and reliability challenges across clusters of thousands of GPUs.</p>\n<p>You&#39;ll also contribute to platform strategy and participate in on-call rotations to ensure critical production systems remain robust and operational.</p>\n<p>The base salary range for this role is $139,000 to $220,000.</p>\n<p>In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>We offer a variety of benefits to support your needs, including medical, dental, and vision insurance, 100% paid for by CoreWeave, company-paid Life Insurance, voluntary supplemental life insurance, short and long-term disability insurance, flexible Spending Account, Health Savings Account, tuition reimbursement, ability to participate in Employee Stock Purchase Program (ESPP), mental wellness benefits through Spring Health, family-forming support provided by Carrot, paid parental leave, flexible, full-service childcare support with Kinside, 401(k) with a generous employer match, flexible PTO, catered lunch each day in our office and data center locations, a casual work environment, and a work culture focused on innovative disruption.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a1ba5c28-9ce","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4554201006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$139,000 to $220,000","x-skills-required":["Go","Python","Kubernetes","containerization","microservices architectures","Helm","YAML-based configurations","automated testing","progressive release strategies","on-call rotations"],"x-skills-preferred":["designing, operating, or scaling logging, metrics, or tracing platforms","data streaming systems for observability pipelines","automating infrastructure provisioning","OpenTelemetry for unified telemetry collection and instrumentation","exposure to modern AI workloads and GPU-based infrastructure"],"datePosted":"2026-04-18T15:51:55.238Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY / Sunnyvale, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Python, Kubernetes, containerization, microservices architectures, Helm, YAML-based configurations, automated testing, progressive release strategies, on-call rotations, designing, operating, or scaling logging, metrics, or tracing platforms, data streaming systems for observability pipelines, automating infrastructure provisioning, OpenTelemetry for unified telemetry collection and instrumentation, exposure to modern AI workloads and GPU-based infrastructure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":139000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c1008b66-0b1"},"title":"Engineering Manager, Integrations","description":"<p>We&#39;re looking for an experienced Engineering Manager to lead our Integrations Platform team. As a key member of our Platform Engineering team, you will own and drive our Integrations Platform, including acquisition pipelines, action/write capabilities, SDK development, and developer experience.</p>\n<p>Your responsibilities will include creating and executing on a compelling vision to build world-class integration systems that empower both internal developers and third-party partners to build integrations with confidence. You will lead, mentor, and grow our talented team of platform engineers, while also coordinating with external consultants who build connectors using our SDK.</p>\n<p>You will champion organizational growth by implementing scalable, reliable integration practices that accelerate product development and delight our customers. Elevate the principles of platform engineering across Lumos.</p>\n<p>The ideal candidate will have 2+ years of proven people management experience leading engineering teams, experience managing teams focused on Integrations and/or Platform Engineering, and demonstrated experience building SDKs, developer platforms, or third-party integration ecosystems with a focus on developer experience and reliability.</p>\n<p>Experience working at a startup, exposure to excellent engineering practices gained at a top-tier technology company, and knowledge and experience with modern integration patterns (REST, GraphQL, webhooks, event-driven architectures) are highly valued.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c1008b66-0b1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Lumos","sameAs":"https://lumos.com","logo":"https://logos.yubhub.co/lumos.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/lumos/jobs/7662233003","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 - $236,000","x-skills-required":["2+ years of proven people management experience leading engineering teams","Experience managing teams focused on Integrations and/or Platform Engineering","Demonstrated experience building SDKs, developer platforms, or third-party integration ecosystems with a focus on developer experience and reliability","Strong background in ETL and data extraction/acquisition pipelines","Proven ability to plan and implement integration architectures that balance reliability, cost efficiency, and developer productivity"],"x-skills-preferred":["Experience working at a startup","Exposure to excellent engineering practices gained at a top-tier technology company","Knowledge and experience with modern integration patterns (REST, GraphQL, webhooks, event-driven architectures)"],"datePosted":"2026-04-18T15:51:51.509Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"2+ years of proven people management experience leading engineering teams, Experience managing teams focused on Integrations and/or Platform Engineering, Demonstrated experience building SDKs, developer platforms, or third-party integration ecosystems with a focus on developer experience and reliability, Strong background in ETL and data extraction/acquisition pipelines, Proven ability to plan and implement integration architectures that balance reliability, cost efficiency, and developer productivity, Experience working at a startup, Exposure to excellent engineering practices gained at a top-tier technology company, Knowledge and experience with modern integration patterns (REST, GraphQL, webhooks, event-driven architectures)","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":236000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6d20be06-d60"},"title":"Software Engineer - Backend","description":"<p>As a software engineer with a backend focus, you will work with your team to build infrastructure for the Databricks platform at scale.</p>\n<p>Our backend teams cover a diverse range of domains, from core compute fabric resource management to service platforms and machine learning infrastructure.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Building foundational infrastructure platforms that enable seamless operation across numerous geographic regions and cloud providers.</li>\n</ul>\n<ul>\n<li>Implementing cloud-agnostic infrastructure abstractions to help Databricks engineers more efficiently manage and operate their services.</li>\n</ul>\n<ul>\n<li>Developing tools and processes that drive engineering efficiency at Databricks.</li>\n</ul>\n<p>We enhance the developer experience for Databricks engineers across various areas, including programming languages, linters, static analysis, IDEs, remote development environments, automated release pipelines, and test automation frameworks.</p>\n<p>To be successful in this role, you will need:</p>\n<ul>\n<li>2+ years of professional software development experience, or alternatively a PhD degree.</li>\n</ul>\n<ul>\n<li>Proficiency in one or more backend languages such as Java, Scala, or Go.</li>\n</ul>\n<ul>\n<li>Hands-on experience in developing and operating backend systems.</li>\n</ul>\n<ul>\n<li>Ability to contribute effectively throughout all project phases, from initial design and development to implementation and ongoing operations, with guidance from senior team members.</li>\n</ul>\n<ul>\n<li>Self-driven and passionate with a strong focus on delivering impact through team collaboration.</li>\n</ul>\n<ul>\n<li>Strong communication skills, both written and verbal, with the ability to produce clear technical documentation that drives alignment and decision-making.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6d20be06-d60","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7737233002","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Java","Scala","Go","backend languages","cloud-agnostic infrastructure abstractions","engineering efficiency"],"x-skills-preferred":["Rust development experience","programming languages","linters","static analysis","IDEs","remote development environments","automated release pipelines","test automation frameworks"],"datePosted":"2026-04-18T15:51:48.785Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Aarhus, Denmark"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Scala, Go, backend languages, cloud-agnostic infrastructure abstractions, engineering efficiency, Rust development experience, programming languages, linters, static analysis, IDEs, remote development environments, automated release pipelines, test automation frameworks"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0396ac1c-dad"},"title":"Senior Staff Engineer, Cloud Economics","description":"<p>Reddit is a community of communities. It&#39;s built on shared interests, passion, and trust, and is home to the most open and authentic conversations on the internet.</p>\n<p>The Ads Foundations organization is responsible for the technical backbone powering Ads Monetization at scale. Within this ecosystem, efficient resource utilization is critical.</p>\n<p>We are seeking a Senior Staff Engineer to serve as the Cloud Resources Technical Owner for the Ads Domain. You will be the primary engineering point of contact for the Senior Director in Ads and Cloud Operations/Resources (COR &amp; Opex) stakeholders.</p>\n<p><strong>Responsibilities</strong></p>\n<p>Technical Vision &amp; Strategy</p>\n<ul>\n<li>Define and drive the technical strategy for Cloud Resource management within Ad first, ensuring that cost accountability is built into the architecture of our systems.</li>\n<li>High-Fidelity Investment Modeling: Elevate cloud estimation from guesswork to a rigorous engineering discipline. You will lead the high-quality forecasting of new cloud investments and efficiency projects, designing data-driven models to validate technical ROI before builds happen</li>\n<li>Design and implement a roadmap for Cost Observability 2.0, moving beyond simple reporting to real-time, service/team-level spend attribution and automated anomaly detection.</li>\n</ul>\n<p>Engineering &amp; Tooling Leadership</p>\n<ul>\n<li>Design and build internal platforms that programmatically enforce PnL accountability. You will engineer (or collaborate with Core Infrastructure partners) to deliver the dashboards, alerts, and governance tools that every Ads team relies on to manage their cloud footprint.</li>\n<li>Architect automated frameworks for validating cost estimates and forecasting, replacing manual spreadsheets with data-driven software solutions.</li>\n</ul>\n<p>Scale &amp; Optimization</p>\n<ul>\n<li>Fight for observability by instrumenting deep telemetry into our cloud infrastructure. You will be hands-on in identifying inefficiencies (e.g., underutilized clusters, uncompressed data flows) and re-architecting critical paths for cost reduction.</li>\n<li>Lead the technical validation of vendor and 3rd-party tool integration, ensuring we extract maximum engineering value from every dollar spent.</li>\n</ul>\n<p>Cultural &amp; Technical Stewardship</p>\n<ul>\n<li>Act as a role model for the Ads domain and the wider company. You will set the standard for how engineering teams think about Cost as a Non Functional Requirement, eventually scaling these patterns to other domains.</li>\n<li>Partner with Finance and Engineering leadership to translate Cloud Spend into actionable engineering tasks (e.g., refactor Service X to use Spot instances).</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>10+ years of software engineering experience, with a strong focus on public cloud infrastructure (AWS/GCP/Azure) and large-scale distributed systems.</li>\n<li>Engineer-First Mindset: You are comfortable writing code (Go, Python, Java) to solve infrastructure problems. You don&#39;t just ask for a report; you build the API that generates it.</li>\n<li>Deep Cloud Expertise: You have mastery over Kubernetes, container orchestration, and cloud-native storage, understanding exactly how architectural choices impact the bottom line.</li>\n<li>Operational Excellence: Proven track record of building observability pipelines (Prometheus, Grafana, Datadog) that drive operational and financial alerts.</li>\n<li>Influential Leader: Skilled at driving clarity in ambiguous spaces. You can convince a Principal Engineer to refactor their service for cost efficiency because you can prove the technical and business value.</li>\n</ul>\n<p><strong>Bonus Points</strong></p>\n<ul>\n<li>Experience building custom FinOps tooling or internal developer platforms.</li>\n<li>Background in performance engineering or capacity planning for high-traffic ad tech environments.</li>\n<li>Contributions to open-source projects related to cloud efficiency or observability.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0396ac1c-dad","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit Inc.","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7628291","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$232,500-$325,500 USD","x-skills-required":["public cloud infrastructure","large-scale distributed systems","Kubernetes","container orchestration","cloud-native storage","observability pipelines","Prometheus","Grafana","Datadog"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:43.900Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"public cloud infrastructure, large-scale distributed systems, Kubernetes, container orchestration, cloud-native storage, observability pipelines, Prometheus, Grafana, Datadog","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":232500,"maxValue":325500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_119ff7f5-669"},"title":"Staff Backend Engineer, Developer Experience","description":"<p>As a Staff Backend Engineer in the Developer Experience group, you&#39;ll work cross-functionally to help us identify and improve the engineering experience across GitLab&#39;s internal engineering teams.</p>\n<p>We&#39;re at a pivotal time in transforming our testing approach and development experience to enable engineers to deliver high-quality code efficiently and with confidence. Developer Experience currently has six, globally distributed teams of Backend Engineers and Software Engineers in Test.</p>\n<p>This role will work across the entire group to help us identify and remove the biggest points of friction that affect development at GitLab. Expect to split your time between hands-on coding in Ruby and Go, designing scalable solutions and helping every team deliver.</p>\n<p>You&#39;ll be instrumental in shaping our engineering culture across GitLab. This is a highly visible role with strong company-wide support. You&#39;ll be instrumental in shifting our culture and practices toward a cohesive Developer Experience Platform strategy,backed by buy-in and appetite for change.</p>\n<p>Some examples of our projects include:</p>\n<p>Identifying and removing the source of flakiness in our tests pipelines Introducing comprehensive test coverage reporting to guide our decisions about what and when to test Standardizing logging libraries and documentation to make it easy for all features and components to provide consistent logs Providing stable, configurable development environments</p>\n<p>What You&#39;ll Do:</p>\n<p>Drive Technical Excellence Across Teams: Provide hands-on engineering contributions and technical leadership across six Developer Experience teams, jumping in to unblock critical work or accelerate high-impact initiatives Champion Engineering Needs: Partner with engineering teams across GitLab to identify pain points, gather feedback, and translate day-to-day frustrations into actionable improvements for the Developer Experience roadmap Transform Testing at Scale: Help us modernize our testing to support GitLab&#39;s growth,ensuring engineers have fast, reliable tools that catch issues early without slowing them down Optimize CI/CD Pipelines: Reduce merge and CI friction by building workflows that deliver fast, actionable feedback,helping engineers iterate confidently and ship faster Break Down Systemic Bottlenecks: Collaborate with Platform, Product, and Infrastructure teams to identify and eliminate the architectural and process friction that slows down development Shape Engineering Culture: Foster a culture of continuous learning, inclusion, and improvement,where great developer experience is a shared value across the organization</p>\n<p>What You&#39;ll Bring:</p>\n<p>Deep experience with developer tooling, DevEx strategies, and testing frameworks at scale Experience leading platform or transformation initiatives across multiple teams Proven ability to influence without authority and collaborate across functions Familiarity with DevOps workflows and CI/CD pipelines Strong written and verbal communication skills in a remote-first environment A product mindset with a passion for empowering developers Alignment with GitLab values, especially collaboration, iteration, and transparency</p>\n<p>The base salary range for this role&#39;s listed level is currently for residents of the United States only. This range is intended to reflect the role&#39;s base salary rate in locations throughout the US. Grade level and salary ranges are determined through interviews and a review of education, experience, knowledge, skills, abilities of the applicant, equity with other team members, alignment with market data, and geographic location. The base salary range does not include any bonuses, equity, or benefits.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_119ff7f5-669","directApply":true,"hiringOrganization":{"@type":"Organization","name":"GitLab","sameAs":"https://about.gitlab.com/","logo":"https://logos.yubhub.co/about.gitlab.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/gitlab/jobs/8490477002","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$131,600-$282,000 USD","x-skills-required":["Ruby","Go","DevEx strategies","testing frameworks","DevOps workflows","CI/CD pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:29.965Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote, Canada; Remote, United Kingdom; Remote, US"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Ruby, Go, DevEx strategies, testing frameworks, DevOps workflows, CI/CD pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":131600,"maxValue":282000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_72ebb09d-b37"},"title":"Staff+ Software Engineer, Observability","description":"<p>We&#39;re seeking talented and experienced Software Engineers to join our Observability team within the Infrastructure organization. The Observability team owns the monitoring and telemetry infrastructure that every engineer and researcher at Anthropic depends on,from metrics and logging pipelines to distributed tracing, error analytics, alerting, and the dashboards and query interfaces that make it all actionable.</p>\n<p>As Anthropic scales its infrastructure across massive GPU, TPU, and Trainium clusters, the volume and complexity of operational data is growing by orders of magnitude. We&#39;re building next-generation observability systems,high-throughput ingest pipelines, cost-efficient columnar storage, unified query layers across signals, and agentic diagnostic tools,to ensure that engineers can detect, diagnose, and resolve issues in minutes rather than hours, even as the systems they operate become exponentially more complex.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build scalable telemetry ingest and storage pipelines for metrics, logs, traces, and error data across Anthropic&#39;s multi-cluster infrastructure</li>\n<li>Own and evolve core observability platforms, driving migrations and architectural improvements that improve reliability, reduce cost, and scale with organisational growth</li>\n<li>Build instrumentation libraries, SDKs, and integrations that make it easy for engineering teams to emit high-quality telemetry from their services</li>\n<li>Drive alerting and SLO infrastructure that enables teams to define, monitor, and respond to reliability targets with minimal noise</li>\n<li>Reduce mean time to detection and resolution by building cross-signal correlation, unified query interfaces, and AI-assisted diagnostic tooling</li>\n<li>Partner with Research, Inference, Product, and Infrastructure teams to ensure observability solutions meet the unique needs of each organisation</li>\n</ul>\n<p>You May Be a Good Fit If You:</p>\n<ul>\n<li>Have 10+ years of relevant industry experience building and operating large-scale observability or monitoring infrastructure</li>\n<li>Have deep experience with at least one observability signal area (metrics, logging, tracing, or error analytics) and familiarity with the others</li>\n<li>Understand high-throughput data pipelines, columnar storage engines, and the tradeoffs involved in ingesting and querying telemetry data at scale</li>\n<li>Have experience operating or building on top of observability platforms such as Prometheus, Grafana, ClickHouse, OpenTelemetry, or similar systems</li>\n<li>Have strong proficiency in at least one of Python, Rust, or Go</li>\n<li>Have excellent communication skills and enjoy partnering with internal teams to improve their operational visibility and incident response capabilities</li>\n<li>Are excited about building foundational infrastructure and are comfortable working independently on ambiguous, high-impact technical challenges</li>\n</ul>\n<p>Strong Candidates May Also Have:</p>\n<ul>\n<li>Experience operating metrics systems at very high cardinality (hundreds of millions of active time series or more)</li>\n<li>Experience with log storage migrations or operating columnar databases (ClickHouse, BigQuery, or similar) for analytics workloads</li>\n<li>Experience with OpenTelemetry instrumentation, collector pipelines, and tail-based sampling strategies</li>\n<li>Experience building or operating alerting platforms, on-call tooling, or SLO frameworks at scale</li>\n<li>Experience with Kubernetes-native monitoring, eBPF-based observability, or continuous profiling</li>\n<li>Interest in applying AI/LLMs to operational workflows such as automated root cause analysis, anomaly detection, or intelligent alerting</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_72ebb09d-b37","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5139910008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["observability","monitoring","telemetry","metrics","logging","tracing","error analytics","alerting","SLO infrastructure","cross-signal correlation","unified query interfaces","AI-assisted diagnostic tooling","Python","Rust","Go","Prometheus","Grafana","ClickHouse","OpenTelemetry"],"x-skills-preferred":["high-throughput data pipelines","columnar storage engines","operating system administration","cloud computing","containerization","DevOps"],"datePosted":"2026-04-18T15:51:29.494Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"observability, monitoring, telemetry, metrics, logging, tracing, error analytics, alerting, SLO infrastructure, cross-signal correlation, unified query interfaces, AI-assisted diagnostic tooling, Python, Rust, Go, Prometheus, Grafana, ClickHouse, OpenTelemetry, high-throughput data pipelines, columnar storage engines, operating system administration, cloud computing, containerization, DevOps","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_22ff82ac-40b"},"title":"Software Engineer, Research Data Platform","description":"<p>We&#39;re looking for engineers who love working directly with users and who excel at building data products. The Research Data Platform team builds the tools that Anthropic&#39;s researchers use every day to manage, query, and analyze the data that goes into training and evaluating frontier models.</p>\n<p>As a software engineer on this team, you will:</p>\n<ul>\n<li>Build and operate data pipelines that extract data from research training runs and land it in storage systems that are easy and fast to query</li>\n<li>Work closely with researchers to design and build APIs, libraries, and web interfaces that support data management, exploration, and analysis</li>\n<li>Develop dataset management, data cataloging, and provenance tooling that researchers use in their day-to-day work</li>\n<li>Embed with research teams to understand their workflows, identify high-leverage tooling opportunities, and ship solutions quickly</li>\n<li>Collaborate with adjacent teams to build on existing systems rather than reinventing them</li>\n</ul>\n<p>You may be a good fit if you have significant software engineering experience, particularly building data-intensive applications or internal tooling. You should enjoy working directly with users, gathering requirements iteratively, and shipping things that get adopted. You should also be results-oriented, with a bias towards flexibility and impact.</p>\n<p>Strong candidates may also have experience with large-scale ETL, columnar storage formats, and query engines, high-volume time series data, data cataloging, lineage, or metadata management systems, ML experiment tracking or metrics platforms, complex data visualization, and full-stack web application development.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_22ff82ac-40b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5191226008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["software engineering","data-intensive applications","internal tooling","data pipelines","storage systems","APIs","libraries","web interfaces","dataset management","data cataloging","provenance tooling","research workflows","adjacent teams"],"x-skills-preferred":["large-scale ETL","columnar storage formats","query engines","high-volume time series data","lineage","metadata management systems","ML experiment tracking","metrics platforms","complex data visualization","full-stack web application development"],"datePosted":"2026-04-18T15:51:29.293Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"software engineering, data-intensive applications, internal tooling, data pipelines, storage systems, APIs, libraries, web interfaces, dataset management, data cataloging, provenance tooling, research workflows, adjacent teams, large-scale ETL, columnar storage formats, query engines, high-volume time series data, lineage, metadata management systems, ML experiment tracking, metrics platforms, complex data visualization, full-stack web application development","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0ae48270-bef"},"title":"Senior Software Engineer, Storage Engineer","description":"<p>The Storage Engine Organisation at CoreWeave is responsible for the product capabilities and data plane function of CoreWeave&#39;s managed storage products.</p>\n<p>We build reliable, scalable storage solutions with segment leading performance. Storage engine works with engineering teams across infrastructure, compute, and platform to ensure our storage services meet the needs of the world&#39;s most demanding AI workloads.</p>\n<p>The role involves designing and implementing distributed storage solutions to support scaling data-intensive AI workloads, contributing to the development of exabyte-scale, S3-compatible object storage, and integrating dedicated storage clusters into diverse customer environments.</p>\n<p>Key responsibilities include working with technologies such as RDMA, GPU Direct Storage, and distributed filesystems protocols like NFS or FUSE to optimise storage performance and efficiency, participating in efforts to improve the reliability, durability, and observability of our storage stack, collaborating with operations teams to monitor, troubleshoot, and improve storage systems in production environments, and helping develop metrics and dashboards to provide visibility into storage performance and health.</p>\n<p>The ideal candidate will have a strong background in storage systems engineering or infrastructure, with experience working with object storage or distributed filesystems in production environments, proficiency in a systems programming language like Go, C, or Rust, and familiarity with storage observability tools and telemetry pipelines.</p>\n<p>As a senior software engineer, you will be responsible for designing, developing, and deploying scalable and efficient storage solutions, working closely with cross-functional teams to ensure seamless integration with other components of the platform, and mentoring junior engineers to help them grow in their roles.</p>\n<p>If you are passionate about building high-performance storage solutions and have a strong background in software engineering, we encourage you to apply for this exciting opportunity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0ae48270-bef","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4643524006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$139,000 to $204,000","x-skills-required":["Storage systems engineering","Infrastructure","Object storage","Distributed filesystems","RDMA","GPU Direct Storage","NFS","FUSE","Systems programming languages (Go, C, Rust)","Storage observability tools","Telemetry pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:26.395Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ/ New York , NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Storage systems engineering, Infrastructure, Object storage, Distributed filesystems, RDMA, GPU Direct Storage, NFS, FUSE, Systems programming languages (Go, C, Rust), Storage observability tools, Telemetry pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":139000,"maxValue":204000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_92d63795-0ea"},"title":"Principal Systems Engineer, M&A","description":"<p>The Infrastructure Engineering organization is seeking an accomplished Principal Systems Engineer to lead our acquisition integration engineering practice. This pivotal role will own the end-to-end infrastructure engineering lifecycle of integrating newly acquired businesses into Anduril&#39;s existing ecosystem.</p>\n<p>As the world enters an era of strategic competition, Anduril is committed to bringing cutting-edge autonomy, AI, computer vision, sensor fusion, and networking technology to the military in months, not years.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Define, establish, and lead the infrastructure integration engineering practice, setting the technical vision and strategy for integrating new entities and technologies.</li>\n<li>Single-threaded ownership of all infrastructure component(s) of acquisition integration from discovery and due diligence through migration execution and hypercare.</li>\n<li>Conduct comprehensive technical assessments of target companies&#39; infrastructure, systems, and operational capabilities, identifying risks and opportunities.</li>\n<li>Develop and present high-level architectural strategies and detailed roadmaps for integrations to executive leadership, founders, and technical teams.</li>\n<li>Design and implement robust, scalable, and secure system architectures for integrated environments, ensuring alignment with Anduril&#39;s overall technology strategy.</li>\n<li>Develop and execute detailed migration plans, managing complex technical challenges and dependencies.</li>\n<li>Provide post-migration hypercare support, ensuring a smooth transition and stabilization of integrated systems.</li>\n<li>Define, document, and continuously improve repeatable processes to accelerate acquisition integration, establishing benchmarks, conducting post-mortems, and implementing lessons learned.</li>\n<li>Identify, evaluate, and implement or scope the development of new tools and technologies to enhance discovery, migration, and testing efficiency.</li>\n<li>Collaborate closely with Security teams to ensure all integrated systems meet Anduril&#39;s stringent security requirements and policies.</li>\n<li>Partner with Client Engineering teams to ensure seamless integration of acquired client and client-facing technologies and services.</li>\n<li>Provide clear, concise, and opinionated technical guidance, and proactively push back on misaligned proposals to ensure successful technical outcomes.</li>\n<li>Act as a technical authority, mentor, and trusted advisor to engineering teams involved in integration efforts.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Minimum of 12 years of progressive experience in Systems Engineering, Network Engineering, and/or IT Infrastructure roles with a focus on complex, enterprise-scale environments.</li>\n<li>Self-sufficient ability to execute in (technical and non-technical) program management, architecture, and hands-on engineering capacities.</li>\n<li>Demonstrated expertise in defining and building engineering practices and repeatable processes.</li>\n<li>Proven ability to operate across the entire engineering lifecycle, from strategic discovery and architecture to hands-on execution and hypercare.</li>\n<li>Exceptional ability to communicate complex technical concepts to diverse audiences, including C-suite executives, founders, and engineering teams.</li>\n<li>Deep understanding of modern cloud architectures (AWS, Azure, GCP), hybrid cloud solutions, and on-premises infrastructure.</li>\n<li>Extensive experience with enterprise networking technologies, including routing, switching, firewalls, VPNs, and load balancing.</li>\n<li>Strong knowledge of server virtualization, containerization technologies (e.g., Docker, Kubernetes), and operating systems (Linux, Windows).</li>\n<li>Experience with identity and access management (IAM) solutions, single sign-on (SSO), and multi-factor authentication (MFA).</li>\n<li>Proficiency in scripting and automation for infrastructure deployment and management (e.g., Python, Ansible, Terraform).</li>\n<li>Strong understanding of security principles, best practices, and common vulnerabilities within systems and networks.</li>\n<li>Familiarity with client engineering principles and technologies.</li>\n<li>Proven experience in identifying tooling gaps and either developing solutions or effectively scoping them for development.</li>\n<li>Excellent analytical, problem-solving, and critical thinking skills.</li>\n<li>Ability to travel for remote deployments and assessments as required.</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Experience with infrastructure-as-code (IaC) principles and tools.</li>\n<li>Familiarity with CI/CD pipelines and DevOps methodologies.</li>\n<li>Experience with data center design and operations.</li>\n<li>Experience in the defense technology or highly regulated industries.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_92d63795-0ea","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5111019007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$220,000-$292,000 USD","x-skills-required":["Systems Engineering","Network Engineering","IT Infrastructure","Cloud Architectures","Hybrid Cloud Solutions","On-Premises Infrastructure","Enterprise Networking Technologies","Server Virtualization","Containerization Technologies","Operating Systems","Identity and Access Management","Single Sign-On","Multi-Factor Authentication","Scripting and Automation","Infrastructure Deployment and Management"],"x-skills-preferred":["Infrastructure-as-Code","CI/CD Pipelines","DevOps Methodologies","Data Center Design and Operations","Defense Technology","Highly Regulated Industries"],"datePosted":"2026-04-18T15:51:18.220Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, Washington, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Systems Engineering, Network Engineering, IT Infrastructure, Cloud Architectures, Hybrid Cloud Solutions, On-Premises Infrastructure, Enterprise Networking Technologies, Server Virtualization, Containerization Technologies, Operating Systems, Identity and Access Management, Single Sign-On, Multi-Factor Authentication, Scripting and Automation, Infrastructure Deployment and Management, Infrastructure-as-Code, CI/CD Pipelines, DevOps Methodologies, Data Center Design and Operations, Defense Technology, Highly Regulated Industries","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":220000,"maxValue":292000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f296b6b0-e66"},"title":"Senior Software Security Engineer","description":"<p>Job Title: Senior Software Security Engineer</p>\n<p>About the Role: The Security Engineering team&#39;s mission is to safeguard our AI systems and maintain the trust of our users and society at large. Whether we&#39;re developing critical security infrastructure, building secure development practices, or partnering with our research and product teams, we are committed to operating as a world-class security organization and keeping the safety and trust of our users at the forefront of everything we do.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Build security for large-scale AI clusters, implementing robust cloud security architecture including IAM, network segmentation, and encryption controls</li>\n</ul>\n<ul>\n<li>Design secure-by-design workflows, secure CI/CD pipelines across our services, help build secure cloud infrastructure, with expertise in various cloud environments, Kubernetes security, container orchestration and identity management</li>\n</ul>\n<ul>\n<li>Ship and operate secure, high-reliability services using Infrastructure-as-Code (IaC) practices and GitOps workflows</li>\n</ul>\n<ul>\n<li>Apply deep expertise in threat modeling and risk assessment to secure complex multi cloud environments</li>\n</ul>\n<ul>\n<li>Mentor engineers and contribute to hiring and growth of the Security team</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>5-15+ years of software engineering experience implementing and maintaining critical systems at scale</li>\n</ul>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science/Software Engineering or equivalent industry experience</li>\n</ul>\n<ul>\n<li>Strong software engineering skills in Python or at least one systems language (Go, Rust, C/C++)</li>\n</ul>\n<ul>\n<li>Experience managing infrastructure at scale with DevOps and cloud automation best practices</li>\n</ul>\n<ul>\n<li>Track record of driving engineering excellence through high standards, constructive code reviews, and mentorship</li>\n</ul>\n<ul>\n<li>Proven ability to lead cross-functional security initiatives and navigate complex organizational dynamics</li>\n</ul>\n<ul>\n<li>Outstanding communication skills, translating technical concepts effectively across all organizational levels</li>\n</ul>\n<ul>\n<li>Demonstrated success in bringing clarity and ownership to ambiguous technical problems</li>\n</ul>\n<ul>\n<li>Strong systems thinking with ability to identify and mitigate risks in complex environments</li>\n</ul>\n<ul>\n<li>Low ego, high empathy engineer who attracts talent and supports diverse, inclusive teams</li>\n</ul>\n<ul>\n<li>Experience supporting fast-paced startup engineering teams</li>\n</ul>\n<ul>\n<li>Passionate about AI safety and alignment, with keen interest in making AI systems more interpretable and aligned with human values</li>\n</ul>\n<p>Salary: The annual compensation range for this role is £240,000-£325,000 GBP.</p>\n<p>Experience Level: senior Employment Type: full-time Workplace Type: hybrid Category: Engineering Industry: Technology Salary Range: £240,000-£325,000 GBP Required Skills:</p>\n<ul>\n<li>Cloud security architecture</li>\n<li>IAM</li>\n<li>Network segmentation</li>\n<li>Encryption controls</li>\n<li>Kubernetes security</li>\n<li>Container orchestration</li>\n<li>Identity management</li>\n<li>Infrastructure-as-Code (IaC)</li>\n<li>GitOps</li>\n<li>Threat modeling</li>\n<li>Risk assessment</li>\n<li>DevOps</li>\n<li>Cloud automation</li>\n<li>Python</li>\n<li>Go</li>\n<li>Rust</li>\n<li>C/C++</li>\n</ul>\n<p>Preferred Skills:</p>\n<ul>\n<li>Secure-by-design workflows</li>\n<li>CI/CD pipelines</li>\n<li>Secure cloud infrastructure</li>\n<li>Cloud environments</li>\n<li>Containerization</li>\n<li>Identity and access management</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f296b6b0-e66","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5022845008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"£240,000-£325,000 GBP","x-skills-required":["Cloud security architecture","IAM","Network segmentation","Encryption controls","Kubernetes security","Container orchestration","Identity management","Infrastructure-as-Code (IaC)","GitOps","Threat modeling","Risk assessment","DevOps","Cloud automation","Python","Go","Rust","C/C++"],"x-skills-preferred":["Secure-by-design workflows","CI/CD pipelines","Secure cloud infrastructure","Cloud environments","Containerization","Identity and access management"],"datePosted":"2026-04-18T15:51:17.687Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Cloud security architecture, IAM, Network segmentation, Encryption controls, Kubernetes security, Container orchestration, Identity management, Infrastructure-as-Code (IaC), GitOps, Threat modeling, Risk assessment, DevOps, Cloud automation, Python, Go, Rust, C/C++, Secure-by-design workflows, CI/CD pipelines, Secure cloud infrastructure, Cloud environments, Containerization, Identity and access management","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":240000,"maxValue":325000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f2c6f765-eca"},"title":"Staff Engineer, Storage Control Plane","description":"<p>We&#39;re looking for a Staff Storage Engineer to play a key role in designing, building, and operating the control plane for our high-performance AI storage platform. You&#39;ll help evolve CoreWeave&#39;s storage systems by building reliable, scalable, and high-throughput solutions that power some of the largest and innovative AI workloads in the world.</p>\n<p>This role involves close collaboration with teams across infrastructure, compute, and platform to ensure our storage services scale automatically and seamlessly while maximizing performance and reliability.</p>\n<p>About the role:</p>\n<ul>\n<li>Design and implement a highly scalable multi-tenant control plane that supports CoreWeave&#39;s growing AI storage and cloud infrastructure needs.</li>\n</ul>\n<ul>\n<li>Contribute to the development of exabyte-scale, S3-compatible object storage, distributed file system and integrate dedicated storage clusters into diverse customer environments.</li>\n</ul>\n<ul>\n<li>Work with technologies such as RDMA, GPU Direct Storage, RoCE, InfiniBand, SPDK, and distributed filesystems to optimize storage performance and efficiency.</li>\n</ul>\n<ul>\n<li>Participate in efforts to improve the reliability, durability, and observability of our storage stack.</li>\n</ul>\n<ul>\n<li>Collaborate with operations teams to monitor, analyze, and optimize storage systems using telemetry, metrics, and dashboards to improve performance, latency, and resilience.</li>\n</ul>\n<ul>\n<li>Work cross-functionally with platform, product, and infrastructure teams to deliver seamless storage capabilities across the stack.</li>\n</ul>\n<ul>\n<li>Share your knowledge and mentor other engineers on best practices in building distributed, high-performance systems.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>Bachelor&#39;s or Master&#39;s degree in Computer Science, Engineering, or a related field.</li>\n</ul>\n<ul>\n<li>10+ years of experience working in storage systems engineering or infrastructure.</li>\n</ul>\n<ul>\n<li>Strong hands-on experience with object storage or distributed filesystems in production environments.</li>\n</ul>\n<ul>\n<li>Experience with one or more storage protocols (e.g. S3, NFS) and file systems such as Ceph, DAOS, or similar.</li>\n</ul>\n<ul>\n<li>Proficiency in a systems programming language such as Go, C++, or Rust.</li>\n</ul>\n<ul>\n<li>Familiarity with storage observability tools and telemetry pipelines (e.g., ClickHouse, Prometheus, Grafana).</li>\n</ul>\n<ul>\n<li>Solid understanding of cloud-native infrastructure, Kubernetes, and scalable system architecture.</li>\n</ul>\n<ul>\n<li>Strong debugging and problem-solving skills in distributed, high-performance environments.</li>\n</ul>\n<ul>\n<li>Clear communicator, able to work collaboratively across teams and share technical insights effectively.</li>\n</ul>\n<p>Wondering if you&#39;re a good fit? We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren&#39;t a 100% skill or experience match. Here are a few qualities we&#39;ve found compatible with our team. If some of this describes you, we&#39;d love to talk.</p>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n</ul>\n<ul>\n<li>Act Like an Owner</li>\n</ul>\n<ul>\n<li>Empower Employees</li>\n</ul>\n<ul>\n<li>Deliver Best-in-Class Client Experiences</li>\n</ul>\n<ul>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!</p>\n<p>The base salary range for this role is $165,000 to $242,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we&#39;ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location. In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n</ul>\n<ul>\n<li>100% paid for by CoreWeave</li>\n</ul>\n<ul>\n<li>Company-paid Life Insurance</li>\n</ul>\n<ul>\n<li>Voluntary supplemental life insurance</li>\n</ul>\n<ul>\n<li>Short and long-term disability insurance</li>\n</ul>\n<ul>\n<li>Flexible Spending Account</li>\n</ul>\n<ul>\n<li>Health Savings Account</li>\n</ul>\n<ul>\n<li>Tuition Reimbursement</li>\n</ul>\n<ul>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n</ul>\n<ul>\n<li>Mental Wellness Benefits through Spring Health</li>\n</ul>\n<ul>\n<li>Family-Forming support provided by Carrot</li>\n</ul>\n<ul>\n<li>Paid Parental Leave</li>\n</ul>\n<ul>\n<li>Flexible, full-service childcare support with Kinside</li>\n</ul>\n<ul>\n<li>401(k) with a generous employer match</li>\n</ul>\n<ul>\n<li>Flexible PTO</li>\n</ul>\n<ul>\n<li>Catered lunch each day in our office and data center locations</li>\n</ul>\n<ul>\n<li>A casual work environment</li>\n</ul>\n<ul>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p>California Consumer Privacy Act - California applicants only</p>\n<p>CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace. All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information. As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship. If reasonable accommodation is needed, please contact: careers@coreweave.com.</p>\n<p>Export Control Compliance</p>\n<p>This position requires access to export controlled information. To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without a required export authorization, or (C) eligible and reasonably likely to obtain the required export authorization from the applicable U.S. government agency. CoreWeave may, for legitimate business reasons, decline to pursue any export licensing process.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f2c6f765-eca","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4669836006","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["object storage","distributed filesystems","storage protocols","file systems","cloud-native infrastructure","Kubernetes","scalable system architecture","systems programming language","Go","C++","Rust","storage observability tools","telemetry pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:51:06.353Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA / Dallas, TX"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"object storage, distributed filesystems, storage protocols, file systems, cloud-native infrastructure, Kubernetes, scalable system architecture, systems programming language, Go, C++, Rust, storage observability tools, telemetry pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4b2edfb8-1c2"},"title":"Senior Software Engineer, Client Platform","description":"<p>We&#39;re looking for a Senior Software Engineer to join our Builder Experience (BIX) team. As a key member of our platform team, you&#39;ll be responsible for designing and implementing the foundations that every product engineer builds on top of. This includes the design system, core UI frameworks, client performance, state management patterns, continuous integration, and the libraries and tooling that keep our codebase healthy and our engineers productive.</p>\n<p>You&#39;ll be working closely with our Design team to evolve and scale our component library, ensuring it&#39;s accessible, composable, and well-documented. You&#39;ll also be responsible for profiling, diagnosing, and fixing client-side performance bottlenecks, establishing performance budgets, and building dashboards to keep the team honest.</p>\n<p>As a force multiplier, you&#39;ll act as a coach and enablement specialist, helping product teams adopt improvements and level up their craft. You&#39;ll write playbooks and docs, deliver tech talks, pair with product engineers, and create local tooling to improve developer speed and quality.</p>\n<p>In this role, you&#39;ll have the opportunity to work on a wide range of challenging projects, from performance optimization to design system evolution. You&#39;ll be part of a flat organizational structure, where everyone is valued and empowered to contribute. And, as a remote-friendly company, you&#39;ll have the flexibility to work from anywhere, with opportunities for in-person collaboration when needed.</p>\n<p>If you&#39;re passionate about frontend platform work, enjoy making an entire engineering organization faster and more effective, and are excited about the prospect of joining a dynamic and growing company, we&#39;d love to hear from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4b2edfb8-1c2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Descript","sameAs":"https://descript.com/","logo":"https://logos.yubhub.co/descript.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/descript/jobs/7668317003","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$195,000–$250,000/year","x-skills-required":["React","Modern React ecosystem (hooks, concurrent features, Suspense)","Client-side performance (profiling tools, rendering optimization, bundle analysis, runtime performance tuning)","TypeScript","Modern frontend build tooling","State management approaches in large React applications","Mentoring and guiding other engineers"],"x-skills-preferred":["Experience working on tooling in a monorepo","Background in accessibility (WCAG, ARIA patterns) and inclusive component design","Familiarity with CI/CD optimization for frontend builds and test pipelines","Experience with Electron or desktop web-hybrid applications","Contributions to open-source design systems, React libraries, or developer tooling"],"datePosted":"2026-04-18T15:51:03.743Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"React, Modern React ecosystem (hooks, concurrent features, Suspense), Client-side performance (profiling tools, rendering optimization, bundle analysis, runtime performance tuning), TypeScript, Modern frontend build tooling, State management approaches in large React applications, Mentoring and guiding other engineers, Experience working on tooling in a monorepo, Background in accessibility (WCAG, ARIA patterns) and inclusive component design, Familiarity with CI/CD optimization for frontend builds and test pipelines, Experience with Electron or desktop web-hybrid applications, Contributions to open-source design systems, React libraries, or developer tooling","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":195000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_76f61aca-ede"},"title":"Software Engineer, Human Data Interface","description":"<p><strong>About the Role</strong></p>\n<p>As a Software Engineer on Anthropic&#39;s Human Data Interfaces team, you&#39;ll own the architecture and execution of our data collection pipelines, designing systems that are both performant at scale and resilient to the rapidly changing needs of our research teams.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Architect and build data collection pipelines that support rapid iteration, balancing data quality and system maintainability</li>\n<li>Think deeply about the experience of the crowdworkers and vendors using these systems, building interfaces that are clear, efficient, and lead to high-quality data</li>\n<li>Collaborate closely with research teams to understand evolving data needs and iterate quickly on collection methods</li>\n<li>Partner with our Human Data Operations team to understand the end-to-end data workflow and design interfaces that make their jobs easier</li>\n<li>Prioritize and juggle multiple workstreams, making trade-off decisions in a fast-moving environment where research priorities can shift quickly</li>\n</ul>\n<p><strong>You May Be a Good Fit If You:</strong></p>\n<ul>\n<li>Believe that advanced AI systems could have a transformative effect on the world and are interested in helping make sure that transformation goes well</li>\n<li>Are a strong full-stack engineer with broad experience across the stack</li>\n<li>Are very good at building internal tools, including working with users of the tools to understand their needs</li>\n<li>Thrive in fast-moving environments where you need to balance speed of iteration with long-term system health</li>\n<li>Are a quick study,this team sits at the intersection of a large number of different complex technical systems that you&#39;ll need to understand (at a high level) to be effective</li>\n</ul>\n<p><strong>Strong Candidates May Also Have:</strong></p>\n<ul>\n<li>Experience building human data labelling interfaces, human-in-the-loop systems, or data collection pipelines</li>\n<li>Familiarity with how preference data and reward models are used in AI model training</li>\n<li>Experience working with researchers who are internal users/customers</li>\n<li>Background in building, and improving the user-experience of user-facing applications, particularly those involving complex UI interactions or annotation workflows</li>\n<li>Strong instincts around system design , building things that evolve gracefully as requirements change</li>\n<li>Experience influencing technical and product direction on a team</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience</li>\n<li>Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience</li>\n<li>Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position</li>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Competitive compensation and benefits</li>\n<li>Optional equity donation matching</li>\n<li>Generous vacation and parental leave</li>\n<li>Flexible working hours</li>\n<li>Lovely office space in which to collaborate with colleagues</li>\n</ul>\n<p><strong>How to Apply</strong></p>\n<p>If you&#39;re interested in this role, please submit your application through our website. We look forward to hearing from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_76f61aca-ede","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5109273008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["full-stack engineering","data collection pipeline design","human data labelling interfaces","human-in-the-loop systems","data collection pipelines"],"x-skills-preferred":["preference data and reward models","AI model training","researcher collaboration","user experience design","system design"],"datePosted":"2026-04-18T15:51:03.093Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"full-stack engineering, data collection pipeline design, human data labelling interfaces, human-in-the-loop systems, data collection pipelines, preference data and reward models, AI model training, researcher collaboration, user experience design, system design","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b2efa219-a4e"},"title":"Senior Software Engineer - Search","description":"<p>We are seeking a Senior Software Engineer to join our Applied AI team at the forefront of advancing AI/ML-powered products. The successful candidate will drive enhancements to our Search Quality, focusing on enhancing search ranking, improving query understanding, building robust evals, and growing the coverage of assets to enable seamless search at scale.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Developing and deploying ML-based search and discovery relevance models and systems integrated with Databricks&#39; products and services.</li>\n<li>Designing and implementing automated ML and NLP pipelines for data preprocessing, query understanding and rewrite, ranking and retrieval, and model evaluation, enabling rapid experimentation and iteration.</li>\n<li>Collaborating with product managers and cross-functional teams to drive technology-first initiatives that enable novel business strategies and product roadmaps for the search and discovery experience.</li>\n<li>Contributing to building a robust framework for evaluating search ranking improvements - both offline and online.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>A Bachelor&#39;s degree in Computer Science or a related field, with a Master&#39;s or PhD preferred.</li>\n<li>5+ years of experience developing search relevance systems at scale in production or in high-impact research environments.</li>\n<li>Experience applying LLM to search relevance.</li>\n<li>Strong understanding of computer science fundamentals.</li>\n<li>Contributions to well-used open-source projects.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b2efa219-a4e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7841782002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["ML","NLP","search relevance","data preprocessing","query understanding","ranking and retrieval","model evaluation","computer science fundamentals"],"x-skills-preferred":["LLM","automated ML pipelines","open-source projects"],"datePosted":"2026-04-18T15:50:57.913Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"ML, NLP, search relevance, data preprocessing, query understanding, ranking and retrieval, model evaluation, computer science fundamentals, LLM, automated ML pipelines, open-source projects"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a9d5360b-229"},"title":"Staff Platform Engineer - Infra + DevOps","description":"<p>We&#39;re looking for a seasoned Platform Engineer to join our team. As a leader in aging care innovation, Honor provides technology, tools, and services that empower older adults to live life on their own terms. Our platform engineering team builds and manages the infrastructure &amp; core services that powers Honor&#39;s Care Platform. We&#39;re seeking someone with at least 6 years of professional experience in a platform engineering team within a product-centric company. You will be responsible for designing, implementing, and maintaining scalable distributed systems &amp; infrastructure. Your expertise should include cloud platforms, advanced software design patterns &amp; architecture, operations and automation, and containerization technologies like Kubernetes. You will be joining a small team of highly-skilled, enthusiastic, and passionate engineers with an opportunity to create an outsized impact in contributing to the future evolution of Honor&#39;s Care Platform.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and implement foundational patterns and libraries for Python applications, across a range of technologies from API services to event processing</li>\n<li>Utilize Infrastructure as Code (IaC) tools to ensure reproducible and scalable environment setups</li>\n<li>Design and implement infrastructure for applications hosted on AWS, supporting event-driven systems, containerized services on Kubernetes, and serverless functions</li>\n<li>Develop and maintain robust CI/CD pipelines using tools such as Jenkins, ArgoCD</li>\n<li>Have experience automating the lifecycle management of code from development through production, including code promotion and configuration management</li>\n<li>Instrument observability through tools such as CloudWatch and DataDog to monitor and optimize application performance across multiple environments</li>\n<li>Scale infrastructure to meet increasing demand while managing cost effectively</li>\n<li>Have experience defining, instrumenting and measuring standards for quality, security, scalability, and availability with a focus on delivering business value</li>\n<li>Have passion for delivering turn-key developer experience for local development</li>\n<li>Keen interest in developing talent through mentorship</li>\n<li>Strong written and verbal communication, tailored to a variety of audiences</li>\n<li>A strategic thinker with a product-first approach and customer obsession</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>At least 6 years of professional experience in a platform engineering team within a product-centric company</li>\n<li>Experience working with an RPC architecture</li>\n<li>Experience working at or having worked at a technology startup and familiar with the challenges of evolving platform maturity</li>\n<li>First-hand experience navigating multiple distributed architecture patterns</li>\n</ul>\n<p>Our range reflects the hiring range for this position. We use national average to determine pay as we are a remote first company. Individual pay is based on a number of factors including qualifications, skills, experience, education, and training. Base pay is just a part of our total rewards program. Honor offers generous equity packages that increase with position level and responsibilities, and a 401K with up to a 4% employer match. We provide medical, dental and vision coverage including zero cost plans for employees. Short Term Disability, Long Term Disability and Life Insurance are fully employer paid with a voluntary additional Life Insurance option. We offer a generous time off program, mental health benefits, wellness program, and discount program.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a9d5360b-229","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Honor Technology","sameAs":"https://www.honortech.com/","logo":"https://logos.yubhub.co/honortech.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/honor/jobs/8297124002","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$200,700-$223,000 USD","x-skills-required":["cloud platforms","advanced software design patterns & architecture","operations and automation","containerization technologies like Kubernetes","Infrastructure as Code (IaC)","AWS","event-driven systems","serverless functions","CI/CD pipelines","Jenkins","ArgoCD","observability","CloudWatch","DataDog","quality","security","scalability","availability"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:55.286Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote Position"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"cloud platforms, advanced software design patterns & architecture, operations and automation, containerization technologies like Kubernetes, Infrastructure as Code (IaC), AWS, event-driven systems, serverless functions, CI/CD pipelines, Jenkins, ArgoCD, observability, CloudWatch, DataDog, quality, security, scalability, availability","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":200700,"maxValue":223000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_062d8648-c7c"},"title":"Anthropic Fellows Program — ML Systems & Performance","description":"<p>About Anthropic\\n\\nAnthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole.\\n\\nApply using this link. The next cohort of Anthropic fellows starts on July 20, 2026. Apply by April 26, 2026 to be considered for this cohort. We will continue accepting applications for later cohorts on a rolling basis. In exceptional circumstances, we may be able to accommodate fellows starting outside of usual cohort timelines.\\n\\nThis page is specific to one of the Anthropic Fellows Workstreams, see also the main Anthropic Fellows posting.\\n\\nAnthropic Fellows Program overview\\n\\nThe Anthropic Fellows Program is designed to foster AI research and engineering talent. We provide funding and mentorship to promising technical talent - regardless of previous experience.\\n\\nFellows will primarily use external infrastructure (e.g. open-source models, public APIs) to work on an empirical project aligned with our research priorities, with the goal of producing a public output (e.g. a paper submission). In one of our earlier cohorts, over 80% of fellows produced papers.\\n\\nWe run multiple cohorts of Fellows each year and review applications on a rolling basis. This application is for cohorts starting in July 2026 and beyond.\\n\\nWhat to expect\\n\\n- 4 months of full-time research \\n\\n- Direct mentorship from Anthropic researchers \\n\\n- Access to a shared workspace (in either Berkeley, California or London, UK)\\n\\n- Connection to the broader AI safety and security research community\\n\\n- Weekly stipend of 3,850 USD / 2,310 GBP / 4,300 CAD + benefits (these vary by country)\\n\\n- Funding for compute (~$15k/month) and other research expenses\\n\\nInterview process\\n\\nThe interview process will include an initial application &amp; reference check, technical assessments &amp; interviews, and a research discussion. \\n\\nCompensation\\n\\nThe expected base stipend for this role is 3,850 USD / 2,310 GBP / 4,300 CAD per week, with an expectation of 40 hours per week for 4 months (with possible extension).\\n\\nFellows workstreams\\n\\nDue to the success of the Anthropic Fellows for AI Safety Research program, we are now expanding it across teams at Anthropic. We expect there to be significant overlap in the types of skills and responsibilities across the roles and will by default consider candidates for all the workstreams.\\n\\nSome of the workstreams may include unique assessment steps; we therefore ask you for workstream preferences in the application. You can see an overview of the current workstreams below:\\n\\n- AI Safety Fellows\\n\\n- AI Security Fellows\\n\\n- ML Systems &amp; Performance Fellows\\n\\n- Reinforcement Learning Fellows\\n\\n- Economics &amp; Societal Impacts Fellows\\n\\nAcross the workstreams, you may be a good fit if you:\\n\\n- Are motivated by making sure AI is safe and beneficial for society as a whole\\n\\n- Are excited to transition into empirical AI research and would be interested in a full-time role at Anthropic\\n\\n- Have a strong technical background in computer science, mathematics, or physics\\n\\n- Thrive in fast-paced, collaborative environments\\n\\n- Can implement ideas quickly and communicate clearly\\n\\nStrong candidates may also have:\\n\\n- Strong background in a discipline relevant to a specific Fellows workstream (e.g. economics, social sciences, or cybersecurity)\\n\\n- Experience in areas of research or engineering related to their workstream\\n\\nCandidates must be:\\n\\n- Fluent in Python programming\\n\\n- Available to work full-time on the Fellows program\\n\\nML Systems &amp; Performance Fellows\\n\\nMentors, research areas, &amp; past projects\\n\\nFellows will undergo a project selection &amp; mentor matching process. Potential mentors include:\\n\\n- Alwin Peng\\n\\n- Zygi Straznickas\\n\\nNote: You may research mentors&#39; prior work, but all applications must go through the official form, not the mentors.\\n\\nFor a past example of an engineering-heavy project, see:\\n\\n- AI agents find $4.6M in blockchain smart contract exploits\\n\\nProjects in this workstream may include:\\n\\n- Building a CPU simulator for accelerator workloads\\n\\n- Adding backends for different accelerators on an open source project\\n\\n- Building on demand infrastructure for other infrastructure heavy fellows projects \\n\\n- Building complex synthetic data or environment pipelines\\n\\nUnique candidate criteria\\n\\nYou might be a particularly great fit for this workstream if you:\\n\\n- Have strong software engineering skills with experience building complex ML systems\\n\\n- Can balance research exploration with engineering rigor and operational reliability\\n\\n- Enjoy collaborating across research and engineering disciplines\\n\\n- Are comfortable working with large-scale distributed systems and high-performance computing (e.g. in trading)\\n\\n- Have experience with training, fine-tuning, or evaluating large language models\\n\\n- Are adept at analyzing and debugging model training processes\\n\\nLogistics\\n\\nLogistics Requirements: To participate in the Fellows program, you must have work authorization in the US, UK, or Canada and be located in that country during the program.\\n\\nWorkspace Locations: We have designated shared workspaces in London and Berkeley where fellows will work from and mentors will visit. We are also open to remote fellows in the UK, US, or Canada. We will ask you about your availability to work from Berkeley or London (full- or part-time) during the program.\\n\\nVisa Sponsorship: We are not currently able to sponsor visas for fellows. To participate in the Fellows program, you need to have or independently obtain full-time work authorization in the UK, the US, or Canada.\\n\\nProgram Duration: The program runs for 4 months, full-time. If you can&#39;t commit to the full duration, please still apply and note your constraints in the application. We review these requests on a case-by-case basis.\\n\\nPlease note: We do not guarantee that we will make any full-time offers to fellows. However, strong performance during the program may indicate that a Fellow would be a good fit for full-time roles at Anthropic. In previous cohorts, 25-50% of fellows received a full-time offer, and we’ve supported many more to go on to do great work on AI safety and security at other organizations.\\n\\nApplications and interviews are managed by Constellation, our official recruiting partner for this program. Constellation also runs the Berkeley workspace that hosts fellows. Clicking &quot;Apply here&quot; will redirect you to Constellation&#39;s application portal. You can expect to receive emails from Constellation with application updates.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_062d8648-c7c","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5183051008","x-work-arrangement":"remote","x-experience-level":null,"x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python programming","AI research","Machine learning","Software engineering","Research and development","Collaboration","Communication","Problem-solving","Analytical thinking"],"x-skills-preferred":["Large language models","Distributed systems","High-performance computing","Trading","Synthetic data","Environment pipelines"],"datePosted":"2026-04-18T15:50:38.472Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK; Ontario, CAN; Remote-Friendly, United States; San Francisco, CA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python programming, AI research, Machine learning, Software engineering, Research and development, Collaboration, Communication, Problem-solving, Analytical thinking, Large language models, Distributed systems, High-performance computing, Trading, Synthetic data, Environment pipelines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ae05b1e7-61b"},"title":"Salesforce Administrator II","description":"<p>Why join us</p>\n<p>Brex is the intelligent finance platform that enables companies to spend smarter and move faster in more than 200 markets. By combining global corporate cards and banking with intuitive spend management, bill pay, and travel software, Brex enables founders and finance teams to accelerate operations, gain real-time visibility, and control spend effortlessly.</p>\n<p>As a Salesforce Administrator II at Brex, you will work as a core member of the Salesforce Engineering team. You will help configure, maintain, and scale the Salesforce ecosystem that powers our Sales, Support, and Operations teams. We’re looking for someone eager to build scalable solutions, automate complex processes, and maintain a high bar for data quality and system health.</p>\n<p>In this role, you will move beyond simple ticket resolution to own features end-to-end. You will collaborate directly with engineers and business stakeholders to translate requirements into technical reality. You’ll be encouraged to think like an engineer,focusing on reliability, scalability, and clean configuration,to help Brex prepare for the new markets we are about to enter.</p>\n<p>Where you’ll work</p>\n<p>This role will be based in our São Paulo office. We are a hybrid environment that combines the energy and connections of being in the office with the benefits and flexibility of working from home. We currently require a minimum of two coordinated days in the office per week, Wednesday and Thursday. Starting February 2, 2026, we will require 3 days per week in the office - Monday, Wednesday, and Thursday. As a perk, we also have up to four weeks per year of fully remote work!</p>\n<p>Responsibilities</p>\n<ul>\n<li>Collaborate with cross-functional teams (Sales, Operations, Finance) to configure and implement new Salesforce features and enhancements.</li>\n<li>Manage day-to-day administration, including user management, roles, profiles, sharing rules, and security controls.</li>\n<li>Own configuration tasks end-to-end, from understanding the user need to testing and deployment.</li>\n<li>Build and maintain complex Flows and automation logic, ensuring they are performant and scalable.</li>\n<li>Participate in the software development lifecycle (SDLC) by managing deployments, validating metadata changes, and maintaining documentation.</li>\n<li>Tune and polish existing configurations to specific engineering standards and assist in technical debt reduction.</li>\n<li>Triaging and resolving bugs and support requests to maintain high system reliability.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>2+ years of professional experience as a Salesforce Administrator.</li>\n<li>Experience with Salesforce Sales Cloud and Service Cloud.</li>\n<li>Strong proficiency with Salesforce declarative development (Flows, Object Manager, Validation Rules) and security models.</li>\n<li>Experience deploying changes through Salesforce DevOps pipelines.</li>\n<li>A high bar for configuration standards, best practices, naming conventions, and documentation.</li>\n<li>Ability to communicate technical concepts clearly to non-technical stakeholders.</li>\n<li>Strong understanding of SOQL and Salesforce&#39;s data model.</li>\n<li>Solid grasp of Salesforce&#39;s Governor Limits.</li>\n</ul>\n<p>Bonus points</p>\n<ul>\n<li>Experience with declarative integrations between Salesforce and third-party applications.</li>\n<li>Familiarity with Agile methodologies.</li>\n<li>Experience with DevOps tools (e.g., Gearset, Copado) or version control (Git).</li>\n<li>Basic understanding of Apex (ability to read/debug code) and programmatic solutions available.</li>\n<li>Experience working within a high-growth technology or financial services company.</li>\n<li>Salesforce certifications: Administrator, Platform App Builder, Sales Cloud Consultant, Service Cloud Consultant, etc.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ae05b1e7-61b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Brex","sameAs":"https://brex.com/","logo":"https://logos.yubhub.co/brex.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/brex/jobs/8398611002","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Salesforce","Sales Cloud","Service Cloud","Flows","Object Manager","Validation Rules","Security Models","DevOps Pipelines","SOQL","Salesforce Data Model","Governor Limits"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:33.799Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"São Paulo, São Paulo, Brazil"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Salesforce, Sales Cloud, Service Cloud, Flows, Object Manager, Validation Rules, Security Models, DevOps Pipelines, SOQL, Salesforce Data Model, Governor Limits"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_550052ff-6a5"},"title":"Director of Product, Growth/AI","description":"<p>Why join us</p>\n<p>Brex is the intelligent finance platform that enables companies to spend smarter and move faster in more than 200 markets. Our platform combines global corporate cards and banking with intuitive spend management, bill pay, and travel software, enabling founders and finance teams to accelerate operations, gain real-time visibility, and control spend effortlessly.</p>\n<p>As a Product Leader at Brex, you will be the driving force behind our Growth Product team , overseeing the thoughtful strategy and execution of team and technical systems to drive customer acquisition and onboarding. You will also collaborate closely with our Go-to-Market (GTM) teams to ensure seamless acquisition and onboarding for customers, particularly those with significant and complex spending needs.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Drive product and systems investments that improve GTM metrics such as CAC payback, pipeline conversion, and efficiency.</li>\n<li>Define and execute a multi-quarter Growth strategy in partnership with Engineering, Design, GTM, Marketing, Ops, and Data to deliver high-impact, scalable products with measurable business impact.</li>\n<li>Embed within Sales, Marketing, and Customer Success to identify workflow bottlenecks and lead automation and optimization initiatives.</li>\n<li>Analyze GTM funnel and conversion metrics to uncover insights and guide product priorities and roadmap decisions.</li>\n<li>Set clear goals, metrics, and success criteria for the Growth team, driving accountability to measurable outcomes.</li>\n<li>Manage and develop a team of Product Managers, fostering clarity, ownership, and a culture of high performance.</li>\n<li>Oversee core Growth Product and Engineering domains, including account scoring, TAM data, outbound automation, and GTM tooling.</li>\n<li>Align stakeholders around an aspirational 6–12 month roadmap that advances GTM efficiency and business growth.</li>\n<li>Ensure successful activation and implementation of customers initiatives by partnering with GTM and Customer Success to operationalize tooling and automations.</li>\n<li>Continuously refine activation based on feedback and performance data to drive faster time-to-value and sustainable adoption.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>Relevant experience: 10+ years of experience in product management or management consulting, ideally with direct responsibility for or experience scaling B2B products and GTM systems, with a strong understanding of growth, GTM, and/or marketing pipelines.</li>\n<li>Analytical sense: You demonstrate a strong aptitude for and actively use data to inform your decision-making. Basic SQL proficiency is strongly preferred.</li>\n<li>Communication: You have proven experience working and communicating effectively with senior leaders and executives, and influencing stakeholders across departments and functions, including Engineering, Design, Marketing, Sales, CRMX/Business Systems, IT, Legal, and more. You excel at building trust and working relationships.</li>\n<li>Systems-thinker: You are a systems thinker, capable of thinking holistically across product, technical, and people systems to identify bottlenecks and opportunities for efficiency, automation, fraud prevention, and product growth within the payables ecosystem.</li>\n</ul>\n<p>Compensation</p>\n<p>The expected salary range for this role is $340,000 - $425,000. However, the starting base pay will depend on a number of factors including the candidate’s location, skills, experience, market demands, and internal pay parity. Depending on the position offered, equity and other forms of compensation may be provided as part of a total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_550052ff-6a5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Brex","sameAs":"https://brex.com/","logo":"https://logos.yubhub.co/brex.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/brex/jobs/8432707002","x-work-arrangement":"hybrid","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":"$340,000 - $425,000","x-skills-required":["product management","management consulting","growth","GTM","marketing pipelines","data analysis","SQL","communication","systems thinking"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:31.761Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Finance","industry":"Finance","skills":"product management, management consulting, growth, GTM, marketing pipelines, data analysis, SQL, communication, systems thinking","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":340000,"maxValue":425000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_372999e8-579"},"title":"Senior Software Engineer II, AI Workload Orchestration","description":"<p>As a Senior Software Engineer II on the AI Workload Orchestration team, you will help build and operate CoreWeave&#39;s Kubernetes-native platform for admitting, scheduling, and operating AI workloads at scale.</p>\n<p>This platform integrates multiple orchestration and scheduling frameworks such as Kueue, Volcano, and Ray to support modern AI training and inference workflows. It complements SUNK (Slurm on Kubernetes) by providing a Kubernetes-first, cloud-native orchestration layer with deep platform integration.</p>\n<p>You will own meaningful components of the platform, drive reliability and performance improvements, and help scale the system as customer demand and workload complexity continue to grow.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and operate Kubernetes-native services for AI workload orchestration and scheduling</li>\n<li>Own one or more platform components end-to-end, including design, implementation, testing, and on-call support</li>\n<li>Improve scheduling latency, cluster utilization, and workload reliability through metrics-driven engineering</li>\n<li>Contribute to architectural discussions across services and influence design decisions within the platform</li>\n<li>Work closely with adjacent teams (CKS, infrastructure, managed inference) to ensure clean interfaces and integrations</li>\n<li>Mentor junior engineers and raise the quality bar for code, design, and operations</li>\n</ul>\n<p>About the role:</p>\n<ul>\n<li>5–8 years of professional software engineering experience in distributed systems, cloud infrastructure, or platform engineering</li>\n<li>Strong experience building production systems in Go (Python or C++ a plus)</li>\n<li>Solid understanding of Kubernetes fundamentals, APIs, controllers, and operating services in production</li>\n<li>Experience working with scheduling, resource management, or quota-based systems</li>\n<li>Proven ability to improve system reliability and performance using data and operational metrics</li>\n<li>Comfortable owning services in production and participating in on-call rotations</li>\n</ul>\n<p>Preferred:</p>\n<ul>\n<li>Experience with Kubernetes-native orchestration frameworks such as Kueue, Volcano, Ray, Kubeflow, or Argo Workflows</li>\n<li>Familiarity with GPU-based workloads, ML training, or inference pipelines</li>\n<li>Knowledge of scheduling concepts such as quota enforcement, pre-emption, and backfilling</li>\n<li>Experience with reliability practices including SLOs, alerting, and incident response</li>\n<li>Exposure to AI infrastructure, HPC, or large-scale distributed compute environments</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>The base salary range for this role is $165,000 to $242,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation. In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.</p>\n<p>In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance - 100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_372999e8-579","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4647595006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["Kubernetes","Go","Distributed systems","Cloud infrastructure","Platform engineering","Scheduling","Resource management","Quota-based systems"],"x-skills-preferred":["Kueue","Volcano","Ray","Kubeflow","Argo Workflows","GPU-based workloads","ML training","Inference pipelines","SLOs","Alerting","Incident response","AI infrastructure","HPC","Large-scale distributed compute environments"],"datePosted":"2026-04-18T15:50:19.636Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, Go, Distributed systems, Cloud infrastructure, Platform engineering, Scheduling, Resource management, Quota-based systems, Kueue, Volcano, Ray, Kubeflow, Argo Workflows, GPU-based workloads, ML training, Inference pipelines, SLOs, Alerting, Incident response, AI infrastructure, HPC, Large-scale distributed compute environments","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_44adf646-ae7"},"title":"OIC Developer","description":"<p>We are looking for an expert Oracle Integration Developer to join our Arsenal (Enterprise Systems) team. Your immediate mission: take ownership of our critical enterprise integrations connecting Oracle Fusion ERP with our upstream and downstream systems. These integrations, built on Oracle Integration Cloud, form the digital backbone that governs how we manage our business operations, from product data and procurement to manufacturing and financial processes. You will be tasked with stabilizing, optimizing, and making them exceptionally robust. Long-term, you will be the subject matter expert responsible for architecting and scaling our enterprise integration landscape. This is a high-impact role for someone who thrives on solving complex data challenges and wants to build the operational foundation that enables Anduril to scale its mission.</p>\n<p>Stabilize &amp; Optimize: Dive deep into existing Oracle Fusion ERP integrations across manufacturing, supply chain, finance, and engineering systems. Diagnose root causes of instability, re-architect weak points, and implement robust error handling and monitoring to achieve mission-critical reliability.</p>\n<p>Architect &amp; Build: Design and develop new, scalable enterprise integrations using Oracle Integration Cloud (OIC). Translate complex business requirements for product data, multi-level Bills of Material (BOMs), procurement, inventory, work orders, and financial transactions into resilient and efficient integration flows.</p>\n<p>Own the Integration Lifecycle: Manage the end-to-end process from design and development through testing (unit, SIT, UAT) and deployment, utilizing CI/CD best practices. Proactively tune and maintain integrations to ensure peak performance as data volumes grow.</p>\n<p>Ensure Data Integrity: Become the trusted expert on data transformation and mapping between systems. Implement rigorous validation and reconciliation logic to guarantee that our enterprise data is flawless across all systems.</p>\n<p>Collaborate &amp; Influence: Act as the key technical partner to our ERP, Manufacturing, Supply Chain, and Finance teams. Clearly articulate technical designs, trade-offs, and progress to both engineering peers and business stakeholders, guiding them toward best-practice integration patterns.</p>\n<p>Leverage Modern Oracle Cloud Tools: Utilize Oracle Visual Builder Cloud Service (VBCS) where appropriate to build lightweight user interfaces that enhance integration workflows, data validation, or operational dashboards.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_44adf646-ae7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5061445007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$126,000-$167,000 USD","x-skills-required":["Oracle Integration Cloud (OIC)","Oracle Fusion ERP","RESTful APIs","SOAP web services","XSLT","XPath","complex data mapping","SQL","relational database concepts"],"x-skills-preferred":["Oracle Visual Builder Cloud Service (VBCS)","Oracle Business Intelligence Cloud Connector (BICC)","Oracle Cloud Infrastructure (OCI) services","PLM systems","Git-based source control","CI/CD pipelines"],"datePosted":"2026-04-18T15:50:18.562Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Atlanta, Georgia, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Oracle Integration Cloud (OIC), Oracle Fusion ERP, RESTful APIs, SOAP web services, XSLT, XPath, complex data mapping, SQL, relational database concepts, Oracle Visual Builder Cloud Service (VBCS), Oracle Business Intelligence Cloud Connector (BICC), Oracle Cloud Infrastructure (OCI) services, PLM systems, Git-based source control, CI/CD pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":126000,"maxValue":167000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_059293a1-afa"},"title":"Systems Engineer, Data","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>We were named to Entrepreneur Magazine’s Top Company Cultures list and ranked among the World’s Most Innovative Companies by Fast Company.</p>\n<p>About the Team</p>\n<p>The Core Data team’s mission is building a centralized data platform for Cloudflare that provides secure, democratized access to data for internal customers throughout the company. We operate infrastructure and craft tools to empower both technical and non-technical users to answer their most important questions. We facilitate access to data from federated sources across the company for dashboarding, ad-hoc querying and in-product use cases. We power data pipelines and data products, secure and monitor data, and drive data governance at Cloudflare.</p>\n<p>Our work enables every individual at the company to act with greater information and make more informed decisions.</p>\n<p>About the Role</p>\n<p>We are looking for a systems engineer with a strong background in data to help us expand and maintain our data infrastructure. You’ll contribute to the technical implementation of our scaling data platform, manage access while accounting for privacy and security, build data pipelines, and develop tools to automate accessibility and usefulness of data. You’ll collaborate with teams including Product Growth, Marketing, and Billing to help them make informed decisions and power usage-based invoicing platforms, as well as work with product teams to bring new data-driven solutions to Cloudflare customers.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Contribute to the design and execution of technical architecture for highly visible data infrastructure at the company.</li>\n<li>Design and develop tools and infrastructure to improve and scale our data systems at Cloudflare.</li>\n<li>Build and maintain data pipelines and data products to serve customers throughout the company, including tools to automate delivery of those services.</li>\n<li>Gain deep knowledge of our data platforms and tools to guide and enable stakeholders with their data needs.</li>\n<li>Work across our tech stack, which includes Kubernetes, Trino, Iceberg, Clickhouse, and PostgreSQL, with software built using Go, Javascript/Typescript, Python, and others.</li>\n<li>Collaborate with peers to reinforce a culture of exceptional delivery and accountability on the team.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>3-5+ years of experience as a software engineer with a focus on building and maintaining data infrastructure.</li>\n<li>Experience participating in technical initiatives in a cross-functional context, working with stakeholders to deliver value.</li>\n<li>Practical experience with data infrastructure components, such as Trino, Spark, Iceberg/Delta Lake, Kafka, Clickhouse, or PostgreSQL.</li>\n<li>Hands-on experience building and debugging data pipelines.</li>\n<li>Proficient using backend languages like Go, Python, or Typescript, along with strong SQL skills.</li>\n<li>Strong analytical skills, with a focus on understanding how data is used to drive business value.</li>\n<li>Solid communication skills, with the ability to explain technical concepts to both technical and non-technical audiences.</li>\n</ul>\n<p>Desirable Skills</p>\n<ul>\n<li>Experience with data orchestration and infrastructure platforms like Airflow and DBT.</li>\n<li>Experience deploying and managing services in Kubernetes.</li>\n<li>Familiarity with data governance processes, privacy requirements, or auditability.</li>\n<li>Interest in or knowledge of machine learning models and MLOps.</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law. We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals with disabilities. Please tell us if you require a reasonable accommodation to apply for a job. Examples of reasonable accommodations include, but are not limited to, changing the application process, providing documents in an alternate format, using a sign language interpreter, or using specialized equipment. If you require a reasonable accommodation to apply for a job, please contact us via e-mail at hr@cloudflare.com or via mail at 101 Townsend St. San Francisco, CA 94107.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_059293a1-afa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7527453","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data infrastructure","data pipelines","data products","Kubernetes","Trino","Iceberg","Clickhouse","PostgreSQL","Go","Javascript/Typescript","Python","SQL"],"x-skills-preferred":["data orchestration","infrastructure platforms","Airflow","DBT","machine learning models","MLOps"],"datePosted":"2026-04-18T15:50:12.541Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data infrastructure, data pipelines, data products, Kubernetes, Trino, Iceberg, Clickhouse, PostgreSQL, Go, Javascript/Typescript, Python, SQL, data orchestration, infrastructure platforms, Airflow, DBT, machine learning models, MLOps"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_854e95b5-76b"},"title":"Sr. Director of Product, Research and Training Infrastructure","description":"<p>CoreWeave is seeking a visionary Sr. Director of Product, Research Training Infrastructure to lead the product strategy and engineering execution for the services that power the most ambitious AI research labs in the world.</p>\n<p>This executive leader will own the product strategy and engineering execution for the Research Training Stack, focusing on the specialized orchestration, evaluation, and iteration tools required for massive-scale pre-training and post-training.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Frontier Orchestration: Oversee the evolution of SUNK (Slurm on Kubernetes) to provide researchers with deterministic, bare-metal performance through a cloud-native interface.</li>\n</ul>\n<ul>\n<li>Holistic Training Services: Drive the development of next-generation orchestrators and automated training-based evaluation frameworks that ensure model quality throughout the lifecycle.</li>\n</ul>\n<ul>\n<li>Post-Training Excellence: Build the infrastructure required for sophisticated Reinforcement Learning (RL) and RLHF pipelines, enabling labs to refine foundation models with maximum efficiency.</li>\n</ul>\n<ul>\n<li>Customer Advocacy: Act as the primary technical partner for lead researchers at global AI labs, translating their &#39;future-state&#39; requirements into actionable product roadmaps.</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>Proven leadership experience in engineering leadership, with at least 5+ years managing large-scale infrastructure at a top-tier research lab or an AI-native cloud provider.</li>\n</ul>\n<ul>\n<li>Deep, hands-on knowledge of Slurm, Kubernetes, and the specific networking requirements (InfiniBand/RDMA) for distributed training clusters.</li>\n</ul>\n<ul>\n<li>Research mindset and understanding of the &#39;pain points&#39; of a research scientist.</li>\n</ul>\n<ul>\n<li>Scaling experience delivering mission-critical services on multi-thousand GPU clusters (H100/Blackwell/Rubin architectures).</li>\n</ul>\n<ul>\n<li>Strategic vision to define &#39;what&#39;s next&#39; in the AI stack, from automated RL loops to specialized sandbox environments.</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>In 2026, CoreWeave is the foundation of the largest infrastructure buildout in human history. We are building AI Factories, not just data centers.</p>\n<ul>\n<li>Silicon-Up Innovation: Work directly with the latest NVIDIA architectures.</li>\n</ul>\n<ul>\n<li>Impact: You will be the architect of the environment that enables the next new discovery.</li>\n</ul>\n<p>Velocity: We move at the speed of the researchers we support, bypassing legacy cloud bottlenecks to deliver raw power.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_854e95b5-76b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4665964006","x-work-arrangement":"hybrid","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":"$233,000 to $341,000","x-skills-required":["Slurm","Kubernetes","InfiniBand/RDMA","Distributed training clusters","GPU clusters","H100/Blackwell/Rubin architectures","Reinforcement Learning (RL)","RLHF pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:50:11.130Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Slurm, Kubernetes, InfiniBand/RDMA, Distributed training clusters, GPU clusters, H100/Blackwell/Rubin architectures, Reinforcement Learning (RL), RLHF pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":233000,"maxValue":341000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b5ce114e-dac"},"title":"Cloud Engineer – Factory Systems and Operational Technology","description":"<p>Anduril Industries is a defence technology company with a mission to transform U.S. and allied military capabilities with advanced technology. By bringing the expertise, technology and business model of the 21st century&#39;s most innovative companies to the defence industry, Anduril is changing how military systems are designed, built and sold.</p>\n<p>The company&#39;s family of systems is powered by Lattice OS, an AI-powered operating system that turns thousands of data streams into a real-time, 3D command and control centre.</p>\n<p>As the world enters an era of strategic competition, Anduril is committed to bringing cutting-edge autonomy, AI, computer vision, sensor fusion and networking technology to the military in months, not years.</p>\n<p>We are seeking a mission-driven Cloud Infrastructure Engineer to take a leading role in designing and implementing world-class defensive controls. This is a high-impact role with the autonomy to shape security architecture and protect the technology that is changing the future of defence.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Design and Own Security Architecture: Architect, build and deploy robust, scalable security controls for our corporate, development and production cloud environments (AWS, Azure, GCP).</li>\n</ul>\n<ul>\n<li>Automate Everything: Develop and automate infrastructure-as-code (IaC) to manage and scale our cloud deployments securely and efficiently.</li>\n</ul>\n<ul>\n<li>Proactively Defend: Continuously monitor, identify and remediate security weaknesses and configuration drift across our entire cloud footprint.</li>\n</ul>\n<ul>\n<li>Be a Force Multiplier: Partner with infrastructure, application and product teams to embed security best practices into their workflows and secure environments holding mission-critical data.</li>\n</ul>\n<ul>\n<li>Enable Scale and Reliability: Engineer systems and processes that ensure our platforms are highly available, resilient and prepared for rapid growth.</li>\n</ul>\n<ul>\n<li>Serve as a Cloud Security Expert: Act as the go-to subject matter expert for teams across Anduril, providing guidance, mentorship and paved-road solutions for building securely in the cloud.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Proven experience building and securing complex cloud environments, typically gained through 3+ years in a Cloud Security, DevOps or SRE role.</li>\n</ul>\n<ul>\n<li>Deep proficiency in at least one major cloud provider (AWS, Azure or GCP).</li>\n</ul>\n<ul>\n<li>Strong hands-on experience with Infrastructure as Code (e.g., Terraform, CloudFormation, Bicep).</li>\n</ul>\n<ul>\n<li>Solid programming/scripting ability in one or more languages (e.g., Python, Go, Rust).</li>\n</ul>\n<ul>\n<li>Firm understanding of public cloud networking principles (e.g., VPCs, subnets, routing, security groups).</li>\n</ul>\n<ul>\n<li>Must be a U.S. Person and eligible to obtain and maintain a U.S. Top Secret security clearance.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience hardening and monitoring Kubernetes clusters (EKS, GKE, AKS).</li>\n</ul>\n<ul>\n<li>Experience with cloud security posture management (CSPM) or threat detection tooling.</li>\n</ul>\n<ul>\n<li>Familiarity with CI/CD pipelines and securing the software supply chain.</li>\n</ul>\n<ul>\n<li>Knowledge of compliance frameworks such as FedRAMP, MRL, SOC 2 or CMMC.</li>\n</ul>\n<ul>\n<li>On-premises network engineering experience.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b5ce114e-dac","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril Industries","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/5087348007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$129,000-$193,000 USD","x-skills-required":["Cloud Security","DevOps","SRE","Infrastructure as Code","Terraform","CloudFormation","Bicep","Python","Go","Rust","Public Cloud Networking","VPCs","Subnets","Routing","Security Groups"],"x-skills-preferred":["Kubernetes","Cloud Security Posture Management","Threat Detection Tooling","CI/CD Pipelines","Software Supply Chain Security","Compliance Frameworks","FedRAMP","MRL","SOC 2","CMMC","On-Premises Network Engineering"],"datePosted":"2026-04-18T15:49:59.253Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Cloud Security, DevOps, SRE, Infrastructure as Code, Terraform, CloudFormation, Bicep, Python, Go, Rust, Public Cloud Networking, VPCs, Subnets, Routing, Security Groups, Kubernetes, Cloud Security Posture Management, Threat Detection Tooling, CI/CD Pipelines, Software Supply Chain Security, Compliance Frameworks, FedRAMP, MRL, SOC 2, CMMC, On-Premises Network Engineering","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":129000,"maxValue":193000,"unitText":"YEAR"}}}]}